Abstract: Effective Human Computer Intelligent Interaction (HCII) requires the information about the userís identity, state and intent which can be extracted from images, so that computers can then react accordingly, e.g. systems behaving according to the emotional state of the person. The most expressive way humans display emotions is through facial expressions. In this paper we have presented the technique of recognizing emotion using facial expressions which is of prime focus in human interactions. By creating machines that can detect and understand emotion, we can enhance the human computer interaction. In this paper, we have discussed a framework for the classification of emotional states, based on images of the face and the implementation details of a real-time facial feature extraction and emotion recognition application. This application automatically detects faces from the captured images and codes them with respect to 7 dimensions in real time: Neutral, Anger, Disgust, Fear, Smile, Sadness, and Surprise. Most interestingly the outputs of the classifier change smoothly as a function of time, providing a possibly worth representation of code facial expression dynamics in a fully automatic and unnoticeable manner. The main objective of our work is the real-time implementation of a facial emotion recognition system.
Keywords: Real Time, Emotion Recognition, Human Computer Interaction (HCI), Feed-Forward Neural Network, Largest Connected Region (LCR), Cubic Bezier Curve.