Sunday, July 26, 2015

Mood Detection: Implementing a facial expression recognition system

 Neeraj Agrawal, Rob Cosgriff and Ritvik Mudur


Introduction
The purpose of this research is to develop a mood detection system which can achieve high recognition rates of facial expression across multiple databases.

Approach
An approach using eigenface masks was developed and implemented to classify facial expressions across publicly available databases. These learned masks along with feature vectors of the image were used to train the SVM (Support Vector Machines).

Method
  • Pre-processing image data by normalizing and applying a simple mask.
  • Extracting certain (facial) features using PCA (Principal Component Analysis) and Gabor filters.
  • Using SVMs for classification and recognition of expressions.

Feature extraction
a.     Normalized pixel intensities (Every image in training set normalized by subtracting the mean of all training set images. The masked region is then converted to a column vector which forms the feature vector).
b.     Gabor filter representations (Gabor filters are often used in image processing and are based on physiological studies of the human visual cortex. The use of Gabor filtered facial images has been shown to result in improved accuracy for facial expression recognition).

Eigenface masks (eigenfaces are used to generate a mask that eliminates pixels that vary little across training samples in different labels. In this system, they modify the approach to generate a separate mask for each expression class).



Strength
When apply Eigenface mask the results were similar for both the normalized pixel value and the gabor filtered feature vectors.



Advantages
  • This approach works reasonably well in capturing expression-relevant facial information across all databases.


Conclusion
The results were similar for both the normalized pixel value and the gabor filtered feature vectors, with neither representation being clearly superior to the other. The performance across databases is indicative of the method’s robustness to variations in facial structure and skin tone when recognizing the expression.
However, both representations of the feature vectors end up being very high dimensional. This hurts the run time of the algorithm hampering the ability to use it in real time, as well as leading to possible over fitting of the data during training. Increasing the number of available training images would help to compensate for over fitting.

Download Research Paper



No comments:

Post a Comment