Publication
- Facial Expression Detection
- Client: IJRTE
- Publication date: Jan, 2020
- Publication URL: volume-8-issue-5
- Leading a team of three, published a paper on various implementation and methodologies of 'Facial Expression Detection' like RGB Model and Comparison Model on IJRTE, Volume-8 Issue-5, January 2020.
- For Comparison Model, Developed and trained a Convolution Neural Network (CNN) through Keras in Python, using fer2013 dataset, which contains almost 30000 face centric RGB images, categorized based on the facial expression.
- Emotion Detection Using Color Image processing, face detection module and Open CV, referred as the RGB Model.
International Journal of Recent Technology and Engineering (IJRTE)
ISSN:
Facial Expression Detection using Artificial
Intelligence
D. Ram Kiran, K. Vinay Kumar, T. Kalyan, K. Ch. Kavya, K. Sarat Kumar
Abstract: The research on the facial expression detection or the
Keywords: Automatic face recognition,
clinical psychology, feeling detection and torment appraisal. There are at most three significant strides in an Expression Recognition System 1. To distinguish the face from the given info picture or video, 2. To remove the facial features like eyes, nose, mouth from the distinguished face and 3. To group the facial expressions into various classes like Happy, Angry, Sad, Fear, Disgust, and Surprise. Face detection is an exceptional instance of object detection. In the proposed system, face detection is executed utilizing skin color detection and division. Additionally, it involves predefined packages in python like Keras, OpenCV and Face Detection module. Let’s talk about these modules in detail.
A. Keras
Keras is like an inside of an software. Keras is an API inside python used to develop neural networks. Its main application is the creation of the convolution neural networks. A neural network is a process which learns and updates itself. This type of
I. INTRODUCTION
Face assumes a significant job in
Revised Manuscript Received on January 15, 2020
D. Ram Kiran, Department of ECE, Koneru Lakshmaiah Education Foundation, Guntur, Email: ramkiran55.devireddy@gmail.com
K .VinayKumar, Department of ECE, Koneru Lakshmaiah Education Foundation, Guntur, Email: vinaykumarkarusala@gmail.com
T. Kalyan, Department of ECE, Koneru Lakshmaiah Education Foundation, Guntur, Email: kalyan6166@gmail.com
Dr. K. Ch. Kavya, Department of ECE, Koneru Lakshmaiah Education Foundation, Guntur, Email: kavya@kluniversity.in
Dr. K. SARAT KUMAR, Department of ECE, Koneru Lakshmaiah Education Foundation, Guntur, Email: kksarat@kluniversity.in
B. OpenCV
OpenCV or the Open Source Computer Vision Library is like an eye for a machine. OpenCV is nothing but the vision to a computer. This library is used for
C. Face detection
Face detection is a computer technology being utilized in an assortment of uses that identifies human faces in computerized pictures. Face identification likewise alludes to the psychological process by which humans find and take care of faces in a visual scene.
II. PREVIOUS WORKS
There are tons of works done in this domain. We took few methodologies as reference for this paper.
A. RGB Model
One of those methodologies is emotion detection using RGB model. In this model firstly, the video capturing is used to capture frames.
|
|
Published By: |
Retrieval Number: E6284018520/2020©BEIESP |
1720 |
Blue Eyes Intelligence Engineering |
DOI:10.35940/ijrte.E6284.018520 |
& Sciences Publication |
Facial Expression Detection using Artificial Intelligence
Once the image is captured, face detection module is applied to extract face from the image. Then comes the major part, the extracted face is divided into segments using RGB processing. Eyes, lips and eyebrows are the major segments considered. From these segments, emotion is classified and using a data write the emotion which is classified is written in the form of a text. Pros for this model is the
detected is compared with the images under these folders and the percentage of matching for each folder is stored in a list. The index of the value which is highest among the list is retrieved. If this highest value is less than the threshold (40%), neutral is given as output. It the highest value is greater than the threshold, its index is taken. If the index value is 0, happy is given as output, if the index is 1, sad is given and so on. The output which is collected is then printed over the image using the data write command. The advantage of this model which we have developed is quite accurate and take’s moderate time for processing. But the disadvantage for this model is, more expressions and feature adding, delays the processing.
B. Comparison Model
The other methodology is a comparison model. In this model, frames capturing is done based on OpenCV package. Once the frames capturing is done, face is separated from the frames of images using any face recognition model. The extracted face is then compared with a dataset. A dataset is collection of images. Collection of these images contain almost all the categories of these images. Once any one of the images is matched with the face with the matching percentage of about 40%, the output is given. The output is nothing but the category under which the image is matched. Advantage for this model is the emotion classification. The comparison is one of the accurate and finest models. But the comparison is not done for the full extent. That is the disadvantage of this model. We have modified this model to be a full comparison model.
III. IMPLEMENTATION
Considering the second model above as the primary model, we have developed a model for the classification of emotion. In this model comparison is done for full extent. In this model firstly, video capturing is done in order to capture frames.
From the captured frames the face is detected and extracted using face recognition modules in python (like the first model). From here the tricky part starts.
The comparison model is used from here. The detected face is then compared to the categories of images. Prior to this implementation a dataset is downloaded and categorized under happy, sad, surprise, angry and fear. The face which is
IV. KEY STEPS
Firstly, import all the pre required libraries. Libraries which are available in python. This is most important and the toughest part. Some of them has to be installed using a pip installer in python. The libraries are OpenCV, NumPy for image processing. Now we need other libraries for face recognition, for giving the output. Hence, we have to import detect faces and draw text packages in python. There after we need to get the required inputs from a dataset. The dataset we choose to be is fer2013 dataset. We need to get all the labels from the dataset into our program. Now fix the window size and face boundaries and import draw_boundingbox. We are now going to start our processing. Firstly, capture images using OpenCV or cv2. Using video capture command read images continuously. After reading these images convert them into grey scale and RGB models for further image processing. Further processing includes applying offsets on the face coordinates and emotion offsets for which we need to import apply offsets package from python. Complete all the logical processing in order to find the emotion or facial expression. Get the emotion text and append it on the face using draw text which have imported. Ensure that the emotion window size should be less than emotion window size. Methodology is shown in the below Figure.
V. DATASET
The fer2013 collection consists of the grey scale images of size 48 x 48. The images have been automatically saved such that the face is at the middle and takes about the same percent of space in every image. The task is to separate every face based on the expression mentioned in to one of seven categories (0=Happy, 1=Sad, 2=Surprise, 3=fear, 4=angry, 5=Neutral). This collection of data set contains the images under every category and can be able to import in to the programming like python. This helps in the separation of images based on their expressions. The set of images under these folders can be compared in order to get the matching percentage. Matching percentage is the amount of overlapping over the image. This measure of overlapping gives values precisely which are then used to estimate the expression. The list which is used to store these percentages has two columns. They are the "expression" column and the "matching percentage" column.
Retrieval Number: E6284018520/2020©BEIESP |
|
Published By: |
|
|
Blue Eyes Intelligence Engineering |
||
DOI:10.35940/ijrte.E6284.018520 |
1721 |
||
& Sciences Publication |
|||
|
|
International Journal of Recent Technology and Engineering (IJRTE) ISSN:
The prediction of emotion is done based on the comparison with the dataset. The dataset has almost 3589 examples and the training set consists of the 28709 examples.
VI. RESULTS AND DISCUSSION
E. Surprise
Table- V: Tabular values for emotion Surprise
Index |
0 |
1 |
2 |
3 |
4 |
values |
42.36% |
42.03% |
49.29% |
44.34% |
41.25% |
|
|
|
|
|
|
The index ‘2’ is highest and is greater than 40%. Thus, Surprise is given as output.
F. Fear
Table- VI: Tabular values for emotion Fear
Index |
0 |
1 |
2 |
3 |
4 |
values |
38.12 |
46.97 |
28.12 |
48.34 |
41.89 |
|
% |
% |
% |
% |
% |
The index ‘3’ is highest and is greater than 40%. Thus, Fear is given as output.
A. Happy
Table- I: Tabular values for emotion Happy
Index |
0 |
1 |
2 |
3 |
4 |
values |
82.4 |
42.67 |
64.93 |
24.35 |
26.48 |
|
% |
% |
% |
% |
% |
The index ‘0’ is highest and is greater than 40%. Thus, Happy is given as output.
B. Sad
Table- II: Tabular values for emotion Sad
VI. FUTURE SCOPE
The future scope for this paper is obviously going to be accuracy which we can never compromise on. The accuracy can be increased by increasing the samples in the dataset. But this also increases the time and complexity. We also need to add more expressions like normally happy, excited, tremendously happy, very sad, bold, cold etc. But this again increases the time of processing and the complexity. Thus, for this model must be altered. One of the modifications that can be done is adding a neural network. A new neural network which alters the dataset after every execution. The neural network must take the image of the user and it must add the image to one of the categories in the dataset. This can achieve high reduction in processing time. Because the users who comes again will be given output immediately without any execution or processing.
Index |
0 |
1 |
2 |
3 |
4 |
values |
62.13 |
72.67 |
43.47 |
69.43 |
58.41 |
|
% |
% |
% |
% |
% |
The index ‘1’ is highest and is greater than 40%. Thus, Sad is given as the output.
C. Neutral
Table- III: Tabular values for emotion Neutral
Index |
0 |
1 |
2 |
3 |
4 |
values |
33.89 |
24.56 |
38.55 |
36.44 |
39.35 |
|
% |
% |
% |
% |
% |
The index ‘4’ is highest and but is less than 40%. Thus, Neutral is given as the output.
D. Angry
Table- IV: Tabular values for emotion Angry
Index |
0 |
1 |
2 |
3 |
4 |
values |
24.32 |
61.54 |
44.93 |
64.91 |
69.42 |
|
% |
% |
% |
% |
% |
The index ‘4’ is maximum and is greater than 40%. Thus, Angry is given as output.
VII. APPLICATIONS
1)One of the major applications is going to be security.
2)This can be added to face unlock, face with only certain emotion will unlock the mobile.
3)The other application is to give a driving alert.
4)While feeling drowsy and driving a vehicle the emotion like sleepy can be recognized and intimidated to the user which ensures safety.
5)With perfect judgment or classification of emotions, we could get instant and genuine feedback from clients which a big factor in marketing products
VIII. CONCLUSION
Therefore, one the best thing about this project is its simplicity. It can be implemented by using python. One of the major applications is going to be security. This can be added to face unlock, face with only certain emotion will unlock the mobile. In other words, with simple emotion detection module there are a lot of advantages. We can get instant feedback which will be helpful
for marketing.
|
|
Published By: |
Retrieval Number: E6284018520/2020©BEIESP |
1722 |
Blue Eyes Intelligence Engineering |
DOI:10.35940/ijrte.E6284.018520 |
& Sciences Publication |
Facial Expression Detection using Artificial Intelligence
This project will lay a great foundation for greater security and safety
REFERENCES
1.A. H. Shaikh, “Human Face Detection and Facial Expression Identification”, Second International Conference on Computing
Methodologies and Communication on IEEE, 2018.
2.Anagha S. Dhavalikar, “Face Detection and Facial Expression Recognition System”, International Conference on Electronics and
Communication System on IEEE, 2014.
3.Hari Prasad Mal, “Facial Expression Detection using Facial Expression Model”, International Conference on Energy, Communication, Data
Analytics and Soft Computing on IEEE, 2017.
4.I. Michael Revina, “Recognition of Facial Expressions using Gaussian based Edge Direction and Texture Descriptor”, International
Conference on Inventive Computing and Informatics on IEEE, 2017.
5.Kenz Ahmed Bozed, “Detection of Facial Expressions based on Morphological Face Features and Minimum Distance Classifier”, 14th international conference on Sciences and Techniques of Automatic control & computer engineering on IEEE, 2013.
6.Khadija Lekdioui, “Facial Expression Recognition Using
Technologies for Signal and Image Processing on IEEE, 2017.
7.Marryam Murtaza, “Facial expression detection using Six Facial Expressions Hexagon (SFEH) model”, IEEE International conference on IEEE, 2019.
8.N. Zainudin, “Facial Expression Change Detection Algorithm Using Optical Flow Technique”, 10th International Conference for Internet
Technology and Secured Transactions on IEEE, 2015.
9.Octavio Arriaga,
Emotion and Gender Classification”, IEEE International conference on IEEE, 2017.
10.
11.Tomas Matlovic, “Emotions Detection Using Facial Expressions Recognition and EEG”, IEEE International conference on IEEE, 2016.
AUTHORS PROFILE
D. Ram Kiran, Department of ECE, Koneru
Lakshmaiah Education Foundation, Guntur, India,
Email: ramkiran55.devireddy@gmail.com
K. Vinay Kumar, Department of ECE, Koneru Lakshmaiah Education Foundation, Guntur, Email: vinaykumarkarusala@gmail.com
T. Kalyan, Department of ECE, Koneru Lakshmaiah
Education Foundation, Guntur, Email: kalyan6166@gmail.com
Dr. K. Ch. Kavya, Department of ECE, Koneru Lakshmaiah Education Foundation, Guntur, Email: kavya@kluniversity.in
.
Dr. K. SARAT KUMAR, Department of ECE, Koneru Lakshmaiah Education Foundation, Guntur, Email: kksarat@kluniversity.in
Retrieval Number: E6284018520/2020©BEIESP |
|
Published By: |
|
|
Blue Eyes Intelligence Engineering |
||
DOI:10.35940/ijrte.E6284.018520 |
1723 |
||
& Sciences Publication |
|||
|
|