Withdraw
Loading…
How deep learning can help emotion recognition
Khorrami, Pooya Rezvani
Loading…
Permalink
https://hdl.handle.net/2142/97284
Description
- Title
- How deep learning can help emotion recognition
- Author(s)
- Khorrami, Pooya Rezvani
- Issue Date
- 2017-03-29
- Director of Research (if dissertation) or Advisor (if thesis)
- Huang, Thomas S.
- Doctoral Committee Chair(s)
- Huang, Thomas S.
- Committee Member(s)
- Hasegawa-Johnson, Mark
- Hoiem, Derek W.
- Liang, Zhi-Pei
- Department of Study
- Electrical & Computer Eng
- Discipline
- Electrical & Computer Engr
- Degree Granting Institution
- University of Illinois at Urbana-Champaign
- Degree Name
- Ph.D.
- Degree Level
- Dissertation
- Keyword(s)
- Emotion recognition
- Deep learning
- Machine learning
- Computer vision
- Facial expression recognition
- Affective computing
- Deep neural networks
- Abstract
- As technological systems become more and more advanced, the need for including the human during the interaction process has become more apparent. One simple way is to have the computer system understand and respond to the human's emotions. Previous works in emotion recognition have focused on improving performance by incorporating domain knowledge into the underlying system either through pre-specified rules or hand-crafted features. However, in the last few years, learned feature representations have experienced a resurgence mainly due to the success of deep neural networks. In this dissertation, we highlight how deep neural networks, when applied to emotion recognition, can learn representations that not only achieve superior accuracy to hand-crafted techniques, but also align with previous domain knowledge. Moreover, we show how these learned representations can generalize to different definitions of emotions and to different input modalities. The first part of this dissertation considers the task of categorical emotion recognition on images. We show how a convolutional neural network (CNN) that achieves state-of-the-art performance can also learn features that strongly correspond to Facial Action Units (FAUs). In the second part, we focus our attention on emotion recognition in video. We take the image-based CNN model and combine it with a recurrent neural network (RNN) in order to do dimensional emotion recognition. We also visualize the portions of the faces that most strongly affect the output prediction by using the gradient as a saliency map. Lastly, we explore the merit of doing multimodal emotion recognition by combining our model with other models trained on audio and physiological data.
- Graduation Semester
- 2017-05
- Type of Resource
- text
- Permalink
- http://hdl.handle.net/2142/97284
- Copyright and License Information
- Copyright 2017 Pooya Rezvani Khorrami
Owning Collections
Graduate Dissertations and Theses at Illinois PRIMARY
Graduate Theses and Dissertations at IllinoisDissertations and Theses - Electrical and Computer Engineering
Dissertations and Theses in Electrical and Computer EngineeringManage Files
Loading…
Edit Collection Membership
Loading…
Edit Metadata
Loading…
Edit Properties
Loading…
Embargoes
Loading…