4.5 Article

Generalized zero-shot emotion recognition from body gestures

Journal

APPLIED INTELLIGENCE
Volume 52, Issue 8, Pages 8616-8634

Publisher

SPRINGER
DOI: 10.1007/s10489-021-02927-w

Keywords

Generalized zero-shot learning; Emotion recognition; Body gesture recognition; Prototype learning

Funding

  1. National Key Research and Development Project of China [2019YFB1310601]
  2. National Key R&D Program of China [2017YFC0820203]
  3. National Natural Science Foundation of China [62103410]

Ask authors/readers for more resources

In this study, a novel emotion recognition framework is designed using Generalized Zero-Shot Learning (GZSL) to recognize both seen and unseen body gesture categories and provide emotion predictions based on the relationship between gestures and emotions. The framework consists of two branches, one for predicting samples of seen gesture categories and the other for predicting samples of unseen gesture categories, with emotion labels ultimately obtained from these results.
In human-human interaction, body language is one of the most important emotional expressions. However, each emotion category contains abundant emotional body gestures, and basic emotions used in most researches are difficult to describe complex and diverse emotional states. It is costly to collect sufficient samples of all emotional expressions, and new emotions or new body gestures that are not included in the training set may appear during testing. To address the above problems, we design a novel mechanism that treats each emotion category as a collection of multiple body gesture categories to make better use of gesture information for emotion recognition. A Generalized Zero-Shot Learning (GZSL) framework is introduced to recognize both seen and unseen body gesture categories with the help of semantic information, and emotion predictions are further provided based on the relationship between gestures and emotions. This framework consists of two branches. The first branch is a Hierarchical Prototype Network (HPN) which learns the prototypes of body gestures and uses them to calculate the emotion attentive prototypes. This branch aims to obtain predictions on samples of the seen gesture categories. The second branch is a Semantic Auto-Encoder (SAE) which utilizes semantic representations to predict samples of unseen gesture categories. Thresholds are further trained to determine which branch result will be used during testing, and the emotion labels are finally obtained from these results. Comprehensive experiments are conducted on an emotion recognition dataset which contains skeleton data of multiple body gestures, and the performance of our framework is superior to both the traditional emotion classifier and state-of-the-art zero-shot learning methods.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available