In the last decade, there has been a growing interest in emotion analysis research, which has been applied in several areas of computer science. Many authors have contributed to the development of emotion recognition algorithms, considering textual or non verbal data as input, such as facial expressions, gestures or, in the case of multi-modal emotion recognition, a combination of them. In this paper, we describe a method to detect emotions from gestures using the skeletal data obtained from Kinect-like devices as input, as well as atextual description of their meaning. The experimental results show that the correlation existing between body movements and spoken user sentence(s) can be used to reveal user’s emotions from gestures.
|Numero di pagine||4|
|Stato di pubblicazione||Published - 2017|
All Science Journal Classification (ASJC) codes
- Computer Science Applications
- Information Systems
- Computer Networks and Communications