Speech Emotion Recognition (SER) is a hot research topic in the field of Human Computer Interaction. In this paper a SER system is developed with the aim of providing a classification of the “state of interest” of a human subject involved in a job interview. Classification of emotions is performed by analyzing the speech produced during the interview. The presented methods and results show just preliminary conclusions, as the work is part of a larger project including also analysis, investigation and classification of facial expressions and body gestures during human interaction. At the current state of the work, investigation is carried out by using software tools already available for free on the web; furthermore, the features extracted from the audio tracks are analyzed by studying their sensitivity to an audio compression stage. The Berlin Database of Emotional Speech (EmoDB) is exploited to provide the preliminary results.
First progresses in evaluation of resonance in staff selection through speech emotion recognition / Bevilacqua, Vitoantonio; Guccione, Pietro; Mascolo, Luigi; Pazienza, Pasquale Pio; Antonio Salatino, Angelo; Pantaleo, Michele (LECTURE NOTES IN COMPUTER SCIENCE). - In: Intelligent Computing Theories and Technology : 9th International Conference, ICIC 2013, Nanning, China, July 28-31, 2013. Proceedings / [a cura di] De-Shuang HuangKang-Hyun JoYong-Quan ZhouKyungsook Han. - STAMPA. - Berlin; Heidelberg : Springer, 2013. - ISBN 978-3-642-39481-2. - pp. 658-671 [10.1007/978-3-642-39482-9_76]
First progresses in evaluation of resonance in staff selection through speech emotion recognition
Vitoantonio Bevilacqua;Pietro Guccione;Luigi Mascolo;Pasquale Pio Pazienza;
2013-01-01
Abstract
Speech Emotion Recognition (SER) is a hot research topic in the field of Human Computer Interaction. In this paper a SER system is developed with the aim of providing a classification of the “state of interest” of a human subject involved in a job interview. Classification of emotions is performed by analyzing the speech produced during the interview. The presented methods and results show just preliminary conclusions, as the work is part of a larger project including also analysis, investigation and classification of facial expressions and body gestures during human interaction. At the current state of the work, investigation is carried out by using software tools already available for free on the web; furthermore, the features extracted from the audio tracks are analyzed by studying their sensitivity to an audio compression stage. The Berlin Database of Emotional Speech (EmoDB) is exploited to provide the preliminary results.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.