MixedEmotions: An open-source toolbox for multi-modal emotion analysis

View/ Open
Date
2018-01-25Author
Buitelaar, Paul
Wood, Ian D.
Negi, Sapna
Arcan, Mihael
McCrae, John P.
Abele, Andrejs
Robin, Cécile
Andryushechkin, Vladimir
Ziad, Housam
Sagha, Hesam
Schmitt, Maximilian
Schuller, Björn W.
Sánchez-Rada, J. Fernando
Iglesias, Carlos A.
Navarro, Carlos
Giefer, Andreas
Heise, Nicolaus
Masucci, Vincenzo
Danza, Francesco A.
Caterino, Ciro
Smrž, Pavel
Hradiš, Michal
Povolný, Filip
Klimeš, Marek
Matějka, Pavel
Tummarello, Giovanni
Metadata
Show full item recordUsage
This item's downloads: 721 (view details)
Cited 10 times in Scopus (view citations)
Recommended Citation
Buitelaar, P. and Wood, I. D. and Negi, S. and Arcan, M. and McCrae, J. P. and Abele, A. and Robin, C. and Andryushechkin, V. and Ziad, H. and Sagha, H. and Schmitt, M. and Schuller, B. W. and Sanchez, J. F. and Iglesias, C. A. and Navarro, C. and Giefer, A. and Heise, N. and Masucci, V. and Danza, F. A. and Caterino, C. and Smrz, P. and Hradis, M. and Povolny, F. and Klimes, M. and Matejka, P. and Tummarello, G. (2018). MixedEmotions: An Open-Source Toolbox for Multimodal Emotion Analysis. IEEE Transactions on Multimedia, 20(9), 2454-2465. doi: 10.1109/TMM.2018.2798287
Published Version
Abstract
Recently, there is an increasing tendency to embed the functionality of recognizing emotions from the user generated contents, to infer richer profile about the users or contents, that can be used for various automated systems such as call-center operations, recommendations, and assistive technologies. However, to date, adding this functionality was a tedious, costly, and time consuming effort, and one should look for different tools that suits one's needs, and should provide different interfaces to use those tools. The MixedEmotions toolbox leverages the need for such functionalities by providing tools for text, audio, video, and linked data processing within an easily integrable plug-and-play platform. These functionalities include: (i) for text processing: emotion and sentiment recognition, (ii) for audio processing: emotion, age, and gender recognition, (iii) for video processing: face detection and tracking, emotion recognition, facial landmark localization, head pose estimation, face alignment, and body pose estimation, and (iv) for linked data: knowledge graph. Moreover, the MixedEmotions Toolbox is open-source and free. In this article, we present this toolbox in the context of the existing landscape, and provide a range of detailed benchmarks on standardized test-beds showing its state-of-the-art performance. Furthermore, three real-world use-cases show its effectiveness, namely emotion-driven smart TV, call center monitoring, and brand reputation analysis.