Authors
Iulia Lefter, Leon JM Rothkrantz, Gertjan Burghouts, Zhenke Yang, Pascal Wiggers
Publication date
2011
Conference
Text, Speech and Dialogue: 14th International Conference, TSD 2011, Pilsen, Czech Republic, September 1-5, 2011. Proceedings 14
Pages
25-32
Publisher
Springer Berlin Heidelberg
Description
Automatic detection of aggressive situations has a high societal and scientific relevance. It has been argued that using data from multimodal sensors as for example video and sound as opposed to unimodal is bound to increase the accuracy of detections. We approach the problem of multimodal aggression detection from the viewpoint of a human observer and try to reproduce his predictions automatically. Typically, a single ground truth for all available modalities is used when training recognizers. We explore the benefits of adding an extra level of annotations, namely audio-only and video-only. We analyze these annotations and compare them to the multimodal case in order to have more insight into how humans reason using multimodal data. We train classifiers and compare the results when using unimodal and multimodal labels as ground truth. Both in the case of audio and video recognizer the …
Total citations
201220132014201520162017201820192020202120222023113211221
Scholar articles
I Lefter, LJM Rothkrantz, G Burghouts, Z Yang… - Text, Speech and Dialogue: 14th International …, 2011