학술논문

Automated Laughter Detection From Full-Body Movements
Document Type
Periodical
Source
IEEE Transactions on Human-Machine Systems IEEE Trans. Human-Mach. Syst. Human-Machine Systems, IEEE Transactions on. 46(1):113-123 Feb, 2016
Subject
Communication, Networking and Broadcast Technologies
Components, Circuits, Devices and Systems
Signal Processing and Analysis
Robotics and Control Systems
Power, Energy and Industry Applications
General Topics for Engineers
Computing and Processing
Feature extraction
Joints
Games
Motion segmentation
Context
Face
Tracking
Automated analysis of full-body movement
body expressivity
detection
laughter
motion capture
multimodal interaction
Language
ISSN
2168-2291
2168-2305
Abstract
In this paper, we investigate the detection of laughter from the user's nonverbal full-body movement in social and ecological contexts. Eight hundred and one laughter and nonlaughter segments of full-body movement were examined from a corpus of motion capture data of subjects participating in social activities that stimulated laughter. A set of 13 full-body movement features was identified, and corresponding automated extraction algorithms were developed. These features were extracted from the laughter and nonlaughter segments, and the resulting dataset was provided as input to supervised machine learning techniques. Both discriminative (radial basis function-support vector machines, k-nearest neighbor, and random forest) and probabilistic (naive Bayes and logistic regression) classifiers were trained and evaluated. A comparison of automated classification with the ratings of human observers for the same laughter and nonlaughter segments showed that the performance of our approach for automated laughter detection is comparable with that of humans. The highest F-score (0.74) was obtained by the random forest classifier, whereas the F-score obtained by human observers was 0.70. Based on the analysis techniques introduced in the paper, a vision-based system prototype for automated laughter detection was designed and evaluated. Support vector machines (SVMs) and Kohonen's self-organizing maps were used for training, and the highest F-score was obtained with SVM (0.73).