Abstract

Laughter detection is an essential aspect towards effective human-computer interaction. This work primarily addresses the problem of laughter detection in a real-time environment. We utilize annotated audio and visual data collected from a Kinect sensor to identify discriminative features for audio and video, separately. We show how the features can be used with classifiers such as support vector machines (SVM). The two modalities are then fused into a single output to form a decision. We test our setup by emulating real-time data with Kinect sensor, and compare the results with the offline version of the setup. Our results indicate that our laughter detection system gives a promising performance for a real-time human-computer interactions.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call