Smile and Laughter Recognition using Speech Processing and Face Recognition from Conversation Video
CW '05 Proceedings of the 2005 International Conference on Cyberworlds
Automatic discrimination between laughter and speech
Speech Communication
Visual learning of texture descriptors for facial expression recognition in thermal imagery
Computer Vision and Image Understanding
A comparison of statistical significance tests for information retrieval evaluation
Proceedings of the sixteenth ACM conference on Conference on information and knowledge management
How to distinguish posed from spontaneous smiles using geometric features
Proceedings of the 9th international conference on Multimodal interfaces
PIT '08 Proceedings of the 4th IEEE tutorial and research workshop on Perception and Interactive Technologies for Speech-Based Systems: Perception in Multimodal Dialogue Systems
Decision-Level Fusion for Audio-Visual Laughter Detection
MLMI '08 Proceedings of the 5th international workshop on Machine Learning for Multimodal Interaction
Detection of Laughter-in-Interaction in Multichannel Close-Talk Microphone Recordings of Meetings
MLMI '08 Proceedings of the 5th international workshop on Machine Learning for Multimodal Interaction
Honest Signals: How They Shape Our World
Honest Signals: How They Shape Our World
Audiovisual laughter detection based on temporal features
ICMI '08 Proceedings of the 10th international conference on Multimodal interfaces
Contrasting emotion-bearing laughter types in multiparticipant vocal activity detection for meetings
ICASSP '09 Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing
Image and Vision Computing
Automatic nonverbal analysis of social interaction in small groups: A review
Image and Vision Computing
Toward Practical Smile Detection
IEEE Transactions on Pattern Analysis and Machine Intelligence
Is this joke really funny? judging the mirth by audiovisual laughter analysis
ICME'09 Proceedings of the 2009 IEEE international conference on Multimedia and Expo
Tools and resources for visualising conversational-speech interaction
Multimodal corpora
Human computing and machine understanding of human behavior: a survey
ICMI'06/IJCAI'07 Proceedings of the ICMI 2006 and IJCAI 2007 international conference on Artifical intelligence for human computing
Particle filtering with factorized likelihoods for tracking facial features
FGR' 04 Proceedings of the Sixth IEEE international conference on Automatic face and gesture recognition
The AMI meeting corpus: a pre-announcement
MLMI'05 Proceedings of the Second international conference on Machine Learning for Multimodal Interaction
Toward Pose-Invariant 2-D Face Recognition Through Point Distribution Models and Facial Symmetry
IEEE Transactions on Information Forensics and Security - Part 1
Audio-visual speech modeling for continuous speech recognition
IEEE Transactions on Multimedia
Audio–Visual Affective Expression Recognition Through Multistream Fused HMM
IEEE Transactions on Multimedia
IEEE Transactions on Multimedia
Audiovisual Discrimination Between Speech and Laughter: Why and When Visual Information Might Help
IEEE Transactions on Multimedia
Editorial: Introduction To The Special Issue On Affect Analysis In Continuous Input
Image and Vision Computing
Bimodal log-linear regression for fusion of audio and visual features
Proceedings of the 21st ACM international conference on Multimedia
Hi-index | 0.00 |
Laughter is clearly an audiovisual event, consisting of the laughter vocalization and of facial activity, mainly around the mouth and sometimes in the upper face. A major obstacle in studying the audiovisual aspects of laughter is the lack of suitable data. For this reason, the majority of past research on laughter classification/detection has focused on audio-only approaches. A few audiovisual studies exist which use audiovisual data from existing corpora of recorded meetings. The main problem with such data is that they usually contain large head movements which make audiovisual analysis very difficult. In this work, we present a new publicly available audiovisual database, the MAHNOB Laughter database, suitable for studying laughter. It contains 22 subjects who were recorded while watching stimulus material, using two microphones, a video camera and a thermal camera. The primary goal was to elicit laughter, but in addition, posed smiles, posed laughter, and speech were recorded as well. In total, 180 sessions are available with a total duration of 3h and 49min. There are 563 laughter episodes, 849 speech utterances, 51 posed laughs, 67 speech-laughs episodes and 167 other vocalizations annotated in the database. We also report baseline experiments for audio, visual and audiovisual approaches for laughter-vs-speech discrimination as well as further experiments on discrimination between voiced laughter, unvoiced laughter and speech. These results suggest that the combination of audio and visual information is beneficial in the presence of acoustic noise and helps discriminating between voiced laughter episodes and speech utterances. Finally, we report preliminary experiments on laughter-vs-speech discrimination based on thermal images.