Correlation-based Feature Selection for Discrete and Numeric Class Machine Learning
ICML '00 Proceedings of the Seventeenth International Conference on Machine Learning
Foundations of human computing: facial expression and emotion
Proceedings of the 8th international conference on Multimodal interfaces
When Human Coders (and Machines) Disagree on the Meaning of Facial Affect in Spontaneous Videos
IVA '09 Proceedings of the 9th International Conference on Intelligent Virtual Agents
Affect Detection: An Interdisciplinary Review of Models, Methods, and Their Applications
IEEE Transactions on Affective Computing
IEEE Transactions on Affective Computing
Affect detection from multichannel physiology during learning sessions with AutoTutor
AIED'11 Proceedings of the 15th international conference on Artificial intelligence in education
New Perspectives on Affect and Learning Technologies
New Perspectives on Affect and Learning Technologies
A dynamic approach for detecting naturalistic affective states from facial videos during HCI
AI'12 Proceedings of the 25th Australasian joint conference on Advances in Artificial Intelligence
Hi-index | 0.00 |
Significant progress has been made in automatic facial expression recognition, yet most state of the art approaches produce significantly better reliabilities on acted expressions than on natural ones. User interfaces that use facial expressions to understand user's affective states need to be most accurate during naturalistic interactions. This paper presents a study where head movement features are used to recognize naturalistic expressions of affect. The International Affective Picture System (IAPS) collection was used as stimulus for triggering different affective states. Machine learning techniques are applied to classify user's expressions based on their head position and skin color changes. The proposed approach shows a reasonable accuracy in detecting three levels of valence and arousal for user-dependent model during naturalistic human-computer interaction.