Sourcebook of automatic identification and data collection
Sourcebook of automatic identification and data collection
Speech and Language Processing (2nd Edition)
Speech and Language Processing (2nd Edition)
Multimodal human-computer interaction: A survey
Computer Vision and Image Understanding
Wave like an Egyptian: accelerometer based gesture recognition for culture specific interactions
BCS-HCI '08 Proceedings of the 22nd British HCI Group Annual Conference on People and Computers: Culture, Creativity, Interaction - Volume 1
Intuitive E-Teaching by Using Combined HCI Devices: Experiences with Wiimote Applications
UAHCI '09 Proceedings of the 5th International Conference on Universal Access in Human-Computer Interaction. Part III: Applications and Services
Vietnamese automatic speech recognition: the FLaVoR approach
ISCSLP'06 Proceedings of the 5th international conference on Chinese Spoken Language Processing
What do you want to do next: a novel approach for intent prediction in gaze-based interaction
Proceedings of the Symposium on Eye Tracking Research and Applications
ZeroTouch: an optical multi-touch and free-air interaction architecture
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Microsoft Kinect Sensor and Its Effect
IEEE MultiMedia
A Multi-User Interaction System Based on Kinect and Wii Remote
ICMEW '12 Proceedings of the 2012 IEEE International Conference on Multimedia and Expo Workshops
Hi-index | 0.00 |
The authors propose a system that allows presenters to control presentations in a natural way by their body gestures and vocal commands. Thus a presentation no longer follows strictly a rigid sequential structure but can be delivered in various flexible and content adapted scenarios. Our proposed system fuses three interaction modules: gesture recognition with Kinect 3D skeletal data, key concepts detection by context analysis from natural speech, and small-scaled hand gesture recognition with haptic data from smart phone sensors. Each module can process in realtime with the accuracy of 95.0%, 91.2%, and 90.1% respectively. The system uses events generated from the three modules to trigger pre-defined scenarios in a presentation to enhance the exciting experience for audiences.