Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns
IEEE Transactions on Pattern Analysis and Machine Intelligence
Smart Sight: A Tourist Assistant System
ISWC '99 Proceedings of the 3rd IEEE International Symposium on Wearable Computers
SNAP&TELL: A Multi-Modal Wearable Computer Interface for Browsing the Environment
ISWC '02 Proceedings of the 6th IEEE International Symposium on Wearable Computers
A Miniature Stereo Vision Machine (MSVM-III) for Dense Disparity Mapping
ICPR '04 Proceedings of the Pattern Recognition, 17th International Conference on (ICPR'04) Volume 1 - Volume 01
Multimodal interaction in an augmented reality scenario
Proceedings of the 6th international conference on Multimodal interfaces
Hand-Gesture Based Text Input for Wearable Computers
ICVS '06 Proceedings of the Fourth IEEE International Conference on Computer Vision Systems
Neural Networks - 2005 Special issue: IJCNN 2005
Multimodal Interaction with a Wearable Augmented Reality System
IEEE Computer Graphics and Applications
VoiceLabel: using speech to label mobile sensor data
ICMI '08 Proceedings of the 10th international conference on Multimodal interfaces
SURF: speeded up robust features
ECCV'06 Proceedings of the 9th European conference on Computer Vision - Volume Part I
Hi-index | 0.00 |
Under wearable environments, it is not convenient to label an object with portable keyboards and mice. This paper presents a multimodal labeling interface to solve this problem with natural and efficient operations. Visual and audio modalities cooperate with each other: an object is encircled by visual tracking of a pointing gesture, and meanwhile its name is obtained by speech recognition. In this paper, we propose a concept of virtual touchpad based on stereo vision techniques. With the touchpad, the object encircling task is achieved by drawing a closed curve on a transparent blackboard. The touch events and movements of a pointing gesture are robustly detected for natural gesture interactions. The experimental results demonstrate the efficiency and usability of our multimodal interface.