The logic of typed feature structures
The logic of typed feature structures
Synthetic experience: a proposed taxonomy
Presence: Teleoperators and Virtual Environments
ISSD-93 Selected papers presented at the international symposium on Spoken dialogue
Artificial Intelligence Review - Special issue on integration of natural language and vision processing: recent advances
Visual tracking for multimodal human computer interaction
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Dynamic registration correction in augmented-reality systems
VRAIS '95 Proceedings of the Virtual Reality Annual International Symposium (VRAIS'95)
Multimodal Interfaces for Multimedia Information Agents
ICASSP '97 Proceedings of the 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP '97) -Volume 1 - Volume 1
An Adaptive Multimodal Interface For Wireless Applications
ISWC '98 Proceedings of the 2nd IEEE International Symposium on Wearable Computers
A framework and toolkit for the construction of multimodal learning interfaces
A framework and toolkit for the construction of multimodal learning interfaces
COLING '98 Proceedings of the 17th international conference on Computational linguistics - Volume 1
A programmable multi-blackboard architecture for dialogue processing systems
ISDS '97 Interactive Spoken Dialog Systems on Bringing Speech and NLP Together in Real Applications
Proceedings of the 2001 conference on Virtual reality, archeology, and cultural heritage
Position-Annotated Photographs: A Geotemporal Web
IEEE Pervasive Computing
The evolution of perspective view in WalkMap
Personal and Ubiquitous Computing
NetGames '03 Proceedings of the 2nd workshop on Network and system support for games
An automatic sign recognition and translation system
Proceedings of the 2001 workshop on Perceptive user interfaces
Robust finger tracking for wearable computer interfacing
Proceedings of the 2001 workshop on Perceptive user interfaces
LingWear: a mobile tourist information system
HLT '01 Proceedings of the first international conference on Human language technology research
Towards automatic sign translation
HLT '01 Proceedings of the first international conference on Human language technology research
TGH: a case study of designing natural interaction for mobile guide systems
Proceedings of the 7th international conference on Human computer interaction with mobile devices & services
Browsing the environment with the SNAP&TELL wearable computer system
Personal and Ubiquitous Computing
CampusGenie --- A Campus-Life Assistant on the Mobile-Phone
PCM '09 Proceedings of the 10th Pacific Rim Conference on Multimedia: Advances in Multimedia Information Processing
Noun phrase generation for situated dialogs
INLG '06 Proceedings of the Fourth International Natural Language Generation Conference
A multimodal labeling interface for wearable computing
Proceedings of the 15th international conference on Intelligent user interfaces
Artificial intelligence
Automatic detection and recognition of Korean text in outdoor signboard images
Pattern Recognition Letters
A trip into the countryside: an experience design for explorative car cruises
CHI '13 Extended Abstracts on Human Factors in Computing Systems
Context management for RFID-based distributed interaction spaces
OCSC'13 Proceedings of the 5th international conference on Online Communities and Social Computing
Distributed user interfaces in public spaces using RFID-based panels
International Journal of Human-Computer Studies
Hi-index | 0.00 |
In this paper, we present our efforts towards developing an intelligent tourist system. The system is equipped with a unique combination of sensors and software. The hardware includes two computers, a GPS receiver, a lapel microphone plus an earphone, a video camera and a head-mounted display. This combination enables a multimodal interface to take advantage of speech and gesture input to provide assistance for a tourist. The software supports natural language processing, speech recognition, machine translation, handwriting recognition and multimodal fusion. A vision module is trained to locate and read written language, is able to adapt to to new environments, and is able to interpret intentions offered by the user, such as a spoken clarification or pointing gesture. We illustrate the applications of the system using two examples.