Evaluating usability of human-computer interfaces: a practical method
Evaluating usability of human-computer interfaces: a practical method
Measuring performance of the information systems function
Journal of Management Information Systems
The role of voice in human-machine communication
Voice communication between humans and machines
Human-computer interaction
Tutorial 2: Designing and evaluating intelligent user interfaces
IUI '98 Proceedings of the 3rd international conference on Intelligent user interfaces
Evaluation of eye gaze interaction
Proceedings of the SIGCHI conference on Human Factors in Computing Systems
Improving Level of Service for Mobile Users Using Context-Awareness
SRDS '99 Proceedings of the 18th IEEE Symposium on Reliable Distributed Systems
A framework for rapid development of multimodal interfaces
Proceedings of the 5th international conference on Multimodal interfaces
Tangible multimodal interfaces for safety-critical applications
Communications of the ACM - Multimodal interfaces that flex, adapt, and persist
The state of the art in ontology learning: a framework for comparison
The Knowledge Engineering Review
Real-Time Gesture Recognition by Learning and Selective Control of Visual Interest Points
IEEE Transactions on Pattern Analysis and Machine Intelligence
Conversing with the user based on eye-gaze patterns
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
Face and body gesture recognition for a vision-based multimodal analyzer
VIP '05 Proceedings of the Pan-Sydney area workshop on Visual information processing
Using Context-Awareness to Support User Interaction withWeb Services
AICT-ICIW '06 Proceedings of the Advanced Int'l Conference on Telecommunications and Int'l Conference on Internet and Web Applications and Services
Transmodal feedback as a new perspective for audio-visual effects
NIME '06 Proceedings of the 2006 conference on New interfaces for musical expression
Multimodal human-computer interaction: A survey
Computer Vision and Image Understanding
A User context awareness Model for mobile environment processing
SERA '07 Proceedings of the 5th ACIS International Conference on Software Engineering Research, Management & Applications
Semantic Layers for Illustrative Volume Rendering
IEEE Transactions on Visualization and Computer Graphics
Ontology-based models in pervasive computing systems
The Knowledge Engineering Review
User Profile Selection by Means of Ontology Reasoning
AICT '08 Proceedings of the 2008 Fourth Advanced International Conference on Telecommunications
An ontology engineering methodology for DOGMA
Applied Ontology - Ontological Foundations of Conceptual Modelling
Ambient Intelligence: A Multimedia Perspective
IEEE MultiMedia
Context-Aware Generation of User Interface Containers for Mobile Devices
ENC '08 Proceedings of the 2008 Mexican International Conference on Computer Science
Improved third-person perspective: a solution reducing occlusion of the 3PP?
VRCAI '08 Proceedings of The 7th ACM SIGGRAPH International Conference on Virtual-Reality Continuum and Its Applications in Industry
A User-centered Approach to Context-awareness in Mobile Computing
MOBIQUITOUS '07 Proceedings of the 2007 Fourth Annual International Conference on Mobile and Ubiquitous Systems: Networking&Services (MobiQuitous)
Personalized Approach for Mobile Search
CSIE '09 Proceedings of the 2009 WRI World Congress on Computer Science and Information Engineering - Volume 05
Visualization learning for visually impaired people
Edutainment'07 Proceedings of the 2nd international conference on Technologies for e-learning and digital entertainment
Advanced mixed reality technologies for surveillance and risk prevention applications
ISCIS'06 Proceedings of the 21st international conference on Computer and Information Sciences
Hi-index | 0.00 |
Nowadays, several techniques exist to render digital content such as graphics, audio, haptic, etc. Unfortunately, they require different faculties that cannot always be applied, e.g. providing a picture to a blind person would be useless. In this paper, we present a new multimodal rendering engine with a server web-connected to other devices to perform ubiquitous computing. In order to take advantage of user capabilities, we defined an ontology populated with the following elements: user, device, and information. Our system, with the help of this ontology, aims to select and launch automatically a rendering application. Several test case applications were implemented to render shape, text, and video information via audio, haptic, and sight channels. Validations demonstrate that our system is flexible, easily extensible, and shows promise.