Model-Based Design and Evaluation of Interactive Applications
Model-Based Design and Evaluation of Interactive Applications
Validating interactive system design through the verification of formal task and system models
Proceedings of the IFIP TC2/WG2.7 Working Conference on Engineering for Human-Computer Interaction
A Petri Net based Environment for the Design of Event-driven Interfaces
Proceedings of the 16th International Conference on Application and Theory of Petri Nets
Lexical and pragmatic considerations of input structures
ACM SIGGRAPH Computer Graphics
Reality-based interaction: a framework for post-WIMP interfaces
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
ACM Transactions on Computer-Human Interaction (TOCHI)
Engineering patterns for multi-touch interfaces
Proceedings of the 2nd ACM SIGCHI symposium on Engineering interactive computing systems
Towards a formalization of multi-touch gestures
ACM International Conference on Interactive Tabletops and Surfaces
A novel taxonomy for gestural interaction techniques based on accelerometers
Proceedings of the 16th international conference on Intelligent user interfaces
Proton: multitouch gestures as regular expressions
Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
GestIT: a declarative and compositional framework for multiplatform gesture definition
Proceedings of the 5th ACM SIGCHI symposium on Engineering interactive computing systems
Hi-index | 0.00 |
The description of a gesture requires temporal analysis of values generated by input sensors and does not fit well the observer pattern traditionally used by frameworks to handle user input. The current solution is to embed particular gesture-based interactions, such as pinch-to-zoom, into frameworks by notifying when a whole gesture is detected. This approach suffers from a lack of flexibility unless the programmer performs explicit temporal analysis of raw sensors data. This paper proposes a compositional, declarative meta-model for gestures definition based on Petri Nets. Basic traits are used as building blocks for defining gestures; each one notifies the change of a feature value. A complex gesture is defined by the composition of other sub-gestures using a set of operators. The user interface behaviour can be associated to the recognition of the whole gesture or to any other sub-component, addressing the problem of granularity for the notification events. The meta-model can be instantiated for different gesture recognition supports and its definition has been validated through a proof of concept library. Sample applications have been developed for supporting multitouch gestures on iOS and full body gestures with Microsoft Kinect.