Graphical definitions: expanding spreadsheet languages through direct manipulation and gestures
ACM Transactions on Computer-Human Interaction (TOCHI)
WYSIWYT testing in the spreadsheet paradigm: an empirical evaluation
Proceedings of the 22nd international conference on Software engineering
Programming by voice, VocalProgramming
Assets '00 Proceedings of the fourth international ACM conference on Assistive technologies
Exception Handling in the Spreadsheet Paradigm
IEEE Transactions on Software Engineering - special section on current trends in exception handling—part II
A methodology for testing spreadsheets
ACM Transactions on Software Engineering and Methodology (TOSEM)
Testing Homogeneous Spreadsheet Grids with the "What You See Is What You Test" Methodology
IEEE Transactions on Software Engineering
Automated test case generation for spreadsheets
Proceedings of the 24th International Conference on Software Engineering
A scalable method for deductive generalization in the spreadsheet paradigm
ACM Transactions on Computer-Human Interaction (TOCHI)
ICSM '01 Proceedings of the IEEE International Conference on Software Maintenance (ICSM'01)
Forms/3: A first-order visual language to explore the boundaries of the spreadsheet paradigm
Journal of Functional Programming
Integrating automated test generation into the WYSIWYT spreadsheet testing methodology
ACM Transactions on Software Engineering and Methodology (TOSEM)
BCS-HCI '08 Proceedings of the 22nd British HCI Group Annual Conference on People and Computers: Culture, Creativity, Interaction - Volume 2
Adding speech recognition support to UML tools
Journal of Visual Languages and Computing
Hi-index | 0.00 |
Visual programming languages have facilitated the application development process, improving our ability to express programs, as well as our ability to view, edit, and interact with them. Yet even in visual programming environments, productivity is often restricted by the primary input sources: the mouse and the keyboard. As an alternative, we investigate a program development interface which responds to the most natural human communication technologies: voice, handwriting, and gesture. Speech- and pen-based systems have yet to find broad acceptance in everyday life because they are insufficiently advantageous to overcome problems with reliability. However, we believe that a visual programming environment with a multimodal user interface properly constrained so as not to exceed the limits of the current technology has the potential to increase programming productivity for not only those people who are manually or visually impaired, but for the general population as well. In this paper we report on such a system.