Touching floating objects in projection-based virtual reality environments

  • Authors:
  • D. Valkov;F. Steinicke;G. Bruder;K. Hinrichs;J. Schöning;F. Daiber;A. Krüger

  • Affiliations:
  • Department of Computer Science, WWU Münster, Germany;Department of Computer Science, WWU Münster, Germany;Department of Computer Science, WWU Münster, Germany;Department of Computer Science, WWU Münster, Germany;German Research Centre for Artificial Intelligence (DFKI), Saarbrücken, Germany;German Research Centre for Artificial Intelligence (DFKI), Saarbrücken, Germany;German Research Centre for Artificial Intelligence (DFKI), Saarbrücken, Germany

  • Venue:
  • EGVE - JVRC'10 Proceedings of the 16th Eurographics conference on Virtual Environments & Second Joint Virtual Reality
  • Year:
  • 2010

Quantified Score

Hi-index 0.00

Visualization

Abstract

Touch-sensitive screens enable natural interaction without any instrumentation and support tangible feedback on the touch surface. In particular multi-touch interaction has proven its usability for 2D tasks, but the challenges to exploit these technologies in virtual reality (VR) setups have rarely been studied. In this paper we address the challenge to allow users to interact with stereoscopically displayed virtual environments when the input is constrained to a 2D touch surface. During interaction with a large-scale touch display a user changes between three different states: (1) beyond the arm-reach distance from the surface, (2) at arm-reach distance and (3) interaction. We have analyzed the user's ability to discriminate stereoscopic display parallaxes while she moves through these states, i. e., if objects can be imperceptibly shifted onto the interactive surface and become accessible for natural touch interaction. Our results show that the detection thresholds for such manipulations are related to both user motion and stereoscopic parallax, and that users have problems to discriminate whether they touched an object or not, when tangible feedback is expected.