Improving static and dynamic registration in an optical see-through HMD

  • Authors:
  • Ronald Azuma;Gary Bishop

  • Affiliations:
  • Department of Computer Science, University of North Carolina at Chapel Hill, CB 3175 Sitterson Hall/ UNC/ Chapel Hill, NC;Department of Computer Science, University of North Carolina at Chapel Hill, CB 3175 Sitterson Hall/ UNC/ Chapel Hill, NC

  • Venue:
  • SIGGRAPH '94 Proceedings of the 21st annual conference on Computer graphics and interactive techniques
  • Year:
  • 1994

Quantified Score

Hi-index 0.00

Visualization

Abstract

In Augmented Reality, see-through HMDs superimpose virtual 3D objects on the real world. This technology has the potential to enhance a user's perception and interaction with the real world. However, many Augmented Reality applications will not be accepted until we can accurately register virtual objects with their real counterparts. In previous systems, such registration was achieved only from a limited range of viewpoints, when the user kept his head still. This paper offers improved registration in two areas. First, our system demonstrates accurate static registration across a wide variety of viewing angles and positions. An optoelectronic tracker provides the required range and accuracy. Three calibration steps determine the viewing parameters. Second, dynamic errors that occur when the user moves his head are reduced by predicting future head locations. Inertial sensors mounted on the HMD aid head-motion prediction. Accurate determination of prediction distances requires low-overhead operating systems and eliminating unpredictable sources of latency. On average, prediction with inertial sensors produces errors 2-3 times lower than prediction without inertial sensors and 5-10 times lower than using no prediction at all. Future steps that may further improve registration are outlined.