Combining sensory information to improve visualization

  • Authors:
  • Marc Ernst;Martin Banks;Felix Wichmann;Laurence Maloney;Heinrich Bülthoff

  • Affiliations:
  • Max-Planck für biol. Kybernetik;University of California, Berkeley;Max-Planck für biol. Kybernetik;New York University;Max-Planck für biol. Kybernetik

  • Venue:
  • Proceedings of the conference on Visualization '02
  • Year:
  • 2002

Quantified Score

Hi-index 0.00

Visualization

Abstract

Seemingly effortlessly the human brain reconstructs the three-dimensional environment surrounding us from the light pattern striking the eyes. This seems to be true across almost all viewing and lighting conditions. One important factor for this apparent easiness is the redundancy of information provided by the sensory organs. For example, perspective distortions, shading, motion parallax, or the disparity between the two eyes' images are all, at least partly, redundant signals which provide us with information about the three-dimensional layout of the visual scene. Our brain uses all these different sensory signals and combines the available information into a coherent percept. In displays visualizing data, however, the information is often highly reduced and abstracted, which may lead to an altered perception and therefore a misinterpretation of the visualized data. In this panel we will discuss mechanisms involved in the combination of sensory information and their implications for simulations using computer displays, as well as problems resulting from current display technology such as cathode-ray tubes.