Combining Sensory Information to Improve Visualization Organizer
暂无分享,去创建一个
Seemingly effortlessly the human brain reconstructs the threedimensional environment surrounding us from the light pattern striking the eyes. This seems to be true across almost all viewing and lighting conditions. One important factor for this apparent easiness is the redundancy of information provided by the sensory organs. For example, perspective distortions, shading, motion parallax, or the disparity between the two eyes’ images are all, at least partly, redundant signals which provide us with information about the three-dimensional layout of the visual scene. Our brain uses all these different sensory signals and combines the available information into a coherent percept. In displays visualizing data, however, the information is often highly reduced and abstracted, which may lead to an altered perception and therefore a misinterpretation of the visualized data. In this panel we will discuss mechanisms involved in the combination of sensory information and their implications for simulations using computer displays, as well as problems resulting from current display technology such as cathode-ray tubes.
[1] Michael F. Cohen,et al. Radiosity and realistic image synthesis , 1993 .
[2] Heinrich H. Bülthoff,et al. Touch can change visual slant perception , 2000, Nature Neuroscience.
[3] M. Ernst,et al. Humans integrate visual and haptic information in a statistically optimal fashion , 2002, Nature.