SAGE Publications, Proceedings of the Human Factors and Ergonomics Society Annual Meeting, 28(54), p. 2427-2431
DOI: 10.1177/154193121005402814
PsycEXTRA Dataset
Full text: Download
The correct spatial registration between virtual and real objects in optical see-through augmented reality implies accurate estimates of the user's eyepoint relative to the location and orientation of the display surface. A common approach is to estimate the display parameters through a calibration procedure involving a subjective alignment exercise. Human postural sway and targeting precision contribute to imprecise alignments, which in turn adversely affect the display parameter estimation resulting in registration errors between virtual and real objects. The technique commonly used has its origin in computer vision, and calibrates stationary cameras using hundreds of correspondence points collected instantaneously in one video frame where precision is limited only by pixel quantization and image blur. Subsequently the input noise level is several order of magnitudes greater when a human operator manually collects correspondence points one by one. This paper investigates the effect of human alignment noise on view parameter estimation in an optical see-through head mounted display to determine how well a standard camera calibration method performs at greater noise levels than documented in computer vision literature. Through Monte-Carlo simulations we show that it is particularly difficult to estimate the user's eyepoint in depth, but that a greater distribution of correspondence points in depth help mitigate the effects of human alignment noise. Copyright 2010 by Human Factors and Ergonomics Society, Inc. All rights reserved.