Neel Joshi, Abhishek Kar, and Michael Cohen
5 May 2012
We present a touch-free interface for viewing large imagery on mobile devices. In particular, we focus on viewing paradigms for 360 degree panoramas, parallax image sequences, and long multi-perspective panoramas. We describe a sensor fusion methodology that combines face tracking using a frontfacing camera with gyroscope data to produce a robust signal that de?nes the viewer's 3D position relative to the display. The gyroscopic data provides both low-latency feedback and allows extrapolation of the face position beyond the the ?eldof-view of the front-facing camera. We also demonstrate a hybrid position and rate control that uses the viewer's 3D position to drive exploration of very large image spaces. We report on the ef?cacy of the hybrid control vs. position only control through a user study.
|Published in||ACM SIGGCHI 2012|
|Publisher||ACM Conference on Human Factors in Computing Systems|