Toward compound navigation tasks on mobiles via spatial manipulation

M Pahud, K Hinckley, S Iqbal, A Sellen… - Proceedings of the 15th …, 2013 - dl.acm.org
Proceedings of the 15th international conference on Human-computer …, 2013dl.acm.org
We contrast the Chameleon Lens, which uses 3D movement of a mobile device held in the
nonpreferred hand to support panning and zooming, with the Pinch-Flick-Drag metaphor of
directly manipulating the view using multi-touch gestures. Lens-like approaches have
significant potential because they can support navigation-selection, navigation-annotation,
and other such compound tasks by off-loading navigation to the nonpreferred hand while the
preferred hand annotates, marks a location, or draws a path on the screen. Our experimental …
We contrast the Chameleon Lens, which uses 3D movement of a mobile device held in the nonpreferred hand to support panning and zooming, with the Pinch-Flick-Drag metaphor of directly manipulating the view using multi-touch gestures. Lens-like approaches have significant potential because they can support navigation-selection, navigation-annotation, and other such compound tasks by off-loading navigation to the nonpreferred hand while the preferred hand annotates, marks a location, or draws a path on the screen. Our experimental results show that the Chameleon Lens is significantly slower than Pinch-Flick-Drag for the navigation subtask in isolation. But our studies also reveal that for navigation between a few known targets the lens performs significantly faster, that differences between the Chameleon Lens and Pinch-Flick-Drag rapidly diminish as users gain experience, and that in the context of a compound navigation-annotation task, the lens performs as well as Pinch-Flick-Drag despite its deficit for the navigation subtask itself.
ACM Digital Library