AR Interaction

Panning & Zooming in 3D Dataspaces

In this project, my colleagues and I investigated 3D pan & zoom using mobile devices in augmented reality information spaces. To this end, we designed a series of interaction techniques that combine touch and spatial interaction and compared them in a user study.

The official video at YouTube.
These pictures show the setup of our study prototype: The AR HMD and phone are tracked with an IR tracking system, and an abstract zoomable data space is located in front of the user.
Overview of the different pan & zoom techniques from our study.

For more details about this project, please take a look at the main project website! Additionally, if you would like to cite this work in your research, please cite our MobileHCI ‘19 paper.



  1. Pan+Zoom2019.jpg
    Wolfgang Büschel, Annett Mitschick, Thomas Meyer, and 1 more author
    In Proceedings of the 21st International Conference on Human-Computer Interaction with Mobile Devices and Services, Oct 2019