In this project, my colleagues and I investigated 3D pan & zoom using mobile devices in augmented reality information spaces. To this end, we designed a series of interaction techniques that combine touch and spatial interaction and compared them in a user study.
The official video at YouTube.
These pictures show the setup of our study prototype: The AR HMD and phone are tracked with an IR tracking system, and an abstract zoomable data space is located in front of the user.
Overview of the different pan & zoom techniques from our study.
For more details about this project, please take a look at the main project website! Additionally, if you would like to cite this work in your research, please cite our MobileHCI ‘19 paper.
In Proceedings of the 21st International Conference on Human-Computer Interaction with Mobile Devices and Services, Oct 2019