ABSTRACT
In many complex tasks, a remote subject-matter expert may need to assist a local user to guide actions on objects in the local user's environment. However, effective spatial referencing and action demonstration in a remote physical environment can be challenging. We introduce two approaches that use Virtual Reality (VR) or Augmented Reality (AR) for the remote expert, and AR for the local user, each wearing a stereo head-worn display. Both approaches allow the expert to create and manipulate virtual replicas of physical objects in the local environment to refer to parts of those physical objects and to indicate actions on them. This can be especially useful for parts that are occluded or difficult to access. In one approach, the expert points in 3D to portions of virtual replicas to annotate them. In another approach, the expert demonstrates actions in 3D by manipulating virtual replicas, supported by constraints and annotations. We performed a user study of a 6DOF alignment task, a key operation in many physical task domains, comparing both approaches to an approach in which the expert uses a 2D tablet-based drawing system similar to ones developed for prior work on remote assistance. The study showed the 3D demonstration approach to be faster than the others. In addition, the 3D pointing approach was faster than the 2D tablet in the case of a highly trained expert.
Supplemental Material
- M. Adcock, D. Ranatunga, R. Smith, and B.H. Thomas. 2014. Object-Based Touch Manipulation for Remote Guidance of Physical Tasks. In Proc. ACM SUI. 113--122. Google ScholarDigital Library
- M. Bauer, G. Kortuem, and Z. Segall. 1999. "Where Are You Pointing At?" A Study of Remote Collaboration in a Wearable Video conference System. In Proc. IEEE ISWC. 151--158. Google ScholarDigital Library
- S. Bottecchia, J. Cieutat, and J. Jessel. 2010. T.A.C: Augmented Reality System for Collaborative Tele-assistance in the Field of Maintenance Through Internet. In Proc. ACM AH. 14:1--14:7. Google ScholarDigital Library
- J. Chastine, K. Nagel, Y. Zhu, and M. Hudachek-Buswell. 2008. Studies on the Effectiveness of Virtual Pointers in Collaborative Augmented Reality. In Proc. IEEE 3DUI. 117--124. Google ScholarDigital Library
- S. Gauglitz, B. Nuernberger, M. Turk, and T. Hollerer. 2014. World-Stabilized Annotations And Virtual Scene Navigation For Remote Collaboration. In Proc. ACM UIST. 449--459. Google ScholarDigital Library
- M. Goto, Y. Uematsu, H. Saito, S. Senda, and A Iketani. 2010. Task Support System by Displaying Instructional Video Onto AR Workspace. In Proc. IEEE ISMAR. 83--90.Google ScholarCross Ref
- J. Heiser, B. Tversky, and M. Silverman. 2004. Sketches for and from Collaboration. Visual and spatial reasoning in design III 3 (2004), 69--78.Google Scholar
- S. Henderson and S. Feiner. 2011. Augmented Reality in the Psychomotor Phase of a Procedural Task. In Proc. IEEE ISMAR. 191--200. Google ScholarDigital Library
- S. Kim, G. Lee, N. Sakata, and M. Billinghurst. 2014. Improving Co-Presence with Augmented Visual Communication Cues for Sharing Experience through Video Conference. In Proc. IEEE ISMAR. 83--92.Google Scholar
- D.S. Kirk and D.S. Fraser. 2005. The Effects of Remote Gesturing on Distance Instruction. In Proc. CSCL. 301--310. Google ScholarDigital Library
- T. Kurata, N. Sakata, M. Kourogi, H. Kuzuoka, and M. Billinghurst. 2004. Remote Collaboration Using a Shoulder-Worn Active Camera/laser. In Proc. IEEE ISWC, Vol. 1. 62--69. Google ScholarDigital Library
- G. Kurillo, R. Bajcsy, K. Nahrsted, and O. Kreylos. 2008. Immersive 3D Environment for Remote Collaboration and Training of Physical Activities. In Proc. IEEE VR. 269--270.Google Scholar
- H. Kuzuoka. 1992. Spatial Workspace Collaboration: A SharedView Video Support System for Remote Collaboration Capability. In Proc. ACM CHI. 533--540. Google ScholarDigital Library
- J. Lanir, R. Stone, B. Cohen, and P. Gurevich. 2013. Ownership and Control of Point of View in Remote Assistance. In Proc. ACM CHI. 2243--2252. Google ScholarDigital Library
- R.A. Newcombe, S. Izadi, O. Hilliges, D. Molyneaux, D. Kim, A.J. Davison, P. Kohli, J. Shotton, S. Hodges, and A. Fitzgibbon. 2011. KinectFusion: Real-time Dense Surface Mapping and Tracking. In Proc. IEEE ISMAR. 127--136. Google ScholarDigital Library
- O. Oda and S. Feiner. 2015. Goblin XNA Framework. (2015). http://goblinxna.codeplex.com/Google Scholar
- J.S. Pierce, B.C. Stearns, and R. Pausch. 1999. Voodoo Dolls: Seamless Interaction at Multiple Scales in Virtual Environments. In Proc. ACM i3D. 141--145. Google ScholarDigital Library
- T. Piumsomboon, D. Altimira, H. Kim, A. Clark, G. Lee, and M. Billinghurst. 2014. Grasp-Shell vs Gesture-Speech: A Comparison of Direct and Indirect Natural Interaction Techniques in Augmented Reality. In Proc. IEEE ISMAR. 73--82.Google Scholar
- R Core Team. 2015. R: A Language and Environment for Statistical Computing. http://www.R-project.org/Google Scholar
- G. Rizzolatti and L. Craighero. 2004. The Mirror-Neuron System. Annu. Rev. Neurosci. 27 (2004), 169--192.Google ScholarCross Ref
- N. Sakata, T. Kurata, and H. Kuzuoka. 2006. Visual Assist with a Laser Pointer and Wearable Display for Remote Collaboration. In Proc CollabTech. 66--71.Google Scholar
- I. Sipiran and B. Bustos. 2011. Harris 3D: A Robust Extension of the Harris Operator for Interest Point Detection on 3D Meshes. Vis. Comput. 27, 11 (2011), 963--976. Google ScholarDigital Library
- R.S. Sodhi, B.R. Jones, D. Forsyth, B.P. Bailey, and G. Maciocci. 2013. BeThere: 3D Mobile Collaboration with Spatial Input. In Proc. ACM CHI. 179--188. Google ScholarDigital Library
- A. Stafford and W. Piekarski. 2008. User Evaluation of God-like Interaction Techniques. In Proc. AUIC. 19--27. Google ScholarDigital Library
- M. Tait and M. Billinghurst. 2014. View independence in remote collaboration using AR. In Proc. IEEE ISMAR. 309--310.Google Scholar
- L. Talmy. 2003. Toward a Cognitive Semantics. Vol. 1--2. MIT press.Google Scholar
- M. Tatzgern, R. Grasset, D. Kalkofen, and D. Schmalstieg. 2014. Transitional Augmented Reality Navigation for Live Captured Scenes. In Proc. IEEE VR. 21--26.Google Scholar
- F. Tecchia, L. Alem, and W. Huang. 2012. 3D Helping Hands: A Gesture Based MR System for Remote Collaboration. In Proc. ACM VRCAI. 323--328. Google ScholarDigital Library
- S.G. Vandenberg and A.R. Kuse. 1978. Mental Rotations, a Group Test of Three-dimensional Spatial Visualization. Perceptual and motor skills 47, 2 (1978), 599--604.Google Scholar
- R. Wang, S. Paris, and J. Popovi'c. 2011. 6D Hands: Markerless Hand-tracking for Computer Aided Design. In Proc. ACM UIST. 549--558. Google ScholarDigital Library
- M. Wexler, S.M. Kosslyn, and A. Berthoz. 1998. Motor Processes in Mental Rotation. Cognition 68, 1 (1998), 77--94.Google ScholarCross Ref
Index Terms
- Virtual Replicas for Remote Assistance in Virtual and Augmented Reality
Recommendations
Remote collaboration in AR and VR using virtual replicas
SIGGRAPH '17: ACM SIGGRAPH 2017 VR VillageIn many complex tasks, a remote subject-matter expert may need to assist a local user, to guide their actions on objects in the local user's environment. However, effective spatial referencing and action demonstration in a remote physical environment ...
Extending Virtual Reality Display Wall Environments Using Augmented Reality
SUI '19: Symposium on Spatial User InteractionTwo major form factors for virtual reality are head-mounted displays and large display environments such as CAVE®and the LCD-based successor CAVE2®. Each of these has distinct advantages and limitations based on how they’re used. This work explores ...
Speed reading on virtual reality and augmented reality
AbstractMany virtual reality (VR) and augmented reality (AR) applications in education require speed reading. The current study aimed to explore whether the reading performance on VR and AR is different from that on traditional desktop display,...
Highlights- We explored performance of speed reading on virtual and augmented reality.
- ...
Comments