ABSTRACT
Modern volumetric projection-based telepresence approaches are capable of providing realistic full-size virtual representations of remote people. Interacting with full-size people may not be desirable due to the spatial constraints of the physical environment, application context, or display technology. However, the miniaturization of remote people is known to create an eye gaze matching problem. Eye-contact is essential to communication as it allows for people to use natural nonverbal cues and improves the sense of “being there”. In this paper we discuss the design space for interacting with volumetric representations of people and present an approach for dynamically manipulating scale, orientation and the position of holograms which guarantees eye-contact. We created a working augmented reality-based prototype and validated it with 14 participants.
- Michael Argyle and Mark Cook. 1976. Gaze and mutual gaze.(1976).Google Scholar
- Michael Argyle and Janet Dean. 1965. Eye-contact, distance and affiliation. Sociometry (1965), 289–304.Google Scholar
- Stephan Beck, Andre Kunert, Alexander Kulik, and Bernd Froehlich. 2013. Immersive group-to-group telepresence. IEEE Transactions on Visualization and Computer Graphics 19, 4(2013), 616–625.Google ScholarDigital Library
- Steve Benford, John Bowers, Lennart E. Fahlén, Chris Greenhalgh, and Dave Snowdon. 1995. User Embodiment in Collaborative Virtual Environments. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems(CHI ’95). ACM Press/Addison-Wesley Publishing Co., New York, NY, USA, 242–249. https://doi.org/10.1145/223904.223935Google ScholarDigital Library
- Bill Buxton. 2009. Mediaspace – Meaningspace – Meetingspace. Springer London, London, 217–231. https://doi.org/10.1007/978-1-84882-483-6_13Google Scholar
- William Buxton. 1992. Telepresence: Integrating shared task and person spaces. In Proceedings of graphics interface, Vol. 92. 123–129.Google ScholarDigital Library
- Susan R. Fussell, Leslie D. Setlock, Jie Yang, Jiazhi Ou, Elizabeth Mauer, and Adam D. I. Kramer. 2004. Gestures over Video Streams to Support Remote Collaboration on Physical Tasks. Hum.-Comput. Interact. 19, 3 (Sept. 2004), 273–309. https://doi.org/10.1207/s15327051hci1903_3Google ScholarDigital Library
- Maia Garau, Mel Slater, Simon Bee, and Martina Angela Sasse. 2001. The Impact of Eye Gaze on Communication Using Humanoid Avatars. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems(CHI ’01). ACM, New York, NY, USA, 309–316. https://doi.org/10.1145/365024.365121Google ScholarDigital Library
- Hiroshi Ishii and Minoru Kobayashi. 1992. ClearBoard: A Seamless Medium for Shared Drawing and Conversation with Eye Contact. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems(CHI ’92). ACM, New York, NY, USA, 525–532. https://doi.org/10.1145/142750.142977Google ScholarDigital Library
- Andrew Jones, Magnus Lang, Graham Fyffe, Xueming Yu, Jay Busch, Ian McDowall, Mark Bolas, and Paul Debevec. 2009. Achieving eye contact in a one-to-many 3D video teleconferencing system. ACM Transactions on Graphics (TOG) 28, 3 (2009), 64.Google ScholarDigital Library
- H. Kimura, A. Asano, I. Fujishiro, A. Nakatani, and H. Watanabe. 2011. True 3D Display. ACM SIGGRAPH 2011 Emerging Technologies, SIGGRAPH’11 (2011). https://doi.org/10.1145/2048259.2048279 cited By 10.Google ScholarDigital Library
- Daniil Kononenko, Yaroslav Ganin, Diana Sungatullina, and Victor Lempitsky. 2017. Photorealistic monocular gaze redirection using machine learning. IEEE transactions on pattern analysis and machine intelligence 40, 11(2017), 2696–2710.Google Scholar
- David Nguyen and John Canny. 2005. MultiView: Spatially Faithful Group Video Conferencing. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems(CHI ’05). ACM, New York, NY, USA, 799–808. https://doi.org/10.1145/1054972.1055084Google ScholarDigital Library
- Sergio Orts-Escolano, Christoph Rhemann, Sean Fanello, Wayne Chang, Adarsh Kowdle, Yury Degtyarev, David Kim, Philip L. Davidson, Sameh Khamis, Mingsong Dou, Vladimir Tankovich, Charles Loop, Qin Cai, Philip A. Chou, Sarah Mennicken, Julien Valentin, Vivek Pradeep, Shenlong Wang, Sing Bing Kang, Pushmeet Kohli, Yuliya Lutchyn, Cem Keskin, and Shahram Izadi. 2016. Holoportation: Virtual 3D Teleportation in Real-time. In Proceedings of the 29th Annual Symposium on User Interface Software and Technology(UIST ’16). ACM, New York, NY, USA, 741–754. https://doi.org/10.1145/2984511.2984517Google ScholarDigital Library
- Tomislav Pejsa, Julian Kantor, Hrvoje Benko, Eyal Ofek, and Andrew Wilson. 2016a. Room2Room: Enabling Life-Size Telepresence in a Projected Augmented Reality Environment. In Proceedings of the 19th ACM Conference on Computer-Supported Cooperative Work & Social Computing(CSCW ’16). ACM, New York, NY, USA, 1716–1725. https://doi.org/10.1145/2818048.2819965Google ScholarDigital Library
- Tomislav Pejsa, Julian Kantor, Hrvoje Benko, Eyal Ofek, and Andrew Wilson. 2016b. Room2Room: Enabling life-size telepresence in a projected augmented reality environment. In Proceedings of the 19th ACM Conference on Computer-Supported Cooperative Work & Social Computing. ACM, 1716–1725.Google ScholarDigital Library
- Nathan Shedroff and Christopher Noessel. 2012. Make it so: Interaction design lessons from science fiction. Rosenfeld Media.Google Scholar
- DE Smalley, E Nygaard, K Squire, J Van Wagoner, J Rasmussen, S Gneiting, K Qaderi, J Goodsell, W Rogers, M Lindsey, 2018. A photophoretic-trap volumetric display. Nature 553, 7689 (2018), 486.Google Scholar
- Maurício Sousa, Daniel Mendes, Rafael Kuffner Dos Anjos, Daniel Medeiros, Alfredo Ferreira, Alberto Raposo, João Madeiras Pereira, and Joaquim Jorge. 2017. Creepy Tracker Toolkit for Context-aware Interfaces. In Proceedings of the 2017 ACM International Conference on Interactive Surfaces and Spaces(ISS ’17). ACM, New York, NY, USA, 191–200. https://doi.org/10.1145/3132272.3134113Google ScholarDigital Library
- Justus Thies, Michael Zollhöfer, Marc Stamminger, Christian Theobalt, and Matthias Nießner. 2018. FaceVR: Real-time gaze-aware facial reenactment in virtual reality. ACM Transactions on Graphics (TOG) 37, 2 (2018), 25.Google ScholarDigital Library
- Shih-En Wei, Jason Saragih, Tomas Simon, Adam Harley, Stephen Lombardi, Michal Perdoch, Alexander Hypes, Dawei Wang, Hernan Badino, and Yaser Sheikh. [n.d.]. VR Facial Animation via Multiview Image Translation. ACM Trans. Graph.([n. d.]).Google Scholar
Recommendations
Volumetric Mixed Reality Telepresence for Real-time Cross Modality Collaboration
CHI '23: Proceedings of the 2023 CHI Conference on Human Factors in Computing SystemsMixed-reality telepresence allows local and remote users feel as if they are present together in the same space. In this paper we report on a mixed-reality volumetric telepresence system that is adaptable, multi-user and cross-modal, i.e. combining ...
A Design Space for Gaze Interaction on Head-mounted Displays
CHI '19: Proceedings of the 2019 CHI Conference on Human Factors in Computing SystemsAugmented and virtual reality (AR/VR) has entered the mass market and, with it, will soon eye tracking as a core technology for next generation head-mounted displays (HMDs). In contrast to existing gaze interfaces, the 3D nature of AR and VR requires ...
Shared space: An augmented reality approach for computer supported collaborative work
Virtual Reality (VR) appears a natural medium for three-dimensional computer supported collaborative work (CSCW). However the current trend in CSCW is to adapt the computer interface to work with the user's traditional tools, rather than separating the ...
Comments