skip to main content
10.1145/3234253.3234317acmotherconferencesArticle/Chapter ViewAbstractPublication PagesvricConference Proceedingsconference-collections
research-article

Nectar: Multi-user Spatial Augmented Reality for everyone: Three live demonstrations of educative applications

Published:04 April 2018Publication History

ABSTRACT

In this demonstration we showcase a new spatial augmented reality device (interactive projection) with three applications: education and experimentation of color models, map exploration for visually impaired people and scientific vulgarization of machine learning.

The first exhibition is an interactive exploration about the nature of light. Visitors can experiment with additive subtractive color models. We engage them with questions, and they have to reply using cards to find out answers. This exhibit is suitable for children.

The second exhibition is about map exploration and creation for Visually Impaired Persons (VIP). VIP generally use tactile maps with braille to learn about an unknown environment. However, these maps are not accessible to the 80% of VIP who don't read braille. Our prototype augments raised-line maps with audio output.

The third exhibition is destined to be used for scientific outreach. It enables the creation of artificial neural networks (ANN) using tangible interfaces. Neurons are represented by laser-cut diamond shaped tokens, and the data to learn is printed on cards. The ANN learns to differentiate shapes, and the whole learning process is made visible and interactive.

These three applications demonstrate the capabilities of our hardware and software development kit in different scenarios. At ReVo, each demonstration will have its own setup and interactive space.

References

  1. Jeremy Albouys-Perrois, Jeremy Laviole, Carine Briant, and Anke Brock. 2018. Towards a Multisensory Augmented Reality Map for Blind and Low Vision People: a Participatory Design Approach. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (CHI '18). ACM, New York, NY, USA. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Oliver Bimber and Ramesh Raskar. 2005. Spatial augmented reality: merging real and virtual worlds. CRC press. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Alberto Crivellaro, Mahdi Rad, Yannick Verdie, Kwang Moo Yi, Pascal Fua, and Vincent Lepetit. 2017. Robust 3D Object Tracking from Monocular Images using Stable Parts. IEEE Transactions on Pattern Analysis and Machine Intelligence (2017).Google ScholarGoogle Scholar
  4. Julie Ducasse, Anke Brock, and Christophe Jouffrais. 2018. Accessible Interactive Maps for Visually Impaired Users. In Mobility in Visually Impaired People - Fundamentals and ICT Assistive Technologies, Edwige Pissaloux and Ramiro Velasquez (Eds.). Springer.Google ScholarGoogle Scholar
  5. Renaud Gervais, Joan Sol Roo, and Martin Hachet. 2016. Tangible viewports: Getting out of flatland in desktop environments. In Proceedings of the TEI'16: Tenth International Conference on Tangible, Embedded, and Embodied Interaction. ACM, 176--184. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. HP. 2017. HP Sprout website. http://store.hp.com/us/en/mdp/desktops/sprout-by-hp-204036--1. (2017). {Online; accessed 1-December-2017}.Google ScholarGoogle Scholar
  7. Hirokazu Kato and Mark Billinghurst. 1999. Marker tracking and hmd calibration for a video-based augmented reality conferencing system. In Augmented Reality, 1999.(IWAR'99) Proceedings. 2nd IEEE and ACM International Workshop on. IEEE, 85--94. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. 2012. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems. 1097--1105. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Sébastien Kubicki, Sophie Lepreux, and Christophe Kolski. 2012. RFID-driven situation awareness on TangiSense, a table interacting with tangible objects. Personal and Ubiquitous Computing 16, 8 (2012), 1079--1094. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Jérémy Laviole and Martin Hachet. 2012. PapARt: interactive 3D graphics and multi-touch augmented paper for artistic creation. In 3D User Interfaces (3DUI), 2012 IEEE Symposium on. IEEE, 3--6.Google ScholarGoogle Scholar
  11. David Lindlbauer, Jörg Mueller, and Marc Alexa. 2017. Changing the Appearance of Real-World Objects By Modifying Their Surroundings. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. ACM, 3954--3965. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. David G Lowe. 1999. Object recognition from local scale-invariant features. In Computer vision, 1999. The proceedings of the seventh IEEE international conference on, Vol. 2. Ieee, 1150--1157. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Paul Milgram and Fumio Kishino. 1994. A taxonomy of mixed reality visual displays. IEICE TRANSACTIONS on Information and Systems 77, 12 (1994), 1321--1329.Google ScholarGoogle Scholar
  14. MotionMagix. 2017. MotionMagix interactive projection. http://www.motionmagix.com/interactive-projection-system-software-price/. (2017). {Online; accessed 1-December-2017}.Google ScholarGoogle Scholar
  15. Jun Rekimoto and Masanori Saitoh. 1999. Augmented surfaces: a spatially continuous work space for hybrid computing environments. In Proceedings of the SIGCHI conference on Human Factors in Computing Systems. ACM, 378--385. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Joan Sol Roo and Martin Hachet. 2017. One Reality: Augmenting How the Physical World is Experienced by combining Multiple Mixed Reality Modalities. In Proceedings of the 30th Annual ACM Symposium on User Interface Software and Technology. ACM, 787--795. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Christian Siegl, Matteo Colaianni, Lucas Thies, Justus Thies, Michael Zollhöfer, Shahram Izadi, Marc Stamminger, and Frank Bauer. 2015. Realtime pixel luminance optimization for dynamic multi-projection mapping. ACM Transactions on Graphics (TOG) 34, 6 (2015), 237. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Sony. 2017. Sony Xperia Touch website. https://www.sonymobile.com/global-en/products/smart-products/xperia-touch/. (2017). {Online; accessed 1-December-2017}.Google ScholarGoogle Scholar
  19. Vertigo systems. 2017. Vertigo living floor. https://www.vertigo-systems.de/en/products/living-floor/. (2017). {Online; accessed 1-December-2017}.Google ScholarGoogle Scholar
  20. You Tech. 2017. Interactive ground game. http://www.dltongyou.com/en/Home/Index/column/column_id/11.html. (2017). {Online; accessed 1-December-2017}.Google ScholarGoogle Scholar
  21. TouchMagix. 2017. TouchMagix interactive floor. http://www.touchmagix.com/interactive-floor-interactive-wall. (2017). {Online; accessed 1-December-2017}.Google ScholarGoogle Scholar
  22. Pierre Wellner. 1993. Interacting with paper on the DigitalDesk. Commun. ACM 36, 7 (1993), 87--96. Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. WHO. 2017. Visual Impairment and blindness Fact Sheet Nr 282. Technical Report. World Health Organization. http://www.who.int/mediacentre/factsheets/fs282/en/Google ScholarGoogle Scholar
  24. Jacob O Wobbrock, Meredith Ringel Morris, and Andrew D Wilson. 2009. User-defined gestures for surface computing. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, 1083--1092. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Paul Wohlhart and Vincent Lepetit. 2015. Learning descriptors for object recognition and 3d pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3109--3118.Google ScholarGoogle ScholarCross RefCross Ref
  26. Robert Xiao, Scott Hudson, and Chris Harrison. 2016. DIRECT: Making Touch Tracking on Ordinary Surfaces Practical with Hybrid Depth-Infrared Sensing. In Proceedings of the 2016 ACM on Interactive Surfaces and Spaces. ACM, 85--94. Google ScholarGoogle ScholarDigital LibraryDigital Library
  1. Nectar: Multi-user Spatial Augmented Reality for everyone: Three live demonstrations of educative applications

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      VRIC '18: Proceedings of the Virtual Reality International Conference - Laval Virtual
      April 2018
      173 pages
      ISBN:9781450353816
      DOI:10.1145/3234253
      • Conference Chair:
      • Simon Richir

      Copyright © 2018 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 4 April 2018

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader