ABSTRACT
ReFlex is a flexible smartphone with bend input and active haptic feedback. ReFlex's features allow the introduction of sensations such as friction or resistance. We report results from an experiment using ReFlex in a targeting task, as well as initial users' reactions to the prototype. We explore both absolute and relative tactile haptic feedback, paired with two types of bend input mappings: position-control and rate-control. We observed that position-controlled cursors paired well with relative bend feedback, while rate-controlled cursors paired well with absolute bend feedback to indicate targets. We also explored an eyes-free condition. Results suggest that while eyes-free, haptic feedback conditions were more error-prone than visual-only conditions, the size of the error was relatively small, and users were able to complete the task in all cases. We present two application scenarios that take advantage of the unique input and output modalities of ReFlex and discuss its potential for within document navigation.
- Motoyuki Akamatsu, I. Scott MacKenzie, and Thierry Hasbrouc. 1995. A comparison of tactile, auditory, and visual feedback in a pointing task using a mouse-type device. Ergonomics 38: 816--827.Google ScholarCross Ref
- Olivier Bau, Ivan Poupyrev, Ali Israr, and Chris Harrison. 2010. TeslaTouch: electrovibration for touch surfaces. In Proceedings of the 23nd annual ACM symposium on User interface software and technology (UIST '10). ACM, New York, NY, USA, 283--292. Google ScholarDigital Library
- Frank Biocca, Jin Kim, and Yung Choi. 2011. Visual Touch in Virtual Environments: An Exploratory Study of Presence, Multi-Modal Interfaces, and Cross-Modal Sensory Illusions. Presence, 10: 247--265. Google ScholarDigital Library
- Nicholas Chen, Francois Guimbretiere, Morgan Dixon, Cassandra Lewis, and Maneesh Agrawala. 2008. Navigation techniques for dual-display e-book readers. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '08). ACM, New York, NY, USA, 1779--1788. Google ScholarDigital Library
- Clifton Forlines and Ravin Balakrishnan. 2008. Evaluating tactile feedback and direct vs. indirect stylus input in pointing and crossing selection tasks. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '08). ACM, New York, NY, USA, 1563--1572. Google ScholarDigital Library
- Audrey Girouard, Aneesh Tarun, and Roel Vertegaal. 2012. DisplayStacks: interaction techniques for stacks of flexible thin-film displays. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '12). ACM, New York, NY, USA, 2431--2440. Google ScholarDigital Library
- Chris Harrison, Zhiquan Yeo, and Scott E. Hudson. 2010. Faster progress bars: manipulating perceived duration with visual augmentations. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '10). ACM, New York, NY, USA, 1545--1548. Google ScholarDigital Library
- Masaharu Hirose, Karin Iwazaki, Kozue Nojiri, Minato Takeda, Yuta Sugiura, and Masahiko Inami. 2015. Gravitamine spice: a system that changes the perception of eating through virtual weight sensation. In Proceedings of the 6th Augmented Human International Conference (AH '15). ACM, New York, NY, USA, 33--40. Google ScholarDigital Library
- David Holman and Roel Vertegaal. 2008. Organic user interfaces: designing computers in any way, shape, or form. Commun. ACM 51, 6 (June 2008), 48--55. Google ScholarDigital Library
- Johan Kildal. 2010. 3D-press: haptic illusion of compliance when pressing on a rigid surface. In International Conference on Multimodal Interfaces and the Workshop on Machine Learning for Multimodal Interaction (ICMI-MLMI '10). ACM, New York, NY, USA, Article 21, 8 pages. Google ScholarDigital Library
- Johan Kildal, Susanna Paasovaara, and Viljakaisa Aaltonen. 2012. Kinetic device: designing interactions with a deformable mobile interface. In CHI '12 Extended Abstracts on Human Factors in Computing Systems (CHI EA '12). ACM, New York, NY, USA, 1871--1876. Google ScholarDigital Library
- Johan Kildal and Graham Wilson. 2012. Feeling it: the roles of stiffness, deformation range and feedback in the control of deformable ui. In Proceedings of the 14th ACM international conference on Multimodal interaction (ICMI '12). ACM, New York, NY, USA, 393--400. Google ScholarDigital Library
- Roberta L. Klatzky and Susan J. Lederman. 2006. The perceived roughness of resistive virtual textures: I. rendering by a force-feedback mouse. ACM Trans. Appl. Percept. 3, 1 (January 2006), 1--14. Google ScholarDigital Library
- Byron Lahey, Audrey Girouard, Winslow Burleson, and Roel Vertegaal. 2011. PaperPhone: understanding the use of bend gestures in mobile devices with flexible electronic paper displays. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '11). ACM, New York, NY, USA, 1303--1312. Google ScholarDigital Library
- Susan J. Lederman and Lynette A. Jones. 2011. Tactile and Haptic Illusions. IEEE Transactions on Haptics 4, 4: 273--294. Google ScholarDigital Library
- Catherine C. Marshall and Sara Bly. 2005. Turning the page on navigation. In Proceedings of the 5th ACM/IEEE-CS joint conference on Digital libraries (JCDL '05). ACM, New York, NY, USA, 225--234. Google ScholarDigital Library
- Yusuke Nakagawa, Akiya Kamimura, and Yoichiro Kawaguchi. 2012. MimicTile: a variable stiffness deformable user interface for mobile devices. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '12). ACM, New York, NY, USA, 745--748. Google ScholarDigital Library
- Donald Norman. 1988. The Psychology of Everyday Things. Basic Books.Google Scholar
- Kenton O'Hara and Abigail Sellen. 1997. A comparison of reading paper and on-line documents. In Proceedings of the ACM SIGCHI Conference on Human factors in computing systems (CHI '97). ACM, New York, NY, USA, 335--342. Google ScholarDigital Library
- Omega Engineering. Strain Gages. 2015. http://www.omega.ca/guides/straingages.htmlGoogle Scholar
- Joseph D. Rutledge and Ted Selker. 1990. Force-to-motion functions for pointing. In Proceedings of the IFIP TC13 Third International Conference on Human-Computer Interaction (INTERACT '90), Dan Diaper, David J. Gilmore, Gilbert Cockton, and Brian Shackel (Eds.). North-Holland Publishing Co., Amsterdam, The Netherlands, The Netherlands, 701--706. Google ScholarDigital Library
- Carsten Schwesig, Ivan Poupyrev, and Eijiro Mori. 2004. Gummi: a bendable computer. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '04). ACM, New York, NY, USA, 263--270. Google ScholarDigital Library
- M.A. Srinivasan, G.L. Beauregard, and D.L. Brock. The Impact of Visual Information on the Haptic Perception of Stiffness in Virtual Environments. In Proceedings of ASME Dynamic Systems and Control Division (1996), 58. 555--559.Google Scholar
- Tactile Labs. Hapuator Mark II. 2012. http://www.tactilelabs.com/products/haptics/haptuator-mark-ii-v2/Google Scholar
- Hsin-yun Yao and Vincent Hayward. 2006. An experiment on length perception with a virtual rolling stone. In Proceedings of EuroHaptics 2006. Springer-Verlag, 325--330.Google Scholar
- Shumin Zhai. 1995. Human Performance in Six Degree of Freedom Input Control. Ph.D Dissertation. University of Toronto, Toronto, ON.Google Scholar
Index Terms
- ReFlex: A Flexible Smartphone with Active Haptic Feedback for Bend Input
Recommendations
WhammyPhone: Exploring Tangible Audio Manipulation Using Bend Input on a Flexible Smartphone
UIST '16 Adjunct: Adjunct Proceedings of the 29th Annual ACM Symposium on User Interface Software and TechnologyWe present WhammyPhone, a novel audio interface that supports physical manipulation of digital audio through bend gestures. WhammyPhone combines a high-resolution flexible display, bend sensors, and a set of intuitive interaction techniques that enable ...
FlexView: an evaluation of depth navigation on deformable mobile devices
TEI '13: Proceedings of the 7th International Conference on Tangible, Embedded and Embodied InteractionWe present FlexView, a set of interaction techniques for Z-axis navigation on touch-enabled flexible mobile devices. FlexView augments touch input with bend to navigate through depth-arranged content. To investigate Z-axis navigation with FlexView, we ...
MagicWand: Exploring Physical Affordances with a Handheld Cylindrical Display Object
CHI EA '16: Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing SystemsWe introduce MagicWand, a cylindrical handheld display object made out of 2 Flexible Organic Light-Emitting Diode (FOLED) screens. Magic Wand features a smartphone running the Android operating system. Gesture recognition allows movements of the wand as ...
Comments