Abstract
Visual graphics are widely spread in digital media and are useful in many contexts of daily life. However, access to this type of graphical information remains a challenging task for people with visual impairments (VI). In this study, we designed and evaluated an on-hand vibrotactile interface that enables users with VI to explore digital graphics presented on tablets. We first conducted a set of exploratory tests with both people with VI and blindfolded (BF) people to investigate several design factors. We then conducted a comparative experiment to verify that on-hand vibrotactile cues (indicating direction and progression) can enhance the non-visual exploration of digital graphics. The results based on 12 participants with VI and 12 BF participants confirmed the usability of the technique and revealed that the visual status of the users does not impact graphics identification and comparison tasks.
- Jérémy Albouys-Perrois, Jérémy Laviole, Carine Briant, and Anke M. Brock. 2018. Towards a Multisensory Augmented Reality Map for Blind and Low Vision People: a Participatory Design Approach. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems (CHI '18). ACM, New York, NY, USA, Paper 629, 14 pages. DOI: https://doi.org/10.1145/3173574.3174203Google Scholar
- David S. Alles. 1970. Information Transmission by Phantom Sensations. IEEE Trans. Man-Machine Syst. 11, 1 (1970), 85--91. DOI:https://doi.org/10.1109/TMMS.1970.299967Google ScholarCross Ref
- Jessalyn Alvina, Shengdong Zhao, Simon T. Perrault, Maryam Azh, Thijs Roumen, and Morten Fjeld. 2015. OmniVib: Towards Cross-body Spatiotemporal Vibrotactile Notifications for Mobile Phones. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems (CHI '15). ACM, New York, NY, USA, 2487--2496. DOI: https://doi.org/10.1145/2702123.2702341Google ScholarDigital Library
- Heller, M.A., 1989. Picture and pattern perception in the sighted and the blind: the advantage of the late blind. Perception, 18(3), pp.379--389Google ScholarCross Ref
- Sandra Bardot, Marcos Serrano, and Christophe Jouffrais. 2016. From tactile to virtual: using a smartwatch to improve spatial map exploration for visually impaired users. In Proceedings of the 18th International Conference on Human-Computer Interaction with Mobile Devices and Services (MobileHCI '16). ACM, New York, NY, USA, 100--111. DOI: https://doi.org/10.1145/2935334.2935342Google ScholarDigital Library
- Anke Brock, Samuel Lebaz, Bernard Oriola, Delphine Picard, Christophe Jouffrais, and Philippe Truillet. 2012. Kin'touch: understanding how visually impaired people explore tactile maps. In CHI '12 Extended Abstracts on Human Factors in Computing Systems (CHI EA '12). ACM, New York, NY, USA, 2471--2476. DOI: https://doi.org/10.1145/2212776.2223821Google ScholarDigital Library
- Emeline Brulé, Brianna J. Tomlinson, Oussama Metatla, Christophe Jouffrais, and Marcos Serrano. 2020. Review of Quantitative Empirical Evaluations of Technology for People with Visual Impairments. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (CHI '20). Association for Computing Machinery, New York, NY, USA, 1--14. DOI:https://doi.org/10.1145/3313831.3376749Google ScholarDigital Library
- Jessica R. Cauchard, Janette L. Cheng, Thomas Pietrzak, and James A. Landay. 2016. ActiVibe: Design and Evaluation of Vibrations for Progress Monitoring. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems (CHI '16). Association for Computing Machinery, New York, NY, USA, 3261--3271. DOI:https://doi.org/10.1145/2858036.2858046Google Scholar
- Cornoldi Cesare, Carla Tinti, Irene C. Mammarella, Anna Maria Re, and Diego Varotto. "Memory for an imagined pathway and strategy effects in sighted and in totally congenitally blind individuals." Acta Psychologica 130, no. 1 (2009): 11--16.Google Scholar
- Angela Chang, Sile O'Modhrain, Rob Jacob, Eric Gunther, and Hiroshi Ishii. 2002. ComTouch: design of a vibrotactile communication device. In Proceedings of the 4th conference on Designing interactive systems: processes, practices, methods, and techniques (DIS '02). ACM, New York, NY, USA, 312--320. DOI=http://dx.doi.org/10.1145/778712.778755Google ScholarDigital Library
- Qin Chen, Simon T. Perrault, Quentin Roy, and Lonce Wyse. 2018. Effect of temporality, physical activity and cognitive load on spatiotemporal vibrotactile pattern recognition. In Proceedings of the 2018 International Conference on Advanced Visual Interfaces (AVI '18). ACM, New York, NY, USA, Article 25, 9 pages. DOI: https://doi.org/10.1145/3206505.3206511Google ScholarDigital Library
- Camors Damien, Damien Appert, Jean-Baptiste Durand, and Christophe Jouffrais. "Tactile Cues for Improving Target Localization in Subjects with Tunnel Vision." Multimodal Technologies and Interaction 3, no. 2 (2019): 26.Google Scholar
- Jacobson, R. Dan. "Navigating maps with little or no sight: An audio-tactile approach." In Content Visualization and Intermedia Representations (CVIR'98). 1998.Google Scholar
- Delphine Picard, Samuel Lebaz, Christophe Jouffrais, and Catherine Monnier. "Haptic recognition of two-dimensional raised-line patterns by early-blind, late-blind, and blindfolded sighted adults." Perception 39, no. 2 (2010): 224--235.Google Scholar
- Julie Ducasse, Anke M Brock, and Christophe Jouffrais. 2018. Accessible Interactive Maps for Visually Impaired Users. In Mobility of Visually Impaired People. Springer, 537--584. DOI:https://doi.org/10.1007/978--3--319--54446--5_17Google Scholar
- Julie Ducasse, Marc J-M Macé, Marcos Serrano, and Christophe Jouffrais. 2016. Tangible Reels: Construction and Exploration of Tangible Maps by Visually Impaired Users. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems (CHI '16). ACM, New York, NY, USA, 2186--2197. DOI: https://doi.org/10.1145/2858036.2858058Google ScholarDigital Library
- Don Samitha Elvitigala, Denys J. C. Matthies, Vipula Dissanayaka, Chamod Weerasinghe, and Suranga Nanayakkara. 2019. 2bit-TactileHand: Evaluating Tactons for On-Body Vibrotactile Displays on the Hand and Wrist. In Proceedings of the 10th Augmented Human International Conference 2019 (AH2019). ACM, New York, NY, USA, Article 3, 8 pages. DOI: https://doi.org/10.1145/3311823.3311832Google ScholarDigital Library
- Nicholas A. Giudice, Hari Prasath Palani, Eric Brenner, and Kevin M. Kramer. 2012. Learning non-visual graphical information using a touch-based vibro-audio interface. In Proceedings of the 14th international ACM SIGACCESS conference on Computers and accessibility (ASSETS '12). ACM, New York, NY, USA, 103--110. DOI: https://doi.org/10.1145/2384916.2384935Google Scholar
- Cagatay Goncu and Kim Marriott. 2011. GraVVITAS: Generic Multi-touch Presentation of Accessible Graphics. In Proceedings of the 13th IFIP TC 13 Conference on Human-Computer Interaction (INTERACT '11),30--48. DOI:https://doi.org/10.1007/978--3--642--23774--4Google ScholarCross Ref
- Jonggi Hong, Alisha Pradhan, Jon E. Froehlich, and Leah Findlater. 2017. Evaluating Wrist-Based Haptic Feedback for Non-Visual Target Finding and Path Tracing on a 2D Surface. In Proceedings of the 19th International ACM SIGACCESS Conference on Computers and Accessibility (ASSETS '17). Association for Computing Machinery, New York, NY, USA, 210--219. DOI:https://doi.org/10.1145/3132525.3132538Google ScholarDigital Library
- Sreekar Krishna, Shantanu Bala, Troy McDaniel, Stephen McGuire, and Sethuraman Panchanathan. 2010. VibroGlove: an assistive technology aid for conveying facial expressions. In CHI '10 Extended Abstracts on Human Factors in Computing Systems (CHI EA '10). ACM, New York, NY, USA, 3637--3642. DOI: https://doi.org/10.1145/1753846.1754031Google ScholarDigital Library
- Zhuoluo Ma, Yue Liu, Dejiang Ye, and Lu Zhao. 2019. Vibrotactile Wristband for Warning and Guiding in Automated Vehicles. In Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems (CHI EA '19). ACM, New York, NY, USA, Paper LBW2220, 6 pages. DOI: https://doi.org/10.1145/3290607.3312819Google Scholar
- Jonatan Martínez, Arturo García, Miguel Oliver, José Pascual Molina, and Pascual González. 2016. Identifying Virtual 3D Geometric Shapes with a Vibrotactile Glove. IEEE Comput. Graph. Appl. 36, 1 (2016), 42--51. DOI:https://doi.org/10.1109/MCG.2014.81Google ScholarDigital Library
- Anita Meier, Denys J. C. Matthies, Bodo Urban, and Reto Wettach. 2015. Exploring vibrotactile feedback on the body and foot for the purpose of pedestrian navigation. In Proceedings of the 2nd International Workshop on Sensor-based Activity Recognition and Interaction, ACM, 1--11. DOI:https://doi.org/10.1145/2790044.2790051Google ScholarDigital Library
- Giuseppe Melfi, Karin Müller, Thorsten Schwarz, Gerhard Jaworek, and Rainer Stiefelhagen. 2020. Understanding what you feel: A Mobile Audio-Tactile System for Graphics Used at Schools with Students with Visual Impairment. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (CHI '20). Association for Computing Machinery, New York, NY, USA, 1--12. DOI:https://doi.org/10.1145/3313831.3376508Google ScholarDigital Library
- Hari P. Palani, Paul D. S. Fink & Nicholas A. Giudice (2020): Design Guidelines for Schematizing and Rendering Haptically Perceivable Graphical Elements on Touchscreen Devices, International Journal of Human--Computer Interaction, 1--22.Google ScholarCross Ref
- Jaeyoung Park, Jaeha Kim, Yonghwan Oh, and Hong Z Tan. 2016. Rendering moving tactile stroke on the palm using a sparse 2d array. In International Conference on Human Haptic Sensing and Touch Enabled Computer Applications. Springer, Cham, 47--56. DOI:https://doi.org/10.1007/978--3--319--42324--1Google ScholarDigital Library
- Benjamin Poppinga, Charlotte Magnusson, Martin Pielot, and Kirsten Rassmus-Gröhn. 2011. TouchOver map: audio-tactile exploration of interactive maps. In Proceedings of the 13th International Conference on Human Computer Interaction with Mobile Devices and Services (MobileHCI '11). ACM, New York, NY, USA, 545--550. DOI: https://doi.org/10.1145/2037373.2037458Google ScholarDigital Library
- R.M. Sakia. The Box-Cox transformation technique: a review. Journal of the Royal Statistical Society: Series D (The Statistician), 41(2), 169--178.Google Scholar
- Elizabeth B-N Sanders. 2002. From user-centered to participatory design approaches. Design and the social sciences: Making connections 1, 8 (2002).Google Scholar
- Choi Seungmoon, and Katherine J. Kuchenbecker. "Vibrotactile display: Perception, technology, and applications." Proceedings of the IEEE 101, no. 9 (2012): 2093--2104.Google Scholar
- Abigale Stangl, Meredith Ringel Morris, and Danna Gurari. 2020. 'Person, Shoes, Tree. Is the Person Naked'? What People with Vision Impairments Want in Image Descriptions. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (CHI '20). Association for Computing Machinery, New York, NY, USA, 1--13. DOI:https://doi.org/10.1145/3313831.3376404Google ScholarDigital Library
- Anil K. Raj, Steven J. Kass, and James F. Perry. 2012. Vibrotactile Displays for Improving Spatial Awareness. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting. Sage CA: Los Angeles, 44, 1 (2012), 181--184. DOI:https://doi.org/10.1177/154193120004400148Google Scholar
- Joe Tekli, Youssef Bou Issa, and Richard Chbeir. 2018. Evaluating touch-screen vibration modality for blind users to access simple shapes and graphics. International Journal of Human-Computer Studies 110, August 2017 (2018), 115--133. DOI:https://doi.org/10.1016/j.ijhcs.2017.10.009Google ScholarDigital Library
- Zheshen Wang, Baoxin Li, Terri Hedgpeth, and Teresa Haven. 2009. Instant tactile-audio map: enabling access to digital maps for people with visual impairment. In Proceedings of the 11th international ACM SIGACCESS conference on Computers and accessibility (Assets '09). ACM, New York, NY, USA, 43--50. DOI: https://doi.org/10.1145/1639642.1639652Google ScholarDigital Library
- Ruolin Wang, Chun Yu, Xing-Dong Yang, Weijie He, and Yuanchun Shi. 2019. EarTouch: Facilitating Smartphone Use for Visually Impaired People in Mobile and Public Scenarios. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (CHI '19). ACM, New York, NY, USA, Paper 24, 13 pages. DOI: https://doi.org/10.1145/3290605.3300254Google ScholarDigital Library
- John S Zelek. 2005. Seeing by touch (haptics) for wayfinding. In International congress series, Vol. 1282. Elsevier, 1108--1112.Google Scholar
Index Terms
- VibHand: On-Hand Vibrotactile Interface Enhancing Non-Visual Exploration of Digital Graphics
Recommendations
Investigating touchscreen accessibility for people with visual impairments
NordiCHI '08: Proceedings of the 5th Nordic conference on Human-computer interaction: building bridgesTouchscreen computing devices such as the iPhone are becoming more common. However this technology is largely inaccessible to people with visual impairments. We present the results of a requirements capture study that illustrates the problems with ...
Perceptions of accessibility and usability by blind or visually impaired persons: a pilot study
ASIST '16: Proceedings of the 79th ASIS&T Annual Meeting: Creating Knowledge, Enhancing Lives through Information & TechnologyThis pilot study utilizes qualitative interviews to explore perceptions of accessibility and usability from the perspective of blind or visually impaired (BVI) persons. Using the frameworks of everyday life information seeking (ELIS) and the sense-...
Rich Representations of Visual Content for Screen Reader Users
CHI '18: Proceedings of the 2018 CHI Conference on Human Factors in Computing SystemsAlt text (short for "alternative text") is descriptive text associated with an image in HTML and other document formats. Screen reader technologies speak the alt text aloud to people who are visually impaired. Introduced with HTML 2.0 in 1995, the alt ...
Comments