skip to main content
10.1145/1978942.1979353acmconferencesArticle/Chapter ViewAbstractPublication PageschiConference Proceedingsconference-collections
research-article

Sasayaki: augmented voice web browsing experience

Published: 07 May 2011 Publication History

Abstract

Auditory user interfaces have great Web-access potential for billions of people with visual impairments, with limited literacy, who are driving, or who are otherwise unable to use a visual interface. However a sequential speech-based representation can only convey a limited amount of information. In addition, typical auditory user interfaces lose the visual cues such as text styles and page structures, and lack effective feedback about the current focus. To address these limitations, we created Sasayaki (from whisper in Japanese), which augments the primary voice output with a secondary whisper of contextually relevant information, automatically or in response to user requests. It also offers new ways to jump to semantically meaningful locations. A prototype was implemented as a plug-in for an auditory Web browser. Our experimental results show that the Sasayaki can reduce the task completion times for finding elements in webpages and increase satisfaction and confidence.

Supplementary Material

index.html (index.html)
Slides from the presentation
Audio only (1979353.mp3)
Video (1979353.mp4)

References

[1]
Internet World Stats, World Internet Users and Population Stats, http://www.internetworldstats.com/stats.htm.
[2]
UNESCO, International Literacy Statistics: A Review of Concepts, Methodology and Current Data, http://www.uis.unesco.org/template/pdf/Literacy/LiteracyReport2008.pdf.
[3]
WHO, Fact sheet of visual impairment and blindness, http://www.who.int/mediacentre/factsheets/fs282/en/.
[4]
ITU. Measuring the Information Society 2010. http://www.itu.int/ITU-D/ict/publications/idi/2010/.
[5]
Takagi, H., Saito, S., Fukuda, K. and Asakawa, C. Analysis of navigability of Web applications for improving blind usability. ACM Trans. Comp.-Hum. Interact 14:3 (2007), 13.
[6]
Barnicle, K. Usability testing with screen reading technology in a Windows environment. In Proc. CUU 2000, ACM Press (2000), 102--109.
[7]
Lazar, J., Allen, A.and Kleinman, J. and Malarkey, C. What Frustrates Screen Reader Users on the Web: A Study of 100 Blind Users. International Journal of Human-Computer Interaction 22:3 (2007), 247--269.
[8]
Maes, P. 1994. Agents that reduce work and information overload. Communications of. ACM 37, 7 (1994), 30--40.
[9]
Bederson, B. B. Audio augmented reality: a prototype automated tour guide.,In Proc. CHI 1995, ACM Press, (1995), 210--211.
[10]
Sawhney, N. and Schmandt, C. Nomadic radio: speech and audio interaction for contextual messaging in nomadic environments. ACM Trans. Comput.-Hum. Interact. (7:3), (2000), 353--383.
[11]
Eckel, G. Immersive Audio-Augmented Environments: The LISTEN Project, Fifth Framework Programme, Creating a user-friendly information society (IST), (2001), 571.
[12]
Kalantari, L., Hatala, M. and Willms, J. Using semantic web approach in augmented audio reality system for museum visitors. In Proc. WWW 2004, ACM Press, (2004), 386--387.
[13]
Miyashita, T., Meier, P., Tachikawa, T., Orlic, S., Eble, T., Scholz, V., Gapel, A., Gerl, O., Arnaudov, S. and Lieberknecht, S. An Augmented Reality museum guide. In Proc. ISMAR 2008, IEEE Computer Society (2008),103--106.
[14]
Shoval, S., Borenstein, J. and Koren, Y. The Navbelt - A Computerized Travel Aid for the Blind Based on Mobile Robotics Technology. IEEE Trans. on Biomedical Engineering (45:11) (1998), 1376--1386.
[15]
Jones, M., Jones, S., Bradley, G., Warren, N., Bainbridge, D. and Holmes, G. ONTRACK: Dynamically adapting music playback to support navigation. Personal and Ubiquitous Computing (12:7), (2008), 513--525.
[16]
Stylos, J., Myers, B. A. and Faulring, A. Citrine: providing intelligent copy-and-paste. In Proc. UIST 2004, ACM Press (2004), 185--188.
[17]
Wagner, E. J. and Lieberman, H. Supporting user hypotheses in problem diagnosis. In Proc. IUI 2004, ACM Press (2004), 30--37.
[18]
Roth, P., Petrucci, L., Pun, T., and Assimacopouls, A. Auditory browser for blind and visually impaired users, In Proc. CHI 1999. ACM Press (1999), 218--219.
[19]
Yu, W., McAllister, G., Strain, P., Kuber, R. and Murphy, E. Improving web accessibility using content-aware plug-ins. In Proc. CHI 2005, ACM Press (2005), 1893--1896.
[20]
Dontcheva, M., Drucker, S. M., Wade, G., Salesin, D., and Cohen, M. F. Summarizing personal web browsing sessions. In Proc. UIST 2006. ACM Press (2006), 115--124.
[21]
Hartmann, M., Schreiber, D. and Muhlhauser, M. AUGUR: providing context-aware interaction support. In Proc. of symposium on Engineering interactive computing systems, ACM Press (2009), 123--132.
[22]
Parente, P. Clique: a conversant, task-based audio display for GUI applications. SIGACCESS Access. Comput. 84 (2006), 34--37.
[23]
Mahmud, J. U., Borodin, Y., and Ramakrishnan, I. V. Csurf: a context-driven non-visual web-browser. In Proc. WWW 2007. ACM Press (2007), 31--40.
[24]
Borodin, Y., Bigham, J. P., Raman, R. and Ramakrishnan, I. V. What's new? Making web page updates accessible. In Proc. ASSETS 2008, ACM Press (2008), 145--152.
[25]
Yesilada, Y., Stevens, R., Harper, S. and Goble, C. Evaluating DANTE: Semantic transcoding for visually disabled users. ACM Trans. Comput.-Hum. Interact. (14:3), (2007), 14.
[26]
Harper, S. and Patel, N. Gist summaries for visually impaired surfers. In Proc. ASSETS 2005, ACM Press (2005), 90--97.
[27]
Miyashita, H., Sato, D., Takagi, H. and Asakawa, C. aiBrowser for multimedia: introducing multimedia content accessibility for visually impaired users. In Proc.ASSETS 2007, ACM Press (2007), 91--98.
[28]
Lunn, D., Bechhofer, S. and Harper, S. The SADIe transcoding platform. In Proceedings of the 2008 international cross-disciplinary conference on Web accessibility (W4A), ACM Press (2008), 128--129.
[29]
Takagi, H., Kawanaka, S., Kobayashi, M., Itoh, T., and Asakawa, C. Social accessibility: achieving accessibility through collaborative metadata authoring. In Proc. ASSETS 2008. ACM Press (2008), 193--200.
[30]
Chen, C. L. and Raman, T. V. AxsJAX: a talking translation bot using google IM: bringing web-2.0 applications to life. In Proc. the 2008 international cross-disciplinary conference on Web accessibility (W4A), ACM Press (2008), 54--56.
[31]
Goble, C., Harper, S., and Stevens, R. The travails of visually impaired web travellers. In Proceedings of the Eleventh ACM on Hypertext and Hypermedia. ACM Press (2000), 1--10.
[32]
Eclipse ACTF Accessibility Internet Browser, http://www.eclipse.org/actf/downloads/tools/aiBrowser/.
[33]
Kawanaka, S., Borodin, Y., Bigham, J. P., Lunn, D., Takagi, H. and Asakawa, C. Accessibility commons: a metadata infrastructure for web accessibility. In Proc. ASSETS 2008, ACM Press (2008), 153--160.
[34]
Kanayama, H., Nasukawa, T. and Watanabe, H. Deeper sentiment analysis using machine translation technology. In Proc. of the 20th international conference on Computational Linguistic', Association for Computational Linguistics (2004), 494.
[35]
Shaojian Zhu, Daisuke Sato, Hironobu Takagi, and Chieko Asakawa. Sasayaki: an augmented voice-based web browsing experience. In Proc. ASSETS 2010. ACM Press (2008), 279--280.
[36]
Wilson, J., Walker, B. N., Lindsay, J., Cambias, C., and Dellaert, F. SWAN: System for Wearable Audio Navigation. In Proc. ISWC 2007. IEEE Computer Society (2007), 1--8.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
CHI '11: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems
May 2011
3530 pages
ISBN:9781450302289
DOI:10.1145/1978942
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 07 May 2011

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Sasayaki
  2. auditory interface
  3. multiple voices
  4. voice augmented browsing
  5. web accessibility

Qualifiers

  • Research-article

Conference

CHI '11
Sponsor:

Acceptance Rates

CHI '11 Paper Acceptance Rate 410 of 1,532 submissions, 27%;
Overall Acceptance Rate 6,199 of 26,314 submissions, 24%

Upcoming Conference

CHI 2025
ACM CHI Conference on Human Factors in Computing Systems
April 26 - May 1, 2025
Yokohama , Japan

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)14
  • Downloads (Last 6 weeks)2
Reflects downloads up to 21 Dec 2024

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media