skip to main content
10.1145/3573381.3596159acmconferencesArticle/Chapter ViewAbstractPublication PagesimxConference Proceedingsconference-collections
research-article

Accessibility Research in Digital Audiovisual Media: What Has Been Achieved and What Should Be Done Next?

Published: 29 August 2023 Publication History

Abstract

The consumption of digital audiovisual media is a mainstay of many people’s lives. However, people with accessibility needs often have issues accessing this content. With a view to addressing this inequality, there exists a wide range of interventions that researchers have explored to bridge this accessibility gap. Despite this work, our understanding of the capability of these interventions is poor. In this paper, we address this through a systematic review of the literature, creating a dataset of and analysing N = 181 scientific papers. We have found that certain areas have accrued a disproportionate amount of attention from the research community – for example, blind and visually impaired and d/Deaf and hard of hearing people account for of papers (N = 170). We describe challenges researchers have addressed, end-user communities of focus, and interventions examined. We conclude by evaluating gaps in the literature and areas that could use more focus on in the future.

Supplementary Material

The supplemental materials contain two files: * the codebook used to code the dataset, with explanations of what every category is and what every code means * an MS Excel spreadsheet with all 181 coded papers in our dataset (supplemental_files.zip)

References

[1]
Diana Affi, Joël Dumoulin, Marco Bertini, Elena Mugellini, Omar Abou Khaled, and Alberto Del Bimbo. 2015. SensiTV: Smart EmotioNal System for Impaired People’s TV. In Proceedings of the ACM International Conference on Interactive Experiences for TV and Online Video (Brussels, Belgium) (TVX ’15). Association for Computing Machinery, New York, NY, USA, 125–130. https://doi.org/10.1145/2745197.2755512
[2]
Chanchal Agrawal and Roshan L Peiris. 2021. I See What You’re Saying: A Literature Review of Eye Tracking Research in Communication of Deaf or Hard of Hearing Users. In The 23rd International ACM SIGACCESS Conference on Computers and Accessibility. ACM, New York, NY, USA, 1–13. https://doi.org/10.1145/3441852.3471209
[3]
Wataru Akahori, Tatsunori Hirai, Shunya Kawamura, and Shigeo Morishima. 2016. Region-of-Interest-Based Subtitle Placement Using Eye-Tracking Data of Multiple Viewers. In Proceedings of the ACM International Conference on Interactive Experiences for TV and Online Video (Chicago, Illinois, USA) (TVX ’16). Association for Computing Machinery, New York, NY, USA, 123–128. https://doi.org/10.1145/2932206.2933558
[4]
Akhter Al Amin, Abraham Glasser, Raja Kushalnagar, Christian Vogler, and Matt Huenerfauth. 2021. Preferences of Deaf or Hard of Hearing Users for Live-TV Caption Appearance. In Universal Access in Human-Computer Interaction. Access to Media, Learning and Assistive Environments, Margherita Antona and Constantine Stephanidis (Eds.). Springer International Publishing, Cham, 189–201. https://doi.org/10.1007/978-3-030-78095-1_15
[5]
Akhter Al Amin, Saad Hassan, and Matt Huenerfauth. 2021. Caption-Occlusion Severity Judgments across Live-Television Genres from Deaf and Hard-of-Hearing Viewers. In Proceedings of the 18th International Web for All Conference (Ljubljana, Slovenia) (W4A ’21). Association for Computing Machinery, New York, NY, USA, Article 26, 12 pages. https://doi.org/10.1145/3430263.3452429
[6]
Akhter Al Amin, Saad Hassan, and Matt Huenerfauth. 2021. Effect of Occlusion on Deaf and Hard of Hearing Users’ Perception of Captioned Video Quality. In Universal Access in Human-Computer Interaction. Access to Media, Learning and Assistive Environments, Margherita Antona and Constantine Stephanidis (Eds.). Springer International Publishing, Cham, 202–220.
[7]
Akhter Al Amin, Saad Hassan, Sooyeon Lee, and Matt Huenerfauth. 2022. Watch It, Don’t Imagine It: Creating a Better Caption-Occlusion Metric by Collecting More Ecologically Valid Judgments from DHH Viewers. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems (New Orleans, LA, USA) (CHI ’22). Association for Computing Machinery, New York, NY, USA, Article 459, 14 pages. https://doi.org/10.1145/3491102.3517681
[8]
Akher Al Amin, Joseph Mendis, Raja Kushalnagar, Christian Vogler, Sooyeon Lee, and Matt Huenerfauth. 2022. Deaf and Hard of Hearing Viewers’ Preference for Speaker Identifier Type in Live TV Programming. In Universal Access in Human-Computer Interaction. Novel Design Approaches and Technologies, Margherita Antona and Constantine Stephanidis (Eds.). Springer International Publishing, Cham, 200–211. https://doi.org/10.1007/978-3-031-05028-2_13
[9]
A. Ando, T. Imai, A. Kobayashi, H. Isono, and K. Nakabayashi. 2000. Real-time transcription system for simultaneous subtitling of Japanese broadcast news programs. IEEE Transactions on Broadcasting 46, 3 (2000), 189–196. https://doi.org/10.1109/11.892155
[10]
Mike Armstrong, Andy Brown, Michael Crabb, Chris J. Hughes, Rhianne Jones, and James Sandford. 2016. Understanding the Diverse Needs of Subtitle Users in a Rapidly Evolving Media Landscape. SMPTE Motion Imaging Journal 125, 9 (2016), 33–41. https://doi.org/10.5594/JMI.2016.2614919
[11]
Ali Selman Aydin, Shirin Feiz, Vikas Ashok, and IV Ramakrishnan. 2020. Towards Making Videos Accessible for Low Vision Screen Magnifier Users. In Proceedings of the 25th International Conference on Intelligent User Interfaces (Cagliari, Italy) (IUI ’20). Association for Computing Machinery, New York, NY, USA, 10–21. https://doi.org/10.1145/3377325.3377494
[12]
Ali Selman Aydin, Shirin Feiz, Vikas Ashok, and I V Ramakrishnan. 2020. A Saliency-Driven Video Magnifier for People with Low Vision. In Proceedings of the 17th International Web for All Conference (Taipei, Taiwan) (W4A ’20). Association for Computing Machinery, New York, NY, USA, Article 6, 2 pages. https://doi.org/10.1145/3371300.3383356
[13]
Ali Selman Aydin, Yu-Jung Ko, Utku Uckun, IV Ramakrishnan, and Vikas Ashok. 2021. Non-Visual Accessibility Assessment of Videos. In Proceedings of the 30th ACM International Conference on Information and Knowledge Management (Virtual Event, Queensland, Australia) (CIKM ’21). Association for Computing Machinery, New York, NY, USA, 58–67. https://doi.org/10.1145/3459637.3482457
[14]
Liam Bannon, Jeffrey Bardzell, and Susanne Bødker. 2018. Reimagining participatory design. Interactions 26, 1 (Dec. 2018), 26–32. https://doi.org/10.1145/3292015
[15]
Gökçe Elif Baykal, Maarten Van Mechelen, and Eva Eriksson. 2020. Collaborative Technologies for Children with Special Needs. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, 1–13. https://doi.org/10.1145/3313831.3376291
[16]
Larwan Berke, Khaled Albusays, Matthew Seita, and Matt Huenerfauth. 2019. Preferred Appearance of Captions Generated by Automatic Speech Recognition for Deaf and Hard-of-Hearing Viewers. In Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems (Glasgow, Scotland Uk) (CHI EA ’19). Association for Computing Machinery, New York, NY, USA, 1–6. https://doi.org/10.1145/3290607.3312921
[17]
Larwan Berke, Matthew Seita, and Matt Huenerfauth. 2020. Deaf and Hard-of-Hearing Users’ Prioritization of Genres of Online Video Content Requiring Accurate Captions. In Proceedings of the 17th International Web for All Conference (Taipei, Taiwan) (W4A ’20). Association for Computing Machinery, New York, NY, USA, Article 3, 12 pages. https://doi.org/10.1145/3371300.3383337
[18]
Alexy Bhowmick and Shyamanta M. Hazarika. 2017. An insight into assistive technology for the visually impaired and blind people: state-of-the-art and future trends. Journal on Multimodal User Interfaces 11, 2 (Jan. 2017), 149–172. https://doi.org/10.1007/s12193-016-0235-6
[19]
Pradipta Biswas, Pat Langdon, Carlos Duarte, and José Coelho. 2011. Multimodal adaptation through simulation for digital TV interface. In Proceedings of the 9th European Conference on Interactive TV and Video. ACM, New York, NY, USA, 231–234. https://doi.org/10.1145/2000119.2000167
[20]
Aditya Bodi, Pooyan Fazli, Shasta Ihorn, Yue-Ting Siu, Andrew T Scott, Lothar Narins, Yash Kant, Abhishek Das, and Ilmi Yoon. 2021. Automated Video Description for Blind and Low Vision Users. In Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI EA ’21). Association for Computing Machinery, New York, NY, USA, Article 230, 7 pages. https://doi.org/10.1145/3411763.3451810
[21]
Eva Brandt, Erling Björgvinsson, Per-Anders Hillgren, Viktor Bergqvist, and Marcus Emilson. 2002. PDA’s, Barcodes and Video-Films for Continuous Learning at an Intensive Care Unit. In Proceedings of the Second Nordic Conference on Human-Computer Interaction (Aarhus, Denmark) (NordiCHI ’02). Association for Computing Machinery, New York, NY, USA, 293–294. https://doi.org/10.1145/572020.572070
[22]
Andy Brown, Rhia Jones, Mike Crabb, James Sandford, Matthew Brooks, Mike Armstrong, and Caroline Jay. 2015. Dynamic Subtitles: The User Experience. In Proceedings of the ACM International Conference on Interactive Experiences for TV and Online Video (Brussels, Belgium) (TVX ’15). Association for Computing Machinery, New York, NY, USA, 103–112. https://doi.org/10.1145/2745197.2745204
[23]
Andy Brown, Jayson Turner, Jake Patterson, Anastasia Schmitz, Mike Armstrong, and Maxine Glancy. 2017. Subtitles in 360-Degree Video. In Adjunct Publication of the 2017 ACM International Conference on Interactive Experiences for TV and Online Video (Hilversum, The Netherlands) (TVX ’17 Adjunct). Association for Computing Machinery, New York, NY, USA, 3–8. https://doi.org/10.1145/3084289.3089915
[24]
Emeline Brulé, Brianna J. Tomlinson, Oussama Metatla, Christophe Jouffrais, and Marcos Serrano. 2020. Review of Quantitative Empirical Evaluations of Technology for People with Visual Impairments. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, 1–14. https://doi.org/10.1145/3313831.3376749
[25]
Denis Burnham, Greg Leigh, William Noble, Caroline Jones, Michael Tyler, Leonid Grebennikov, and Alex Varley. 2008. Parameters in Television Captioning for Deaf and Hard-of-Hearing Adults: Effects of Caption Rate Versus Text Reduction on Comprehension. The Journal of Deaf Studies and Deaf Education 13, 3 (03 2008), 391–404. https://doi.org/10.1093/deafed/enn003 arXiv:https://academic.oup.com/jdsde/article-pdf/13/3/391/1070640/enn003.pdf
[26]
Virginia Pinto Campos, Luiz Marcos G. Goncalves, and Tiago Maritan U. de Araujo. 2017. Applying audio description for context understanding of surveillance videos by people with visual impairments. In 2017 14th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS). IEEE, New York, NY, USA, 1–5. https://doi.org/10.1109/avss.2017.8078530
[27]
Marina Ramos Caro. 2016. Testing audio narration: the emotional impact of language in audio description. Perspectives 24, 4 (March 2016), 606–634. https://doi.org/10.1080/0907676x.2015.1120760
[28]
Patrick Carrington, Amy Hurst, and Shaun K. Kane. 2014. Wearables and chairables. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, 3103––3112. https://doi.org/10.1145/2556288.2557237
[29]
Johnny Carroll and Kevin McLaughlin. 2005. Closed Captioning in Distance Education. J. Comput. Sci. Coll. 20, 4 (apr 2005), 183–189.
[30]
Jade Cartwright and Kym A. E. Elliott. 2009. Promoting strategic television viewing in the context of progressive language impairment. Aphasiology 23, 2 (2009), 266–285. https://doi.org/10.1080/02687030801942932
[31]
Daniel Carvalho, Telmo Silva, and Jorge Abreu. 2021. TV Remote Control and Older Adults: A Systematic Literature Review. In Communications in Computer and Information Science. Springer International Publishing, Cham, 119–133. https://doi.org/10.1007/978-3-030-81996-5_9
[32]
Teresa Chambel, Sérgio Neves, Celso Sousa, and Rafael Francisco. 2010. Synesthetic Video: Hearing Colors, Seeing Sounds. In Proceedings of the 14th International Academic MindTrek Conference: Envisioning Future Media Environments (Tampere, Finland) (MindTrek ’10). Association for Computing Machinery, New York, NY, USA, 130–133. https://doi.org/10.1145/1930488.1930515
[33]
Pierre-Antoine Champin, Benoît Encelle, Nicholas W. D. Evans, Magali O.-Beldame, Yannick Prié, and Raphaël Troncy. 2010. Towards Collaborative Annotation for Video Accessibility. In Proceedings of the 2010 International Cross Disciplinary Conference on Web Accessibility (W4A) (Raleigh, North Carolina) (W4A ’10). Association for Computing Machinery, New York, NY, USA, Article 17, 4 pages. https://doi.org/10.1145/1805986.1806010
[34]
Claude Chapdelaine. 2010. In-Situ Study of Blind Individuals Listening to Audio-Visual Contents. In Proceedings of the 12th International ACM SIGACCESS Conference on Computers and Accessibility (Orlando, Florida, USA) (ASSETS ’10). Association for Computing Machinery, New York, NY, USA, 59–66. https://doi.org/10.1145/1878803.1878816
[35]
Claude Chapdelaine. 2012. Specialized DVD Player to Render Audio Description and Its Usability Performance. In Proceedings of the 14th International ACM SIGACCESS Conference on Computers and Accessibility (Boulder, Colorado, USA) (ASSETS ’12). Association for Computing Machinery, New York, NY, USA, 203–204. https://doi.org/10.1145/2384916.2384954
[36]
Claude Chapdelaine and Langis Gagnon. 2009. Accessible Videodescription On-Demand. In Proceedings of the 11th International ACM SIGACCESS Conference on Computers and Accessibility (Pittsburgh, Pennsylvania, USA) (Assets ’09). Association for Computing Machinery, New York, NY, USA, 221–222. https://doi.org/10.1145/1639642.1639685
[37]
Mario Montagud Climent, Olga Soler-Vilageliu, Isaac Fraile Vila, and Sergi Fernandez Langa. 2021. VR360 Subtitling: Requirements, Technology and User Experience. IEEE Access 9 (2021), 2819–2838. https://doi.org/10.1109/access.2020.3047377
[38]
Daniel Costa and Carlos Duarte. 2019. Personalized and Accessible TV Interaction for People with Visual Impairments. In Proceedings of the 16th International Web for All Conference (San Francisco, CA, USA) (W4A ’19). Association for Computing Machinery, New York, NY, USA, Article 24, 4 pages. https://doi.org/10.1145/3315002.3317566
[39]
Rostand Costa, Tiago Maritan, Renan Soares, Vinicius Veríssimo, Suanny Vieira, Alexandre Santos, Manuella Aschoff, and Guido Lemos. 2018. An Open and Extensible Platform for Machine Translation of Spoken Languages into Sign Languages. In Applications and Usability of Interactive Television, María José Abásolo, Jorge Abreu, Pedro Almeida, and Telmo Silva (Eds.). Springer International Publishing, Cham, 161–176.
[40]
Enrique Costa-Montenegro, Fátima M. García-Doval, Jonathan Juncal-Martínez, and Belén Barragáns-Martínez. 2015. SubTitleMe, subtitles in cinemas in mobile devices. Universal Access in the Information Society 15, 3 (June 2015), 461–472. https://doi.org/10.1007/s10209-015-0420-5
[41]
Francisco M. Costela, Stephanie M. Reeves, and Russell L. Woods. 2021. The Effect of Zoom Magnification and Large Display on Video Comprehension in Individuals With Central Vision Loss. Translational Vision Science and Technology 10, 8 (July 2021), 30. https://doi.org/10.1167/tvst.10.8.30
[42]
Francisco M Costela, Stephanie M Reeves, and Russell L Woods. 2021. An implementation of Bubble Magnification did not improve the video comprehension of individuals with central vision loss. Ophthalmic and Physiological Optics 41, 4 (March 2021), 842–852. https://doi.org/10.1111/opo.12797
[43]
Michael Crabb, Rhianne Jones, Mike Armstrong, and Chris J. Hughes. 2015. Online News Videos: The UX of Subtitle Position. In Proceedings of the 17th International ACM SIGACCESS Conference on Computers and Accessibility (Lisbon, Portugal) (ASSETS ’15). Association for Computing Machinery, New York, NY, USA, 215–222. https://doi.org/10.1145/2700648.2809866
[44]
Humphrey Curtis, Timothy Neate, and Carlota Vazquez Gonzalez. 2022. State of the Art in AAC: A Systematic Review and Taxonomy. In The 24th International ACM SIGACCESS Conference on Computers and Accessibility. ACM, New York, NY, USA, 1–22. https://doi.org/10.1145/3517428.3544810
[45]
Tiago Maritan U. de Araújo, Felipe L.S. Ferreira, Danilo A.N.S. Silva, Leonardo D. Oliveira, Eduardo L. Falcão, Leonardo A. Domingues, Vandhuy F. Martins, Igor A.C. Portela, Yurika S. Nóbrega, Hozana R.G. Lima, Guido L. Souza Filho, Tatiana A. Tavares, and Alexandre N. Duarte. 2014. An approach to generate and embed sign language video tracks into multimedia contents. Information Sciences 281 (2014), 762–780. https://doi.org/10.1016/j.ins.2014.04.008 Multimedia Modeling.
[46]
Alex de Souza Vieira, Álan Lívio V. Guedes, Daniel de Sousa Moraes, Lucas Ribeiro Madeira, Sérgio Colcher, and Carlos de S. Soares Neto. 2020. ListeningTV: Accessible Video using Interactive Audio Descriptions. In Anais Estendidos do XXVI Simpósio Brasileiro de Sistemas Multimídia e Web (WebMedia 2020). Sociedade Brasileira de Computação - SBC, Brazil, 71–74. https://doi.org/10.5753/webmedia_estendido.2020.13065
[47]
Patricio Domingues, Ruben Nogueira, José Carlos Francisco, and Miguel Frade. 2020. Post-mortem digital forensic artifacts of TikTok Android App. In Proceedings of the 15th International Conference on Availability, Reliability and Security. ACM, New York, NY, USA, 1–8. https://doi.org/10.1145/3407023.3409203
[48]
Joël Dumoulin, Diana Affi, Elena Mugellini, Omar Abou Khaled, Marco Bertini, and Alberto Del Bimbo. 2015. Movie’s Affect Communication Using Multisensory Modalities. In Proceedings of the 23rd ACM International Conference on Multimedia (Brisbane, Australia) (MM ’15). Association for Computing Machinery, New York, NY, USA, 739–740. https://doi.org/10.1145/2733373.2807965
[49]
Benoît Encelle, Magali Ollagnier Beldame, and Yannick Prié. 2013. Towards the Usage of Pauses in Audio-Described Videos. In Proceedings of the 10th International Cross-Disciplinary Conference on Web Accessibility (Rio de Janeiro, Brazil) (W4A ’13). Association for Computing Machinery, New York, NY, USA, Article 31, 4 pages. https://doi.org/10.1145/2461121.2461130
[50]
Benoît Encelle, Pierre-Antoine Champin, Yannick Prié, and Olivier Aubert. 2011. Models for Video Enrichment. In Proceedings of the 11th ACM Symposium on Document Engineering (Mountain View, California, USA) (DocEng ’11). Association for Computing Machinery, New York, NY, USA, 85–88. https://doi.org/10.1145/2034691.2034710
[51]
Benoît Encelle, Magali Ollagnier-Beldame, Stéphanie Pouchot, and Yannick Prié. 2011. Annotation-Based Video Enrichment for Blind People: A Pilot Study on the Use of Earcons and Speech Synthesis. In The Proceedings of the 13th International ACM SIGACCESS Conference on Computers and Accessibility (Dundee, Scotland, UK) (ASSETS ’11). Association for Computing Machinery, New York, NY, USA, 123–130. https://doi.org/10.1145/2049536.2049560
[52]
Jose Enrique Garcia, Alfonso Ortega, Eduardo Lleida, Tomas Lozano, Emiliano Bernues, and Daniel Sanchez. 2009. Audio and text synchronization for TV news subtitling based on Automatic Speech Recognition. In 2009 IEEE International Symposium on Broadband Multimedia Systems and Broadcasting. IEEE, New York, NY, USA, 1–6. https://doi.org/10.1109/ISBMSB.2009.5133758
[53]
Maria Federico and Marco Furini. 2012. Enhancing Learning Accessibility through Fully Automatic Captioning. In Proceedings of the International Cross-Disciplinary Conference on Web Accessibility (Lyon, France) (W4A ’12). Association for Computing Machinery, New York, NY, USA, Article 40, 4 pages. https://doi.org/10.1145/2207016.2207053
[54]
Deborah I. Fels, John Patrick Udo, Jonas E. Diamond, and Jeremy I. Diamond. 2006. A Comparison of Alternative Narrative Approaches to Video Description for Animated Comedy. Journal of Visual Impairment and Blindness 100, 5 (May 2006), 295–305. https://doi.org/10.1177/0145482x0610000507
[55]
Itamar Rocha Filho, Felipe Honorato, J. Wallace Lucena, J. Pedro Teixeira, and Tiago Maritan. 2021. An Approach for Automatic Description of Characters for Blind People. In Proceedings of the Brazilian Symposium on Multimedia and the Web (Belo Horizonte, Minas Gerais, Brazil) (WebMedia ’21). Association for Computing Machinery, New York, NY, USA, 53–56. https://doi.org/10.1145/3470482.3479617
[56]
Shadiqin Firdus, Wan Fatimah Wan Ahmad, and Josefina Barnachea Janier. 2012. Development of Audio Video Describer using narration to visualize movie film for blind and visually impaired children. In 2012 International Conference on Computer and Information Science (ICCIS). IEEE, New York, NY, USA, 1068–1072. https://doi.org/10.1109/iccisci.2012.6297184
[57]
Joseph L. Fleiss and Jacob Cohen. 1973. The Equivalence of Weighted Kappa and the Intraclass Correlation Coefficient as Measures of Reliability. Educational and Psychological Measurement 33, 3 (Oct. 1973), 613–619. https://doi.org/10.1177/001316447303300309
[58]
Louise Fryer and Jonathan Freeman. 2013. Cinematic language and the description of film: keeping AD users in the frame. Perspectives 21, 3 (Sept. 2013), 412–426. https://doi.org/10.1080/0907676x.2012.693108
[59]
Louise Fryer and Jonathan Freeman. 2013. Visual Impairment and Presence: Measuring the Effect of Audio Description. In Proceedings of the 2013 Inputs-Outputs Conference: An Interdisciplinary Conference on Engagement in HCI and Performance (Brighton, United Kingdom) (Inputs-Outputs ’13). Association for Computing Machinery, New York, NY, USA, Article 4, 5 pages. https://doi.org/10.1145/2557595.2557599
[60]
Liam Funnell, Isabel Garriock, Ben Shirley, and Tracey Williamson. 2019. Dementia-friendly design of television news broadcasts. Journal of Enabling Technologies 13, 3 (Sept. 2019), 137–149. https://doi.org/10.1108/jet-02-2018-0009
[61]
L. Gagnon, C. Chapdelaine, D. Byrns, S. Foucher, M. Héritier, and V. Gupta. 2010. A computer-vision-assisted system for Videodescription scripting. In 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops. IEEE, New York, NY, USA, 41–48. https://doi.org/10.1109/CVPRW.2010.5543575
[62]
Ángel García-Crespo, José Luis López-Cuadrado, and Israel González-Carrasco. 2016. Accesibility on VoD Platforms via Mobile Devices. In Communications in Computer and Information Science. Springer International Publishing, Cham, 149–160. https://doi.org/10.1007/978-3-319-38907-3_12
[63]
Angel Garcia-Crespo, Jose Luis Lopez-Cuadrado, and Israel Gonzalez-Carrasco. 2016. Accesibility on VoD Platforms via Mobile Devices. In Applications and Usability of Interactive TV, Maria Jose Abasolo, Francisco J. Perales, and Antoni Bibiloni (Eds.). Springer International Publishing, Cham, 149–160.
[64]
Angel García-Crespo, Mariuxi Montes-Chunga, Carlos Alberto Matheus-Chacin, and Ines Garcia-Encabo. 2018. Increasing the autonomy of deafblind individuals through direct access to content broadcasted on digital terrestrial television. Assistive Technology 32, 5 (Dec. 2018), 268–276. https://doi.org/10.1080/10400435.2018.1543219
[65]
Angel García-Crespo, Mariuxi Montes-Chunga, Carlos Alberto Matheus-Chacin, and Ines Garcia-Encabo. 2020. Increasing the autonomy of deafblind individuals through direct access to content broadcasted on digital terrestrial television. Assistive Technology 32, 5 (2020), 268–276. https://doi.org/10.1080/10400435.2018.1543219 arXiv:https://doi.org/10.1080/10400435.2018.1543219PMID: 30592435.
[66]
D. Gaw, D. Morris, and K. Salisbury. 2006. Haptically Annotated Movies: Reaching Out and Touching the Silver Screen. In 2006 14th Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems. IEEE, New York, NY, USA, 287–288. https://doi.org/10.1109/haptic.2006.1627106
[67]
Olivia Gerber-Morón, Agnieszka Szarkowska, and Bencie Woll. 2018. The impact of text segmentation on subtitle reading. Journal of Eye Movement Research 11, 4 (jun 2018), 1–18. https://doi.org/10.16910/11.4.2
[68]
Cagatay Goncu and Daniel J. Finnegan. 2021. ‘Did You See That!?’ Enhancing the Experience of Sports Media Broadcast for Blind People. In Human-Computer Interaction – INTERACT 2021, Carmelo Ardito, Rosa Lanzilotti, Alessio Malizia, Helen Petrie, Antonio Piccinno, Giuseppe Desolda, and Kori Inkpen (Eds.). Springer International Publishing, Cham, 396–417.
[69]
I. Gonzalez-Carrasco, L. Puente, B. Ruiz-Mezcua, and J. L. Lopez-Cuadrado. 2019. Sub-Sync: Automatic Synchronization of Subtitles in the Broadcasting of True Live programs in Spanish. IEEE Access 7 (2019), 60968–60983. https://doi.org/10.1109/access.2019.2915581
[70]
Benjamin M. Gorman, Michael Crabb, and Michael Armstrong. 2021. Adaptive Subtitles: Preferences and Trade-Offs in Real-Time Media Adaption. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI ’21). Association for Computing Machinery, New York, NY, USA, Article 733, 11 pages. https://doi.org/10.1145/3411764.3445509
[71]
Michael Gower, Brent Shiver, Charu Pandhi, and Shari Trewin. 2018. Leveraging Pauses to Improve Video Captions. In Proceedings of the 20th International ACM SIGACCESS Conference on Computers and Accessibility (Galway, Ireland) (ASSETS ’18). Association for Computing Machinery, New York, NY, USA, 414–416. https://doi.org/10.1145/3234695.3241023
[72]
Trisha Greenhalgh and Richard Peacock. 2005. Effectiveness and efficiency of search methods in systematic reviews of complex evidence: audit of primary sources. BMJ 331, 7524 (Oct. 2005), 1064–1065. https://doi.org/10.1136/bmj.38636.593461.68
[73]
Brian Grellmann, Timothy Neate, Abi Roper, Stephanie Wilson, and Jane Marshall. 2018. Investigating Mobile Accessibility Guidance for People with Aphasia. In Proceedings of the 20th International ACM SIGACCESS Conference on Computers and Accessibility. ACM, New York, NY, USA, 410–413. https://doi.org/10.1145/3234695.3241011
[74]
Darren Guinness, Annika Muehlbradt, Daniel Szafir, and Shaun K. Kane. 2018. The Haptic Video Player: Using Mobile Robots to Create Tangible Video Annotations. In Proceedings of the 2018 ACM International Conference on Interactive Surfaces and Spaces (Tokyo, Japan) (ISS ’18). Association for Computing Machinery, New York, NY, USA, 203–211. https://doi.org/10.1145/3279778.3279805
[75]
S.R. Gulliver and G. Ghinea. 2002. Impact of captions on deaf and hearing perception of multimedia video clips. In Proceedings. IEEE International Conference on Multimedia and Expo. IEEE, New York, NY, USA, 753–756. https://doi.org/10.1109/icme.2002.1035891
[76]
S.R. Gulliver and G. Ghinea. 2003. How level and type of deafness affect user perception of multimedia video clips. Universal Access in the Information Society 2, 4 (Nov. 2003), 374–386. https://doi.org/10.1007/s10209-003-0067-5
[77]
William A. Hamilton, Oliver Garretson, and Andruid Kerne. 2014. Streaming on twitch. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, 1315––1324. https://doi.org/10.1145/2556288.2557048
[78]
Zdenek Hanzlicek, Jindrich Matousek, and Daniel Tihelka. 2008. Towards automatic audio track generation for Czech TV broadcasting: Initial experiments with subtitles-to-speech synthesis. In 2008 9th International Conference on Signal Processing. IEEE, New York, NY, USA, 2721–2724. https://doi.org/10.1109/icosp.2008.4697710
[79]
Richang Hong, Meng Wang, Mengdi Xu, Shuicheng Yan, and Tat-Seng Chua. 2010. Dynamic Captioning: Video Accessibility Enhancement for Hearing Impairment. In Proceedings of the 18th ACM International Conference on Multimedia (Firenze, Italy) (MM ’10). Association for Computing Machinery, New York, NY, USA, 421–430. https://doi.org/10.1145/1873951.1874013
[80]
Richang Hong, Meng Wang, Xiao-Tong Yuan, Mengdi Xu, Jianguo Jiang, Shuicheng Yan, and Tat-Seng Chua. 2011. Video Accessibility Enhancement for Hearing-Impaired Users. ACM Trans. Multimedia Comput. Commun. Appl. 7S, 1, Article 24 (nov 2011), 19 pages. https://doi.org/10.1145/2037676.2037681
[81]
Elfed Howells and David Jackson. 2021. Object-Based Media Report.
[82]
Yongtao Hu, Jan Kautz, Yizhou Yu, and Wenping Wang. 2015. Speaker-Following Video Subtitles. ACM Trans. Multimedia Comput. Commun. Appl. 11, 2, Article 32 (jan 2015), 17 pages. https://doi.org/10.1145/2632111
[83]
Yun Huang, Yifeng Huang, Na Xue, and Jeffrey P. Bigham. 2017. Leveraging Complementary Contributions of Different Workers for Efficient Crowdsourcing of Video Captions. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (Denver, Colorado, USA) (CHI ’17). Association for Computing Machinery, New York, NY, USA, 4617–4626. https://doi.org/10.1145/3025453.3026032
[84]
Chris Hughes, Mario Montagud Climent, and Peter tho Pesch. 2019. Disruptive Approaches for Subtitling in Immersive Environments. In Proceedings of the 2019 ACM International Conference on Interactive Experiences for TV and Online Video (Salford (Manchester), United Kingdom) (TVX ’19). Association for Computing Machinery, New York, NY, USA, 216–229. https://doi.org/10.1145/3317697.3325123
[85]
Chris J. Hughes, Mike Armstrong, Rhianne Jones, and Michael Crabb. 2015. Responsive Design for Personalised Subtitles. In Proceedings of the 12th International Web for All Conference (Florence, Italy) (W4A ’15). Association for Computing Machinery, New York, NY, USA, Article 8, 4 pages. https://doi.org/10.1145/2745555.2746650
[86]
Sushant Kafle, Becca Dingman, and Matt Huenerfauth. 2021. Deaf and Hard-of-Hearing Users Evaluating Designs for Highlighting Key Words in Educational Lecture Videos. ACM Trans. Access. Comput. 14, 4, Article 20 (oct 2021), 24 pages. https://doi.org/10.1145/3470651
[87]
Sushant Kafle and Matt Huenerfauth. 2017. Evaluating the Usability of Automatically Generated Captions for People Who Are Deaf or Hard of Hearing. In Proceedings of the 19th International ACM SIGACCESS Conference on Computers and Accessibility (Baltimore, Maryland, USA) (ASSETS ’17). Association for Computing Machinery, New York, NY, USA, 165–174. https://doi.org/10.1145/3132525.3132542
[88]
Sushant Kafle, Peter Yeung, and Matt Huenerfauth. 2019. Evaluating the Benefit of Highlighting Key Words in Captions for People Who Are Deaf or Hard of Hearing. In Proceedings of the 21st International ACM SIGACCESS Conference on Computers and Accessibility (Pittsburgh, PA, USA) (ASSETS ’19). Association for Computing Machinery, New York, NY, USA, 43–55. https://doi.org/10.1145/3308561.3353781
[89]
Claire Kearney-Volpe, Shannon Holloway, and Amy Hurst. 2019. Entertainment for All: Understanding Media Streaming Accessibility. In Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems (Glasgow, Scotland Uk) (CHI EA ’19). Association for Computing Machinery, New York, NY, USA, 1–6. https://doi.org/10.1145/3290607.3312882
[90]
Masatomo Kobayashi, Kentarou Fukuda, Hironobu Takagi, and Chieko Asakawa. 2009. Providing Synthesized Audio Description for Online Videos. In Proceedings of the 11th International ACM SIGACCESS Conference on Computers and Accessibility (Pittsburgh, Pennsylvania, USA) (Assets ’09). Association for Computing Machinery, New York, NY, USA, 249–250. https://doi.org/10.1145/1639642.1639699
[91]
Masatomo Kobayashi, Tohru Nagano, Kentarou Fukuda, and Hironobu Takagi. 2010. Describing Online Videos with Text-to-Speech Narration. In Proceedings of the 2010 International Cross Disciplinary Conference on Web Accessibility (W4A) (Raleigh, North Carolina) (W4A ’10). Association for Computing Machinery, New York, NY, USA, Article 29, 2 pages. https://doi.org/10.1145/1805986.1806025
[92]
Masatomo Kobayashi, Trisha O’Connell, Bryan Gould, Hironobu Takagi, and Chieko Asakawa. 2010. Are Synthesized Video Descriptions Acceptable?. In Proceedings of the 12th International ACM SIGACCESS Conference on Computers and Accessibility (Orlando, Florida, USA) (ASSETS ’10). Association for Computing Machinery, New York, NY, USA, 163–170. https://doi.org/10.1145/1878803.1878833
[93]
Dimitrios Konstantinidis, Kosmas Dimitropoulos, Kiriakos Stefanidis, Thanassis Kalvourtzis, Salim Gannoum, Nikolaos Kaklanis, Konstantinos Votis, Petros Daras, Sara Rovira-Esteva, Pilar Orero, Silvia Uribe, Francisco Moreno, Alvaro Llorente, Pablo Calleja, Maria Poveda-Villalón, Pasquale Andriani, Giuseppe Vitolo, Giuseppa Caruso, Nicolamaria Manes, Fabrizio Giacomelli, Jordi Fabregat, Francesc Mas, Jordi Mata, Stavros Skourtis, Chrysostomos Bourlis, Giuliano Frittelli, Emilio Ferreiro Lago, and Federico Alvarez. 2020. Developing Accessibility Multimedia Services: The Case of EasyTV. In Proceedings of the 13th ACM International Conference on PErvasive Technologies Related to Assistive Environments (Corfu, Greece) (PETRA ’20). Association for Computing Machinery, New York, NY, USA, Article 38, 8 pages. https://doi.org/10.1145/3389189.3397973
[94]
Kuno Kurzhals, Emine Cetinkaya, Yongtao Hu, Wenping Wang, and Daniel Weiskopf. 2017. Close to the Action: Eye-Tracking Evaluation of Speaker-Following Subtitles. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems (Denver, Colorado, USA) (CHI ’17). Association for Computing Machinery, New York, NY, USA, 6559–6568. https://doi.org/10.1145/3025453.3025772
[95]
Kuno Kurzhals, Fabian Göbel, Katrin Angerbauer, Michael Sedlmair, and Martin Raubal. 2020. A View on the Viewer: Gaze-Adaptive Captions for Videos. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu, HI, USA) (CHI ’20). Association for Computing Machinery, New York, NY, USA, 1–12. https://doi.org/10.1145/3313831.3376266
[96]
Raja Kushalnagar and Kesavan Kushalnagar. 2018. SubtitleFormatter: Making Subtitles Easier to Read for Deaf and Hard of Hearing Viewers on Personal Devices. In Computers Helping People with Special Needs, Klaus Miesenberger and Georgios Kouroupetroglou (Eds.). Springer International Publishing, Cham, 211–219.
[97]
Raja Kushalnagar, Matthew Seita, and Abraham Glasser. 2017. Closed ASL Interpreting for Online Videos. In Proceedings of the 14th International Web for All Conference. ACM, New York, NY, USA, 1–4. https://doi.org/10.1145/3058555.3058578
[98]
Raja S. Kushalnagar, Gary W. Behm, Joseph S. Stanislow, and Vasu Gupta. 2014. Enhancing Caption Accessibility through Simultaneous Multimodal Information: Visual-Tactile Captions. In Proceedings of the 16th International ACM SIGACCESS Conference on Computers and Accessibility (Rochester, New York, USA) (ASSETS ’14). Association for Computing Machinery, New York, NY, USA, 185–192. https://doi.org/10.1145/2661334.2661381
[99]
Raja S. Kushalnagar, Walter S. Lasecki, and Jeffrey P. Bigham. 2013. Captions versus Transcripts for Online Video Content. In Proceedings of the 10th International Cross-Disciplinary Conference on Web Accessibility (Rio de Janeiro, Brazil) (W4A ’13). Association for Computing Machinery, New York, NY, USA, Article 32, 4 pages. https://doi.org/10.1145/2461121.2461142
[100]
Raja S. Kushalnagar, John J. Rivera, Warrance Yu, and Daniel S. Steed. 2014. AVD-LV: An Accessible Player for Captioned STEM Videos. In Proceedings of the 16th International ACM SIGACCESS Conference on Computers and Accessibility (Rochester, New York, USA) (ASSETS ’14). Association for Computing Machinery, New York, NY, USA, 287–288. https://doi.org/10.1145/2661334.2661353
[101]
Walter S. Lasecki, Raja Kushalnagar, and Jeffrey P. Bigham. 2014. Helping Students Keep up with Real-Time Captions by Pausing and Highlighting. In Proceedings of the 11th Web for All Conference (Seoul, Korea) (W4A ’14). Association for Computing Machinery, New York, NY, USA, Article 39, 8 pages. https://doi.org/10.1145/2596695.2596701
[102]
Daniel G. Lee, Deborah I. Fels, and John Patrick Udo. 2007. Emotive Captioning. Comput. Entertain. 5, 2, Article 11 (apr 2007), 15 pages. https://doi.org/10.1145/1279540.1279551
[103]
Hae-Na Lee and Vikas Ashok. 2021. Towards Enhancing Blind Users’ Interaction Experience with Online Videos via Motion Gestures. In Proceedings of the 32nd ACM Conference on Hypertext and Social Media (Virtual Event, USA) (HT ’21). Association for Computing Machinery, New York, NY, USA, 231–236. https://doi.org/10.1145/3465336.3475116
[104]
Franklin Mingzhe Li, Cheng Lu, Zhicong Lu, Patrick Carrington, and Khai N. Truong. 2022. An Exploration of Captioning Practices and Challenges of Individual Content Creators on YouTube for People with Hearing Impairments. Proceedings of the ACM on Human-Computer Interaction 6, CSCW1 (March 2022), 1–26. https://doi.org/10.1145/3512922
[105]
Xiangbo Li, Mohsen Amini Salehi, and Magdy Bayoumi. 2016. High Performance On-demand Video Transcoding Using Cloud Services. In 2016 16th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGrid). IEEE Computer Society, New York, NY, USA, 600–603. https://doi.org/10.1109/CCGrid.2016.50
[106]
Xingyu Liu, Patrick Carrington, Xiang ’Anthony’ Chen, and Amy Pavel. 2021. What Makes Videos Accessible to Blind and Visually Impaired People?. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI ’21). Association for Computing Machinery, New York, NY, USA, Article 272, 14 pages. https://doi.org/10.1145/3411764.3445233
[107]
Peter Ljunglof, Sandra Derbring, and Maria Olsson. 2012. A Free and Open-Source Tool That Reads Movie Subtitles Aloud. In Proceedings of the Third Workshop on Speech and Language Processing for Assistive Technologies (Montreal, Canada) (SLPAT ’12). Association for Computational Linguistics, USA, 1–4.
[108]
Mariana Lopez, Gavin Kearney, and Krisztian Hofstadter. 2021. Enhancing Audio Description: Inclusive Cinematic Experiences Through Sound Design. Journal of Audiovisual Translation 4, 1 (Oct. 2021), 157–182. https://doi.org/10.47476/jat.v4i1.2021.154
[109]
Kelly Mack, Emma McDonnell, Dhruv Jain, Lucy Lu Wang, Jon E. Froehlich, and Leah Findlater. 2021. What Do We Mean by “Accessibility Research”?. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, 1–18. https://doi.org/10.1145/3411764.3445412
[110]
Alejandro Martín, Israel González-Carrasco, Victor Rodriguez-Fernandez, Mónica Souto-Rico, David Camacho, and Belén Ruiz-Mezcua. 2021. Deep-Sync: A novel deep learning-based tool for semantic-aware subtitling synchronisation. Neural Computing and Applications 33, 1 (Feb. 2021), 1–15. https://doi.org/10.1007/s00521-021-05751-y
[111]
Anna Matamala and Aline Remael. 2014. Audio-description reloaded: An analysis of visual scenes in 2012 and Hero. Translation Studies 8, 1 (Aug. 2014), 63–81. https://doi.org/10.1080/14781700.2014.943678
[112]
Jindrich Matousek, Zdenek Hanzlicek, Daniel Tihelka, and Martin Mener. 2010. Automatic dubbing of TV programmes for the hearing impaired. In IEEE 10th INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING PROCEEDINGS. IEEE, New York, NY, USA, 589–592. https://doi.org/10.1109/icosp.2010.5655861
[113]
Tara Matthews, Janette Fong, and Jennifer Mankoff. 2005. Visualizing Non-Speech Sounds for the Deaf. In Proceedings of the 7th International ACM SIGACCESS Conference on Computers and Accessibility (Baltimore, MD, USA) (Assets ’05). Association for Computing Machinery, New York, NY, USA, 52–59. https://doi.org/10.1145/1090785.1090797
[114]
Troy McDaniel, Lakshmie Narayan Viswanathan, and Sethuraman Panchanathan. 2013. An evaluation of haptic descriptions for audio described films for individuals who are blind. In 2013 IEEE International Conference on Multimedia and Expo (ICME). IEEE, New York, NY, USA, 1–6. https://doi.org/10.1109/ICME.2013.6607554
[115]
Mark McGill, Florian Mathis, Mohamed Khamis, and Julie Williamson. 2020. Augmenting TV Viewing Using Acoustically Transparent Auditory Headsets. In ACM International Conference on Interactive Media Experiences (Cornella, Barcelona, Spain) (IMX ’20). Association for Computing Machinery, New York, NY, USA, 34–44. https://doi.org/10.1145/3391614.3393650
[116]
Mark McGill, John H. Williamson, and Stephen Brewster. 2016. Examining The Role of Smart TVs and VR HMDs in Synchronous At-a-Distance Media Consumption. ACM Transactions on Computer-Human Interaction 23, 5 (Nov. 2016), 1–57. https://doi.org/10.1145/2983530
[117]
Apostolos Meliones and Stavros Maidonis. 2020. DALÍ. In Proceedings of the 13th ACM International Conference on PErvasive Technologies Related to Assistive Environments. ACM, New York, NY, USA, 1–9. https://doi.org/10.1145/3389189.3397972
[118]
Bogdan Mocanu and Ruxandra Tapu. 2021. Automatic Subtitle Placement Through Active Speaker Identification in Multimedia Documents. In 2021 International Conference on e-Health and Bioengineering (EHB). IEEE, New York, NY, USA, 1–4. https://doi.org/10.1109/ehb52898.2021.9657604
[119]
Bogdan Mocanu, Ruxandra Tapu, and Titus Zaharia. 2019. Enhancing the Accessibility of Hearing Impaired to Video Content through Fully Automatic Dynamic Captioning. In 2019 E-Health and Bioengineering Conference (EHB). IEEE, New York, NY, USA, 1–4. https://doi.org/10.1109/ehb47216.2019.8970038
[120]
Mario Montagud, Fernando Boronat, Juan González, and Javier Pastor. 2017. Web-Based Platform for Subtitles Customization and Synchronization in Multi-Screen Scenarios. In Adjunct Publication of the 2017 ACM International Conference on Interactive Experiences for TV and Online Video (Hilversum, The Netherlands) (TVX ’17 Adjunct). Association for Computing Machinery, New York, NY, USA, 81–82. https://doi.org/10.1145/3084289.3089913
[121]
Mario Montagud, Fernando Boronat, Javier Pastor, and Dani Marfil. 2020. Web-based platform for a customizable and synchronized presentation of subtitles in single- and multi-screen scenarios. Multimedia Tools and Applications 79, 29-30 (May 2020), 21889–21923. https://doi.org/10.1007/s11042-020-08955-x
[122]
Mario Montagud, Issac Fraile, Juan A. Nuñez, and Sergi Fernández. 2018. ImAc: Enabling Immersive, Accessible and Personalized Media Experiences. In Proceedings of the 2018 ACM International Conference on Interactive Experiences for TV and Online Video (SEOUL, Republic of Korea) (TVX ’18). Association for Computing Machinery, New York, NY, USA, 245–250. https://doi.org/10.1145/3210825.3213570
[123]
Mario Montagud, Jaume Segura-Garcia, J. Antonio De Rus, and Rafael Fayos Jordán. 2020. Towards an Immersive and Accessible Virtual Reconstruction of Theaters from the Early Modern: Bringing Back Cultural Heritage from the Past. In ACM International Conference on Interactive Media Experiences (Cornella, Barcelona, Spain) (IMX ’20). Association for Computing Machinery, New York, NY, USA, 143–147. https://doi.org/10.1145/3391614.3399390
[124]
Francisco Montero, Víctor López-Jaquero, and Pascual González. 2015. Accessibility: An Open Issue in the Interactive Digital Television. In Applications and Usability of Interactive TV, María José Abásolo and Raoni Kulesza (Eds.). Springer International Publishing, Cham, 105–119.
[125]
Martez Mott, John Tang, Shaun Kane, Edward Cutrell, and Meredith Ringel Morris. 2020. “I just went into it assuming that I wouldn't be able to have the full experience”. In The 22nd International ACM SIGACCESS Conference on Computers and Accessibility. ACM, New York, NY, USA, 1–13. https://doi.org/10.1145/3373625.3416998
[126]
Rosiana Natalie, Jolene Loh, Huei Suen Tan, Joshua Tseng, Ian Luke Yi-Ren Chan, Ebrima H Jarjue, Hernisa Kacorri, and Kotaro Hara. 2021. The Efficacy of Collaborative Authoring of Video Scene Descriptions. In Proceedings of the 23rd International ACM SIGACCESS Conference on Computers and Accessibility (Virtual Event, USA) (ASSETS ’21). Association for Computing Machinery, New York, NY, USA, Article 17, 15 pages. https://doi.org/10.1145/3441852.3471201
[127]
Timothy Neate, Abi Roper, Stephanie Wilson, Jane Marshall, and Madeline Cruice. 2020. CreaTable Content and Tangible Interaction in Aphasia. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. ACM, New York, NY, USA, 1–14. https://doi.org/10.1145/3313831.3376490
[128]
Tam V. Nguyen, Mengdi Xu, Guangyu Gao, Mohan Kankanhalli, Qi Tian, and Shuicheng Yan. 2013. Static Saliency vs. Dynamic Saliency: A Comparative Study. In Proceedings of the 21st ACM International Conference on Multimedia (Barcelona, Spain) (MM ’13). Association for Computing Machinery, New York, NY, USA, 987–996. https://doi.org/10.1145/2502081.2502128
[129]
Simon Nielsen and Hans-Heinrich Bothe. 2007. SUBPAL: A Device for Reading Aloud Subtitles from Television and Cinema. In Proceedings of the Conference and Workshop on Assistive Technologies for People with Vision and Hearing Impairments: Assistive Technology for All Ages (CVHI-2007), Granada, Spain, 28th - 31th August, 2007(CEUR Workshop Proceedings, Vol. 415), Marion A. Hersh and James Ohene-Djan (Eds.). CEUR-WS.org, Leibniz, Germany, 1–7. https://ceur-ws.org/Vol-415/paper17.pdf
[130]
Francisco Nunes, Maureen Kerwin, and Paula Alexandra Silva. 2012. Design Recommendations for Tv User Interfaces for Older Adults: Findings from the ECAALYX Project. In Proceedings of the 14th International ACM SIGACCESS Conference on Computers and Accessibility (Boulder, Colorado, USA) (ASSETS ’12). Association for Computing Machinery, New York, NY, USA, 41–48. https://doi.org/10.1145/2384916.2384924
[131]
J. Ohene-Djan and R. Shipsey. 2006. E- Subtitles: Emotional Subtitles as a Technology to Assist the Deaf and Hearing-Impaired when Learning from Television and Film. In Sixth IEEE International Conference on Advanced Learning Technologies (ICALT ’06). IEEE, New York, NY, USA, 464–466. https://doi.org/10.1109/icalt.2006.1652472
[132]
Rita Oliveira, Jorge Ferraz de Abreu, and Ana Margarida Almeida. 2014. Promoting IPTV Accessibility for Visually Impaired Users: Implementation of an Adapted Service. Procedia Computer Science 27 (2014), 113–122. https://doi.org/10.1016/j.procs.2014.02.014 5th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-exclusion, DSAI 2013.
[133]
Rita Oliveira, Jorge Ferraz de Abreu, and Ana Margarida Almeida. 2016. Audio Description in Interactive Television (iTV): Proposal of a Collaborative and Voluntary Approach. Procedia Computer Science 100 (2016), 935–940. https://doi.org/10.1016/j.procs.2016.09.252 International Conference on ENTERprise Information Systems/International Conference on Project MANagement/International Conference on Health and Social Care Information Systems and Technologies, CENTERIS/ProjMAN / HCist 2016.
[134]
Rita Oliveira, Jorge Ferraz de Abreu, and Ana Margarida Almeida. 2016. Promoting interactive television (iTV) accessibility: an adapted service for users with visual impairments. Universal Access in the Information Society 16, 3 (July 2016), 533–544. https://doi.org/10.1007/s10209-016-0482-z
[135]
Rita Oliveira, Jorge Ferraz de Abreu, and Ana Margarida Almeida. 2018. Audio Description of Television Programs: A Voluntary Production Approach. In Applications and Usability of Interactive Television, María José Abásolo, Jorge Abreu, Pedro Almeida, and Telmo Silva (Eds.). Springer International Publishing, Cham, 150–160.
[136]
Rita Oliveira, Jorge Ferraz de Abreu, and Ana Margarida Almeida. 2022. Audiovisual translation models for visually impaired users of Interactive Television (iTV). Procedia Computer Science 196 (2022), 647–654. https://doi.org/10.1016/j.procs.2021.12.060 International Conference on ENTERprise Information Systems / ProjMAN - International Conference on Project MANagement / HCist - International Conference on Health and Social Care Information Systems and Technologies 2021.
[137]
Rita Oliveira, Jorge Ferraz de Abreu, and Ana Margarida Pisco Almeida. 2011. An iTV Audio Description Service: Suggesting Requirements and Features for Visually Impaired Users. In ENTERprise Information Systems, Maria Manuela Cruz-Cunha, João Varajão, Philip Powell, and Ricardo Martinho (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 59–68. https://doi.org/10.1007/978-3-642-24352-3_7
[138]
Rita Oliveira, Jorge Ferraz de Abreu, Margarida Almeida, and Bernardo Cardoso. 2016. Inclusive Approaches for Audiovisual Translation Production in Interactive Television (ITV). In Proceedings of the 7th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-Exclusion (Vila Real, Portugal) (DSAI 2016). Association for Computing Machinery, New York, NY, USA, 146–153. https://doi.org/10.1145/3019943.3019965
[139]
Rita Oliveira, Jorge Ferraz de Abreu, and Ana Margarida Almeida. 2011. An Approach to Identify Requirements for an ITV Audio Description Service. In Proceedings of the 9th European Conference on Interactive TV and Video (Lisbon, Portugal) (EuroITV ’11). Association for Computing Machinery, New York, NY, USA, 227–230. https://doi.org/10.1145/2000119.2000166
[140]
Rita Oliveira, Telmo Silva, Jorge Ferraz de Abreu, and Margarida Almeida. 2016. Automatic Identification in Accessible ITV Services: Proposal of an IPTV Interface. In Proceedings of the 7th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-Exclusion (Vila Real, Portugal) (DSAI 2016). Association for Computing Machinery, New York, NY, USA, 384–391. https://doi.org/10.1145/3019943.3019998
[141]
Pilar Orero, Marta Brescia-Zapata, and Chris Hughes. 2020. Evaluating subtitle readability in media immersive environments. In 9th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-exclusion. ACM, New York, NY, USA, 51–54. https://doi.org/10.1145/3439231.3440602
[142]
Pilar Orero, Marta Brescia-Zapata, and Chris Hughes. 2021. Evaluating Subtitle Readability in Media Immersive Environments. In 9th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-Exclusion (Online, Portugal) (DSAI 2020). Association for Computing Machinery, New York, NY, USA, 51–54. https://doi.org/10.1145/3439231.3440602
[143]
Andrew D. Ouzts, Nicole E. Snell, Prabudh Maini, and Andrew T. Duchowski. 2013. Determining Optimal Caption Placement Using Eye Tracking. In Proceedings of the 31st ACM International Conference on Design of Communication (Greenville, North Carolina, USA) (SIGDOC ’13). Association for Computing Machinery, New York, NY, USA, 189–190. https://doi.org/10.1145/2507065.2507100
[144]
Cristhian Pachacama and Diego Villamarín. 2020. Implementation of a Brain Computer Interface System to Allow TV Remote Control for People with Physical Disabilities. In Communications in Computer and Information Science. Springer International Publishing, Cham, 119–133. https://doi.org/10.1007/978-3-030-56574-9_8
[145]
Cristian Pamparău and Radu-Daniel Vatavu. 2022. The User Experience of Journeys in the Realm of Augmented Reality Television. In ACM International Conference on Interactive Media Experiences. ACM, New York, NY, USA, 161––174. https://doi.org/10.1145/3505284.3529969
[146]
Muralidhar Pantula and K. S. Kuppusamy. 2019. AuDIVA: A tool for embedding Audio Descriptions to enhance Video Accessibility for Persons with Visual Impairments. Multimedia Tools and Applications 78, 14 (Feb. 2019), 20005–20018. https://doi.org/10.1007/s11042-019-7363-4
[147]
Afra Pascual, Mireia Ribera, and Toni Granollers. 2014. Impact of Web Accessibility Barriers on Users with Hearing Impairment. In Proceedings of the XV International Conference on Human Computer Interaction (Puerto de la Cruz, Tenerife, Spain) (Interacción ’14). Association for Computing Machinery, New York, NY, USA, Article 8, 2 pages. https://doi.org/10.1145/2662253.2662261
[148]
Amy Pavel, Gabriel Reyes, and Jeffrey P. Bigham. 2020. Rescribe: Authoring and Automatically Editing Audio Descriptions. In Proceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology (Virtual Event, USA) (UIST ’20). Association for Computing Machinery, New York, NY, USA, 747–759. https://doi.org/10.1145/3379337.3415864
[149]
Eli Peli. 2005. Recognition performance and perceived quality of video enhanced for the visually impaired. Ophthalmic and Physiological Optics 25, 6 (Nov. 2005), 543–555. https://doi.org/10.1111/j.1475-1313.2005.00340.x
[150]
Yi-Hao Peng, Jeffrey P Bigham, and Amy Pavel. 2021. Slidecho: Flexible Non-Visual Exploration of Presentation Videos. In Proceedings of the 23rd International ACM SIGACCESS Conference on Computers and Accessibility (Virtual Event, USA) (ASSETS ’21). Association for Computing Machinery, New York, NY, USA, Article 24, 12 pages. https://doi.org/10.1145/3441852.3471234
[151]
Bridget Pettitt, Katharine Sharpe, and Steven Cooper. 1996. AUDETEL: Enhancing television for visually impaired people. British Journal of Visual Impairment 14, 2 (May 1996), 48–52. https://doi.org/10.1177/026461969601400202
[152]
Silvia Pfeiffer and Conrad Parker. 2009. Accessibility for the HTML5 <Video> Element. In Proceedings of the 2009 International Cross-Disciplinary Conference on Web Accessibililty (W4A) (Madrid, Spain) (W4A ’09). Association for Computing Machinery, New York, NY, USA, 98–100. https://doi.org/10.1145/1535654.1535679
[153]
Mark Priestley, Martha Stickings, Ema Loja, Stefanos Grammenos, Anna Lawson, Lisa Waddington, and Bjarney Fridriksdottir. 2016. The political participation of disabled people in Europe: Rights, accessibility and activism. Electoral Studies 42 (June 2016), 1–9. https://doi.org/10.1016/j.electstud.2016.01.009
[154]
Dhevi J. Rajendran, Andrew T. Duchowski, Pilar Orero, Juan Martínez, and Pablo Romero-Fresco. 2013. Effects of text chunking on subtitling: A quantitative and qualitative examination. Perspectives 21, 1 (March 2013), 5–21. https://doi.org/10.1080/0907676x.2012.722651
[155]
Silvia Ramis, Francisco J. Perales, Cristina Manresa-Yee, and Antoni Bibiloni. 2015. Usability Study of Gestures to Control a Smart-TV. In Communications in Computer and Information Science. Springer International Publishing, Cham, 135–146. https://doi.org/10.1007/978-3-319-22656-9_10
[156]
Marina Ramos. 2015. The emotional experience of films: does Audio Description make a difference?The Translator 21, 1 (Jan. 2015), 68–94. https://doi.org/10.1080/13556509.2014.994853
[157]
Anni Rander and Peter Olaf Looms. 2010. The Accessibility of Television News with Live Subtitling on Digital Television. In Proceedings of the 8th European Conference on Interactive TV and Video (Tampere, Finland) (EuroITV ’10). Association for Computing Machinery, New York, NY, USA, 155–160. https://doi.org/10.1145/1809777.1809809
[158]
Parthasarathy Ranganathan, Sarita Adve, and Norman P. Jouppi. 1999. Performance of Image and Video Processing with General-Purpose Processors and Media ISA Extensions. In Proceedings of the 26th Annual International Symposium on Computer Architecture (Atlanta, Georgia, USA) (ISCA ’99). IEEE Computer Society, USA, 124–135. https://doi.org/10.1145/300979.300990
[159]
Raisa Rashid, Jonathan Aitken, and Deborah I. Fels. 2006. Expressing Emotions Using Animated Text Captions. In Lecture Notes in Computer Science. Springer Berlin Heidelberg, Berlin, Heidelberg, 24–31. https://doi.org/10.1007/11788713_5
[160]
Raisa Rashid, Quoc Vy, Richard G. Hunt, and Deborah I. Fels. 2007. Dancing with Words. In Proceedings of the 6th ACM SIGCHI Conference on Creativity and Cognition (Washington, DC, USA) (CandC ’07). Association for Computing Machinery, New York, NY, USA, 269–270. https://doi.org/10.1145/1254960.1255007
[161]
Liliana Reis, Hilma Caravau, Telmo Silva, and Pedro Almeida. 2017. Automatic Creation of TV Content to Integrate in Seniors Viewing Activities. In Applications and Usability of Interactive TV, Maria Jose Abasolo, Pedro Almeida, and Joaquin Pina Amargos (Eds.). Springer International Publishing, Cham, 32–46.
[162]
Melissa L. Rethlefsen, Shona Kirtley, Siw Waffenschmidt, Ana Patricia Ayala, David Moher, Matthew J. Page, and Jonathan B. Koffel. 2021. PRISMA-S: an extension to the PRISMA Statement for Reporting Literature Searches in Systematic Reviews. Systematic Reviews 10, 1 (Jan. 2021), 1–19. https://doi.org/10.1186/s13643-020-01542-z
[163]
Jacob M. Rigby, Duncan P. Brumby, Anna L. Cox, and Sandy J. J. Gould. 2016. Watching movies on netflix. In Proceedings of the 18th International Conference on Human-Computer Interaction with Mobile Devices and Services Adjunct. ACM, New York, NY, USA, 714–721. https://doi.org/10.1145/2957265.2961843
[164]
Kathryn E. Ringland, Jennifer Nicholas, Rachel Kornfield, Emily G. Lattie, David C. Mohr, and Madhu Reddy. 2019. Understanding Mental Ill-health as Psychosocial Disability. In The 21st International ACM SIGACCESS Conference on Computers and Accessibility. ACM, New York, NY, USA, 156––170. https://doi.org/10.1145/3308561.3353785
[165]
Agebson Rocha Façanha, Adonias Caetano de Oliveira, Marcos Vinicius de Andrade Lima, Windson Viana, and Jaime Sánchez. 2016. Audio Description of Videos for People with Visual Disabilities. In Universal Access in Human-Computer Interaction. Users and Context Diversity, Margherita Antona and Constantine Stephanidis (Eds.). Springer International Publishing, Cham, 505–515.
[166]
Sunil Rodger, Dan Jackson, John Vines, Janice McLaughlin, and Peter Wright. 2019. JourneyCam: Exploring Experiences of Accessibility and Mobility among Powered Wheelchair Users through Video and Data. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (Glasgow, Scotland Uk) (CHI ’19). Association for Computing Machinery, New York, NY, USA, 1–15. https://doi.org/10.1145/3290605.3300860
[167]
C. Rodrigo and A. García-Serrano. 2015. Social Accessibility Action to Improve Quality of OER User-generated Video-classes and Associated Resources. Procedia Computer Science 67 (2015), 19–27. https://doi.org/10.1016/j.procs.2015.09.245 Proceedings of the 6th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-exclusion.
[168]
Aitor Rodriguez-Alsina, Guillermo Talavera, Pilar Orero, and Jordi Carrabina. 2012. Subtitle Synchronization across Multiple Screens and Devices. Sensors 12, 7 (June 2012), 8710–8731. https://doi.org/10.3390/s120708710
[169]
Anna Rohrbach, Atousa Torabi, Marcus Rohrbach, Niket Tandon, Christopher Pal, Hugo Larochelle, Aaron Courville, and Bernt Schiele. 2017. Movie Description. International Journal of Computer Vision 123, 1 (Jan. 2017), 94–120. https://doi.org/10.1007/s11263-016-0987-1
[170]
Dagfinn Rømen and Dag Svanæs. 2008. Evaluating web site accessibility. In Proceedings of the 5th Nordic conference on Human-computer interaction: building bridges. ACM, New York, NY, USA, 535–538. https://doi.org/10.1145/1463160.1463238
[171]
Pablo Romero-Fresco and Louise Fryer. 2013. Could Audio-Described Films Benefit from Audio Introductions? An Audience Response Study. Journal of Visual Impairment and Blindness 107, 4 (July 2013), 287–295. https://doi.org/10.1177/0145482x1310700405
[172]
Johana Maria Rosas Villena, Rudinei Goularte, and Renata Pontin de Mattos Fortes. 2014. A User Test with Accessible Video Player Looking for User Experience. In Universal Access in Human-Computer Interaction. Design for All and Accessibility Practice, Constantine Stephanidis and Margherita Antona (Eds.). Springer International Publishing, Cham, 623–633. https://doi.org/10.1007/978-3-319-07509-9_59
[173]
Sylvia Rothe, Kim Tran, and Heinrich Hußmann. 2018. Dynamic Subtitles in Cinematic Virtual Reality. In Proceedings of the 2018 ACM International Conference on Interactive Experiences for TV and Online Video (SEOUL, Republic of Korea) (TVX ’18). Association for Computing Machinery, New York, NY, USA, 209–214. https://doi.org/10.1145/3210825.3213556
[174]
Sylvia Rothe, Kim Tran, and Heinrich Hussmann. 2018. Positioning of Subtitles in Cinematic Virtual Reality. In ICAT-EGVE 2018 - International Conference on Artificial Reality and Telexistence and Eurographics Symposium on Virtual Environments, Gerd Bruder, Shunsuke Yoshimoto, and Sue Cobb (Eds.). The Eurographics Association, Germany, 1–8. https://doi.org/10.2312/egve.20181307
[175]
Andreas Sackl, Franziska Graf, Raimund Schatz, and Manfred Tscheligi. 2020. Ensuring Accessibility: Individual Video Playback Enhancements for Low Vision Users. In Proceedings of the 22nd International ACM SIGACCESS Conference on Computers and Accessibility (Virtual Event, Greece) (ASSETS ’20). Association for Computing Machinery, New York, NY, USA, Article 67, 4 pages. https://doi.org/10.1145/3373625.3417997
[176]
Andreas Sackl, Raimund Schatz, Bruno Gardlo, and Manfred Tscheligi. 2021. Enhancing Video Communication Experience for Low Vision Users. In Proceedings of the 18th International Web for All Conference (Ljubljana, Slovenia) (W4A ’21). Association for Computing Machinery, New York, NY, USA, Article 15, 5 pages. https://doi.org/10.1145/3430263.3452419
[177]
Andreas Sackl, Raimund Schatz, Stefan Suette, and Manfred Tscheligi. 2019. From Low Vision to High Quality: Video QoE Enhancement for Visually Impaired Users. In 2019 Eleventh International Conference on Quality of Multimedia Experience (QoMEX). IEEE, New York, NY, USA, 1–6. https://doi.org/10.1109/qomex.2019.8743323
[178]
Andrew Salway and Mike Graham. 2003. Extracting Information about Emotions in Films. In Proceedings of the Eleventh ACM International Conference on Multimedia (Berkeley, CA, USA) (MULTIMEDIA ’03). Association for Computing Machinery, New York, NY, USA, 299–302. https://doi.org/10.1145/957013.957076
[179]
José Francisco Saray Villamizar, Benoît Encelle, Yannick Prié, and Pierre-Antoine Champin. 2011. An Adaptive Videos Enrichment System Based on Decision Trees for People with Sensory Disabilities. In Proceedings of the International Cross-Disciplinary Conference on Web Accessibility (Hyderabad, Andhra Pradesh, India) (W4A ’11). Association for Computing Machinery, New York, NY, USA, Article 7, 4 pages. https://doi.org/10.1145/1969289.1969299
[180]
Tim Schlippe, Shaimaa Alessai, Ghanimeh El-Taweel, Matthias Wölfel, and Wajdi Zaghouani. 2020. Visualizing Voice Characteristics with Type Design in Closed Captions for Arabic. In 2020 International Conference on Cyberworlds (CW). IEEE, New York, NY, USA, 196–203. https://doi.org/10.1109/CW49994.2020.00039
[181]
Emilie Schmeidler and Corinne Kirchner. 2001. Adding Audio Description: Does it Make a Difference?Journal of Visual Impairment and Blindness 95, 4 (April 2001), 197–212. https://doi.org/10.1177/0145482x0109500402
[182]
Remo Schneider, Tobias Ableitner, and Gottfried Zimmermann. 2022. Layered Audio Descriptions for Videos. In Computers Helping People with Special Needs, Klaus Miesenberger, Georgios Kouroupetroglou, Katerina Mavrou, Roberto Manduchi, Mario Covarrubias Rodriguez, and Petr Penáz (Eds.). Springer International Publishing, Cham, 51–63. https://doi.org/10.1007/978-3-031-08645-8_7
[183]
Niels Seidel. 2015. Making Web Video Accessible: Interaction Design Patterns for Assistive Video Learning Environments. In Proceedings of the 20th European Conference on Pattern Languages of Programs (Kaufbeuren, Germany) (EuroPLoP ’15). Association for Computing Machinery, New York, NY, USA, Article 17, 16 pages. https://doi.org/10.1145/2855321.2855339
[184]
Ben Shirley, Melissa Meadows, Fadi Malak, James Woodcock, and Ash Tidball. 2017. Personalized Object-Based Audio for Hearing Impaired TV Viewers. Journal of the Audio Engineering Society 65, 4 (April 2017), 293–303. https://doi.org/10.17743/jaes.2017.0005
[185]
Ben Shirley and Rob Oldfield. 2015. Clean Audio for TV broadcast: An Object-Based Approach for Hearing-Impaired Viewers. Journal of the Audio Engineering Society 63, 4 (April 2015), 245–256. https://doi.org/10.17743/jaes.2015.0017
[186]
Brent N. Shiver and Rosalee J. Wolfe. 2015. Evaluating Alternatives for Better Deaf Accessibility to Selected Web-Based Multimedia. In Proceedings of the 17th International ACM SIGACCESS Conference on Computers and Accessibility (Lisbon, Portugal) (ASSETS ’15). Association for Computing Machinery, New York, NY, USA, 231–238. https://doi.org/10.1145/2700648.2809857
[187]
Andy P. Siddaway, Alex M. Wood, and Larry V. Hedges. 2019. How to Do a Systematic Review: A Best Practice Guide for Conducting and Reporting Narrative Reviews, Meta-Analyses, and Meta-Syntheses. Annual Review of Psychology 70, 1 (Jan. 2019), 747–770. https://doi.org/10.1146/annurev-psych-010418-102803
[188]
Rodrigo L. S. Silva and Frâncila Weidt Neiva. 2016. Systematic Literature Review in Computer Science - A Practical Guide., 8 pages. https://doi.org/10.13140/RG.2.2.35453.87524
[189]
Telmo Silva, Hilma Caravau, Martinho Mota, Liliana Reis, and Carlos Hernandez. 2018. A Process to Design a Video Library for Senior Users of iTV. In Applications and Usability of Interactive Television. Springer International Publishing, Cham, 105–116. https://doi.org/10.1007/978-3-319-90170-1_8
[190]
Laurianne Sitbon, Ross Brown, and Lauren Fell. 2019. Turning Heads: Designing Engaging Immersive Video Experiences to Support People with Intellectual Disability When Learning Everyday Living Skills. In Proceedings of the 21st International ACM SIGACCESS Conference on Computers and Accessibility (Pittsburgh, PA, USA) (ASSETS ’19). Association for Computing Machinery, New York, NY, USA, 171–182. https://doi.org/10.1145/3308561.3353787
[191]
Joel Snyder. 2005. Audio description: The visual made verbal. International Congress Series 1282 (2005), 935–939. https://doi.org/10.1016/j.ics.2005.05.215 Vision 2005.
[192]
Mark Springett, Richard N Griffiths, and Martin Mihajlov. 2017. Issues faced by vision-impaired users of interactive TV search facilities. In Proceedings of the Australasian Computer Science Week Multiconference. ACM, New York, NY, USA, 1–9. https://doi.org/10.1145/3014812.3014824
[193]
Mark V. Springett and Richard N. Griffiths. 2007. Accessibility of Interactive Television for Users with Low Vision: Learning from the Web. In Interactive TV: a Shared Experience, Pablo Cesar, Konstantinos Chorianopoulos, and Jens F. Jensen (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 76–85.
[194]
Ruxandra Tapu, Bogdan Mocanu, and Titus Zaharia. 2019. DEEP-HEAR: A Multimodal Subtitle Positioning System Dedicated to Deaf and Hearing-Impaired People. IEEE Access 7 (2019), 88150–88162. https://doi.org/10.1109/ACCESS.2019.2925806
[195]
Ruxandra Tapu, Bogdan Mocanu, and Titus Zaharia. 2019. Dynamic Subtitles: A Multimodal Video Accessibility Enhancement Dedicated to Deaf and Hearing Impaired Users. In 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW). IEEE, New York, NY, USA, 2558–2566. https://doi.org/10.1109/ICCVW.2019.00313
[196]
Frank Thorn and Sondra Thorn. 1996. Television Captions for Hearing-Impaired People: A Study of Key Factors that Affect Reading Performance. Human Factors: The Journal of the Human Factors and Ergonomics Society 38, 3 (Sept. 1996), 452–463. https://doi.org/10.1518/001872096778702006
[197]
Irene Tor-Carroggio and Sara Rovira-Esteva. 2020. Watching TV Through Your Ears. Journal of Audiovisual Translation 3, 1 (Oct. 2020), 1–25. https://doi.org/10.47476/jat.v2i1.2020.105
[198]
Michael D. Tyler, Caroline Jones, Leonid Grebennikov, Greg Leigh, William Noble, and Denis Burnham. 2009. Effect of Caption Rate on the Comprehension of Educational Television Programmes by Deaf School Students. Deafness and Education International 11, 3 (Sept. 2009), 152–162. https://doi.org/10.1179/146431509790559606
[199]
J.P. Udo, B. Acevedo, and D.I. Fels. 2010. Horatio audio-describes Shakespeare’s Hamlet. British Journal of Visual Impairment 28, 2 (May 2010), 139–156. https://doi.org/10.1177/0264619609359753
[200]
David Unbehaun, Konstantin Aal, Daryoush Daniel Vaziri, Peter David Tolmie, Rainer Wieching, David Randall, and Volker Wulf. 2020. Social technology appropriation in dementia: Investigating the role of caregivers in engaging people with dementia with a videogame-based training system. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu HI USA). ACM, New York, NY, USA, 1–15. https://doi.org/10.1145/3313831.3376648
[201]
Radu-Daniel Vatavu. 2021. Accessibility of Interactive Television and Media Experiences: Users with Disabilities Have Been Little Voiced at IMX and TVX. In ACM International Conference on Interactive Media Experiences (Virtual Event, USA) (IMX ’21). Association for Computing Machinery, New York, NY, USA, 218–222. https://doi.org/10.1145/3452918.3465485
[202]
Maarten Verboom, David Crombie, Evelien Dijk, and Mildred Theunisz. 2002. Spoken Subtitles: Making Subtitled TV Programmes Accessible. In Computers Helping People with Special Needs, Klaus Miesenberger, Joachim Klaus, and Wolfgang Zagler (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 295–302.
[203]
Anna Vilaro, Aitor Rodriguez-Alsina, Pilar Orero, and Jordi Carrabina. 2012. Evaluation of Emerging Audio Description Systems for Broadcast TV. In Ambient Assisted Living and Home Care, José Bravo, Ramón Hervás, and Marcela Rodríguez (Eds.). Springer Berlin Heidelberg, Berlin, Heidelberg, 270–277. https://doi.org/10.1007/978-3-642-35395-6_37
[204]
Johana M. Rosas Villena, Bruno C. Ramos, Renata P.M. Fortes, and Rudinei Goularte. 2014. An Accessible Video Player for Older People: Issues from a User Test. Procedia Computer Science 27 (2014), 168–175. https://doi.org/10.1016/j.procs.2014.02.020 5th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-exclusion, DSAI 2013.
[205]
Johana María Rosas Villena, Bruno Costa Ramos, Renata Pontin M. Fortes, and Rudinei Goularte. 2014. Web Videos - Concerns About Accessibility based on User Centered Design. Procedia Computer Science 27 (2014), 481–490. https://doi.org/10.1016/j.procs.2014.02.052 5th International Conference on Software Development and Technologies for Enhancing Accessibility and Fighting Info-exclusion, DSAI 2013.
[206]
Vinoba Vinayagamoorthy, Maxine Glancy, Paul Debenham, Alastair Bruce, Christoph Ziegler, and Richard Schäffer. 2018. Personalising the TV Experience with Augmented Reality Technology: Synchronised Sign Language Interpretation. In Proceedings of the 2018 ACM International Conference on Interactive Experiences for TV and Online Video (SEOUL, Republic of Korea) (TVX ’18). Association for Computing Machinery, New York, NY, USA, 179–184. https://doi.org/10.1145/3210825.3213562
[207]
Vinoba Vinayagamoorthy, Maxine Glancy, Christoph Ziegler, and Richard Schäffer. 2019. Personalising the TV Experience Using Augmented Reality: An Exploratory Study on Delivering Synchronised Sign Language Interpretation. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems (Glasgow, Scotland Uk) (CHI ’19). Association for Computing Machinery, New York, NY, USA, 1–12. https://doi.org/10.1145/3290605.3300762
[208]
Lakshmie Narayan Viswanathan, Troy McDaniel, and Sethuraman Panchanathan. 2011. Audio-Haptic Description in Movies. In HCI International 2011 – Posters’ Extended Abstracts, Constantine Stephanidis (Ed.). Springer Berlin Heidelberg, Berlin, Heidelberg, 414–418.
[209]
Evangelos Vlachogiannis, Damianos Gavalas, Christos Anagnostopoulos, and George E. Tsekouras. 2008. Towards ITV Accessibility: The MPEG-21 Case. In Proceedings of the 1st International Conference on PErvasive Technologies Related to Assistive Environments (Athens, Greece) (PETRA ’08). Association for Computing Machinery, New York, NY, USA, Article 33, 6 pages. https://doi.org/10.1145/1389586.1389626
[210]
Quoc V. Vy and Deborah I. Fels. 2009. Using Avatars for Improving Speaker Identification in Captioning. In Human-Computer Interaction – INTERACT 2009. Springer Berlin Heidelberg, Berlin, Heidelberg, 916–919. https://doi.org/10.1007/978-3-642-03658-3_110
[211]
Quoc V. Vy and Deborah I. Fels. 2010. Using Placement and Name for Speaker Identification in Captioning. In Lecture Notes in Computer Science. Springer Berlin Heidelberg, Berlin, Heidelberg, 247–254. https://doi.org/10.1007/978-3-642-14097-6_40
[212]
Takahiro Wakao, Terumasa Ehara, Eiji Sawamura, Ichiro Maruyama, and Katsuhiko Shirai. 1998. Project for Production of Closed-Caption TV Programs for the Hearing Impaired. In Proceedings of the 36th Annual Meeting of the Association for Computational Linguistics and 17th International Conference on Computational Linguistics - Volume 2 (Montreal, Quebec, Canada) (ACL ’98/COLING ’98). Association for Computational Linguistics, USA, 1340–1344. https://doi.org/10.3115/980691.980787
[213]
Agnieszka Walczak. 2017. Audio description on smartphones: making cinema accessible for visually impaired audiences. Universal Access in the Information Society 17, 4 (Aug. 2017), 833–840. https://doi.org/10.1007/s10209-017-0568-2
[214]
Agnieszka Walczak. 2017. Measuring immersion in audio description with Polish blind and visually impaired audiences. https://doi.org/10.13137/2421-6763/17350
[215]
Agnieszka Walczak and Louise Fryer. 2017. Creative description: The impact of audio description style on presence in visually impaired audiences. British Journal of Visual Impairment 35, 1 (2017), 6–17. https://doi.org/10.1177/0264619616661603 arXiv:https://doi.org/10.1177/0264619616661603
[216]
Agnieszka Walczak and Louise Fryer. 2018. Vocal delivery of audio description by genre: measuring users’ presence. Perspectives 26, 1 (2018), 69–83. https://doi.org/10.1080/0907676X.2017.1298634 arXiv:https://doi.org/10.1080/0907676X.2017.1298634
[217]
Mitzi Waltz and Alice Schippers. 2021. Politically disabled: barriers and facilitating factors affecting people with disabilities in political life within the European Union. Disabil. Soc. 36, 4 (April 2021), 517–540.
[218]
Fangzhou Wang, Hidehisa Nagano, Kunio Kashino, and Takeo Igarashi. 2017. Visualizing Video Sounds With Sound Word Animation to Enrich User Experience. IEEE Transactions on Multimedia 19, 2 (2017), 418–429. https://doi.org/10.1109/TMM.2016.2613641
[219]
Yujia Wang, Wei Liang, Haikun Huang, Yongqi Zhang, Dingzeyu Li, and Lap-Fai Yu. 2021. Toward Automatic Audio Description Generation for Accessible Videos. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems (Yokohama, Japan) (CHI ’21). Association for Computing Machinery, New York, NY, USA, Article 277, 12 pages. https://doi.org/10.1145/3411764.3445347
[220]
Yu-Huan Wang, Tian-Jun Gu, and Shyang-Yuh Wang. 2019. Causes and Characteristics of Short Video Platform Internet Community Taking the TikTok Short Video Application as an Example. In 2019 IEEE International Conference on Consumer Electronics - Taiwan (ICCE-TW). IEEE, New York, NY, USA, 1–2. https://doi.org/10.1109/icce-tw46550.2019.8992021
[221]
Lauren Ward and Ben Shirley. 2019. Personalization in Object-based Audio for Accessibility: A Review of Advancements for Hearing Impaired Listeners. Journal of the Audio Engineering Society 67, 7/8 (Aug. 2019), 584–597. https://doi.org/10.17743/jaes.2019.0021
[222]
Lauren Ward, Ben Shirley, and Jon Francombe. 2018. Accessible object-based audio using hierarchical narrative importance metadata. Journal of the Audio Engineering Society 145, 395 (2018), 1–10. https://aes.org/e-lib/browse.cfm?elib=19742
[223]
Lauren A. Ward. 2017. Accessible Broadcast Audio Personalisation for Hard of Hearing Listeners. In Adjunct Publication of the 2017 ACM International Conference on Interactive Experiences for TV and Online Video (Hilversum, The Netherlands) (TVX ’17 Adjunct). Association for Computing Machinery, New York, NY, USA, 105–108. https://doi.org/10.1145/3084289.3084293
[224]
Jennifer Wehrmeyer. 2014. Eye-tracking Deaf and hearing viewing of sign language interpreted news broadcasts. Journal of Eye Movement Research 7, 1 (mar 2014), 1–16. https://doi.org/10.16910/jemr.7.1.3
[225]
Maarten Wijnants, Eva Geurts, Hendrik Lievens, Peter Quax, and Wim Lamotte. 2021. Stay Tuned! An Investigation of Content Substitution, the Listener as Curator and Other Innovations in Broadcast Radio. In ACM International Conference on Interactive Media Experiences. ACM, New York, NY, USA, 120–133. https://doi.org/10.1145/3452918.3458793
[226]
Jacob O. Wobbrock and Julie A. Kientz. 2016. Research contributions in human-computer interaction. Interactions 23, 3 (April 2016), 38–44. https://doi.org/10.1145/2907069
[227]
Claes Wohlin. 2014. Guidelines for Snowballing in Systematic Literature Studies and a Replication in Software Engineering. In Proceedings of the 18th International Conference on Evaluation and Assessment in Software Engineering (London, England, United Kingdom) (EASE ’14). Association for Computing Machinery, New York, NY, USA, Article 38, 10 pages. https://doi.org/10.1145/2601248.2601268
[228]
Donghee Yvette Wohn and Guo Freeman. 2020. Audience Management Practices of Live Streamers on Twitch. In ACM International Conference on Interactive Media Experiences. ACM, New York, NY, USA, 106–116. https://doi.org/10.1145/3391614.3393653
[229]
James S. Wolffsohn, Ditipriya Mukhopadhyay, and Martin Rubinstein. 2007. Image Enhancement of Real-Time Television to Benefit the Visually Impaired. American Journal of Ophthalmology 144, 3 (Sept. 2007), 436–440. https://doi.org/10.1016/j.ajo.2007.05.031
[230]
Beste F. Yuksel, Pooyan Fazli, Umang Mathur, Vaishali Bisht, Soo Jung Kim, Joshua Junhee Lee, Seung Jung Jin, Yue-Ting Siu, Joshua A. Miele, and Ilmi Yoon. 2020. Human-in-the-Loop Machine Learning to Increase Video Accessibility for Visually Impaired and Blind Users. In Proceedings of the 2020 ACM Designing Interactive Systems Conference (Eindhoven, Netherlands) (DIS ’20). Association for Computing Machinery, New York, NY, USA, 47–60. https://doi.org/10.1145/3357236.3395433
[231]
Beste F Yuksel, Soo Jung Kim, Seung Jung Jin, Joshua Junhee Lee, Pooyan Fazli, Umang Mathur, Vaishali Bisht, Ilmi Yoon, and Joshua A Siu, Yue-Ting and. 2020. Increasing video accessibility for visually impaired users with human-in-the-loop machine learning. In Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu HI USA). ACM, New York, NY, USA, 1–9. https://doi.org/10.1145/3334480.3382821
[232]
Beste F. Yuksel, Soo Jung Kim, Seung Jung Jin, Joshua Junhee Lee, Pooyan Fazli, Umang Mathur, Vaishali Bisht, Ilmi Yoon, Yue-Ting Siu, and Joshua A. Miele. 2020. Increasing Video Accessibility for Visually Impaired Users with Human-in-the-Loop Machine Learning. In Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems (Honolulu, HI, USA) (CHI EA ’20). Association for Computing Machinery, New York, NY, USA, 1–9. https://doi.org/10.1145/3334480.3382821
[233]
Yongqing Zhu and Shuqiang Jiang. 2019. Attention-Based Densely Connected LSTM for Video Captioning. In Proceedings of the 27th ACM International Conference on Multimedia (Nice, France) (MM ’19). Association for Computing Machinery, New York, NY, USA, 802–810. https://doi.org/10.1145/3343031.3350932

Cited By

View all

Index Terms

  1. Accessibility Research in Digital Audiovisual Media: What Has Been Achieved and What Should Be Done Next?

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    IMX '23: Proceedings of the 2023 ACM International Conference on Interactive Media Experiences
    June 2023
    465 pages
    ISBN:9798400700286
    DOI:10.1145/3573381
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 29 August 2023

    Permissions

    Request permissions for this article.

    Check for updates

    Badges

    • Honorable Mention

    Author Tags

    1. Accessibility
    2. audiovisual
    3. dataset
    4. digital media
    5. systematic literature review

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Conference

    IMX '23

    Acceptance Rates

    Overall Acceptance Rate 69 of 245 submissions, 28%

    Upcoming Conference

    IMX '25

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)199
    • Downloads (Last 6 weeks)19
    Reflects downloads up to 09 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)"I Wish You Could Make the Camera Stand Still": Envisioning Media Accessibility Interventions with People with AphasiaProceedings of the 26th International ACM SIGACCESS Conference on Computers and Accessibility10.1145/3663548.3675598(1-17)Online publication date: 27-Oct-2024
    • (2024)Making Short-Form Videos Accessible with Hierarchical Video SummariesProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642839(1-17)Online publication date: 11-May-2024
    • (2024)“It’s Kind of Context Dependent”: Understanding Blind and Low Vision People’s Video Accessibility Preferences Across Viewing ScenariosProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642238(1-20)Online publication date: 11-May-2024
    • (2024)Lights, Camera, Access: A Closeup on Audiovisual Media Accessibility and AphasiaProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3641893(1-17)Online publication date: 11-May-2024

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media