Abstract
Multimedia content is growing at an increasing pace. Making this content accessible to people with impairments is not only paramount but also a growing challenge in itself. Access services for people with visual or hearing impairments have been studied and refined over the last years, resulting in standards and laws to ensure that minimum amounts of accessible content are produced. Professional content producers have the knowledge and skills required to do so. However, individuals creating multimedia content to publish on the Internet usually lack both. In this chapter, we review existing access services, comprised of subtitles, sign language and audio descriptions. We complement this with a summary of research efforts that could assist both in the production and consumption of access services. Finally, we discuss how emerging technologies and techniques, like machine learning or crowdsourcing, can help us tackle the sheer amount of access services that need to be created to ensure all have equal access to produced content.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Discussed in the ‘Situationally-Induced Impairments and Disabilities’ chapter in the first part of this book.
- 2.
As defined in the Web Content Accessibility Guidelines 2.0 (https://www.w3.org/TR/2006/WD-WCAG20-20060427/appendixA.html).
- 3.
For example, the National Association for the Deaf publishes a list of captioning tools (https://dcmp.org/learn/213).
- 4.
As defined in the Web Content Accessibility Guidelines 2.0 (https://www.w3.org/TR/2006/WD-WCAG20-20060427/appendixA.html).
- 5.
As defined in the Web Content Accessibility Guidelines 2.0 (https://www.w3.org/TR/2006/WD-WCAG20-20060427/appendixA.html).
- 6.
- 7.
- 8.
- 9.
- 10.
- 11.
- 12.
- 13.
- 14.
Accessible Rich Internet Applications (WAI-ARIA) (https://www.w3.org/TR/2014/REC-wai-aria-20140320/).
- 15.
37 Staggering Video Marketing Statistics for 2018 | Wordstream (https://www.wordstream.com/blog/ws/2017/03/08/video-marketing-statistics).
References
Allman T, Dhillon RK, Landau MA, Kurniawan SH (2009) Rock vibe: rock band®computer games for people with no or limited vision. In: Proceedings of the 11th international ACM SIGACCESS conference on computers and accessibility, Assets ’09, ACM, New York, USA, pp 51–58
Azuma M, Hiruma N, Sumiyoshi H, Uchida T, Miyazaki T, Umeda S, Kato N, Yamanouchi Y (2018) Development and evaluation of system for automatically generating sign-language CG animation using meteorological information. In: Miesenberger K, Kouroupetroglou G (eds) Computers helping people with special needs. Springer International Publishing, Cham, pp 233–238
de Araújo TMU, Ferreira FL, Silva DA, Oliveira LD, Falcão EL, Domingues LA, Martins VF, Portela IA, Nóbrega YS, Lima HR, Filho GLS, Tavares TA, Duarte AN (2014) An approach to generate and embed sign language video tracks into multimedia contents. Inf Sci 281:762–780. Multimedia modeling
Federico M, Furini M (2012) Enhancing learning accessibility through fully automatic captioning. In: Proceedings of the international cross-disciplinary conference on web accessibility - W4A ’12, ACM Press, New York, USA, p 1
González M, Moreno L, Martínez P, Iglesias A (2011) Web accessibility requirements for media players. In: Campos P, Graham N, Jorge J, Nunes N, Palanque P, Winckler M (eds) Human-computer interaction - INTERACT 2011. Springer, Berlin, pp 669–674
Guerreiro JA, Gonçalves D (2016) Scanning for digital content: how blind and sighted people perceive concurrent speech. ACM Trans Access Comput 8(1):2:1–2:28
Huang Y, Huang Y, Xue N, Bigham JP (2017) Leveraging complementary contributions of different workers for efficient crowdsourcing of video captions. In: Proceedings of the 2017 CHI conference on human factors in computing systems, CHI ’17, ACM, New York, USA, pp 4617–4626
Ichiki M, Shimizu T, Imai A, Takagi T, Iwabuchi M, Kurihara K, Miyazaki T, Kumano T, Kaneko H, Sato S, Seiyama N, Yamanouchi Y, Sumiyoshi H (2018) Study on automated audio descriptions overlapping live television commentary. In: Miesenberger K, Kouroupetroglou G (eds) Computers helping people with special needs. Springer International Publishing, Cham, pp 220–224
Kacorri H, Huenerfauth M, Ebling S, Patel K, Willard M (2015) Demographic and experiential factors influencing acceptance of sign language animation by deaf users. In: Proceedings of the 17th international ACM SIGACCESS conference on computers and accessibility, ASSETS ’15, ACM, New York, USA, pp 147–154
Khan TM (2010) The effects of multimedia learning on children with different special education needs. Procedia - Soc Behav Sci 2(2):4341–4345. Innovation and creativity in education. http://www.sciencedirect.com/science/article/pii/S1877042810007305
Kim S-W, Li Z-X, Aoki Y (2004) On intelligent avatar communication using Korean, Chinese and Japanese sign-languages: an overview. In: Proceedings of the 8th control, automation, robotics and vision conference (ICARCV’04), vol 1, pp 747–752
Knoop-van Campen CAN, Segers E, Verhoeven L (2018) The modality and redundancy effects in multimedia learning in children with dyslexia. Dyslexia 24(2):140–155. https://onlinelibrary.wiley.com/doi/abs/10.1002/dys.1585
Krüger M (2008) Accessible flash is no oxymoron: a case study in E-learning for blind and sighted users. In: Miesenberger K, Klaus J, Zagler W, Karshmer A (eds) Computers helping people with special needs. ICCHP 2008. Springer, Berlin, pp 362–369
Kubitschke L, Cullen K, Dolphin C, Laurin S, Cederbom A (2013) Study on assessing and promoting E-accessibility, Technical report, European Commission DG Communications Networks, Content and technology
Kushalnagar R, Seita M, Glasser A (2017) Closed ASL interpreting for online videos. In: Proceedings of the 14th web for all conference on the future of accessible work - W4A ’17, ACM Press, New York, USA, pp 1–4
Maybury M (2007) Searching conversational speech. In: Proceedings of the SIGIR workshop on searching spontaneous conversational speech (SSCS’07)
McCade D, Savage G, Naismith SL (2011) Review of emotion recognition in mild cognitive impairment. Dement Geriatr Cogn Disord 32(4):257–266. https://www.karger.com/DOI/10.1159/000335009
Oh J, Jeon S, Kim M, Kwon H, Kim I (2014) An avatar-based weather forecast sign language system for the hearing-impaired. In: Iliadis L, Maglogiannis I, Papadopoulos H (eds) Artificial intelligence applications and innovations. Springer, Berlin, pp 519–527
Peng YH, Hsi M-W, Taele P, Lin T-Y, Lai P-E, Hsu L, Chen TC, Wu, T-Y, Chen Y-A, Tang H-H, Chen MY (2018) Speechbubbles: enhancing captioning experiences for deaf and hard-of-hearing people in group conversations. In: Proceedings of the conference on human factors in computing systems (CHI ’18), ACM, New York, USA, pp 293:1–293:10
Shiver BN, Wolfe RJ (2015) Evaluating alternatives for better deaf accessibility to selected web-based multimedia. In: Proceedings of the 17th international ACM SIGACCESS conference on computers accessibility - ASSETS ’15, ACM Press, New York, USA, pp 231–238
Sloan D, Stratford J, Gregor P (2006) Using multimedia to enhance the accessibility of the learning environment for disabled students: reflections from the skills for access project. ALT-J 14(1):39–54. https://doi.org/10.1080/09687760500479936
Smith RG, Nolan B (2016) Emotional facial expressions in synthesised sign language avatars: a manual evaluation. Univers Access Inf Soc 15(4):567–576
Thompson T (2015) Video for all: accessibility of video content and universal design of a media player, 2nd edn. Harvard Education Press, Boston, pp 259–273
Thompson, T (2018) Media player accessibility: summary of insights from interviews and focus groups. J Technol Pers Disabil 6:325–335
Toledo J, Torres J, Alonso S, Toledo P, González EJ (2005) SLOAS: hearing with the eyes. In: Proceedings of the 4th WSEAS international conference on electronic, signal processing and control, pp 27:1–27:5
Wang J, Dawson K, Saunders K, Ritzhaupt AD, Antonenko PP, Lombardino L, Keil A, Agacli-Dogan N, Luo W, Cheng L, Davis RO (2018) Investigating the effects of modality and multimedia on the learning performance of college students with dyslexia. J Spec Educ Technol 33(3):182–193. https://doi.org/10.1177/0162643418754530
WHO (2011) World report on disability, World Health Organization
Wolfe R, Efthimiou E, Glauert J, Hanke T, McDonald J, Schnepp J (2016) Special issue: recent advances in sign language translation and avatar technology. Univers Access Inf Soc 15(4):485–486
Acknowledgements
This work was supported by national funds through Fundação para a Ciência e Tecnologia, under LASIGE Strategic Project—PEst-OE/EEI/UI0408/2014.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer-Verlag London Ltd., part of Springer Nature
About this chapter
Cite this chapter
Duarte, C., Fonseca, M.J. (2019). Multimedia Accessibility. In: Yesilada, Y., Harper, S. (eds) Web Accessibility. Human–Computer Interaction Series. Springer, London. https://doi.org/10.1007/978-1-4471-7440-0_25
Download citation
DOI: https://doi.org/10.1007/978-1-4471-7440-0_25
Published:
Publisher Name: Springer, London
Print ISBN: 978-1-4471-7439-4
Online ISBN: 978-1-4471-7440-0
eBook Packages: Computer ScienceComputer Science (R0)