Online Presentations with PowerPoint Present Live Real-Time Automated Captions and Subtitles: Perceptions of Faculty and Administrators

Anymir Orellana, Georgina Arguello, Elda Kanzki-Veloso


Captioning of recorded videos is beneficial to many and a matter of compliance with accessibility regulations and guidelines. Like recorded captions, real-time captions can also be means to implement the Universal Design for Learning checkpoint to offer text-based alternatives to auditory information. A cost-effective solution to implement the checkpoint for live online presentations is to use speech recognition technologies to generate automated captions. In particular, Microsoft PowerPoint Present Live (MSPL) is an application that can be used to present with real-time automated captions and subtitles in multiple languages, allowing individuals to follow the presentation in their preferred language. The purpose of this study was to identify challenges that participants could encounter when using the MSPL feature of real-time automated captions/subtitles, and to determine what they describe as potential uses, challenges, and benefits of the feature. Participants were full-time faculty and administrators with a faculty appointment in a higher education institution. Data from five native English speakers and five native Spanish speakers were analyzed. Activities of remote usability testing and interviews were conducted to collect data. Overall, participants did not encounter challenges that they could not overcome and described MSPL as an easy-to-use and useful tool to present with captions/subtitles for teaching or training and to reach English and Spanish-speaking audiences. The themes that emerged as potential challenges were training, distraction, and technology. Findings are discussed and further research is recommended.


online presentation; real-time; captions; subtitles; speech recognition; universal design for learning

Full Text:



PlayMedia. (n.d.). The ultimate guide to closed captioning.

Barnum, C. M. (2010). Usability testing essentials. Morgan Kaufmann.

Bureau of Internet Accessibility. (2019, April). Checklist for creating accessible videos.

CAST. (2018a). Universal Design for Learning guidelines version 2.2.

CAST. (2018b). Universal Design for Learning guidelines version 2.2., Checkpoint 1.2: Offer alternatives for auditory information.

Creswell, J. W., & Poth, C. N. (2018). Qualitative inquiry and research design: Choosing among five approaches. Sage.

Dallas, B. K., McCarthy, A. K., & Long, G. (2016). Examining the educational benefits of and attitudes toward closed captioning among undergraduate students. Journal of the Scholarship of Teaching and Learning, 16(2), 56-65.

De Bleecker, I., & Okoroji, R. (2018). Remote usability testing: Actionable insights in user behavior across geographies and time zones. Packt Publishing.

Denzin, N. K. (2009). The research act: A theoretical introduction to sociological methods. Routledge.

Enamorado, S. (2019a, June 3). How accurate is your transcription service? service/#:~:text=The%20industry%20standard%20for%20caption,is%20a%2099%25%20accuracy%20rate

Enamorado, S. (2019b, October 7). What is 99% accuracy, really? Why caption quality matters.

Federal Communications Commission. (2020). Broadband speed guide.

Gernsbacher, M. A. (2015). Video captions benefit everyone. Policy Insights from the Behavioral and Brain Sciences, 2(1), 195–202.

Gibbs, G. R. (2018). Analyzing qualitative data (2nd ed.). SAGE Publications Ltd.

Google. (n.d.). Docs editor help: Present slides with captions.

Huang, Y. M., Liu, C. L., Shadiev, R., Shen, M. H., & Hwang, W. Y. (2015). Investigating an application of speech-to-text recognition: A study on visual attention and learning behaviour. Journal of Computer Assisted Learning, 31(6), 529–545.

Huang, Y. M., Shadiev, R., & Hwang, W. Y. (2016). Investigating the effectiveness of speech-to-text recognition applications on learning performance and cognitive load. Computers & Education, 101(1), 15–28.

IBM Cloud Education. (2020, September 2). Speech recognition.

Johnson, D. (2020, June 17). Live Presentations is now generally available.

Lincoln, Y. S., & Guba, E. G. (1985). Naturalistic inquiry. Sage.

Linder, K. (2016). Student uses and perceptions of closed captions and transcripts: Results from a national study. Corvallis, OR: Oregon State University Ecampus Research Unit.

Malterud, K., Siersma, V. D., & Guassora, A. D. (2016). Sample size in qualitative interview studies: Guided by information power. Qualitative Health Research, 26, 1753-1760.

McCracken, D. G. (1988). The long interview. Sage.

Meyer, A., Rose, D. H., & Gordon, D. (2014). Universal design for learning: Theory and practice. CAST.

Microsoft Education. (2020, January 6). Engage your audience with Live Presentations in PowerPoint [Video]. YouTube.

Microsoft Education Team. (2019, January 23). What’s New in EDU Live: Bett day 1. Microsoft Education Blog.

Microsoft. (n.d.-a). Present Live: Engage your audience with live presentations.

Microsoft. (n.d.-b). Present with real-time, automatic captions or subtitles in PowerPoint.

Morris, K. K., Frechette, C., Dukes III, L., Stowell, N., Topping, N. E., &

Brodosi, D. (2016). Closed captioning matters: Examining the value of closed captions for all students. Journal of Postsecondary Education and Disability, 29(3), 231-238.

Myers, E. (2019, January 9). Closed captions & subtitles: Which should you use?

PowerPoint Team. (2018, December 3). Present more inclusively with live captions and subtitles in PowerPoint.

Revuelta, P., Jiménez, J., Sánchez, J. M., & Ruiz, B. (2010). Automatic speech recognition to enhance learning for disabled students. In J. Zhao, P. Ordoñez De Pablos, & R. Tennyson (Eds.), Technology enhanced learning for people with disabilities: Approaches and applications (pp. 89-104). IGI Global. ProQuest Ebook Central.

Shadiev, R., Huang, Y-M., & Hwang, J-P. (2017). Investigating the effectiveness of speech-to-text recognition applications on learning performance, attention, and meditation. Educational Technology Research and Development, 65(5), 1239-1261.

Shadiev, R., Sun, A., & Huang, Y-M. (2019). A study of the facilitation of cross‐cultural understanding and intercultural sensitivity using speech‐enabled language translation technology. British Journal of Educational Technology, 50(3), 1415-1433.

Shadiev, R., Wu T-T., Sun A., & Huang Y-M. (2018). Applications of speech-to-text recognition and computer-aided translation for facilitating cross-cultural learning through a learning activity: Issues and their solutions. Educational Technology Research and Development, 66(1), 191-214.

Take Note. (n.d.). Closed captioning vs. subtitles: How to make the right choice

Thomas, D. R. (2006). A general inductive approach for analyzing qualitative evaluation data. American Journal of Evaluation, 27, 237-246.

United States Department of Justice Civil Rights Division. (n.d.). Information and technical assistance on the Americans with Disabilities Act.

U.S. General Service Administration. (n.d.). IT accessibility laws and policies.

World Wide Web Consortium. (n.d.). Web Content Accessibility Guidelines (WCAG) 2.0.

Young, D. A., Casey, E. A. (2019). An examination of the sufficiency of small qualitative samples. Social Work Research, 43(1), 53–58.

Zoom Video Communications. (n.d.). System requirements for Windows, macOS, and Linux.


Copyright (c) 2022 Anymir Orellana, Georgina Arguello, Elda Kanzki-Veloso

License URL: