1,141
Views
7
CrossRef citations to date
0
Altmetric
Articles

The effects of video-annotated learning and reviewing system with vocabulary learning mechanism on English listening comprehension and technology acceptance

ORCID Icon, &
Pages 1557-1593 | Published online: 06 Oct 2020
 

Abstract

This study is aimed to design a novel vocabulary learning mechanism (VLM) in the previously developed video-annotated learning and reviewing system (VALRS) that allows learners to identify unfamiliar or unknown words when listening to the video and generate personalized input enhancement from English subtitles for vocabulary learning. It is expected that the VALRS with VLM (VALRS-VLM) can help learners improve their English listening comprehension in the recognition of speech sounds of individual vocabulary words and meaning understanding of vocabulary words and oral sentences. To investigate the learning effectiveness and learners’ experiences of the newly developed VLM, this study compared the outcomes of English listening comprehension performance and technology acceptance between the experimental group using the VALRS-VLM and the control group using the VALRS without the VLM (VALRS-NVLM). Analytical results show that the learners using the VALRS-VLM achieved remarkably better in both English listening learning outcomes and learning retention for vocabulary learning and overall English listening comprehension performance than those who used the VALRS-NVLM. Besides, according to the results of the questionnaire surveyed after the experiment, both the VALRS-VLM and VALRS-NVLM groups showed good technology acceptance. This study confirms that the proposed VALRS-VLM could effectively facilitate learners’ English listening learning as well as provide them with a positive learning experience.

Disclosure statement

No potential conflict of interest was reported by the authors.

Additional information

Notes on contributors

Chih-Ming Chen

Chih-Ming Chen is currently a professor in the Graduate Institute of Library, Information and Archival Studies at National Chengchi University, Taipei, Taiwan. He received B.Sc. and M.Sc. degree from the Department of Industrial Education at National Taiwan Normal University in 1992 and 1997, and received Ph.D. degree from the Department of Electronic Engineering at National Taiwan University of Science and Technology in 2002. His research interests include computer assisted language learning, e-learning, digital library, data mining, machine learning and intelligent agents on the web.

Ming-Chaun Li

Ming-Chaun Li is currently a postdoctoral researcher in the Graduate Institute of Library, Information and Archival Studies at National Chengchi University, Taipei, Taiwan. He received M.Sc. degree from the Department of Child and Family Studies in University of Wisconsin-Madison in 1999, and received Ph.D. degree from the Graduate Institute of Applied Science and Technology in National Taiwan University of Science and Technology in 2017. His research interests include game-based learning, problem solving, and mini-flipped game-based learning.

Mei-Feng Lin

Mei-Feng Lin is currently a master student in the E-learning Master Program of Library and Information Studies at National Chengchi University, Taipei, Taiwan. Her research interests include computer assisted language learning and e-learning.

Log in via your institution

Log in to Taylor & Francis Online

PDF download + Online access

  • 48 hours access to article PDF & online version
  • Article PDF can be downloaded
  • Article PDF can be printed
USD 53.00 Add to cart

Issue Purchase

  • 30 days online access to complete issue
  • Article PDFs can be downloaded
  • Article PDFs can be printed
USD 339.00 Add to cart

* Local tax will be added as applicable

Related Research

People also read lists articles that other readers of this article have read.

Recommended articles lists articles that we recommend and is powered by our AI driven recommendation engine.

Cited by lists all citing articles based on Crossref citations.
Articles with the Crossref icon will open in a new tab.