Evaluating the Watching-Based Learning Model for Elementary School Students: A Case Study in Muhammadiyah Bandongan

Authors

  • Yuli Wahyuningsih Universitas Muhammadiyah Magelang, Indonesia
  • Ikhwanuddin Abdul Majid International Islamic University Malaysia, Malaysia
  • Faisal Efendi Sekolah Tinggi Agama Islam Balaiselasa YPPTI Pesisir Selatan, Indonesia
  • Arif Wiyat Purnanto Universitas Muhammadiyah Magelang, Indonesia

DOI:

https://doi.org/10.57255/jemast.v4i1.1466

Keywords:

Watching Learning, Multimodal Literacy, Elementary Education, Visual

Abstract

The purpose of this study is to evaluate the implementation of the Watching-Based Learning Model as a multimodal learning approach in Grade IV at SD IT Muhammadiyah Bandongan. This research employed a qualitative design using naturalistic observation supported by interviews and field notes collected over one month. The findings indicate that the multimodal learning model has been implemented at a level of 40%. Within this framework, visual representation accounts for 80% and audiovisual representation for 20%. In terms of multimodal literacy, visual literacy contributes 60% while critical multimodal literacy constitutes 40%. The visual literacy phase is identified as the initial stage, where students demonstrate comprehension of simple and familiar multimodal texts with predictable structures. In contrast, the critical multimodal literacy phase is positioned at the exploratory stage, where students begin to integrate strategies to interpret the content, purpose, and form of multimodal texts. The model is supported by three dominant media: textbooks (61%), PowerPoint (31%), and wall crafts (8%). This study highlights the relevance of multimodal learning in enhancing elementary students’ literacy development by integrating diverse media and literacy phases. The implications suggest that a balanced incorporation of visual and critical multimodal literacy can strengthen students’ comprehension and interpretation skills, offering valuable insights for curriculum development and classroom practice in primary education.

References

Aoki, T., Yamada, A., Kato, Y., Saito, H., Tsuboi, A., Nakada, A., … Tada, T. (2021). Automatic detection of various abnormalities in capsule endoscopy videos by a deep learning-based system: a multicenter study. Gastrointestinal Endoscopy, 93(1), 165-173.e1. https://doi.org/10.1016/j.gie.2020.04.080

Chen, G. (2020). A visual learning analytics (VLA) approach to video-based teacher professional development: Impact on teachers’ beliefs, self-efficacy, and classroom talk practice. Computers & Education, 144, 103670. https://doi.org/10.1016/j.compedu.2019.103670

Colognesi, S., Coppe, T., & Lucchini, S. (2023). Improving the oral language skills of elementary school students through video-recorded performances. Teaching and Teacher Education, 128, 104141. https://doi.org/10.1016/j.tate.2023.104141

Comsa, I.-S., Muntean, G.-M., & Trestian, R. (2021). An Innovative Machine-Learning-Based Scheduling Solution for Improving Live UHD Video Streaming Quality in Highly Dynamic Network Environments. IEEE Transactions on Broadcasting, 67(1), 212–224. https://doi.org/10.1109/TBC.2020.2983298

Deng, R., & Gao, Y. (2023). A review of eye tracking research on video-based learning. Education and Information Technologies, 28(6), 7671–7702. https://doi.org/10.1007/s10639-022-11486-7

Gordillo, A., Lopez-Fernandez, D., & Tovar, E. (2022). Comparing the Effectiveness of Video-Based Learning and Game-Based Learning Using Teacher-Authored Video Games for Online Software Engineering Education. IEEE Transactions on Education, 65(4), 524–532. https://doi.org/10.1109/TE.2022.3142688

Hapsari, A. S., Hanif, M., Gunarhadi, G., & Roemintoyo, R. (2019). Motion Graphic Animation Videos to Improve the Learning Outcomes of Elementary School Students. European Journal of Educational Research, volume-8-2(volume8-issue4.html), 1245–1255. https://doi.org/10.12973/eu-jer.8.4.1245

Hsu, F.-H., Lin, I.-H., Yeh, H.-C., & Chen, N.-S. (2022). Effect of Socratic Reflection Prompts via video-based learning system on elementary school students’ critical thinking skills. Computers & Education, 183, 104497. https://doi.org/10.1016/j.compedu.2022.104497

Ismail, A., Elpeltagy, M., S. Zaki, M., & Eldahshan, K. (2021). A New Deep Learning-Based Methodology for Video Deepfake Detection Using XGBoost. Sensors, 21(16), 5413. https://doi.org/10.3390/s21165413

Khaire, P., & Kumar, P. (2022). A semi-supervised deep learning based video anomaly detection framework using RGB-D for surveillance of real-world critical environments. Forensic Science International: Digital Investigation, 40, 301346. https://doi.org/10.1016/j.fsidi.2022.301346

Kim, B., & Lee, J. (2019). A Video-Based Fire Detection Using Deep Learning Models. Applied Sciences, 9(14), 2862. https://doi.org/10.3390/app9142862

Lee, C.-I., Huang, Y.-C., & Lin, Y.-C. (2016). A personal word-hiding video caption system on English vocabulary learning for elementary school students. In 2016 International Conference on Advanced Materials for Science and Engineering (ICAMSE) (pp. 128–131). IEEE. https://doi.org/10.1109/ICAMSE.2016.7840255

Lei, J., Luan, Q., Song, X., Liu, X., Tao, D., & Song, M. (2019). Action Parsing-Driven Video Summarization Based on Reinforcement Learning. IEEE Transactions on Circuits and Systems for Video Technology, 29(7), 2126–2137. https://doi.org/10.1109/TCSVT.2018.2860797

Li, A., Miao, Z., Cen, Y., Zhang, X.-P., Zhang, L., & Chen, S. (2020). Abnormal event detection in surveillance videos based on low-rank and compact coefficient dictionary learning. Pattern Recognition, 108, 107355. https://doi.org/10.1016/j.patcog.2020.107355

Li, S.-R., & Chang, Y.-J. (2024). Effectiveness of a nutrition education intervention using simulation videos and encouragement of parental involvement for elementary school students. Journal of Nutritional Science, 13, e35. https://doi.org/10.1017/jns.2024.41

Liu, Y., Feng, C., Yuan, X., Zhou, L., Wang, W., Qin, J., & Luo, Z. (2022). Clip-aware expressive feature learning for video-based facial expression recognition. Information Sciences, 598, 182–195. https://doi.org/10.1016/j.ins.2022.03.062

Maredia, M. K., Reyes, B., Ba, M. N., Dabire, C. L., Pittendrigh, B., & Bello-Bravo, J. (2018). Can mobile phone-based animated videos induce learning and technology adoption among low-literate farmers? A field experiment in Burkina Faso. Information Technology for Development, 24(3), 429–460. https://doi.org/10.1080/02681102.2017.1312245

Masmoudi, M., Friji, H., Ghazzai, H., & Massoud, Y. (2021). A Reinforcement Learning Framework for Video Frame-Based Autonomous Car-Following. IEEE Open Journal of Intelligent Transportation Systems, 2, 111–127. https://doi.org/10.1109/OJITS.2021.3083201

Mitra, A., Mohanty, S. P., Corcoran, P., & Kougianos, E. (2020). A Novel Machine Learning based Method for Deepfake Video Detection in Social Media. In 2020 IEEE International Symposium on Smart Electronic Systems (iSES) (Formerly iNiS) (pp. 91–96). IEEE. https://doi.org/10.1109/iSES50453.2020.00031

Nishani, E., & Cico, B. (2017). Computer vision approaches based on deep learning and neural networks: Deep neural networks for video analysis of human pose estimation. In 2017 6th Mediterranean Conference on Embedded Computing (MECO) (pp. 1–4). IEEE. https://doi.org/10.1109/MECO.2017.7977207

Pitarch, R. C. (2018). An Approach to Digital Game-based Learning: Video-games Principles and Applications in Foreign Language Learning. Journal of Language Teaching and Research, 9(6), 1147. https://doi.org/10.17507/jltr.0906.04

Rezaee, K., Rezakhani, S. M., Khosravi, M. R., & Moghimi, M. K. (2024). A survey on deep learning-based real-time crowd anomaly detection for secure distributed video surveillance. Personal and Ubiquitous Computing, 28(1), 135–151. https://doi.org/10.1007/s00779-021-01586-5

Riyanto, A., Murwani, R., Sulistiyani, S., & Rahfiludin, M. (2017). Food Safety Education Using Book Covers and Videos to Improve Street Food Safety Knowledge, Attitude, and Practice of Elementary School Students. Current Research in Nutrition and Food Science Journal, 5(2), 116–125. https://doi.org/10.12944/CRNFSJ.5.2.08

Salman, A., Siddiqui, S. A., Shafait, F., Mian, A., Shortis, M. R., Khurshid, K., … Schwanecke, U. (2020). Automatic fish detection in underwater videos by a deep neural network-based hybrid motion learning system. ICES Journal of Marine Science, 77(4), 1295–1307. https://doi.org/10.1093/icesjms/fsz025

Sharma, V., Gupta, M., Pandey, A. K., Mishra, D., & Kumar, A. (2022). A Review of Deep Learning-based Human Activity Recognition on Benchmark Video Datasets. Applied Artificial Intelligence, 36(1). https://doi.org/10.1080/08839514.2022.2093705

Su, C.-Y., & Chiu, C.-H. (2021). Perceived Enjoyment and Attractiveness Influence Taiwanese Elementary School Students’ Intention to Use Interactive Video Learning. International Journal of Human–Computer Interaction, 37(6), 574–583. https://doi.org/10.1080/10447318.2020.1841423

Valdez-Castro, K. (2025). Uso problemático de los videojuegos como predictor del desarrollo de habilidades sociales en estudiantes de educación básica. Gaceta Médica de Caracas, 133(3), 803–812. https://doi.org/10.47307/GMC.2025.133.3.15

Wang, M., Kumar, S. S., & Cheng, J. C. P. (2021). Automated sewer pipe defect tracking in CCTV videos based on defect detection and metric learning. Automation in Construction, 121, 103438. https://doi.org/10.1016/j.autcon.2020.103438

Wu, J., Guo, R., Wang, Z., & Zeng, R. (2021). Integrating spherical video-based virtual reality into elementary school students’ scientific inquiry instruction: effects on their problem-solving performance. Interactive Learning Environments, 29(3), 496–509. https://doi.org/10.1080/10494820.2019.1587469

Wu, W.-L., Hsu, Y., Yang, Q.-F., & Chen, J.-J. (2021). A Spherical Video-Based Immersive Virtual Reality Learning System to Support Landscape Architecture Students’ Learning Performance during the COVID-19 Era. Land, 10(6), 561. https://doi.org/10.3390/land10060561

Xie, L., Zhang, X., & Guo, Z. (2018). CLS: A Cross-user Learning based System for Improving QoE in 360-degree Video Adaptive Streaming. In Proceedings of the 26th ACM international conference on Multimedia (pp. 564–572). New York, NY, USA: ACM. https://doi.org/10.1145/3240508.3240556

Zhang, H., Luo, C., Wang, Q., Kitchin, M., Parmley, A., Monge-Alvarez, J., & Casaseca-de-la-Higuera, P. (2018). A novel infrared video surveillance system using deep learning based techniques. Multimedia Tools and Applications, 77(20), 26657–26676. https://doi.org/10.1007/s11042-018-5883-y

Zhang, Y., Zhang, T., Wang, S., & Yu, P. (2025). An Efficient Perceptual Video Compression Scheme Based on Deep Learning-Assisted Video Saliency and Just Noticeable Distortion. Engineering Applications of Artificial Intelligence, 141, 109806. https://doi.org/10.1016/j.engappai.2024.109806

Published

2025-05-10

How to Cite

Wahyuningsih, Y. ., Abdul Majid, . I., Efendi, F., & Purnanto, A. W. (2025). Evaluating the Watching-Based Learning Model for Elementary School Students: A Case Study in Muhammadiyah Bandongan. Journal of Educational Management and Strategy, 4(1), 55–65. https://doi.org/10.57255/jemast.v4i1.1466

Citation Check