Evaluating the Watching-Based Learning Model for Elementary School Students: A Case Study in Muhammadiyah Bandongan
DOI:
https://doi.org/10.57255/jemast.v4i1.1466Keywords:
Watching Learning, Multimodal Literacy, Elementary Education, VisualAbstract
The purpose of this study is to evaluate the implementation of the Watching-Based Learning Model as a multimodal learning approach in Grade IV at SD IT Muhammadiyah Bandongan. This research employed a qualitative design using naturalistic observation supported by interviews and field notes collected over one month. The findings indicate that the multimodal learning model has been implemented at a level of 40%. Within this framework, visual representation accounts for 80% and audiovisual representation for 20%. In terms of multimodal literacy, visual literacy contributes 60% while critical multimodal literacy constitutes 40%. The visual literacy phase is identified as the initial stage, where students demonstrate comprehension of simple and familiar multimodal texts with predictable structures. In contrast, the critical multimodal literacy phase is positioned at the exploratory stage, where students begin to integrate strategies to interpret the content, purpose, and form of multimodal texts. The model is supported by three dominant media: textbooks (61%), PowerPoint (31%), and wall crafts (8%). This study highlights the relevance of multimodal learning in enhancing elementary students’ literacy development by integrating diverse media and literacy phases. The implications suggest that a balanced incorporation of visual and critical multimodal literacy can strengthen students’ comprehension and interpretation skills, offering valuable insights for curriculum development and classroom practice in primary education.
References
Aoki, T., Yamada, A., Kato, Y., Saito, H., Tsuboi, A., Nakada, A., … Tada, T. (2021). Automatic detection of various abnormalities in capsule endoscopy videos by a deep learning-based system: a multicenter study. Gastrointestinal Endoscopy, 93(1), 165-173.e1. https://doi.org/10.1016/j.gie.2020.04.080
Chen, G. (2020). A visual learning analytics (VLA) approach to video-based teacher professional development: Impact on teachers’ beliefs, self-efficacy, and classroom talk practice. Computers & Education, 144, 103670. https://doi.org/10.1016/j.compedu.2019.103670
Colognesi, S., Coppe, T., & Lucchini, S. (2023). Improving the oral language skills of elementary school students through video-recorded performances. Teaching and Teacher Education, 128, 104141. https://doi.org/10.1016/j.tate.2023.104141
Comsa, I.-S., Muntean, G.-M., & Trestian, R. (2021). An Innovative Machine-Learning-Based Scheduling Solution for Improving Live UHD Video Streaming Quality in Highly Dynamic Network Environments. IEEE Transactions on Broadcasting, 67(1), 212–224. https://doi.org/10.1109/TBC.2020.2983298
Deng, R., & Gao, Y. (2023). A review of eye tracking research on video-based learning. Education and Information Technologies, 28(6), 7671–7702. https://doi.org/10.1007/s10639-022-11486-7
Gordillo, A., Lopez-Fernandez, D., & Tovar, E. (2022). Comparing the Effectiveness of Video-Based Learning and Game-Based Learning Using Teacher-Authored Video Games for Online Software Engineering Education. IEEE Transactions on Education, 65(4), 524–532. https://doi.org/10.1109/TE.2022.3142688
Hapsari, A. S., Hanif, M., Gunarhadi, G., & Roemintoyo, R. (2019). Motion Graphic Animation Videos to Improve the Learning Outcomes of Elementary School Students. European Journal of Educational Research, volume-8-2(volume8-issue4.html), 1245–1255. https://doi.org/10.12973/eu-jer.8.4.1245
Hsu, F.-H., Lin, I.-H., Yeh, H.-C., & Chen, N.-S. (2022). Effect of Socratic Reflection Prompts via video-based learning system on elementary school students’ critical thinking skills. Computers & Education, 183, 104497. https://doi.org/10.1016/j.compedu.2022.104497
Ismail, A., Elpeltagy, M., S. Zaki, M., & Eldahshan, K. (2021). A New Deep Learning-Based Methodology for Video Deepfake Detection Using XGBoost. Sensors, 21(16), 5413. https://doi.org/10.3390/s21165413
Khaire, P., & Kumar, P. (2022). A semi-supervised deep learning based video anomaly detection framework using RGB-D for surveillance of real-world critical environments. Forensic Science International: Digital Investigation, 40, 301346. https://doi.org/10.1016/j.fsidi.2022.301346
Kim, B., & Lee, J. (2019). A Video-Based Fire Detection Using Deep Learning Models. Applied Sciences, 9(14), 2862. https://doi.org/10.3390/app9142862
Lee, C.-I., Huang, Y.-C., & Lin, Y.-C. (2016). A personal word-hiding video caption system on English vocabulary learning for elementary school students. In 2016 International Conference on Advanced Materials for Science and Engineering (ICAMSE) (pp. 128–131). IEEE. https://doi.org/10.1109/ICAMSE.2016.7840255
Lei, J., Luan, Q., Song, X., Liu, X., Tao, D., & Song, M. (2019). Action Parsing-Driven Video Summarization Based on Reinforcement Learning. IEEE Transactions on Circuits and Systems for Video Technology, 29(7), 2126–2137. https://doi.org/10.1109/TCSVT.2018.2860797
Li, A., Miao, Z., Cen, Y., Zhang, X.-P., Zhang, L., & Chen, S. (2020). Abnormal event detection in surveillance videos based on low-rank and compact coefficient dictionary learning. Pattern Recognition, 108, 107355. https://doi.org/10.1016/j.patcog.2020.107355
Li, S.-R., & Chang, Y.-J. (2024). Effectiveness of a nutrition education intervention using simulation videos and encouragement of parental involvement for elementary school students. Journal of Nutritional Science, 13, e35. https://doi.org/10.1017/jns.2024.41
Liu, Y., Feng, C., Yuan, X., Zhou, L., Wang, W., Qin, J., & Luo, Z. (2022). Clip-aware expressive feature learning for video-based facial expression recognition. Information Sciences, 598, 182–195. https://doi.org/10.1016/j.ins.2022.03.062
Maredia, M. K., Reyes, B., Ba, M. N., Dabire, C. L., Pittendrigh, B., & Bello-Bravo, J. (2018). Can mobile phone-based animated videos induce learning and technology adoption among low-literate farmers? A field experiment in Burkina Faso. Information Technology for Development, 24(3), 429–460. https://doi.org/10.1080/02681102.2017.1312245
Masmoudi, M., Friji, H., Ghazzai, H., & Massoud, Y. (2021). A Reinforcement Learning Framework for Video Frame-Based Autonomous Car-Following. IEEE Open Journal of Intelligent Transportation Systems, 2, 111–127. https://doi.org/10.1109/OJITS.2021.3083201
Mitra, A., Mohanty, S. P., Corcoran, P., & Kougianos, E. (2020). A Novel Machine Learning based Method for Deepfake Video Detection in Social Media. In 2020 IEEE International Symposium on Smart Electronic Systems (iSES) (Formerly iNiS) (pp. 91–96). IEEE. https://doi.org/10.1109/iSES50453.2020.00031
Nishani, E., & Cico, B. (2017). Computer vision approaches based on deep learning and neural networks: Deep neural networks for video analysis of human pose estimation. In 2017 6th Mediterranean Conference on Embedded Computing (MECO) (pp. 1–4). IEEE. https://doi.org/10.1109/MECO.2017.7977207
Pitarch, R. C. (2018). An Approach to Digital Game-based Learning: Video-games Principles and Applications in Foreign Language Learning. Journal of Language Teaching and Research, 9(6), 1147. https://doi.org/10.17507/jltr.0906.04
Rezaee, K., Rezakhani, S. M., Khosravi, M. R., & Moghimi, M. K. (2024). A survey on deep learning-based real-time crowd anomaly detection for secure distributed video surveillance. Personal and Ubiquitous Computing, 28(1), 135–151. https://doi.org/10.1007/s00779-021-01586-5
Riyanto, A., Murwani, R., Sulistiyani, S., & Rahfiludin, M. (2017). Food Safety Education Using Book Covers and Videos to Improve Street Food Safety Knowledge, Attitude, and Practice of Elementary School Students. Current Research in Nutrition and Food Science Journal, 5(2), 116–125. https://doi.org/10.12944/CRNFSJ.5.2.08
Salman, A., Siddiqui, S. A., Shafait, F., Mian, A., Shortis, M. R., Khurshid, K., … Schwanecke, U. (2020). Automatic fish detection in underwater videos by a deep neural network-based hybrid motion learning system. ICES Journal of Marine Science, 77(4), 1295–1307. https://doi.org/10.1093/icesjms/fsz025
Sharma, V., Gupta, M., Pandey, A. K., Mishra, D., & Kumar, A. (2022). A Review of Deep Learning-based Human Activity Recognition on Benchmark Video Datasets. Applied Artificial Intelligence, 36(1). https://doi.org/10.1080/08839514.2022.2093705
Su, C.-Y., & Chiu, C.-H. (2021). Perceived Enjoyment and Attractiveness Influence Taiwanese Elementary School Students’ Intention to Use Interactive Video Learning. International Journal of Human–Computer Interaction, 37(6), 574–583. https://doi.org/10.1080/10447318.2020.1841423
Valdez-Castro, K. (2025). Uso problemático de los videojuegos como predictor del desarrollo de habilidades sociales en estudiantes de educación básica. Gaceta Médica de Caracas, 133(3), 803–812. https://doi.org/10.47307/GMC.2025.133.3.15
Wang, M., Kumar, S. S., & Cheng, J. C. P. (2021). Automated sewer pipe defect tracking in CCTV videos based on defect detection and metric learning. Automation in Construction, 121, 103438. https://doi.org/10.1016/j.autcon.2020.103438
Wu, J., Guo, R., Wang, Z., & Zeng, R. (2021). Integrating spherical video-based virtual reality into elementary school students’ scientific inquiry instruction: effects on their problem-solving performance. Interactive Learning Environments, 29(3), 496–509. https://doi.org/10.1080/10494820.2019.1587469
Wu, W.-L., Hsu, Y., Yang, Q.-F., & Chen, J.-J. (2021). A Spherical Video-Based Immersive Virtual Reality Learning System to Support Landscape Architecture Students’ Learning Performance during the COVID-19 Era. Land, 10(6), 561. https://doi.org/10.3390/land10060561
Xie, L., Zhang, X., & Guo, Z. (2018). CLS: A Cross-user Learning based System for Improving QoE in 360-degree Video Adaptive Streaming. In Proceedings of the 26th ACM international conference on Multimedia (pp. 564–572). New York, NY, USA: ACM. https://doi.org/10.1145/3240508.3240556
Zhang, H., Luo, C., Wang, Q., Kitchin, M., Parmley, A., Monge-Alvarez, J., & Casaseca-de-la-Higuera, P. (2018). A novel infrared video surveillance system using deep learning based techniques. Multimedia Tools and Applications, 77(20), 26657–26676. https://doi.org/10.1007/s11042-018-5883-y
Zhang, Y., Zhang, T., Wang, S., & Yu, P. (2025). An Efficient Perceptual Video Compression Scheme Based on Deep Learning-Assisted Video Saliency and Just Noticeable Distortion. Engineering Applications of Artificial Intelligence, 141, 109806. https://doi.org/10.1016/j.engappai.2024.109806
Downloads
Published
How to Cite
Issue
Section
Citation Check
License
Copyright (c) 2025 Yuli Wahyuningsih, Ikhwanuddin Abdul Majid, Faisal Efendi, Arif Wiyat Purnanto

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.







