Abstract
One type of visualization of data from digital learning environments focuses on students’ interaction with the educational content. Students may, for example, answer questions, read texts, or solve problems. We can represent these interactions as a matrix, where rows correspond to students, columns to educational items, and values to some aspect of student activity (e.g., the correctness of answers, response times, the order of actions). Visualizing this matrix is useful for several purposes. For teachers, it can provide an understanding of the skill and behavior of their students. For system developers, it can provide insight into the behavior of both students and adaptive algorithms, and it can also help detect suspicious activity. For researchers, it can provide an understanding of the properties of datasets used in experiments and valuable warnings about biases that are present in data. However, suitable visualization of the student-item interactions is nontrivial. To facilitate the design of the visualization, we provide a systematic discussion of approaches to student-item matrix visualization. Using data from an introductory programming exercise, we also provide specific illustrations of different visualization designs.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Baker, R., Walonoski, J., Heffernan, N., Roll, I., Corbett, A., & Koedinger, K. (2008). Why students engage in “gaming the system” behavior in interactive learning environments. Journal of Interactive Learning Research, 19(2), 185–224.
Bowers, A. J. (2010). Analyzing the longitudinal K-12 grading histories of entire cohorts of students: Grades, data driven decision making, dropping out and hierarchical cluster analysis. Practical Assessment, Research, and Evaluation, 15(1), 7.
Brusilovsky, P., Somyürek, S., Guerra, J., Hosseini, R., Zadorozhny, V., & Durlach, P. J. (2016). Open social student modeling for personalized learning. IEEE Transactions on Emerging Topics in Computing, 4(3), 450–461.
Čechák, J., & Pelánek, R. (2019). Item ordering biases in educational data. In Proceedings of Artificial Intelligence in Education (pp. 48–58). Springer.
Coffrin, C., Corrin, L., de Barba, P., & Kennedy, G. (2014). Visualizing patterns of student engagement and performance in MOOCs. In Proceedings of Learning Analytics and Knowledge (pp. 83–92).
Confrey, J., Gianopulos, G., McGowan, W., Shah, M., & Belcher, M. (2017). Scaffolding learner-centered curricular coherence using learning maps and diagnostic assessments designed around mathematics learning trajectories. ZDM, 49(5), 717–734.
Costagliola, G., Fuccella, V., Giordano, M., & Polese, G. (2008). Monitoring online tests through data visualization. IEEE Transactions on Knowledge and Data Engineering, 21(6), 773–784.
Desmarais, M., & Lemieux, F. (2013). Clustering and visualizing study state sequences. In Proceedings of Educational Data Mining.
Eagle, M., & Barnes, T. (2014). Survival analysis on duration data in intelligent tutors. In Proceedings of Intelligent Tutoring Systems (pp. 178–187). Springer.
Effenberger, T., & Pelánek, R. (2021). Validity and reliability of student models for problem-solving activities. In Proceedings of Learning Analytics and Knowledge (pp. 1–11).
France, L., Heraud, J.-M., Marty, J.-C., Carron, T., & Heili, J. (2006). Monitoring virtual classroom: Visualization techniques to observe student activities in an e-learning system. In Proceedings of Advanced Learning Technologies (pp. 716–720). IEEE.
Fu, X., Shimada, A., Ogata, H., Taniguchi, Y., & Suehiro, D. (2017). Real-time learning analytics for C programming language courses. In Proceedings of the Learning Analytics & Knowledge (pp. 280–288).
Janssenswillen, G., Depaire, B., Swennen, M., Jans, M., & Vanhoof, K. (2019). BupaR: Enabling reproducible business process analysis. Knowledge-Based Systems, 163, 927–930.
Johnson, M., Eagle, M., & Barnes, T. (2013). Invis: An interactive visualization tool for exploring interaction networks. In Proceedings of Educational Data Mining.
Khajah, M., Wing, R., Lindsey, R., & Mozer, M. (2014). Integrating latent-factor and knowledge-tracing models to predict individual differences in learning. In Proceedings of Educational Data Mining.
Lee, J. E., Recker, M., Bowers, A., & Yuan, M. (2016). Hierarchical cluster analysis heatmaps and pattern analysis: An approach for visualizing learning management system interaction data. In Proceedings of Educational Data Mining (pp. 603–604).
Mazza, R., & Dimitrova, V. (2007). CourseVis: A graphical student monitoring tool for supporting instructors in web-based distance courses. International Journal of Human-Computer Studies, 65(2), 125–139.
Molenaar, I., & Knoop-van Campen, C. (2017). Teacher dashboards in practice: Usage and impact. In European Conference on Technology Enhanced Learning (pp. 125–138). Springer.
Monti, D., Rizzo, G., & Morisio, M. (2019). Visualizing ratings in recommender system datasets. In IntRS@ RecSys (pp. 60–64).
Nixon, T., Fancsali, S., & Ritter, S. (2013). The complex dynamics of aggregate learning curves. In Proceedings of Education Data Mining (pp. 338–339).
Northcutt, C. G., Ho, A. D., & Chuang, I. L. (2016). Detecting and preventing “multiple-account” cheating in massive open online courses. Computers & Education, 100, 71–80.
Pelánek, R. (2017). Bayesian knowledge tracing, logistic models, and beyond: An overview of learner modeling techniques. User Modeling and User-Adapted Interaction, 27(3–5), 313–350.
Pelánek, R. (2018). The details matter: Methodological nuances in the evaluation of student models. User Modeling and User-Adapted Interaction, 28(3), 207–235.
Pelánek, R., & Effenberger, T. (2020). Beyond binary correctness: Classification of students’ answers in learning systems. User Modeling and User-Adapted Interaction, 27(1), 89–118.
Reddy, S., Labutov, I., Banerjee, S., & Joachims, T. (2016). Unbounded human learning: Optimal scheduling for spaced repetition. In Proceedings of Knowledge Discovery and Data Mining (pp. 1815–1824).
Řihák, J., & Pelánek, R. (2016). Choosing a student model for a real world application. In Building ITS bridges across frontiers (ITS workshop).
Sedrakyan, G., Snoeck, M., & De Weerdt, J. (2014). Process mining analysis of conceptual modeling behavior of novices–empirical study using JMermaid modeling and experimental logging environment. Computers in Human Behavior, 41, 486–503.
Song, M., & van der Aalst, W. M. (2007). Supporting process mining by showing events at a glance. In Proceedings of Workshop on Information Technologies and Systems (pp. 139–145).
Trcka, N., Pechenizkiy, M., & van der Aalst, W. (2010). Process mining from educational data. Handbook of Educational Data Mining, 123–142.
van der Aalst, W. M., Guo, S., & Gorissen, P. (2013). Comparative process mining in education: An approach based on process cubes. In International symposium on data-driven process discovery and analysis (pp. 110–134). Springer.
Wallner, G., & Kriglstein, S. (2013). Visualization-based analysis of gameplay data–a review of literature. Entertainment Computing, 4(3), 143–155.
Wang, C.-H., & Chen, C.-P. (2013). Employing online SP diagnostic table for qualitative comments on test results. Electronic Journal of E-Learning, 11(3), 263–271.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Effenberger, T., Pelánek, R. (2021). Visualization of Student-Item Interaction Matrix. In: Sahin, M., Ifenthaler, D. (eds) Visualizations and Dashboards for Learning Analytics. Advances in Analytics for Learning and Teaching. Springer, Cham. https://doi.org/10.1007/978-3-030-81222-5_20
Download citation
DOI: https://doi.org/10.1007/978-3-030-81222-5_20
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-81221-8
Online ISBN: 978-3-030-81222-5
eBook Packages: EducationEducation (R0)