Abstract
Because of the flexibility of online learning courses, students organise and manage their own learning time by choosing where, what, how, and for how long they study. Each individual has their unique learning habits that characterise their behaviours and distinguish them from others. Nonetheless, to the best of our knowledge, the temporal dimension of student learning has received little attention on its own. Typically, when modelling trends, a chosen configuration is set to capture various habits, and a cluster analysis is undertaken. However, the selection of variables to observe and the algorithm used to conduct the analysis is a subjective process that reflects the researcher’s thoughts and ideas. To explore how students behave over time, we present alternative ways of modelling student temporal behaviour. Our real-world data experiments reveal that the generated clusters may or may not differ based on the selected profile and unveil different student learning patterns.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
- 2.
Both courses were structured with Assignment (11,4), Book (1,1), Database (1,1), Feedback(1,1), File(6,24), Forum(7,6), Glossary(1,1), H5P(6,9), Lesson(5,8), Page(2,10), Quiz(5,10), Survey(6,4), URL(10,14), Wiki(1,1). The numbers represent the quantity of resources/activities available in course A and B.
- 3.
K-means has been implemented in Python with scikit-learn, Bisecting K-means with pyclustering, and Hierarchical clustering with scipy. For K-Means we select the number of cluster k analysing the SSE curve. For Bisecting K-Means we vary the parameters controlling the split. For the hierarchical we obtain the clusters by cutting the hierarchy w.r.t. the median value of the distance matrix.
References
Beaudoin, M.F.: Learning or lurking?: tracking the “invisible’’ online student. Internet High. Educ. 5(2), 147–155 (2002)
Bovo, A.: Clustering moodle data as a tool for profiling students. In: ICEEE (2013)
Chen, B., Knight, S., Wise, A.: Critical issues in designing and implementing temporal analytics. J. Learn. Anal. (2018)
Credé, M., et al.: Study habits, skills, and attitudes: the third pillar supporting collegiate academic performance. Perspect. Psychol. Sci. 3(6), 425–453 (2008)
Dermy, O., Brun, A.: Can we take advantage of time-interval pattern mining to model students activity? In: EDM (2020)
Dixson, M.D.: Measuring student engagement in the online course: the online student engagement scale (OSE). Online Learn. 19(4), n4 (2015)
Dunlosky, J., et al.: Improving students’ learning with effective learning techniques: promising directions from cognitive and educational psychology. Psychol. Sci. Public Interest 14(1), 4–58 (2013)
Fleming, N.D., Mills, C.: Not another inventory, rather a catalyst for reflection. Improve Acad. 11(1), 137–155 (1992)
Goda, Y., et al.: Procrastination and other learning behavioral types in e-learning and their relationship with learning outcomes. Learn. Indiv. Diff. 37, 72–80 (2015)
Hart, C.: Factors associated with student persistence in an online program of study: a review of the literature. J. Interact. Online Learn. (2012)
Hecking, T., Ziebarth, S., Hoppe, H.U.: Analysis of dynamic resource access patterns in a blended learning course. In: LAK (2014)
Henrie, C.R., Halverson, L.R., Graham, C.R.: Measuring student engagement in technology-mediated learning: a review. Comput. Educ. 90, 36–53 (2015)
Lee, Y.: Effect of uninterrupted time-on-task on students’ success in massive open online courses (MOOCs). Comput. Hum. Behav. 86, 174–180 (2018)
Macfadyen, L.P., Dawson, S.: Mining LMS data to develop an “early warning system’’ for educators: a proof of concept. Comput. Educ. 54(2), 588–599 (2010)
Riel, J., Lawless, K.A., Brown, S.W.: Timing matters: approaches for measuring and visualizing behaviours of timing and spacing of work in self-paced online teacher professional development courses. JLA 5(1), 25–40 (2018)
Rotelli, D., Monreale, A.: Time-on-task estimation by data-driven outlier detection based on learning activities. In: LAK22 (2022)
Schmitz, B., et al.: New perspectives for the evaluation of training sessions in self-regulated learning: time-series analyses of diary data. Contemp. Educ. Psychol. 31(1), 64–96 (2006)
Sherin, B.: Using computational methods to discover student science conceptions in interview data. In: LAK (2012)
Theobald, M., et al.: Identifying individual differences using log-file analysis: distributed learning as mediator between conscientiousness and exam grades. Learn. Indiv. Dif. 65, 112–122 (2018)
Thindwa, H.: The role of technology in improving quality of teaching in higher education: an international perspective. Teach. Educ. (2016)
Weinstein, C.E., Palmer, D.R., Schultz, A.: Lassi. User’s Manual for those administering Learning and Study Strategies Inventory (2002)
Acknowledgments
This work is supported by the EU H2020 Program under the scheme H2020-INFRAIA-2019-1: Research Infrastructure G.A. 871042 SoBigData++.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Rotelli, D., Monreale, A., Guidotti, R. (2022). Uncovering Student Temporal Learning Patterns. In: Hilliger, I., Muñoz-Merino, P.J., De Laet, T., Ortega-Arranz, A., Farrell, T. (eds) Educating for a New Future: Making Sense of Technology-Enhanced Learning Adoption. EC-TEL 2022. Lecture Notes in Computer Science, vol 13450. Springer, Cham. https://doi.org/10.1007/978-3-031-16290-9_25
Download citation
DOI: https://doi.org/10.1007/978-3-031-16290-9_25
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-16289-3
Online ISBN: 978-3-031-16290-9
eBook Packages: Computer ScienceComputer Science (R0)