Monitoring Learning Activities in PLE Using Semantic Modelling of Learner Behaviour
- 5 Citations
- 1 Mentions
- 3.3k Downloads
Abstract
We report on the reflection of learning activities and revealing hidden information based on tracked user behaviour in our widget based PLE (Personal Learning Environment) at Graz University of Technology. Our reference data set includes information of more then 4000 active learners for a period of around two years. We have modelled activity and usage traces using domain specific ontologies like Activity Ontology and Learning Context Ontology from the IntelLEO EU project. Generally we distinguish three different metrics: user centric, learning object (widget) centric and activity centric. We have used Semantic Web query languages like SPARQL and representation formats like RDF to implement a human and machine readable web service along with a learning analytics dashboard for metrics visualization. The results offer a quick overview of learning habits, preferred set-ups of learning objects (widgets) and overall reflection of usages and activity dynamics in the PLE platform over time. The architecture delivers insights for intervening and recommending as closure of a learning analytics cycle[1] to optimize confidence in the PLE.
Keywords
PLE Semantic Web Learning Analytics Reflection RDF SPARQLPreview
Unable to display preview. Download preview PDF.
References
- 1.Clow, D.: The learning analytics cycle: closing the loop effectively. In: Proceedings of the 2nd International Conference on Learning Analytics and Knowledge, LAK 2012, pp. 134–138. ACM, New York (2012)Google Scholar
- 2.Ebner, M., Holzinger, A., Maurer, H.A.: Web 2.0 technology: Future interfaces for technology enhanced learning? In: Stephanidis, C. (ed.) HCI 2007. LNCS, vol. 4556, pp. 559–568. Springer, Heidelberg (2007)CrossRefGoogle Scholar
- 3.Pohl, M., Holzinger, A., Rester, M., Motschnig, R., Ebner, M., Leitner, G.: Gestaltung von innovativen technologiegesttzten lernsystemen am beispiel von web 2.0-anwendungen. eine herausforderung fr hci. OCG Journal (33), 20–23 (2008)Google Scholar
- 4.Holzinger, A., Searle, G., Wernbacher, M.: The effect of previous exposure to technology on acceptance and its importance in usability and accessibility engineering. Universal Access in the Information Society 10, 245–260 (2011)CrossRefGoogle Scholar
- 5.Reinhardt, W., Mletzko, C., Drachsler, H., Sloep, P.: Awesome: A widget-based dashboard for awareness-support in research networks. In: Learning, PLE Conference, pp. 1–15 (2011)Google Scholar
- 6.Santos Odriozola, J.L., Verbert, K., Govaerts, S., Duval, E.: Visualizing PLE usage. In: Proceedings of EFEPLE 2011 1st Workshop on Exploring the Fitness and Evolvability of Personal Learning Environments, CEUR WS, pp. 34–38 (August 2011)Google Scholar
- 7.Pardo, A., Kloos, C.D.: Stepping out of the box: towards analytics outside the learning management system. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 163–167. ACM, New York (2011)Google Scholar
- 8.Mazza, R., Milani, C.: Exploring usage analysis in learning systems: Gaining insights from visualisations. In: Workshop on Usage Analysis in Learning Systems, Proceedings of Artificial Intelligence in Education, Amsterdam (2005)Google Scholar
- 9.Drachsler, H., Bogers, T., Vuorikari, R., Verbert, K., Duval, E., Manouselis, N., Beham, G., Lindstaedt, S., Stern, H., Friedrich, M., Wolpers, M.: Issues and considerations regarding sharable data sets for recommender systems in technology enhanced learning. Procedia Computer Science 1(2), 2849–2858 (2010); Proceedings of the 1st Workshop on Recommender Systems for Technology Enhanced Learnin (RecSysTEL 2010)Google Scholar
- 10.Rosen, D., Miagkikh, V., Suthers, D.: Social and semantic network analysis of chat logs. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 134–139. ACM, New York (2011)Google Scholar
- 11.Santos, J.L., Govaerts, S., Verbert, K., Duval, E.: Goal-oriented visualizations of activity tracking: a case study with engineering students. In: Proceedings of the 2nd International Conference on Learning Analytics and Knowledge, LAK 2012, pp. 143–152. ACM, New York (2012)Google Scholar
- 12.Bakharia, A., Dawson, S.: Snapp: a bird’s-eye view of temporal participant interaction. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 168–173. ACM, New York (2011)Google Scholar
- 13.Niemann, K., Schmitz, H.C., Scheffel, M., Wolpers, M.: Usage contexts for object similarity: exploratory investigations. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 81–85. ACM, New York (2011)Google Scholar
- 14.Lockyer, L., Dawson, S.: Learning designs and learning analytics. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 153–156. ACM, New York (2011)Google Scholar
- 15.Richards, G., DeVries, I.: Revisiting formative evaluation: dynamic monitoring for the improvement of learning activity design and delivery. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 157–162. ACM, New York (2011)Google Scholar
- 16.Blikstein, P.: Using learning analytics to assess students’ behavior in open-ended programming tasks. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 110–116. ACM, New York (2011)Google Scholar
- 17.Sharkey, M.: Academic analytics landscape at the university of phoenix. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 122–126. ACM, New York (2011)Google Scholar
- 18.Fancsali, S.E.: Variable construction for predictive and causal modeling of online education data. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 54–63. ACM, New York (2011)Google Scholar
- 19.Verbert, K., Drachsler, H., Manouselis, N., Wolpers, M., Vuorikari, R., Duval, E.: Dataset-driven research for improving recommender systems for learning. In: Proceedings of the 1st International Conference on Learning Analytics and Knowledge, LAK 2011, pp. 44–53. ACM, New York (2011)Google Scholar
- 20.Verbert, K., Manouselis, N., Drachsler, H., Duval, E.: Dataset-driven research to support learning and knowledge analytics. Educational Technology & Society 15(3), 133–148 (2012)Google Scholar
- 21.Prinsloo, P., Slade, S., Galpin, F.: Learning analytics: challenges, paradoxes and opportunities for mega open distance learning institutions. In: Proceedings of the 2nd International Conference on Learning Analytics and Knowledge, LAK 2012, pp. 130–133. ACM, New York (2012)Google Scholar
- 22.Mulwad, V., Finin, T., Syed, Z., Joshi, A.: Using linked data to interpret tables. In: Hartig, O., Harth, A., Sequeda, J. (eds.) COLD. CEUR Workshop Proceedings, vol. 665. CEUR-WS.org (2010)Google Scholar
- 23.Jeremic, Z., Jovanovic, J., Gasevic, D.: Personal learning environments on the social semantic web. Semantic Web 4(1), 23–51 (2013)Google Scholar
- 24.Siadaty, M., Jovanovic, J., Pata, K., Holocher-Ertl, T., Gasevic, D., Milikic, N.: A semantic web-enabled tool for self-regulated learning in the workplace. In: ICALT, pp. 66–70. IEEE Computer Society (2011)Google Scholar
- 25.Kirchberg, M., Ko, R.K.L., Lee, B.S.: From linked data to relevant data – time is the essence. CoRR abs/1103.5046 (2011)Google Scholar