Skip to main content

The Proof of the Pudding: Examining Validity and Reliability of the Evaluation Framework for Learning Analytics

  • Conference paper
  • First Online:
Book cover Data Driven Approaches in Digital Education (EC-TEL 2017)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 10474))

Included in the following conference series:

Abstract

While learning analytics (LA) is maturing from being a trend to being part of the institutional toolbox, the need for more empirical evidences about the effects for LA on the actual stakeholders, i.e. learners and teachers, is increasing. Within this paper we report about a further evaluation iteration of the Evaluation Framework for Learning Analytics (EFLA) that provides an efficient and effective measure to get insights into the application of LA in educational institutes. For this empirical study we have thus developed and implemented several LA widgets into a MOOC platform’s dashboard and evaluated these widgets using the EFLA as well as the framework itself using principal component and reliability analysis. The results show that the EFLA is able to measure differences between widget versions. Furthermore, they indicate that the framework is highly reliable after slightly adapting its dimensions.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    https://ecolearning.eu.

References

  1. Arnold, K.E., Lonn, S., Pistilli, M.D.: An exercise in institutional reflection: the learning analytics readiness instrument (LARI). In: Proceedings of the 4th International Conference on Learning Analytics and Knowledge, LAK 2014, pp. 163–167. ACM, New York (2014)

    Google Scholar 

  2. Beheshitha, S., Hatala, M., Gašević, D., Joksimovic, S.: The role of achievement goal orientations when studying effect of learning analytics visualizations. In: Proceedings of the 6th International Conference on Learning Analytics and Knowledge, LAK 2016, pp. 54–63. ACM, New York (2016)

    Google Scholar 

  3. Brooke, J.: SUS: a quick and dirty usability scale. In: Jordan, P.W., Weerdmeester, B., Thomas, A., Mclelland, I.L. (eds.) Usability evaluation in industry. Taylor and Francis, London (1996)

    Google Scholar 

  4. Butler, D., Winne, P.: Feedback and self-regulated learning: a theoretical synthesis. Rev. Educ. Res. 65(3), 245–281 (1995)

    Article  Google Scholar 

  5. Cobo, A., Rocha, R., Rodriguez-Hoyos, C.: Evaluation of the interactivity of students in virtual learning environments using a multicriteria approach and data mining. Behav. Inf. Technol. 33(10), 1000–1012 (2014)

    Article  Google Scholar 

  6. Drachsler, H., Kalz, M.: The MOOC and learning analytics innovation cycle (MOLAC): a reflective summary of ongoing research and its challenges. J. Comput. Assist. Learn. 32(3), 281–290 (2016)

    Article  Google Scholar 

  7. Endsley, M.R.: Toward a theory of situation awareness in dynamic systems. Hum. Factors 37(1), 32–64 (1995)

    Article  Google Scholar 

  8. Ferguson, R., Clow, D.: Learning analytics community exchange: evidence hub. In: Proceedings of the 6th International Conference on Learning Analytics and Knowledge, LAK 2016, pp. 520–521. ACM, New York (2016)

    Google Scholar 

  9. Gašević, D., Dawson, S., Mirriahi, N., Long, P.: Learning analytics - a growing field and community engagement. J. Learn. Anal. 2(1), 1–6 (2015)

    Article  Google Scholar 

  10. Gašević, D., Dawson, S., Siemens, G.: Let’s not forget: learning analytics are about learning. TechTrends 59(1), 64–71 (2015)

    Article  Google Scholar 

  11. Greller, W., Drachsler, H.: Translating learning into numbers: a generic framework for learning analytics. Educ. Technol. Soc. 15(3), 42–57 (2012)

    Google Scholar 

  12. Khan, I., Pardo, A.: Data2U: scalable real time student feedback in active learning environments. In: Proceedings of the 6th International Conference on Learning Analytics and Knowledge, LAK 2016, pp. 249–253. ACM, New York (2016)

    Google Scholar 

  13. Kim, J., Jo, I.H., Park, Y.: Effects of learning analytics dashboard: analyzing the relations among dashboard utilization, satisfaction, and learning achievement. Asia Pac. Educ. Rev. 17(1), 13–24 (2016)

    Article  Google Scholar 

  14. Long, P., Siemens, G.: Penetrating the fog: analytics in learning and education. EDUCAUSE Rev. 46(5), 31–40 (2011)

    Google Scholar 

  15. Lonn, S., Aguilar, S., Teasley, S.: Investigating student motivation in the context of a learning analytics intervention during a summer bridge program. Comput. Hum. Behav. 47, 90–97 (2015)

    Article  Google Scholar 

  16. Persico, D., Pozzi, F.: Informing learning design with learning analytics to improve teacher inquiry. Br. J. Educ. Technol. 46(2), 230–248 (2014)

    Article  Google Scholar 

  17. Scheffel, M., Drachlser, H., Specht, M.: Developing an evaluation framework of quality indicators for learning analytics. In: Proceedings of the 5th International Conference on Learning Analytics and Knowledge, LAK 2015, pp. 16–20. ACM, New York (2015)

    Google Scholar 

  18. Scheffel, M., Drachsler, H., Kreijns, K., de Kraker, J., Specht, M.: Widget, widget as you lead, i am performing well indeed!: using results from an exploratory offline study to inform an empirical online study about a learning analytics widget in a collaborative learning environment. In: Proceedings of the 7th International Conference on Learning Analytics and Knowledge, LAK 2017, pp. 289–298. ACM, New York (2017)

    Google Scholar 

  19. Scheffel, M., Drachsler, H., Stoyanov, S., Specht, M.: Quality indicators for learning analytics. Educ. Technol. Soc. 17(4), 117–132 (2014)

    Google Scholar 

  20. Schön, D.: The Reflective Practitioner: How Professionals Think in Action. Temple Smith, London (1983)

    Google Scholar 

  21. Siemens, G., Dawson, S., Lynch, G.: Improving the quality and productivity of the higher education sector - policy and strategy for system-level deployment of learning analytics. Discussion paper for the Australian Government, Society for Learning Analytics Research (SoLAR) (2013)

    Google Scholar 

  22. Verbert, K., Govaerts, S., Duval, E., Santos, J.L., Assche, F., Parra, G., Klerkx, J.: Learning dashboards: an overview and future research opportunities. Pers. Ubiquit. Comput. 18(6), 1499–1514 (2014)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Maren Scheffel .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Scheffel, M., Drachsler, H., Toisoul, C., Ternier, S., Specht, M. (2017). The Proof of the Pudding: Examining Validity and Reliability of the Evaluation Framework for Learning Analytics. In: Lavoué, É., Drachsler, H., Verbert, K., Broisin, J., Pérez-Sanagustín, M. (eds) Data Driven Approaches in Digital Education. EC-TEL 2017. Lecture Notes in Computer Science(), vol 10474. Springer, Cham. https://doi.org/10.1007/978-3-319-66610-5_15

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-66610-5_15

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-66609-9

  • Online ISBN: 978-3-319-66610-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics