2. (Social) Learning Analytics
“LA is the measurement, collection, analysis and reporting of data about learners
and their contexts, for purposes of understanding and optimising learning and the
environments in which it occurs” (LAK 2011)
Social LA “focuses on how learners build knowledge together in their cultural
and social settings” (Ferguson & Buckingham Shum, 2012)
Ferguson, R., & Buckingham Shum, S. (2012). Social learning analytics: five approaches. 2nd International Conference on Learning Analytics and Knowledge, Vancouver, British Columbia, Canada.
3. Dyckhoff, A. L., Zielke, D., Bültmann, M., Chatti, M. A., & Schroeder, U. (2012). Design and Implementation of a Learning Analytics Toolkit for Teachers. Journal of Educational Technology & Society, 15(3), 58-76.
4. Dyckhoff, A. L., Zielke, D., Bültmann, M., Chatti, M. A., & Schroeder, U. (2012). Design and Implementation of a Learning Analytics Toolkit for Teachers. Journal of Educational Technology & Society, 15(3), 58-76.
5. Adeniji, B. (2019). A Bibliometric Study on Learning Analytics. Long Island University. Retrieved from https://digitalcommons.liu.edu/post_fultext_dis/16/
6.
7. “In the UK the Open University (OU) is a world leader in the collection, intelligent analysis and use of large scale student
analytics. It provides academic staff with systematic and high quality actionable analytics for student, academic and
institutional benefit (Rienties, Nguyen, Holmes, Reedy, 2017). Rienties and Toetenel’s, 2016 study (Rienties & Toetenel,
2016) identifies the importance of the linkage between LA outcomes, student satisfaction, retention and module learning
design. These analytics are often provided through dashboards tailored for each of academics and students
(Schwendimann et al., 2017).
The OU’s world-class Analytics4Action initiative (Rienties, Boroowa, Cross, Farrington-Flint et al., 2016) supports the
university-wide approach to LA. In particular, the initiative provided valuable insights into the identification of students and
modules where interventions would be beneficial, analysing over 90 large-scale modules over a two-year period…
The deployment of LA establishes the need and opportunity for student and module interventions (Clow, 2012). The
study concludes that the faster the feedback loop to students, the more effective the outcomes. This is often an iterative
process allowing institutions to understand and address systematic issues.
Legal, ethical and moral considerations in the deployment of LA and interventions are key challenges to institutions.
They include informed consent, transparency to students, the right to challenge the accuracy of data and resulting analyses
and prior consent to intervention processes and their execution (Slade & Tait, 2019)”
Wakelam, E., Jefferies, A., Davey, N., & Sun, Y. (2020). The potential for student performance
prediction in small cohorts with minimal available attributes. British Journal of Educational
Technology, 51(2), 347-370. doi: 10.1111/bjet.12836
9. Leading global distance learning, delivering high-quality education to anyone, anywhere, anytime
The Open University
Largest
University
in Europe
No formal
entry
requirements
enter with one
A-level or less
33%
38%
of part-time
undergraduates
taught by OU in UK
173,927 formal
students
55%
of students are
'disadvantaged'
FTSE 100 have
sponsored staff on OU
courses in 2017/8
60%
66%
of new
undergraduates
are 25+ 1,300
Open University students
has a disability (23,630)
1 in 8
Students are
already in work
3 in 4
employers use
OU learning
solutions to
develop
workforce
10. A special thanks to Vaclav Bayer, Avinash Boroowa, Shi-Min Chua, Simon Cross, Doug Clow, Chris Edwards, Rebecca Ferguson, Mark Gaved, Christothea Herodotou, Martin Hlosta, Wayne
Holmes, Garron Hillaire, Simon Knight, Nai Li, Vicky Marsh, Kevin Mayles, Jenna Mittelmeier, Vicky Murphy, Mark Nichols, Quan Nguygen, Tom Olney, Lynda Prescott, John Richardson, Saman
Rizvi, Jekaterina Rogaten, Matt Schencks, Mike Sharples, Dirk Tempelaar, Belinda Tynan, Lisette Toetenel, Thomas Ullmann, Denise Whitelock, Zdenek Zdrahal, and others…
11. What we have learned in six years at the OU
Change is slow, but can be enhanced with:
1. Clear senior management support
2. Bottom-up support from teachers and researchers who are
willing to take a risk
3. Evidence-based research can gradually change perspectives
and narratives
4. You quickly forget about the small/medium/large successes
and fail to realise that you are making a real impact
5. Large-scale innovation takes substantial time and effort
6. It is all about people…
12. So what do you want me to focus on?
1) How the OU uses predictive learning
analytics at scale for the last 6 years
2) How the OU uses learning analytics to
improve our learning design
3) What do OU practitioners want as next steps
for learning analytics and where should
distance learning institutions be going?
13. Predictive analytics and professional development
Kuzilek, J., Hlosta, M., Herrmannova, D., Zdrahal, Z., & Wolff, A. (2015). OU Analyse: analysing at-risk students at The Open University LACE Learning Analytics Review (Vol. LAK15-1). Milton Keynes: Open University.
Kuzilek, J., Hlosta, M., & Zdrahal, Z. (2017). Open University Learning Analytics dataset. Scientific Data, 4, 170171. doi: 10.1038/sdata.2017.171
Wolff, A., Zdrahal, Z., Herrmannova, D., Kuzilek, J., & Hlosta, M. (2014). Developing predictive models for early detection of at-risk students on distance learning modules, Workshop: Machine Learning and Learning Analytics
Paper presented at the Learning Analytics and Knowledge (2014), Indianapolis.
14. Start
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
Pass Fail No submit TMA-1time
VLE opens
Start
Activity space
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
15. FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
Start
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
Pass Fail No submit TMA-1time
VLE opens
Start
VLE trail: successful
student
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
16. FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
Start
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
FSF RFSOFS ORFN O SRFROF OR ORSORFS OS RS
Pass Fail No submit TMA-1time
VLE opens
Start
VLE trail: student who
did not submit
19. Herodotou, C., Rienties, B., Hlosta, M., Boroowa, A., Mangafa, C., Zdrahal, Z., (2020). Scalable implementation of predictive learning analytics at a distance learning university:
Insights from a longitudinal case study. Internet and Higher Education, 45, 100725.
20. Herodotou, C., Rienties, B., Hlosta, M., Boroowa, A., Mangafa, C., Zdrahal, Z., (2020). Scalable implementation of predictive learning analytics at a distance learning university:
Insights from a longitudinal case study. Internet and Higher Education, 45, 100725.
Amongst the factors shown to be critical to the scalable PLA implementation were: Faculty's
engagement with OUA, teachers as “champions”, evidence generation and dissemination,
digital literacy, and conceptions about teaching online.
22. Magic of learning design (does not come easy)
“Research on the relationship between learning design and learning
analytics has also been a focus in European research in recent years. For
example, in their research at the Open University UK, Toetenel and
Rienties combine learning design and learning analytics where learning
design provides context to empirical data about OU courses enabling the
learning analytics to give insight into learning design decisions. This
research is important as it attempts to close the virtuous cycle
between learning design to improve courses and enhancing the
quality of learning, something that has been lacking in the research
literature. For example, they study the impact of learning design on
pedagogical decision-making and on future course design, and the
relationship between learning design and student behaviour and outcomes
(Toetenel and Rienties 2016; Rienties and Toetenel 2016; Rienties et al.
2015).”
Wasson, B., & Kirschner, P. A. (2020). Learning Design: European Approaches. TechTrends, 1-13.
23. McAndrew, P., Nadolski, R. and Little, A., 2005. Developing an approach for Learning Design Players. Journal of Interactive Media in Education, 2005(1), p.Art.
15. DOI: http://doi.org/10.5334/2005-14
24. Assimilative Finding and
handling
information
Communication Productive Experiential Interactive/
Adaptive
Assessment
Type of activity Attending to
information
Searching for
and processing
information
Discussing
module related
content with at
least one other
person (student
or tutor)
Actively
constructing an
artefact
Applying
learning in a
real-world
setting
Applying
learning in a
simulated
setting
All forms of
assessment,
whether
continuous, end
of module, or
formative
(assessment for
learning)
Examples of
activity
Read, Watch,
Listen, Think
about, Access,
Observe,
Review, Study
List, Analyse,
Collate, Plot,
Find, Discover,
Access, Use,
Gather, Order,
Classify, Select,
Assess,
Manipulate
Communicate,
Debate, Discuss,
Argue, Share,
Report,
Collaborate,
Present,
Describe,
Question
Create, Build,
Make, Design,
Construct,
Contribute,
Complete,
Produce, Write,
Draw, Refine,
Compose,
Synthesise,
Remix
Practice, Apply,
Mimic,
Experience,
Explore,
Investigate,
Perform,
Engage
Explore,
Experiment,
Trial, Improve,
Model, Simulate
Write, Present,
Report,
Demonstrate,
Critique
Conole, G. (2012). Designing for Learning in an Open World. Dordrecht: Springer.
Rienties, B., Toetenel, L., (2016). The impact of learning design on student behaviour, satisfaction and performance: a cross-institutional comparison across 151
modules. Computers in Human Behavior, 60 (2016), 333-341
Open University Learning Design Initiative (OULDI)
25. Toetenel, L., Rienties, B. (2016). Analysing 157 Learning Designs using Learning Analytic approaches as a means to evaluate the impact of pedagogical
decision-making. British Journal of Educational Technology, 47(5), 981–992.
26.
27. Merging big data sets
• Learning design data (>300 modules mapped)
• VLE data
• >140 modules aggregated individual data weekly
• >37 modules individual fine-grained data daily
• Student feedback data (>140)
• Academic Performance (>140)
• Predictive analytics data (>40)
• Data sets merged and cleaned
• 111,256 students undertook these modules
28. Nguyen, Q., Rienties, B., Toetenel, L., Ferguson, R., Whitelock, D. (2017). Examining the designs of computer-based assessment and its impact on student
engagement, satisfaction, and pass rates. Computers in Human Behavior. DOI: 10.1016/j.chb.2017.03.028.
69% of what students are
doing in a week is
determined by us, teachers!
29. Constructivist
Learning Design
Assessment
Learning Design
Productive
Learning Design
Socio-construct.
Learning Design
VLE Engagement
Student
Satisfaction
Student
retention
150+ modules
Week 1 Week 2 Week30
+
Rienties, B., Toetenel, L., (2016). The impact of learning design on student behaviour, satisfaction and performance: a cross-institutional comparison across 151
modules. Computers in Human Behavior, 60 (2016), 333-341
Nguyen, Q., Rienties, B., Toetenel, L., Ferguson, R., Whitelock, D. (2017). Examining the designs of computer-based assessment and its impact on student
engagement, satisfaction, and pass rates. Computers in Human Behavior. DOI: 10.1016/j.chb.2017.03.028.
Communication
32. Design
Inquiry First Module 2nd Module
NthModule
QualificationLife long learning
Design
Design
Rienties, B., Olney, T., Nichols, M., Herodotou, C. (2020). Effective usage of Learning Analytics: What do practitioners want and where should distance learning institutions be going? Open Learning,
35(2), 178-195
33. STUDENT SUCCESS ANALYTICS
33
O R G AN I S AT I O N AL C APAB I LT I E S
Productionised
output and MI
Strategic
analysis
Modelling /
AI
Data
collection
Data storage
and access
Technology
architecture
Learning
design and
delivery
Student
lifecycle
managemen
t
Continuous
improvemen
t and
innovation
Creation of
actionable
insight
Availability
of data
Impact the
student
experience
Adapted from Barton and Court (2012) - https://hbr.org/2012/10/making-advanced-analytics-work-for-you , updated by Kevin Mayles (2019)
34. Design Inquiry First Module 2nd Module
NthModule
QualificationLife long learning
Communication Integrated Design
Personalisation Evidence based
Integrated learning analytics solution
Alumni
Rienties, B., Olney, T., Nichols, M., Herodotou, C. (2020). Effective usage of Learning Analytics: What do practitioners want and where should distance learning institutions be going? Open Learning,
35(2), 178-195
35. What have I learned in six years at the OU
Change is slow, but can be enhanced with:
1. Clear senior management support
2. Bottom-up support from teachers and researchers who are
willing to take a risk
3. Evidence-based research can gradually change perspectives
and narratives
4. You quickly forget about the small/medium/large successes
and fail to realise that you are making a real impact
5. Large-scale innovation takes substantial time and effort
6. It is all about people…
36. Further reflections
1. Who owns the data?
2. What about the ethics?
3. What about professional development?
4. Are we optimising the record player?
38. References: see oro.open.ac.uk/
• Adeniji, B. (2019). A Bibliometric Study on Learning Analytics Long Island University]. https://digitalcommons.liu.edu/post_fultext_dis/16/
• Conole, G. (2012). Designing for Learning in an Open World [Book]. Springer.
• Dyckhoff, A. L., Zielke, D., Bültmann, M., Chatti, M. A., & Schroeder, U. (2012). Design and Implementation of a Learning Analytics Toolkit for Teachers [Article].
Journal of Educational Technology & Society, 15(3), 58-76.
http://libezproxy.open.ac.uk/login?url=http://search.ebscohost.com/login.aspx?direct=true&db=ehh&AN=79816990&site=ehost-live&scope=site
• Ferguson, R., & Buckingham Shum, S. (2012). Social learning analytics: five approaches. 2nd International Conference on Learning Analytics and Knowledge,
Vancouver, British Columbia, Canada.
• Herodotou, C., Rienties, B., Hlosta, M., Boroowa, A., Mangafa, C., & Zdrahal, Z. (2020). The scalable implementation of predictive learning analytics at a
distance learning university: Insights from a longitudinal case study. The Internet and Higher Education, 45, 100725. https://doi.org/10.1016/j.iheduc.2020.100725
• Kuzilek, J., Hlosta, M., Herrmannova, D., Zdrahal, Z., & Wolff, A. (2015). OU Analyse: analysing at-risk students at The Open University. Learning Analytics
Review, 1-16. http://oro.open.ac.uk/42529/1/__userdata_documents5_ajj375_Desktop_analysing-at-risk-students-at-open-university.pdf
• McAndrew, P., Nadolski, R., & & Little, A. (2005). Developing an approach for Learning Design Players. Journal of Interactive Media in Education, 2005(1).
https://doi.org/10.5334/2005-14
• Nguyen, Q., Rienties, B., Toetenel, L., Ferguson, F., & Whitelock, D. (2017). Examining the designs of computer-based assessment and its impact on student
engagement, satisfaction, and pass rates. Computers in Human Behavior, 76(November 2017), 703-714. https://doi.org/10.1016/j.chb.2017.03.028
• Rienties, B., Olney, T., Nichols, M., & Herodotou, C. (2020). Effective usage of Learning Analytics: What do practitioners want and where should distance
learning institutions be going? Open Learning, 35(2), 178-195.
• Rienties, B., & Toetenel, L. (2016). The impact of learning design on student behaviour, satisfaction and performance: a cross-institutional comparison across
151 modules. Computers in Human Behavior, 60, 333-341. https://doi.org/10.1016/j.chb.2016.02.074
• Toetenel, L., & Rienties, B. (2016). Analysing 157 Learning Designs using Learning Analytic approaches as a means to evaluate the impact of pedagogical
decision-making. British Journal of Educational Technology, 47(5), 981–992. https://doi.org/10.1111/bjet.12423
• Wakelam, E., Jefferies, A., Davey, N., & Sun, Y. (2019). The potential for student performance prediction in small cohorts with minimal available attributes. British
Journal of Educational Technology, 0(0). https://doi.org/10.1111/bjet.12836
• Wasson, B., & Kirschner, P. A. (2020). Learning Design: European Approaches. TechTrends, 1-13. https://link.springer.com/content/pdf/10.1007/s11528-020-
00498-0.pdf
• Wolff, A., Zdrahal, Z., Nikolov, A., & Pantucek, M. (2013). Improving retention: predicting at-risk students by analysing clicking behaviour in a virtual learning
environment. Proceedings of the Third International Conference on Learning Analytics and Knowledge, Indianapolis.
Case-based reasoning (reasoning from precedents, k-Nearest Neighbours)
Based on demographic data
Based on VLE activities
Classification and Regression Trees (CART)
Bayes networks (naïve and full)
Final verdict decided by voting
Explain seven categories
For each module, the learning design team together with module chairs create activity charts of what kind of activities students are expected to do in a week.
5131 students responded – 28%, between 18-76%
Cluster analysis of 40 modules (>19k students) indicate that module teams design four different types of modules: constructivist, assessment driven, balanced, or socio-constructivist. The LAK paper by Rienties and colleagues indicates that VLE engagement is higher in modules with socio-constructivist or balanced variety learning designs, and lower for constructivist designs. In terms of learning outcomes, students rate constructivist modules higher, and socio-constructivist modules lower. However, in terms of student retention (% of students passed) constructivist modules have lower retention, while socio-constructivist have higher. Thus, learning design strongly influences behaviour, experience and performance. (and we believe we are the first to have mapped this with such a large cohort).
‘Creating better opportunities with and for learners, enabling more to achieve their goals’
Creating– includes our content, people, systems and support services
Better – means continuous improvement, always striving to be better than we were yesterday
Opportunities – different learning options to meet different needs
With and For Learners– all learners from all backgrounds etc. which recognises their diversity and emphasising partnership
Enabling – making it easy to engage with us and our products
More – again, continuous improvement, striving to attract more students
To achieve their goals – which may be a new job or career, a qualification, an accreditation or learning for pleasure.