Download slides – ppt : 2.9GB - Effective Learning Analytics

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
June 2016
Learning Analytics – UDD & DPA Update
Unified Data Definition v1.2.4 & our evolving data processing agreement
Data refresher…
• Historical data (12-18 months minimum) for a Learning Analytics implementation
• Predictive Model Creation (12-36months – ideal!) & Pilot
• (Portable) Predictive Model Validation (12months minimum) & Pilot
• Engagement & activity (xAPI) data – VLE (Bb, Moodle), LMS, AMS and more…
• No (or not good enough!) historical data?
›
›
›
›
Confirm/ refine retention and data recycling measures
Guidance on data quality, reduce gaps etc.
Use other predictive models derived from the project – validate now, create later
Evaluate, influence and co-design other LA service offerings – SSP, Student App, dashboards…
• Individual assignment (coursework) and examination marks required per module – with dates!
• Data transformation: Tribal, in-house, using off-the-shelf software tool (Kettle, Alteryx…)
Tribal Student Insight - Overview
Create models
Data collection
and
Data mining
Predict and understand
Patterns
Relationships
Behaviours
Trends
LA Data @ June 2016
•
Openly available: UDD v1.2.4 - https://github.com/jiscdev/analytics-udd
•
UDD v1.3.x – in development - no release for a while!
•
Evaluation & trials of updated VLE plugins:
• BlackBoard Learn – Bb xAPI (live) & Bb Historical plugins;
• Moodle – xAPI (live) plugin & historical log extraction (*new* ULCC)
•
UDD data validation via LRW APIs (batch & incremental)
•
Preparation of LRW for large data submissions for predictive modelling
•
Student App – Beta v1.0 release imminent (iOS/ Android)
•
Student App evaluations – currently planned for 3 HEIs (Q3 2016)
Data Processing Agreement (DPA)
•
Model Creation/ Validation uses identifiable or non-identifiable data
•
Solution dependent - and also HEI technical capability!
•
Pilot Service (business as usual) = uses identifiable as-is student data
•
Concerns around data protection, security, exposure, liability, vendors, storage, brexit……
•
Jisc & Pinsent Masons LLP – Make way for the DPA!
•
Consultation with several HEIs – May and June 2016
•
Feedback varies – by region, size of HEI
•
Jisc support & guidance provided per HEI
•
Peace of mind – for all
Get involved!
• Do you foresee any issues or problems with near-LIVE student data capture
or timely data extraction for Learning Analytics?
• Are there any resource implications for you, in making Learning Analytics
happen at your institution?
• Can student assessment information be easily extracted from your system(s)
per module/ student? Is the information time indexed? Need Turnitin
integration?
• What other sources/ good indicators of student engagement or activity exist?
• Attendance monitoring systems, PC logins, any others key footprints?
Press Coverage – Times Higher (Feb 2016)
Press Coverage – The Independent (June 2016)
Press Coverage – Scottish Herald (May 2016)
Thank you
Rob Wyn Jones – Consultant (Special Projects)
[email protected]
Related documents