Download slides – ppt : 2.9GB - Effective Learning Analytics

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
June 2016
Learning Analytics – UDD & DPA Update
Unified Data Definition v1.2.4 & our evolving data processing agreement
Data refresher…
• Historical data (12-18 months minimum) for a Learning Analytics implementation
• Predictive Model Creation (12-36months – ideal!) & Pilot
• (Portable) Predictive Model Validation (12months minimum) & Pilot
• Engagement & activity (xAPI) data – VLE (Bb, Moodle), LMS, AMS and more…
• No (or not good enough!) historical data?
›
›
›
›
Confirm/ refine retention and data recycling measures
Guidance on data quality, reduce gaps etc.
Use other predictive models derived from the project – validate now, create later
Evaluate, influence and co-design other LA service offerings – SSP, Student App, dashboards…
• Individual assignment (coursework) and examination marks required per module – with dates!
• Data transformation: Tribal, in-house, using off-the-shelf software tool (Kettle, Alteryx…)
Tribal Student Insight - Overview
Create models
Data collection
and
Data mining
Predict and understand
Patterns
Relationships
Behaviours
Trends
LA Data @ June 2016
•
Openly available: UDD v1.2.4 - https://github.com/jiscdev/analytics-udd
•
UDD v1.3.x – in development - no release for a while!
•
Evaluation & trials of updated VLE plugins:
• BlackBoard Learn – Bb xAPI (live) & Bb Historical plugins;
• Moodle – xAPI (live) plugin & historical log extraction (*new* ULCC)
•
UDD data validation via LRW APIs (batch & incremental)
•
Preparation of LRW for large data submissions for predictive modelling
•
Student App – Beta v1.0 release imminent (iOS/ Android)
•
Student App evaluations – currently planned for 3 HEIs (Q3 2016)
Data Processing Agreement (DPA)
•
Model Creation/ Validation uses identifiable or non-identifiable data
•
Solution dependent - and also HEI technical capability!
•
Pilot Service (business as usual) = uses identifiable as-is student data
•
Concerns around data protection, security, exposure, liability, vendors, storage, brexit……
•
Jisc & Pinsent Masons LLP – Make way for the DPA!
•
Consultation with several HEIs – May and June 2016
•
Feedback varies – by region, size of HEI
•
Jisc support & guidance provided per HEI
•
Peace of mind – for all
Get involved!
• Do you foresee any issues or problems with near-LIVE student data capture
or timely data extraction for Learning Analytics?
• Are there any resource implications for you, in making Learning Analytics
happen at your institution?
• Can student assessment information be easily extracted from your system(s)
per module/ student? Is the information time indexed? Need Turnitin
integration?
• What other sources/ good indicators of student engagement or activity exist?
• Attendance monitoring systems, PC logins, any others key footprints?
Press Coverage – Times Higher (Feb 2016)
Press Coverage – The Independent (June 2016)
Press Coverage – Scottish Herald (May 2016)
Thank you
Rob Wyn Jones – Consultant (Special Projects)
[email protected]
Related documents