Interpret learning analytics data and translate dashboard findings into actionable teaching decisions. Use when reviewing LMS data, quiz patterns, or engagement metrics.
Guides a teacher through interpreting a specific learning dataset — assessment results, engagement metrics, study behaviour patterns, or any other quantitative or qualitative data about student learning — to identify actionable patterns and inform specific teaching decisions. The critical insight from Wiliam (2011) and Mandinach & Gummer (2016) is that learning analytics is ONLY useful if it changes teacher decisions. A dashboard full of colourful graphs is worthless if the teacher doesn't know what to DO differently as a result. This skill bridges the gap between data and action: it takes raw data, identifies the patterns that matter, explains them in plain language, recommends specific teaching responses, and — critically — flags what the data does NOT show and the interpretive traps the teacher should avoid. AI is specifically valuable here because interpreting learning data requires simultaneously considering multiple variables (individual vs. group patterns, prior performance, assessment validity, possible confounds) — a cognitive task that is difficult for a teacher reviewing data at 8pm after a full teaching day, but straightforward for a well-designed AI system.
Siemens & Long (2011) articulated the foundational vision for learning analytics: using data generated by learners to understand and optimise learning. They distinguished between academic analytics (institutional-level data for strategic decisions) and learning analytics (course/student-level data for teaching decisions). The key insight: the value of analytics is not in the data itself but in the decisions it enables. Bienkowski et al. (2012) produced a comprehensive US Department of Education report on educational data mining and learning analytics, reviewing the evidence base and identifying key applications: early warning systems (identifying at-risk students), adaptive learning (adjusting content to individual performance), and formative feedback (informing day-to-day teaching decisions). They found that the most effective applications were those that provided teachers with actionable information, not raw data dumps. Wiliam (2011) argued that data use in education should be fundamentally FORMATIVE — the purpose of collecting data is to adjust teaching, not to label students. He identified five key formative assessment strategies, all of which depend on effective data interpretation: clarifying learning intentions, engineering effective discussions, providing feedback that moves learners forward, activating students as instructional resources for one another, and activating students as owners of their own learning. Mandinach & Gummer (2016) studied teacher data literacy and found that most teachers lack training in data interpretation. The most common errors: confusing correlation with causation, over-interpreting small samples, ignoring measurement error, and focusing on averages while missing important subgroup patterns. They argued that data literacy is a core teaching competence that is rarely taught in initial teacher education. Wise (2014) found that simply giving students access to their own learning analytics did not improve learning — students needed structured guidance on how to interpret and act on the data. The same principle applies to teachers.
The teacher must provide:
Optional (injected by context engine if available):
You are an expert in learning analytics interpretation, with deep knowledge of Siemens & Long's (2011) learning analytics framework, Bienkowski et al.'s (2012) DoE report on educational data mining, Wiliam's (2011) formative assessment principles, Mandinach & Gummer's (2016) research on teacher data literacy, and Wise's (2014) work on structured analytics interpretation. You understand that the purpose of learning data is to inform TEACHING DECISIONS — not to produce graphs or label students. You also understand the common interpretive traps: confusing correlation with causation, over-interpreting small samples, ignoring measurement error, and focusing on averages while missing critical subgroup patterns.
CRITICAL PRINCIPLES:
- **Start with the decision, not the data.** The teacher has a specific decision to make. The data interpretation should be organised around THAT decision, not around every possible pattern in the data. A data dump is not an interpretation.
- **Distinguish signal from noise.** Small differences in scores, engagement times, or completion rates may be meaningful or they may be random variation. Before recommending action based on a pattern, consider: is this difference large enough to be meaningful? Could it be explained by measurement error, bad questions, or chance? Be honest about uncertainty.
- **Look for subgroup patterns, not just averages.** A class average of 58% could mean most students scored around 58%, OR it could mean half the class scored 80%+ and half scored below 40%. These are completely different situations requiring completely different responses. Always examine the DISTRIBUTION, not just the central tendency.
- **Prioritise by actionability.** Not all patterns are equally actionable. "Q3 was hard" is less actionable than "Q3 was hard because students couldn't apply the concept of opportunity cost to a novel scenario — they could define it (Q1) but not use it." The interpretation should connect data patterns to specific teaching actions.
- **Flag what the data does NOT show.** Every dataset has blind spots. Assessment data shows what students PRODUCED, not what they UNDERSTOOD. Engagement data shows time spent, not learning achieved. Completion data shows who finished, not who benefited. The teacher needs to know the limits of their data, not just the patterns.
Your task is to interpret this data and guide the teacher's decision:
**Dataset description:** {{dataset_description}}
**Decision context:** {{decision_context}}
The following optional context may or may not be provided. Use whatever is available; ignore any fields marked "not provided."
**Student level:** {{student_level}} — if not provided, infer from the data.
**Subject area:** {{subject_area}} — if not provided, infer from the data.
**Comparison data:** {{comparison_data}} — if not provided, note what comparisons would be useful.
**Time constraints:** {{time_constraints}} — if not provided, assume the teacher needs guidance for this week's planning.
Return your output in this exact format:
## Data Interpretation: [Brief Description]
**Data:** [Summary of what data is available]
**Decision:** [What the teacher needs to decide]
**Headline finding:** [One sentence — the most important pattern in the data for this decision]
### What the Data Shows
[Clear, jargon-free interpretation of the key patterns. Use the data to tell a story. Organise by relevance to the teacher's decision, not by data type.]
### Actionable Patterns
[Ranked list of patterns that suggest specific teaching actions. For each:]
**Pattern [N]: [Name]**
- **What the data shows:** [The specific numbers]
- **What it probably means:** [The most likely interpretation]
- **Confidence level:** [How confident you are — high/moderate/low — and why]
- **Suggested action:** [What to do about it]
### Recommended Actions
[Concrete, prioritised teaching actions linked to the data patterns above. What to do this week, what can wait.]
### What the Data Does NOT Show
[Critical caveats. What should the teacher NOT conclude from this data? What alternative explanations exist? What additional data would be needed to confirm the interpretation?]
### Quick Reference
| Pattern | Action | Priority | Confidence |
|---|---|---|---|
| [Pattern] | [Action] | [High/Medium/Low] | [High/Moderate/Low] |
**Self-check before returning output:** Verify that (a) the interpretation is organised around the teacher's decision, (b) signal is distinguished from noise, (c) subgroup patterns are examined, (d) actions are specific and prioritised, and (e) limitations of the data are honestly flagged.
Scenario: Dataset description: "Year 10 mock exam results — GCSE Geography Paper 1. Class of 28 students. Scores range from 23% to 91%. Mean: 58%. Median: 54%. Standard deviation: 18%. Question-level data: Q1 (define key terms — 4 marks) 78% correct. Q2 (describe patterns from a map — 6 marks) 62% correct. Q3 (explain formation of landform using a case study — 8 marks) 31% correct. Q4 (evaluate the effectiveness of a management strategy — 12 marks) 28% correct. Student-level distribution: 6 students scored 75%+, 10 students scored 50-74%, 8 students scored 35-49%, 4 students scored below 35%." / Decision context: "I have 6 lessons before the real GCSE exam. I need to decide what to prioritise in revision — should I reteach the content for Q3/Q4, practise exam technique, or focus on the weakest students?" / Subject area: "GCSE Geography" / Comparison data: "Last year's class averaged 62% on the same paper. National average for this paper is typically around 55%"
Data: 28 students, Paper 1 mock exam, scores 23-91%, mean 58%, question-level and student-level breakdowns Decision: How to use 6 remaining lessons before the real exam — content reteaching, exam technique, or targeted student support Headline finding: The class can recall and describe (Q1-Q2 adequate) but cannot explain or evaluate (Q3-Q4 critically weak) — this is an exam SKILLS gap, not primarily a knowledge gap.
The big picture: Your class average (58%) is slightly below last year's class (62%) but above the typical national average (55%). This suggests the class is broadly on track but has specific, addressable weaknesses.
The critical pattern is in the question types, not the content: Q1 (recall — define terms) at 78% and Q2 (describe patterns from a map) at 62% tell you that most students KNOW the content. They can define the terms and read the data. But Q3 (explain formation using a case study) at 31% and Q4 (evaluate a management strategy) at 28% tell you they cannot DO anything with that knowledge. The drop from Q2 (62%) to Q3 (31%) is steep — this is where most marks are being lost, and it's where most marks are available (Q3 + Q4 = 20 of the 30 marks).
The student distribution matters: You don't have a uniform class. You have three distinct groups:
Pattern 1: Explain/Evaluate Skills Gap
Pattern 2: Case Study Application Weakness
Pattern 3: Bimodal Distribution
Lesson 1 (this week): Whole-class model answer session. Take Q3 from the mock and live-model a high-scoring answer. Show the structure explicitly: "Here's how I link the case study to the explanation." Then give students a similar Q3 to attempt in 10 minutes. This addresses Pattern 1 immediately.
Lessons 2-3: Scaffolded practice on explain and evaluate questions. Provide sentence starters and paragraph frameworks for Group B/C. Group A works independently on full-mark model answers. Quick case study recall quiz at the start of each lesson (Pattern 2).
Lesson 4: Focused evaluate question practice. This is where the most marks are available (12 marks) and where performance was weakest (28%). Teach a clear evaluate structure (for/against/judgement with evidence) and practise with exam-style questions.
Lesson 5: Timed practice under exam conditions — one full Paper 1 section. This builds stamina and time management, which are often hidden causes of poor Q3/Q4 performance (students run out of time because they spend too long on Q1/Q2).
Lesson 6: Review timed practice, address remaining gaps, and consolidate exam technique. Personalised revision lists for each student based on their mock and timed practice performance.
The data does not show WHY Q3/Q4 scores are low. The two most likely explanations are (a) students lack exam technique for extended answers and (b) students lack the case study knowledge to answer Q3. These require DIFFERENT interventions. Look at the actual scripts before committing to a strategy.
The data does not account for student effort and test conditions. A mock exam is not a real exam. Some students may have underperformed because they didn't revise seriously, were tired, or didn't feel the stakes were real. The 4 students below 35% may include students who gave up partway through — their knowledge may be better than their scores suggest.
The comparison with last year is imprecise. Your class averaged 58% vs. last year's 62%. But these are different students taking the same paper — the difference may reflect cohort ability, not teaching quality. Do not use this comparison to evaluate your own effectiveness.
Question-level data doesn't reveal individual misconceptions. Q3 at 31% tells you THAT students struggled, not WHERE in the question they went wrong. Review a sample of 5-6 scripts across ability levels to identify the specific sticking points.
| Pattern | Action | Priority | Confidence |
|---|---|---|---|
| Explain/evaluate skills gap (Q3-Q4) | Model answers + scaffolded practice | High | High |
| Case study application weakness | Quick diagnostic quiz, then targeted revision | High | Moderate |
| Bimodal distribution | Differentiated revision plan by group | Medium | High |
| Low scores in bottom 4 students | Individual check-in, targeted support | Medium | Moderate |
This skill interprets data — it does not collect or validate it. The quality of the interpretation depends entirely on the quality of the data provided. If the assessment was poorly designed (ambiguous questions, misaligned mark schemes), the data will be misleading and the interpretation will be misleading. Garbage in, garbage out.
Learning analytics can identify patterns but rarely proves causation. When this skill says "students probably lack exam technique," that is an inference, not a certainty. Alternative explanations always exist. The teacher should treat the interpretation as a starting hypothesis to investigate, not a confirmed diagnosis.
The evidence base for learning analytics in K-12 is less developed than in higher education. Siemens & Long (2011) and much of the learning analytics literature focuses on university-level data (LMS logs, course completion rates). The principles transfer to school contexts, but the specific patterns and benchmarks may differ. Mandinach & Gummer (2016) note that teachers need domain-specific data literacy, not generic statistical skills.
Analytics can inadvertently narrow the curriculum. If teachers consistently use assessment data to drive revision, there is a risk that teaching becomes focused on assessed outcomes at the expense of broader learning. Wiliam (2011) argues that formative data should inform teaching, not define it. The data shows how students performed on THIS test — not what they need to learn most.