The cost of bad data is the illusion of knowledge – Stephen Hawking
What’s more likely to kill you? A shark or a hot water tap? We’ve all heard stories of killer sharks, but as yet Spielberg hasn’t made a thriller about killer plumbing. We reason based on the information most readily available to us. We assume that the risk of dying in a plane crash is greater than the risk of dying on our sofa because plane crashes are so much more dramatic. But we’re wrong.
This is the availability bias. We make decisions based on the most readily available information in the belief that because it’s readily available it’s more likely to be accurate. Sometimes the information we can draw to mind might be accurate, but sometimes it’s not. Ignorance isn’t bliss; it’s scary. We are often most terrified by the unknown. But if something feels familiar, no matter how bad it is, we can cope. We prefer erroneous information to no information at all.
So in order to feel like we know what we’re doing, we surround ourselves with data. But although you can’t have information without data, you can most definitely have data without information. Data is uniquely comforting because it’s just so quantifiable. If you can turn something into a percentage or a bar graph it must be objectively true. The problem is, it’s remarkably easy to make up data. This leads us to do all kinds of foolish things in schools.
Consider this entirely fictitious scenario: a school leadership group is considering moving away from lesson grading in the light of a landslide of disconfirming evidence. They accept that lesson grading is invalid and unreliable and that taking a lesson study approach is more likely to support the professional development of teachers. But, and it’s a big but, what about Ofsted? Those pesky inspectors are expecting to see a neat spreadsheet which shows the percentage breakdown of teaching which is outstanding, good and requiring improvement. How will they react if this data is not on hand? How can we be accountable without numbers? And that’s the problem.
If we accept the findings of the MET Project which demonstrates that if a lesson is observed by two observers they will give very different grades. If one observer gave a top grade, there would be likelihood of about 70% that the second observer would give a different grade. And if one observer gave a bottom grade, the likelihood that the second observer would give a different grade was almost 90%! Learning is invisible. Any judgement made in the class room about how pupils are learning is guesswork at best. Any attempt to turn this information into data is witchcraft.
But it’s soooo comforting. Many school leaders have been seduced into the easy certainties of grading lesson observations, aggregate the grades, and then proudly declare that teaching in their school is 80% good or better. But this is meaningless. Assigning numerical values to our preferences and biases gives them the power of data, but they’re still just made up.
Hunt’s illusion of levels
This is one of the many reasons why I’m delighted to see the back of National Curriculum levels. Tristram Hunt inexplicably referred to this decision at Policy Exchange’s recent conference What should the political parties promise on education in 2015? as a “spectacular own goal”. When challenged on this he said something along the lines of, ‘Well, when I’ve spoken to teachers that’s what they’ve told me.’
But NC levels are made-up data at their worst. Someone (often a teacher) is asked to assign a numerical value to students’ work on a regular basis. We then pore over these tea leaves as if they are an objective reality instead of someone’s best guess about how a student may have performed in a particular task on a particular day. And then we write reports saying with absolute certainty that ‘Emily is a 4b in writing’ and ‘Isaac is a 5c in maths’. But what does this actually mean? What can we know about what Emily or Isaac can actually do? These numbers provide the illusion of knowledge. And on this illusory foundation we build a house of cards with which to hold schools and teachers to account.
Always remember, target grades are made up.
We fall into the same bear pits when setting students targets. We tell them they need to know their target grades as if they are cast iron certainties. But while they may not be simply plucked from the air like lesson observation grades, they’re based on statistical probabilities that may have some validity when applied to large cohorts but which are reduced to meaningless nonsense when applied to individuals. I’ve blogged before about how pernicious this practice is. Possibly, the most useful thing we can do is to subvert these targets to harness the power of the growth mindset.
But no data is bad in and of itself. Just as guns don’t kill people, data doesn’t distort the curriculum or warp decisions about what to teach: we do that. We are comforted by the illusion of knowing. But we really don’t know. And any accountability system that allows people to either input numbers they’ve made up or extrapolate data which is wrestled into meaning something it was never intended to mean is doomed to fail. But what’s worse is that many schools, teachers, parents and children aren’t even aware of the failure.
So what can we do?
Next time someone shows you a spreadsheet, try asking the following questions:
- If this data is the solution, what’s the problem?
- Is there a different way of interpreting the data?
- How can I verify the quality of the data I’m being shown, and what are its margins for error?
- What are the limitations of this data – what doesn’t it show?
- How is this data likely to affect my decision making? What would I do differently if I didn’t have this data?
In this way maybe, just maybe, we can avoid some of the potential pitfalls associated with availability bias.
Further reading
Jack Marwood’s Using Data Properly: Ditch the Cargo Cult Data for Actual Data
[…] Read more on The Learning Spy… […]
Hi David, I’m sorry to say that I cannot find anything to disagree with you about on this occasion. However, in my experience, what I can say is that it’s dangerous to be right when the established authority is wrong (or at least on shaky ground). Put another way, as I think Corporal Jones one put it. “They don’t like it up ’em” . Things tend not to go your way once you have poked the bear (did I use too many metaphors?).
You can’t go around slaying these educational dragons as you are doing, you’ll upset someone 😉
Add “what precisely is the problem for which this data suggests a solution?”
Thanks Andrew – I’ve incorporated that in No. 1 now.
OK – I’ll try some appeasement next. Cheers.
My Q6 would be this; Who gave me this data, and is their primary motivation to aid learning or is it accountability ? Or something worse, personal ambition?
Thank you James – always important to spot vested interest
I think the second half of 3 could be strengthened. What are its limitations? would cover not only the limits of significance, but also what the data is not aimed .
I’m also not sure on the significance of the 4th question, I’d make the second half of 3 the 4th question. Doesn’t 5 ask the same question as 4?
Thank you Stuart – have amended. Better?
The key when looking at data is to only ever use it as a pointer towards some questions worth looking at further. I said that at interview and have stuck with it throughout my time on SLT. Too many people want data to distill something very complex into a single number that makes the world look good or bad but, in reality, all it can do is make it slightly less murky!
In terms of lesson observations & grading there is another way to ‘capture’ the quality of teaching and any changes in it over time and identify school-wide need.
Define, as a staff, the things that we believe are important for effective learning – behaviour in lessons, subject knowledge, setting of and student completion of appropriate homework, …. and get staff, on a randomly chosen day, to grade each class they teach that day on each of those strands on a simple 4 point scale (basically) – particularly strong; as expected; not quite where I want it and need some help with this.
That data is captured anonymously other than for the key stage of the group so there is no incentive for staff to worry about being honest.
That gives a measure for the quality of teaching/learning in the school from the staff view point; cross reference with evidence from student results and, tentatively thinking, some sort of similar capture on the same day of student views (but with no possible way of assigning the views back to named teachers) and you’ve got some evidence of how things are progressing over time.
If those homework scores are noticeably worse at KS3 than KS4 then it signals that it may be worth prioritising homework at KS3 etc. – if behaviour scores are slipping over time then questions should be asked about how to address this etc.
Hi Steve – this sounds fascinating. Is this what you do at your school? Can you give me any more details?
Thanks, DD
It’s not – but it is a proposal I’ve been working on the last few months (with feedback from various folks on twitter I have shared it with including some OFSTED inspectors) to iron out the kinks.
The bit I am working on currently is the student input. I think adding that level of info adds to the strength of the package – there will always be some idiots who take the mickey but most would take it seriously.
It’s not about the student grades and staff grades matching up but more about the trends over time – what if students say things are getting better but staff say worse – why might that be? what might be going on? More likley is that both will trend in the same direction!
My biggest hurdle is the notion of students being seen as ‘grading’ staff even though the capture would ensure that there is no way of linking it back to a named member of staff. It’s about making sure that the ‘pitch’ of that element is clear about the role of the student data.
Plan is to take it to SLT before the end of term for hopeful launch next year.
Sounds fascinating. Would you mind if I emailed you?
Not at all.
Hello Steve, I am a member of SLT and very interested in this idea. I have been getting teachers to use student voice as a pilot this term, in order to assess their own development needs.
http://dominicsalles.edublogs.org/2014/04/
It has been very positive so far.
I am definitely going to explore your idea further. Good luck.
Steve, whilst I do think that both teacher and student voice are very important – and I think your proposal above merits exploration I would caution to tread carefully on the “the things we believe are important for effective learning” and make sure that these are backed up with some research / evidence and David (and many others) have talked about before what we believe is not always what is evidential – just think of all those who believe in VAK (or 4bs!).
mmiweb – yes I 100% agree.
My list would be very broad headings (behaviour in the classroom; teacher subject knowledge; effective planning; effective assessment) rather than detail (such as VAK).
I think so, but I wouldn’t add “what does it not show?” but “what are the limits of what it can show?” Otherwise “what does it not show?” could be a never ending answer.
Hmmm. I think I see your point but the two statements seem remarkably similar, no?
They have very different answers.
OK, let’s try this:
No 1 – What do this not show? It doesn’t show x, y and z
No 2 – What are the limits of what it can show? It can only show a, b and c
Of the two questions, I tend to think that No1 provokes the more interesting responses because it forces you to go beyond the data whereas No2 makes you focus ON the data. Or am I deluded?
Yes, that makes sense presented like that. I think that the implications presented by No2 are that it’s actually asking “what might it appear to show but doesn’t?” or even “what is the likelihood of this sample of data (which all data is by its nature) being representative of the population.
A very interesting and well timed article. I was observed by SLT last week after a long medical absence. Whilst my teaching was judged to be outstanding, I was told I was ‘requires improvement’ as the data shows my kids have not made 14 points progress since year 7.
Interesting, I have only taught 4 of them since year 7 and the others had made at least 6 points progress in year 8 and had made another 2-4 points by the time I was off.
It begs the question, how can someone be held accountable for lack of progress when they are signed off and why has my SLT decided that I am to blame when the data shows great progress when I and my kids are actually in the room together?!
Hi Vicky – the good news is that anyone who judges your performances as requiring improvement for lack of progress whilst off sick is an idiot. The bad news is the judgement of your teaching as equally as meaningless.
Thanks, David
Couldn’t agree more David! The only opinion I genuinely care about is that of the kids I teach.
I’ve taught for 16 very happy years but I am now seriously considering my options.
David, I too am (mostly) glad that levels are going. I was at my daughter’s parents evening recently and was ‘proudly’ told she was a level 6a in PE and that she has reached her target grade – when I challenged the teacher to define what this meant she could actually do – and what did she now need to do there was a bit of a spluttering explanation. I later challenged the HT about this and he agreed that the levelling was irrelevant but … yes the O word came along.
However anecdote aside the idea of the NC levels as ladders of progression was I think very valid however once they were hijacked by the “assessment wonks” they lost most of that validity – but I feel very stingily that the assessment tail now wags the learning dog so strongly that it is hard to see how we will get back to this.
I do not see the end of levels as the end of this – the “wonks” will still want to see the deceptively pretty graphs.
Your points on the arbitrary and unreliable use of levels/grades is well made. One of the few lessons that really stuck with me from my PGCE year was the stats session on education data and its reliability. And in particular the importance of the margin of error. I would love to see the impact of requiring inclusion of the standard error on all these evidence based debates that are currently in vogue.
I would also like to see the margin of error on exams publicised more widely too. I do remember that analysis of the A level physics marking reliability in the year I took that exam showed that candidates awarded a B had an equal probability of being awarded a C or even a D (very narrow band grades that year!) Always a comfort as one who had actually been graded at a D. 🙂
But I suspect that more widespread knowledge of the relative unreliability of education data would be far too threatening for the establishment to permit?
[…] 5 questions to guard against availability bias and made up data Perverse incentives and how to counter them A request: Have you experienced any craziness in your school? […]
This needs to go to Governors, Ofsted inspectors and HMIs
[…] 5 questions to guard against availability bias and made-up data […]
[…] Full text […]
[…] curriculum time spent on? If teachers absolutely must be held account for meeting targets based on spurious data or triple mark their books, what will they be required to stop […]
[…] curriculum time spent on? If teachers absolutely must be held account for meeting targets based on spurious data or triple mark their books, what will they be required to stop […]
[…] decision-making. In this post I discussed five techniques we can use to overcome overconfidence; in this post I suggested ways we can take a more objective view of data, and in this post I recommended a […]
[…] 5 questions to guard against availability bias and made up data […]