Pray look better, Sir … those things yonder are no giants, but windmills.
Cervantes
Does it matter if students like their teachers? Is it worth knowing if students don’t maths or hate PE? Should students be asked to evaluate the quality of their lessons? It sometimes seems that the clamour of ‘what students want’ drowns out even the presumed demands of ‘what Ofsted want’. Students’ opinions might be interesting but should they be used to judge the effectiveness of teachers? Certainly some school leaders appear to think so.
An anonymous blog on the Labour Teachers site* reveals the extent of the rise of this sort of managerialism:
This year has seen our graded lesson observations replaced with something else entirely. Pupil Learning Interviews. Fresh from the mouth of an ex-Inspector-now-consultant, this new form of monitoring is where it’s at. Here’s how it works:
1. A small number of children are chosen in advance: underperforming boys, PP, school refusers – you get the idea. Obviously being a nice and hard working girl is not good enough for your views to be polled
2. On the day of the interview, two observers arrive at the door and seat themselves within the vicinity of these lucky children
3. The children are watched – not the teacher. In fact, what happens in the actual lesson itself has no relevance at all
4. The children’s books are studied, so they cannot complete any work in the lesson
5. The children are then interviewed either in the lesson, whilst everything is going on or they are asked to step out into the corridor.
After the lesson children are asked questions like, What have you been learning this year? I see here you have been learning about rivers, what can you tell me about that? What do you find hard in this subject? What do you find easy?How does the teacher help you? What sorts of things do you learn about?
In a Schools Week article about pupil interviews, opinion seems to be divided. Christine Blower general secretary of the NUT, is against “snapshots” being used to judge teaching: “Attempts to do so risk turning our schools into exam factories, and they ignore deeper learning and valuable learning experiences.”
But Brian Lightman from the ACSL thinks it’s fine because schools have been doing it for years(!) He is reported as saying, “They are an informative and useful form of school self-evaluation designed to understand the learning experience from the pupils’ viewpoint… The feedback and insight gathered is useful in identifying how pupils can be helped to do as well as possible.” And maybe in the right
And maybe in the right hands such feedback really could be useful in improving how students make progress, but I doubt it will in this case. As Nick Rose says in this blog, “These kinds of data can serve a useful purpose so long as it is not being used to make superficial judgements about teacher performance.” In another post, Rose talks about how students surveys might be used well. He concludes by saying, “As an accountability measure, I think student surveys would be stressful and unfair – but as a coaching tool to get inside your own teaching, I think it has merit.”
There’s no doubt that students routinely perceive differences in their teachers and that they see some teachers as better than others. One of the most widely researched student survey systems, Tripod, reveals that students of teachers with the top 25% of Tripod score, make 4.6 months more progress in mathematics than those in the classrooms taught by teachers with Tripod scores in the lowest quartile. On average, the most effective 25% of teachers are around 130% more effective than the least effective 25%. Clearly students’ perceptions are worth something.
This research from the MET Project shows the sorts of questions students can be usefully asked:
You might have noticed that these items are rather different to the questioned being asked in the school in question. The authors of the MET Project report, Asking Students about Teaching say,
For results to be useful, they must be correct. It’s of no benefit if survey responses reflect misunderstanding of what’s being asked, or simply what students think others want them to say. Worse yet are results attributed to the wrong teacher. Honest feedback requires carefully worded items and assurances of confidentiality. Standardized procedures for how surveys are distributed, proctored, and collected are a must. Accurate attribution requires verification… (p. 11)
I wonder to what extent these considerations were taken into account before deciding to judge teachers on pupil interviews? The report continues:
Consistency builds confidence. Teachers want to know that any assessment of their practice reflects what they typically do in their classrooms, and not some quirk derived from who does the assessing, when the assessment is done, or which students are in the class at the time. They want results to be reliable. (p. 14)
The suggestions set out to ensure consistency and reliability appear to be in complete opposition to the way interviews are conducted in the school above. But then, we only have one teacher’s view – maybe they just need to be a little more trusting? Maybe, but as the report clearly states,
No policy can succeed without the trust and understanding of those who implement it. Although often ill-defined, collaboration is key to effective implementation of new practices that have consequences for students and teachers. Every system referenced in this brief has made stakeholder engagement central to its rollout of student surveys—involving teachers in the review of survey items, administration procedures, and plans for using results in feedback and evaluation. (p. 21)
I suspect considerations of trust and understanding may have been dismissed as unnecessary.
The other big problem with relying on student surveys is the Dr Fox effect. The original experiment was conducted by Donald Naftulin, John Ware, and Frank Donnelly at the University Of Southern California School Of Medicine in 1970. Two speakers gave lectures to 3 different audiences of MDs and PhDs. The first audience listened to a genuine professor and the other audiences both listened to an actor posing as Dr Myron L. Fox. When ‘Dr Fox’ was instructed to speak in a dull monotone, the audience rightly dismissed him as a know-nothing ignoramus, but when he delivered the same meaningless material with enthusiasm and humour, the students rated Dr. Fox as highly as the genuine professor. This lack of correlation between content-coverage and ratings under conditions of high expressiveness is well-known; even very knowledgeable, expert students are fooled by a good speaker, how much more likely are school students to be swayed by ignorant yet charismatic teaching?
As Naftulin, Ware and Donnelly concluded by saying, these results ought to suggest to teachers that “the extent to which his students are satisfied with his teaching, and even the degree to which they feel they have learned, reflects little more than their illusions of having learned.”
All this should convince us that while it may be very useful to survey students on their perceptions of lessons, this should never be used as a crude, high-stakes accountability measure to judge teacher effectiveness.
*The teacher who wrote the Labour Teachers blog is well-known to me and someone I respect and have every reason to believe.
Clearly there are benefits. However, I’m always worried that teaching will end up being some kind of personality contest and children will grow up thinking that normal adult behaviour should always be extremely extroverted or concerned with feelings and emotions (Hmmm, do they already think this?). I think this consequence might be more likely in primary schools who use student feedback as part of the performance management process because younger children lack the maturity to know what’s in their long term interest.
I really hope my school doesn’t start doing this because it’s enough that we have outside consultants, SLT, HT observe us and feedback sought from TAs and parents about how us teachers are doing without having the kids reporting on how fun our lessons are. It’s all a bit much really.
Sounds like things are bad enough already :\
In my last school we ran a small trial with using the Tripod survey. Nine teachers used it and as the head I crunched all the data for them and then handed it over. It was not part of the appraisal process (unless individual teachers wanted to refer to it in their meetings) and all I asked is that after the second round, six months later, that they told me how they used it and whether it was of benefit.
The nine teachers were a real mix, from NQTs to a teacher with 40 years experience. Two teachers who were good friends came to see me one day (Friday after school) and we just talked for about an hour and half about what we thought the survey said. I went home thinking it had been one of the most interesting, detailed and fascinating meetings I have ever had about teaching. The feedback at the end was that it was a useful experience. But perhaps the best part was people trusted the process, because all the way through we insisted it was their data. One interesting moment was when a middle leader came and asked me if he could see the data, as he was appraising one of the trial teachers. I said no, he would have to get the teacher to ask me to release it before I could hand it over. Trust is about integrity and you need to have both in any system.
I am now at a new school as head and this year I am using the survey tool on myself, together with video lesson observation. My hope is at the end of the year to present my research on myself to the staff as one way to self-manage CPD.
Student surveys are, for me, a really useful resource. What they should never be is part of a high stakes system – surely our role as schools is to put in place the best possible tools for professional development and then let teachers pick what works for them. The ‘tight’ side is simply that we can all get better, and should engage in that process – the ‘loose’ side is provide resources, time and space and let teachers find their own way to develop.
Thank you for the link to the Dr Fox effect – I’ll try to weave that into my final presentation to the staff.
Thom – you sound like a model for intelligent leadership – I’d love to hear how you get on
I had better start being nice to EVERY student no matter what behaviour and attitude towards learning is displayed.
The ref to Dr Fox is useful to me IF I am accused of lack of enthusiasm in my next observed lesson ( my lack of enthusiasm by the way will be because I know I will get a 3 no matter what I do )
Seriously, if that’s true you ought to involve your union.
It is true but could so easily be described as all in my mind. That’s why I’m having to suffer and wait for more evidence of bullying before I approach governors or unions. Certainly it disgusts me that the bully will pretend all of this is on the name of education and helping students. How can persecuting an experience enthusiastic conscientious teacher be good for my students?
Also David – thank you. Your blogs are helping to keep me sane. If only my head would read your blog before making some of her decisions and policies.
If only… 🙂
We use student surveys all the time and especially after prelim exams and before they leave. They ARE anonymous and our Boss takes great care to go through and remove any staff names. This stops kids saying Mrs So and So is a great teacher based on her popularity because she keeps bringing in cupcakes 😎 This way we get to focus on the aspects rather than the personalities and where we need to improve as a department. The Boss obviously also gets an insight into areas where individuals might need a bit of help or advice.
We also work on the assumption that the kids don’t always tell the whole truth. 😎 For example “Mrs So and So isn’t teaching us right because Miss So and So does it a different way in her class and they say they really know what they’re doing when we don’t.” Teachers teach in different ways and some kids don’t understand this which can undermine these surveys (and staff) when done by the Senior Leadership Team. We’ve also noticed recently that so called mentoring/coaching sessions where individual kids are farmed out to SLT are being asked about their teachers rather than being coached.
NB: I always advise my senior kids to attend after school study before exams to hear from the Boss on Close Reading or from Mr X on Shakespeare because I know they might hear them say something slightly differently to how I describe/teach a technique and Lo! a lightbulb moment. Popular teachers may, as you say, not be the best ones. One of the most unpopular teachers in the history of teaching was hated by her students but years later they freely admit she is the one who got them through the exams and the top jobs/universities they ended up in.
Thanks for your comment: I can’t really work out whether you think students surveys are a boon or not. Are you saying they’re a bit of both?
David, thought provoking as usual.
In the day job I see the responses of hundreds of pupil surveys and the problem is deeper than just poorly constructed questions – the problem is magnified by totally inappropriate data analysis and then assuming you can draw conclusions.
The main culprit is Likert type scales – you know the questions that ask students to judge their thoughts somewhere on the continuum from totally disagree to totally agree. Many people then convert this psychological opinion based scale into 1 to 10 (for example) and then proceed to calculate the mean for a particular question (You feel safe in school) – and pronounce that on average, students feel 4.7 safe and convert this back to “agree” from the scale. Totally inappropriate use of data handling.
I blogged about these thoughts on handling survey data here: http://glengilchrist.co.uk/post/handling-survey-data
Not only do we need to reflect on what are we actually trying to find out, we need to acknowledge that they way we ask students questions inevitably influences their responses and we often don’t capture what we thought we were. That, and that the processing / summarising of the data isn’t as straightforward as we would like to believe; just because you can process an average, doesn’t mean that that figure has any value / sense in the real world.
Schools need to be data / statistics savvy and the nuances of unbiased questionnaire design are many – accept that to get the most from these things, professional input is often needed.
However, very often firing off a quick Survey Monkey seems to the be solution to box ticking the student voice issue.
Cheers David – you hot the spot once again.
Glen
We’ve just piloted a student survey. Crucially it is not collecting data about named teachers nor their quality but rather it collects information about what they do (which we then compare, as an institution, against the ‘core principles’ that the we designed as a staff as a whole).
I think this approach has merit and (some) usefulness – but let’s not overplay it. It is just one thing for us to consider as a staff.
Any system that is used to ‘judge’ or ‘grade’ staff is bunkum.
“Any system that is used to ‘judge’ or ‘grade’ staff is bunkum”: spot on