Jack Marwood's Icing on the Cake - An education blog
  • Home
  • Blog
  • MiniBlog
  • About
  • Contact

Using Data Properly: Ditch the Cargo Cult Data for Actual Data

26/5/2014

6 Comments

 
Picture
Cargo Cult Science is the term popularised by Richard Feynman in his 1985 book ‘Surely You're Joking, Mr. Feynman!’, referring to activity which has the trappings of real science, but which lacks the rigour which we might reasonably expect of real science. Tom Bennett, in his book 'Teacher Proof', gives a set of signs to look for ‘Cargo Cult’ research. Tom’s list of klaxons include: Research carried out by interested parties, small sample sizes, a lack of a control group, the absence of double-blind testing, confirmation bias, illusory correlation, mistaking cause and effect, the Hawthorne effect and the appeal to novelty.

Much of the progress-tracking ‘data’ which has become all too important in English education could actually be more correctly described as Cargo Cult Data. Cargo Cult Data has the appearance of real data, without having any of the requirements of statistically valid actual data from which one could reasonably draw inferences. It doesn’t pass the simple Fair Test most primary school children would recognise, much less the expectations of those who wish to Teacher Proof the data they are working with.

Michael Tidd has made the clear case that tracking is not the same as assessment, a point which is often lost in schools. As Michael notes, “Inevitably, the way that Ofsted works has meant that schools have been forced to use their assessments in the form of National Curriculum levels to demonstrate that they are tracking progress towards the end-of-key-stage expectations. However, in doing so we have all but divorced the act of assessment from the processes of teaching and learning.”

All teachers assess, all the time. We all have a good idea which of the children we teach are thinking hard, making progress, struggling with the demands of school, not in a position to learn or actively disrupting the learning of others, and so on. Assessment is an instinctive thing to do if you are trying to help children to learn.

Most data used to track progress, however, is simply guesswork, and often fairly random guesswork at that since, as Daniel Willingham is fond of saying, we can’t get into children’s heads to find out what they are learning or have learnt. We can ask children to write things down, but this isn’t the same as tracking their progress. It is pretty much the source of all progress-tracking data, however.

Where the progress-tracking data originates from high stakes written tests, at best it indicates how well children can answer (or be taught to answer) a written test. Where the progress-tracking data originates from what has become know as ‘teacher assessment’, it is hugely compromised by what Owen Elton refers to as the Teacher’s Dilemma, the effects of targets and many other distortions. The end result is that much of the progress-tracking ‘data’ used to assess learning, schools and teachers is simply ‘Cargo Cult Data’ – it looks like it can be used in the way statisticians use data, but its inherent flaws mean that this simply isn’t the case.

How did we end up in this mess?

Whilst many teachers will know the history of progress-tracking data, many of those not working in schools might not know how we came to be where we are. It’s worth looking at a bit of background.

Up to the mid-1980s, English schools worked in splendid isolation, for the most part. Teachers taught, children learned, but no one really had any overview of what teachers were teaching or children were learning in different schools below the age of 16, and there was no data which could be used to compare schools. That all changed when SATs were introduced, examination results began to be published, and the National Curriculum introduced ‘levels’ in the early 1990s. At that point schools began to develop progress-tracking Cargo Cult Data, taking guesses about the ‘level’ a child might be working at any given point in their education.

Levels were first used as indicators, and if that was all they were being used for, there would be some argument for them being useful. After all, it is helpful to have some information on where children are and where they need to go next. Many argue that teachers have always done this, and were doing it before the National Curriculum levels were introduced. Having a rough linear(ish) plan of development across the curriculum clearly makes sense. What doesn’t make sense is what happened next.

Within schools, teachers began to balk at the assumptions which were being made about the Cargo Cult Data which was rapidly being generated from the broad stages of development which the NC levels outlined. A figure of two levels of progress across a key stage was plucked from the air. The levels, which were already very broad, very subjective descriptions of learning, were subdivided into three sublevels.

In Primary Schools, with 4 years in Key Stage 2, this soon transmogrified into an expectation that a child would make two sub-levels of progress each year to be deemed to be make ‘good progress.’ Sublevels had numbers attached to them, which were labelled ‘point scores’. These point scores look, to those who simply don't understand the assumptions underlying actual data, misleadingly like something which has been measured, rather than simply guessed.

Picture
The point scores had become measures, at which point the true progress-tracking Cargo Cult Data was born. Driven by companies such as RM plc, the numbered point scores were treated if they were accurate measurements, which could then be subjected to statistical analysis. RM’s targeting offshoot, the FFT, then extrapolated this Cargo Cult Data into ‘estimates’ of future progress, in a way of which the financial services industry would be ashamed – past performance being no indication of future performance, as we all now surely know.

The attempt to find out how children were progressing in school spiralled downward into an unholy mess which ended up eating itself – or did it?

Enter politicians and their well thought out ideas

In a bizarre final twist, we are now on the threshold of a brave new world, since the current government have officially abolished levels. What’s that, I hear you ask? They’ve been abolished? Really? How did that happen? Well, with all of the criticism of the levels system, it was fairly obvious to all concerned that whatever levels were supposed to do, they weren’t very good at it. So they’ve gone.

Except they haven’t. Kevin Bartle wrote an excellent article about this this time last year (Spirit Levels: Exorcising The Ghost of Assessment Past), and Joe Kirby followed it up last November with thoughts on Life after Levels.

Liz Truss, Minister for Schools, speaking in April 2014, said that, "The old system for tracking a child’s progress was called ‘levels’. Levels became an end in themselves. So in 2012, we decided to scrap levels." And in one bound we were all free, to do what we want, any old time.

The Dead Levels message hasn’t got through to many schools, however. My school, for example, has a numeracy policy entirely based on levels, and children’s literacy work continues to be graded into sublevels each half term. The school I left in December had levels embedded deeply within all its progress-tracking. When OFSTED judged my school last year, they did so entirely based on progress-tracking Cargo Cult Data based on levels – Achievement of Pupils (based on levels) lead to a Quality of Teaching grade which lead to Overall Effectiveness grade, as it does in (bar one or two exceptions) every OFSTED report.

So what do we do now?

Liz Truss still seems to think that progress-tracking Cargo Cult Data can be used effectively in the brave new No-Level world. Here she is again in that speech in April, “Children will get a score. If they get 100, they’ve hit the expectations for their age. Above that score - and they’re ahead. Below, and they’re behind. It’s consistent across year groups. This is how the new end of key stage 2 tests will work - and schools can decide whether to have similar tests for the other years.”

So at 11 years of age (or significantly younger in some children's cases), children's knowledge will be guessed at and given a number on a scale. As to what schools do to show the progress demanded of by government accountability system 'for the other years', well, who knows. The government, based on Liz Truss's remarks above, clearly has no idea. Conferences are being held, and schools are trying to figure out what to do.

My strong advice is that we ditch anything which is Cargo Cult Data and only allow actual data to be used to track progress. I know. I can but hope, and keep chipping way at the flawed foundations of the Data Disaster.

What actual data can we collect and analyse?

The truth is, not very much. Knowledge is too complicated to be reduced to numbers, on the whole. Tests assess test-taking, and Cargo Cult Data is likely to rear its ugly head again. That said, some knowledge is fundamental to making progress in education, and that knowledge can be assessed and tracked numerically.  

The actual data I collect and analyse is:

  • Number Bond knowledge
  • Times Table Knowledge

This, added to the birthdate information I wrote about in my last article, is just about all I can see can be reasonably described as 'actual data'. I'd be interested to hear what those education think that they collect which isn't Cargo Cult Data. In particular, I'd be interested to hear what data secondary teachers collect. Do secondary maths teachers track times tables and number bond knowledge? Do you track information which is actual data?

I assess number bond and times table knowledge on a weekly basis with all the children I teach. The children take a two minute test, and I record children’s results and progress over time. The process takes around ten minutes a week and it gives me excellent information which I can then use in my teaching.

For the record, I also keep an eye on the following:
 
  • Knowledge of the Alphabet
  • Knowledge of phonemes and graphemes
  • Knowledge of punctuation marks

This is harder to track in any meaningful 'numerical data' way, but it is core knowledge which I can and do monitor. For some children, I do record this using numbers, particularly when I've spotted gaps and am attempting to fill them.

Beyond this, I record indicators of children’s progress. My guess is that, currently, this is what most teachers actually do, even if the 'data' then ends up being used as Cargo Cult Data by others. I find the National Curriculum levels quite useful as indicators. Yes, using the NC levels is hugely subjective, and subject to all kinds of biases, and I probably under record some children and over record others. Although, as long as no-one tries to subject this information to techniques used to summarise actual data (a forlorn hope in most cases), I have no problem with it, and I find it useful.

I'd be interested to hear what actual data other teachers and schools have found to be useful, and how this data is collected and used. Please comment below or contact me at [email protected].
6 Comments
Christopher Pipe
26/5/2014 06:14:38 pm

This is similar to Oliver Rackham's description of "factoids" in his discussion of pseudo-history ("Trees and woodland in the British landscape", 1976, pp. 23-25)): "A factoid looks like a fact, is respectd as a fact, and has all the properties of a fact except that it is not true. . . . Authors copy one another; each repeats his predecessors' errors and adds a few of his own. . . . Pseudo-history . . . is all the history that most of the public - and some Cabinet Ministers - ever read."

Reply
chemistrypoet
27/5/2014 02:37:10 pm

......Truthiness....

Reply
Mike Bostock link
8/6/2014 02:43:13 pm

There is much here to agree with. There is an artificiality with the use of Levels, but I believe that we could have made them more meaningful by emphasising the knowledge, skills and understanding associated with each level in each subject. To be fair, the Programmes of Study performed this task. Now that Levels have been scrapped and the National Curriculum replaced with one that more resembles a list of topics to be taught, and which has no indicators of progress whatsoever, we are not in a stronger position than before, however much people may wish to jump on the grave of Levels.
I have written about these things in my blog at http://mikebostock.com/a-unified-approach-to-assessment/

Now that there is scope to start with a blank sheet, the question is "How do we adopt a diagnostic approach to monitoring what pupils are learning, such that we can be sure they are making the progress that we expect of them?"
As Mr Marwood suggests, starting again is difficult, but we do need to replace a system of Levels with something, so as to create a feedback loop so that teaching can be adapted and focussed.

I think we are all in agreement that data can be misused, and that we need something better than Levels, and that there is a massive job ahead in defining and managing progression in the new National Curriculum. We now need to focus or energies on finding solutions to these problems. Data isn't inherently bad. We need it so we can turn it into useful information that we can understand, weigh up, and act upon.
And yes, we do need to watch out for people who seek to misuse information, like those people who quote statistics in Parliament for example. Within schools there is far more sensible use of data than we give praise for.

Reply
A follower from Aldgate
11/9/2014 04:57:52 pm

Interesting things to measure
1. Using those infant handwriting books - how many Y1 children can do their handwriting practise on the correct line...or how many Y1 children are actually forming their letters starting in the correct place. Measure each child at least once a week (measurement = yes/no). Then do something.
2. How many Y3 and above girls scream when playing football/standing near to a football game. Each scream results in a small deduction in PE teacher's salary. Get those girls to love practicing headers until they realise that screaming doesn't have to be a Pavlovian response to seeing a ball.
3. Ask Y6/Y7+ kids how many of them think they are crap at sport. Then starting with basic skills begin to teach something small (how to catch a ball - start with a beach ball and work down to a leather cricket ball)
4.I really enjoyed this post and agreed 100%. I'm going to continue making a list of useful things to measure.
By the way - could you add a search function to your blog please? It's hard to re-read old posts cause I can't always find them.
Also, I've been deeply thinking about the 'seat them by birthday and observe them' I think I may have treated your idea too simplisitically as I have been pondering about it all month (or at least the last 11 days.)

Reply
From Aldgate
11/9/2014 05:01:04 pm

Have you read 'How to Measure Anything'. I got lost about half-way through because a) the maths was hard b) I was listening to it on audio book. But was interesting.

Reply
Jack.Marwood
13/9/2014 05:21:58 am

I haven't read How to Measure Everything, but I did find this (http://www.howtomeasureanything.com) and this (http://lesswrong.com/lw/i8n/how_to_measure_anything/), both of which make me want to read the book.


Your comment will be posted after it is approved.


Leave a Reply.

    Author

    Me? I work in primary education and have done for 18 years. I also have children in school. I love teaching, but I think that school is a thin layer of icing on top of a very big cake, and that the misunderstanding of test scores is killing the love of teaching and learning.

    Archives

    March 2021
    February 2021
    January 2021
    April 2020
    December 2019
    March 2019
    February 2018
    September 2017
    July 2017
    April 2017
    October 2016
    September 2016
    July 2016
    June 2016
    April 2016
    March 2016
    February 2016
    January 2016
    December 2015
    November 2015
    October 2015
    September 2015
    August 2015
    July 2015
    June 2015
    May 2015
    April 2015
    March 2015
    February 2015
    January 2015
    December 2014
    November 2014
    October 2014
    September 2014
    August 2014
    July 2014
    June 2014
    May 2014
    April 2014
    March 2014
    February 2014

    Categories

    All
    Data
    FFT
    Follow Up Post
    Monday Article
    Ofsted
    Ofsted-schools-data-dashboard
    Ofsted-schools-data-dashboard
    OSDD
    Performance Tables
    Proposals

    RSS Feed

Powered by Create your own unique website with customizable templates.