Must read for goverment officals everywhere.
What can The Simpsons teach us about stats algorithms? Glad you asked...
When his class is asked to give an example of a paradox in The Simpsons, Bart offers: "You're damned if ya' do, and you're damned if ya' don't." The dictionary defines a paradox as an absurd or seemingly absurd or contradictory statement that might prove to be true and when it comes to data a seemingly contradictory situation …
-
-
Wednesday 28th May 2014 09:01 GMT Michael H.F. Wilkinson
Not just government officials, but many others in big data. Too often they assume that arbitrary aggregation will result in better statistics (e.g., because the standard error in the mean is reduced), whereas all too often piling up data from different sources in fact obscures certain effects.
Well-written article again.
-
-
-
-
-
Thursday 29th May 2014 03:57 GMT Archimedes_Circle
Re: So... what's the options, then?
You could do multilevel modelling as a solution. Multilevel modelling is, in general, a solution to when subjects are clustered, like when you want to explore the effects of spending on student performance. Typically you would run a regression, however in this case, if you have multiple schools, the spending levels are not independent observations: they are correlated with each school and that typically inflates the wald test results, which makes smaller p-values.
-
-
-
Wednesday 28th May 2014 12:11 GMT fch
"prove causation" != "observe a [certain amount of] correlation"
If even the author of the article makes that basic mistake, then there's nothing left to conclude but you're correct - the purpose of statistics is to prove someone's point / back up someone's claim, not to ... learn anything from the data.
Go ... [massage] figure[s] !
-
Wednesday 28th May 2014 12:36 GMT MrT
Re: "prove causation" != "observe a [certain amount of] correlation"
Bang goes the secret behind soooo many tabloid headlines...
Still, while we're having fun with correlations, have a look at this collection... the one I like best is 'German passenger cars sold in the US' correlating with 'Suicides by crashing of motor vehicle' - "It's just not as good as everyone told me - goodbye cruel world!" - although the Divorce rate in Maine having anything to do with Per capita consumption of margarine (US) surely must have some mileage with the dairy industry...
-
This post has been deleted by its author
-
Wednesday 28th May 2014 18:05 GMT Anonymous Coward
Re: "prove causation" != "observe a [certain amount of] correlation"
Hey, sucka! I saw that on the BBC a couple of days ago. My favourite was "Per capita consumption of cheese (US)" correlates with "Number of people who died by becoming tangled in their bedsheets" with a correlation of 0.95.
But that is causation - it's well known that eating cheese at night causes bad dreams and so more tossing and turning and so more chance of getting caught up
-
-
-
Wednesday 28th May 2014 13:19 GMT LionelB
Re: "prove causation" != "observe a [certain amount of] correlation"
To be fair, I think the author of the article is quite aware that causation != correlation - hence the statement "Let’s take it as read that not only is there a correlation ... but that causation is also at work", which I read as him saying that causation has been established by some other (unstated) means.
-
Wednesday 28th May 2014 13:57 GMT fch
Re: "prove causation" != "observe a [certain amount of] correlation"
I'm almost tempted to say "QED" ... as I've quoted without context to make my point more obvious. The mere mention of "cause" in the same context [ sentence ] as "statistical correlation" is huge b*llsh*t honeypot. People have gotten it into their heads somehow that statistics prove causation, and the explicit mention (as you did) about first needing a testable theory with a claim of causation and a prediction of measurable changes - that's so conveniently left out.
I do wonder who started down that slippery slope ... when I retire [ never ... ] I may do a PhD on the history of politics/economics to find out :-)
-
Wednesday 28th May 2014 16:34 GMT LionelB
Re: "prove causation" != "observe a [certain amount of] correlation"
"I do wonder who started down that slippery slope ... "
Probably cavemen... X seems to be associated with Y, ergo X causes Y (or Y causes X, so e.g. being fat causes you to eat a lot).
Anyway, there are statistical techniques for dealing with causation - they inevitably involve time precedence (as in cause precedes effect), and inevitably generate controversy as to their interpretation/usefulness. Surprisingly (or perhaps not), there is little consensus in philosophy/science/mathematics/statistics over what "causation" even means. Here's a neat quote from Clive Granger, economist, statistician, Nobel laureate and inventor of the "Granger causality" statistic, (from his Nobel acceptance address):
"At that time, I had little idea that so many people had very fixed ideas about causation, but they did agree that my definition was not “true causation” in their eyes, it was only “Granger causation.” I would ask for a definition of true causation, but no one would reply. However, my definition was pragmatic and any applied researcher with two or more time series could apply it, so I got plenty of citations. Of course, many ridiculous papers appeared."
-
Thursday 29th May 2014 17:24 GMT Michael Wojcik
Re: "prove causation" != "observe a [certain amount of] correlation"
Probably cavemen... X seems to be associated with Y, ergo X causes Y (or Y causes X, so e.g. being fat causes you to eat a lot).
Ample psychological research1 provides strong evidence that mistaking correlation for causation, along with related logical fallacies such as post hoc ergo propter hoc, are basic features of human cognition. Plenty of people have argued that these are useful survival traits2 and so were selected by evolutionary processes, though that's just speculation.
Certainly, given the apparently universal nature of the habit, it looks unlikely to be cultural.
On a related note, McRaney's nice collection of short essays on various forms of fallacy, mental error, delusion, etc, from his website, have been collected into a book, and there is now a sequel. At least the first volume should be required reading for, well, everyone. (Even if you're already familiar with all the phenomena he describes, it's full of nice examples for demonstrating them to other people.)
1Conducted in about as methodologically-sound a fashion as is possible when you're dealing with human subjects, and we're not likely to have anything better than that anytime soon.
2Usually with some variation of the "shortcut" argument: reasoning takes a lot of time and biological resources, and postpones decision-making, so the human mind opts for a quick heuristic decision that may be revised later (but often isn't).
-
-
-
-
-
Wednesday 28th May 2014 09:43 GMT Cliff
Tree ring plus measured?
Is this what caused the climate change denial camp's recent offensive? Something to do with the way tree ring data and directly measured temperature data was used after the 1961 correlation problems? By stealth, did this paradox damage the reputation of climate science just by existing?
-
Wednesday 28th May 2014 18:19 GMT T. F. M. Reader
Re: Tree ring plus measured?
I must have slipped in my vigilance - not sure what denialist offensive you have in mind. Have not heard anything about tree rings for years.
I recall reading the first and then the second paper on tree rings as a proxy for historic temperature measurements. I am too lazy to check, and my memory may be faulty after all these years, but if it isn't the first sample consisted of 3 stumps, and the second - of 21 or so. Both samples were from basically the same place. I decided to discount all the conclusions that could be drawn from either sample (or both - it well may be that the samples were similar enough that the Simpson paradox would not manifest itself) regarding the temperature history for the planet as a whole at that point.
-
-
Wednesday 28th May 2014 09:48 GMT GumboKing
I disagree with the initial analysis of the first two graphs
While I understand the paradox and agree the two sets should not be merged, the initial analysis of the first two graphs does not "prove that the act of spending more on advertising really is directly causing the product to sell better". It seems to show that an increase in spending is causing it to sell better up to a certain point, and past that point it will actually decrease.
For Client A they should cap spending at 45 pounds and Client B should cap spending at 350.
For Client A, two campaigns at 30 pounds should have generated around 600K in results, three 20 pound campaigns would have generated 750K whereas the 60 pound campaign only generated around 325K.
As an imaginary Client, I would be pretty ticked if I would have had much better results running 2 or 3 campaigns of lesser value instead of just one expensive less effective campaign.
P.S. (Paradox Sidenote) in Back to the Future II there is a Pair of Docs Paradox.
-
Wednesday 28th May 2014 11:13 GMT Anonymous Coward
Re: I disagree with the initial analysis of the first two graphs
Quite right, it assumes the existence of a cause; it doesn't prove it.
There's also an assumption that to take an 'average' of 2-dimensional scatter data you can simply average the X and Y components separately. For those who want to try it for themselves, it's a very interesting and educational exercise to 'prove' the formula for averaging from first principles, then expand the solution into the 2- and N- dimensional cases. It's only when you take the same approach to second and higher order statistical methods (standard deviation, etc.) that it starts to go a bit wibbly. But if you've genuinely got multi-dimensional data (i.e. vectors, not independent values) you really should be having a crack at it.
[Helicopter icon, because they churn out vectors like there's no tomorrow].
-
-
Wednesday 28th May 2014 09:59 GMT John Tserkezis
It's Freakonomics all over again.
It happens all the time, all over the place. One example I cite frequently, is in the "Freakonomics" "what's in a name" chapter.
There is quite clear statistical difference in the performance of a child, and their name. So how does a name affect one's school exam scores? Easy, it doesn't. It doesn't make a squat of difference. The difference is in genetics, upbringing, and if the child was an accident or planned.
The most likely answer is because well educated parents have their kid later in life, where it was planned, will probably name their kid some sensible, ordinary, conservative name. On the other side of the street, the type of mother who names her kid "temptress" is probably on the lower end of the socioeconomic scale, had her kid at 16, and was too busy lighting farts instead of worrying about school exams. The name isn't the cause, it's the effect.
Yet, even today, parents are increasingly naming their brats ever-exotic names in the at least subconscious effort that it's going to make a difference.
Which brings me to my point (yeah I know it took a while): It's easy to assume the average plob doesn't know anything about statisics, but the people who analyse databases, come from the same group. They're at the same risk of mistaking cause and effect.
-
Wednesday 28th May 2014 11:03 GMT kmac499
Downhill all the way ??
Sounds a lot like the way I was taught algebra "you can't add Apples and Oranges to get one answer, same goes for 'x' and 'y' "
What woud be even more worrying if there is a reverse Simpsons where two negative correlations sum to make a positive.. (Drug Trials, Diet Foods, etc...)
-
Wednesday 28th May 2014 11:18 GMT Anonymous Coward
Re: Downhill all the way ??
I don't see any reason why there wouldn't be. If one set that showed a negative trend was generally higher than another set that showed a negative trend, when combined, the resultant graph could trend upwards. It's just people aren't usually that interested in negative correlations to try combining them. Unless there was devious intent afoot...
-
-
-
Wednesday 28th May 2014 15:46 GMT Jimbo 6
Re: For real expertise...
The whole article reminded me of a favourite Dilbert cartoon -
Boss : Use the CRS database to size the market.
Dilbert : That data is wrong.
Boss : Then use the SIBS database.
Dilbert : That data is also wrong.
Boss : Can you average them ?
Dilbert : Sure. I can multiply them too.
(That and many more at http://stats.stackexchange.com/questions/423/what-is-your-favorite-data-analysis-cartoon )
-
Thursday 29th May 2014 17:28 GMT Michael Wojcik
Re: For real expertise...
Or check out Huff's classic text How to Lie with Statistics. Full of fun projects for the whole family!
-
-
-
Thursday 29th May 2014 01:59 GMT veti
Re: Statistics...
If by "the Bard" you mean Shakespeare, I'm pretty sure the word "statistics" occurs precisely zero times in the whole of his collected works.
That bon mot comes from someone in the 19th century, though it's by no means clear who.
-
-
Wednesday 28th May 2014 12:23 GMT Mark Simon
The moral or the story …
Never use averages as the source of your data. Anything which combines data has already lost important detail.
This also applies to democratic elections.
Government is won by the party which gains the most seats. However, the majority of a majority is not always a majority. For example 60% of 60% is only 36%.
At least twice in recent Australian political history, the government had the majority of seats but the opposition had the majority overall popular vote. On another occasion, one minor party had 10% of the popular vote, but not one seat.
Isn’t statistics wonderful? Lies and Damned Lies …
-
Thursday 29th May 2014 02:43 GMT Frumious Bandersnatch
Re: The moral or the story …
Never use averages as the source of your data. Anything which combines data has already lost important detail.
Oh, I don't know about that. While reading the first article in the series (and again, with the German tank problem) I was slightly disappointed not to see Little's Law listed. Now there's an interesting (and valid) application of averages...
-
Wednesday 28th May 2014 14:45 GMT Anonymous Coward
Normalization???
To get numbers normalized when appropriate works wonders. Spending x% more yields y% more revenue should correlate in both. But for cripe sakes don't just add numbers to the same pool.
With statistics and fallacy you can prove smoking is healthier, or my favorite: 30% of car crashes are caused by drunk drivers... so you'd rather be drunk than be the cause of the other 70% of crashes. And don't ever call them accidents when a drunk driver is involved, because *it was no accident*.
-
Wednesday 28th May 2014 21:45 GMT Herby
Lies, Damn lies, and Statistics.
All bad. When my mom took a statistics class, the professor went further. He had a bunch of examples of "this is how you can alter the results to your liking". It happens all the time. Of course EVERY politician uses these techniques, so nothing is new in the world.
Of course, my dad also told me something: "Don't average averages, it doesn't work". You can't average 100 MPH and hope to catch up doing 300MPH to compete with a 200MPH vehicle. It just doesn't work.
Mom & Dad were smart! MBAs do that to you (yes, they both had MBAs)!
-
Wednesday 28th May 2014 21:51 GMT J.G.Harston
I saw this all the time observing the election counts last week.
"Ooo! we're 20% ahead in Box D, it's looking good for us!"
Ah, but only 60 people voted in Box D, about 600 people voted in each of Boxes A, B and C and the other candidate only has to be 0.666% ahead in all three other boxes to win.
-
Wednesday 28th May 2014 21:58 GMT Joe Loughry
colour coding is brilliant
Thank you for this. It's the first time I really understood Simpson's Paradox, despite having seen it before. The breakthrough was the colour-coded plot in Figure 4. Brilliant way of explaining it.
Would you care to step into a time machine and take the place of my *first* statistics teacher in 1982?
-
Thursday 29th May 2014 02:39 GMT Frumious Bandersnatch
Doh
The author mustn't have got the memo at Vulture Towers. I thought the current rule was "no Simpsons jokes, please – we're adults here..." [paragraph 3],