## Science and the Shortcomings of Statistics 429

Posted
by
samzenpus

from the 14%-of-people-know-that-statistics-can-prove-anything dept.

from the 14%-of-people-know-that-statistics-can-prove-anything dept.

Kilrah_il writes

*"The linked article provides a short summary of the problems scientists have with statistics. As an intern, I see it many times: Doctors do lots of research but don't have a clue when it comes to statistics — and in the social science area, it's even worse. From the article: 'Even when performed correctly, statistical tests are widely misunderstood and frequently misinterpreted. As a result, countless conclusions in the scientific literature are erroneous, and tests of medical dangers or treatments are often contradictory and confusing.'"*
## Lies, Damned Lies, and Statistics. (Score:5, Informative)

In other news math may not lie but people still can, all the honesty and good statistics in the world doesnt help end-user stupidity, and there are statistically two popes per square kilometer in the vatican.

## Re:Lies, Damned Lies, and Statistics. (Score:3, Informative)

## Re:Maths anxiety (Score:5, Informative)

## Re:Example: Standard Deviation (Score:5, Informative)

## Re:No surprise here (Score:3, Informative)

It's perfectly reasonable that someone use a calculator for sales tax (if an exact answer is desired).

Also, sales tax is multiplication - not algebra.

## Re:No surprise here (Score:1, Informative)

Arithmetic is not algebra. Arithmetic is "What's 10% of $24.45?" Algebra would be "On a given day i, John sells n_i apples to Peter at x_i dollars each, and this price includes sales tax which is a constant proportion 0p1. Let x_1= .. x_2= ... ... What is the tax on the apples sold on days 1 to 12 inclusive?"

The difference is 24.45 . 10/100 versus p\sum_{i=1}^{12} n_ix_i. Granted, there isn't much difference there really, but come on, there is a time and a place for everything, calculators included.

## Re:Long winded troll (Score:3, Informative)

## Re:Example: Standard Deviation (Score:2, Informative)

sigma = population standard deviation = sqrt((sum(x-mu)^2)/N), where mu is the mean

s is approximately equal to (highestValue-lowestValue)/4, range rule of thumb

Unusual values are outside +/- 2 standard deviations

Z = ((x-mu)/sigma) where Z is in terms of standard deviations.

## What it actually said (Score:5, Informative)

And lots of others. It then suggests Bayesian reasoning as an alternative to traditional statistical tests.

Most post-PhD scientists are aware of the common mistakes, but being aware that we make mistakes doesn't necessarily stop us from making them. If you chose a random set of conference proceedings, it is almost certain you will find at least one paper (and I suspect usually a dozen or more) that have statistical mistakes in them.

## Re:Statistical assumptions are often ignored (Score:4, Informative)

and IAAB (biologist) and I can tell you that most scientists don't have access to statisticians or don't have the grant money to pay for them. I also don't have time to learn SAS and code my own tests, therefore I use stuff like SPSS or Genstat (both of which do allow you to code your own tests as well). Just because they are easy to use doesn't mean I do or do not understand the tests, the assumptions or their results. I would say my grasp of stats is above average for my peer group, below where I would like it to be and obviously limited.

One thing that is interesting to me is that throughout my education and career I have been warned off using multiple means comparisons and LSD in particular (I understand why and have avoided where I can and the latter always). Yet the only actual statisticians I have dealt with in recent years have recommended me to use LSD on means comparisons with 10s of means. I would be hard pressed to publish those results.

In summary, whilst statisticians like to blame easy to use stats programs for bad stats the reality is they are just a tool and if statisticians can't agree on the acceptable use of the simplest procedures I'm not sure what chance the rest of us have of getting it right.

## The problem is with statistics itself (Score:3, Informative)

Statistics is changing slowly (mostly because computers and R make non-classical statistics more practical) but the way it's taught still leads to problems.

## Re:Summery? (Score:3, Informative)

That would be Muphry's law [wikipedia.org].

For details on Muphry's law, click on the above hyperlink. For more fun laws, click on the below hyperlink.

More fun here. [wikipedia.org]

## Re:Personal experience (Score:3, Informative)

It's the approach that you can just pump the numbers into SPSS or Statistica, and then call on a battery of tests until you get a "significant" result that results in the kind of errors the article (and a disturbing number of /. readers) fall into.

Unless you're dealing with large samples, all z and t tests assume normality in the population, with insignificant skew or kurtosis. Yet by definition, if we have enough data to be sure we have a normal population, we have enough data that the central limit theorem makes the differences moot. Even more extreme, if we have a complete description of the population (a census) we have no need to use any inferential statistics.

Meanwhile students are told to test the data for normality, homoscedacticity and linearity, to the point where the repeated tests on a single data set make the chance of a Type II error better than even. But by saying "SPSS said so" and burying assumptions beneath a mountain of waffle and misunderstood jargon we can still get these "results" published.

No-one who can't perform a balanced block design ANOVA by hand, or explain what transforming data does to residuals under assumptions of a linear additive model, should be allowed near statistical software in my opinion. And the so-called statistics packages in popular spreadsheets should be banned, and any student relying on them should be failed.

## Re:Long winded troll (Score:5, Informative)

## Re:Example: Standard Deviation (Score:4, Informative)

You're mixing up psychiatrists, psychologists and psychotherapists.

A psychiatrist went to med school, got a doctors degree and specialized in problems with the brain. A psychologist went to university to learn the study of behavior of people. This involves a lot of statistics and many of them probably do consider it something they didn't go to college for, but it's a study that is supposed to follow the scientific method and prepare students for doing research, not therapy.

A psychotherapist is anyone who feels like calling themselves that. As a preparation they may have studied psychology at university, or they may have spent 20 years meditating in the Himalayas, or followed a short course at a religious group such as an institute of multiple personality disorder therapists or scientology.

## Re:Long winded troll (Score:2, Informative)

"Nature journals receive many more submissions than they can publish. Therefore, we ask peer-reviewers to keep in mind that every paper that is accepted means that another good paper must be rejected. To be published in a Nature journal, a paper should meet four general criteria:

....[snip]...

*

Provides strong evidence for its conclusions.* Novel (we do not consider meeting report abstracts and preprints on community servers to compromise novelty).

* Of extreme importance to scientists in the specific field.

* Ideally, interesting to researchers in other related disciplines."

"The editors then make a decision based on the reviewers' advice, from among several possibilities:

* Accept, with or without editorial revisions

* Invite the authors to revise their manuscript to address specific concerns before a final decision is reached

* Reject, but indicate to the authors that further work might justify a resubmission

* Reject outright, typically on grounds of specialist interest, lack of novelty, insufficient conceptual advance or

major technical and/or interpretational problems"## Re:Summery? (Score:5, Informative)

## Re:Looking for a good book on statistics (Score:3, Informative)

Devore's

Probability and Statistics for Engineering and the Sciencesis probably the best one-volume, undergrad-level intro to statistics out there. Get a copy (I think it's on the sixth or seventh edition now; you can pick up a fifth edition for cheap) and work your way through that, and you'll have a pretty good idea of where all those formulae come from and how they're used. Get a copy of R [r-project.org] and check out the "Devore*" packages in the package list [r-project.org] too. If you want to learn more after that, I recommend Kutner et al.'sApplied Linear Statistical Modelsfor applications, and Casella and Berger'sStatistical Inferencefor theory.The Wikipedia stats pages are pretty good for most things, but many of them are written with the assumption of a lot of background knowledge. If you open up a page on a particular stats subject and you comprehend it, great; if not, be prepared to do a lot of digging outside of Wikipedia, because trying to figure out the subject from the links to other WP pages is an exercise in circularity.

## Re:Example: Standard Deviation (Score:1, Informative)

Are you referring to PSYCHOLOGY or PSYCHIATRY (or both?).

The former do a lot more stats than the latter during their training. Psychology students in Australia (me) do at least 4 years of statistics (and often 6) before working in the field. Of course we all have a love-hate relationship with stats, but there is no way to get through the course(s) without working hard and actually learning the material.

Psychiatry students essentially study medicine (and do intermediate stats), and then specialise afterward. They require a completely different skill-set to Psychologists, and have different training.

## Re:Example: Standard Deviation (Score:3, Informative)

## MY common conversation (Score:5, Informative)

The largest demographic in american prisons are black americans. Real statistic but is it true?

Given a particular sample that indicates blacks are 60% of the prison population this would appear to be true.

But what if I said: "The largest demographic in prison is minority, non-whites." Suddenly the % jumps from 60% (black) to 80% (minority). Which is more right? This is the problem with statistics. Context.

Now I can say readily that the largest demographic in prison is actually right-handed people. The % now jumps to 90%.

But wait! There is more! The largest demographic is prison is actually people who prior to arrest were below the poverty line which jumps to 99% of the population. Again, all of the above are accurate based on a sample but which is MORE correct? Linear Algebra is coming into play here quickly....

When that kind of issue comes into play, it is the classic "Correlation != Causation" confusion. The majority of people in prison are in there because of "Being black? Being a minority? being right handed? or being poor?" None of the above. The majority of them are in there because they were convicted of a crime and sentenced. That is the causation of their imprisonment, the rest is correlation which may have a direct causation on the conviction or sentencing, but no direct causation on being in prison. (e.g. You cannot be thrown into prison for being poor, black, minority, right handed)

Same with medical research, politics, economics, etc. The price of oil rising 10% and a subsequent 5% drop in shipping orders. Measuring the significance of regessors is important but oddly never reported most of the time. Many factors get masked or shadowed by higher level regressors (e.g. being a minority masks a variety of other social and economic factors. In addition it can distort statistical work by being too broad. Asians have a variety of different economic and social factors as north american blacks versus even african immigrants.)

Back to the orignal subject:

We can take 100 prisoners and 100 non-prisoners and figure out rather quickly if being black is statistically significant in prison population. Non-prison population blacks would account for 25%-45% of the population (Depending on location). We can see that 60% of prisoners are black. There is a 20+% deviation from the norm. We can test to see the significance of that. Same with minorities. Now we find something quickly that right handed is insignificant because it doesn't deviate from the norm. We can test left-handed and right-handed populations and rule out the handed-ness of a convict being significant.

We can find the economic status is considerable MORE significant then minority or black as a status. We can determine that the reason minorities or blacks are disporotinally more prevelant in prison is that blacks and minorities have higher rates of poverty. We can extract and determine the statistical weight of POVERTY in regards to imprisonment (Since we find a high % of white in prison that are poor compared to the normal population.) Once we figure that out we can remove that and continue an investigation and figure out what weight minority and black has once we have removed POVERTY from the model (Residual analysis).

The problem in reporting is without providing the whole, comprehensive analysis you can miss important things. For instance to correct the injustice in sentencing, without reporting the weight POVERTY has in contrast to BLACK or MINORITY you may lose sight that you may have better success addressing POVERTY to normalize sentencing rather then MINORITY or BLACK (or not).

The same happens in medical reasearch. Given a cocktail of drugs wirthout having the whole analysis you may end up providing more of Medicine A versus B but lose sight that A & B are limited by the dosage of Medicine C.

Satistics are not bullshit, rather mearly observations with no intrinsic agenda or even implication of truth. Purely amoral, like a hand gun.. useful to both the good and evil.

Statistics don't lie, nor do they tell the truth. They simple show the relationship of the data as it stands. The Truth or Thruthiness of it is subjective and vulnerable to context.

## Re:Lies, Damned Lies, and Statistics. (Score:3, Informative)

The actual truth, as usual, is a bit more complex than the bit we all remember and quote.

Where a correlation occurs there are four distinct possible reasons:

Let say a correlation that during the time when X is known to have increased, Y showed a corelatory increase. then

1) It is possible that this is because X caused Y - e.g. the causation that way isn't implied - yes it's a possible implication.

2) It is possible that X in fact caused Y (e.g. the causation is in fact in the opposite direction of what the quoter of the stat is trying to say).

3) It is possible that X and Y were both caused by an unknown third factor Z.

4) It is possible that X and Y were caused by completely different factors and their correlation is purely coincidental.

The mere existence of a correlation does not imply any of these four possibilities more strongly than the other - they are equally likely unless additional data is presented to corroborate one.

The example from my philosophy textbook (which I'm shamelessly citing here) was this:

Between the period 1955 to 1965 the number of schools in the US where sex-ed was given increased by 75%, during the same period the amount of teenage pregnancies increased by nearly 80% (Both compared to the decade before that). Conclusion - giving sex-ed led to more teenage pregnancies.

This citation is classic example of the correlation/causation mistake in that it assumes option 1.

In this case option 2 actually seems quite likely - if teenage pregnancies were going up, that would put pressure on schools to give sex-ed to try and reverse this trend.

But what if we consider more available data. Specifically that the pill came on the market in 1953 sparking the sexual revolution.

If we consider that the pill let to a more relaxed attitude among teenagers about sex, but that this attitude probably spread a lot faster than actual usage of the pill then it explains the increase in teen pregnancies which combined with the known presence of this attitude would put pressure on the schools to give sex ed.

So then it suggests that in fact we have options 2,3 and 4 happening in a commonly reinforcing manner. The only conclusion that isn't supported by the data at all is option one.

With each bit of additional data added, including comparison with other times where there was a sharp increase in teenage pregnancy (like the early years of the current decade under the Bush administration) we find that the likelihood of X actually causing Y in this example gets smaller and smaller and in fact becomes statistically insignificantly small.

But that doesn't mean option 1 is never the right answer. Sometimes a correlation really is due to causation. You just cannot assume it without further evidence.

## Re:Long winded troll (Score:3, Informative)

Also, in your quote you highlighted 2 different lines. The first has to do with the soundness of the conclusions. This is most definitely a role of peer review, but not related to accuracy. It doesn't mean that they verify that your conclusions are correct. Conclusions are not objective. The data gives you objective facts from which to draw subjective conclusions. This line indicates that your discussion will be evaluated for how well the data (yours and previous literature) supports your conclusions. If you extrapolate, or ignore important results then your paper will be rejected.

The second bolded section just indicates that if serious errors are found (using insufficiently large sample size, extrapolating results, etc.) then the paper will be rejected. That's totally understandable to reject, but obviously serious errors of this sort are uncommon. Most errors are much harder to detect, and are not picked up by the peer review process in my experience.