Statistical Inference Versus Substantive Inference
[The Cult of Statistical Significance: How Standard Error Costs Us Jobs, Justice, and Lives, by Stephen T. Ziliak and Deirdre N. McCloskey (Ann Arbor: University of Michigan Press, ISBN-13: 978-472-05007-9, 2007) http://www.cs.trinity.edu/~rjensen/temp/DeirdreMcCloskey/StatisticalSignificance01.htm ]
[ Page 206]
Ò Like scientists today in medical and economic and other sizeless [sic] sciences, Pearson mistook a large sample size for the definite, substantive significance — evidence as Hayek put it — of "wholes." But it was, as Hayek declared, "just an illusion." Pearson's columns of sparkling asterisks, though quantitative in appearance and as appealing as is the simple truth of the sky, signified nothing.Ó
A scholar with the commentary
name Centurian comments as follows following the following article:
"One Economist's Mission to Redeem the Field of Finance," by Robert Schiller, Chronicle of Higher Education, April 8, 2012 --- http://chronicle.com/article/Robert-Shillers-Mission-to/131456/
Economics as a "science" is no different than Sociology,
Psychology, Criminal Justice, Political Science, etc.,etc.. To those in the
"hard sciences" [physics, biology, chemistry, mathematics], these
"soft sciences" are dens of thieves. Thieves who have stolen the
"scientific method" and abused it.
These soft sciences all apply the scientific method to biased and
insufficient data sets, then claim to be "scientific", then assert
their opinions and biases as scientific results. They point to
"correlations". Correlations which are made even though they know
they do not know all the forces/factors involved nor the ratio of effect from
the forces/factors.
They
know their mathematical formulas and models are like taking only a few pieces
of evidence from a crime scene and then constructing an elaborate "what
happened" prosecution and defense. Yet neither side has any real idea,
other than in the general sense, what happened. They certainly have no idea
what all the factors or human behaviors were involved, nor the true motives.
Hence the growing awareness of the limitations of all the quantitative
models that led to the financial crisis/financial WMDs going off.
Take for example the now thoroughly discredited financial and economic
models that claimed validity through the use of the same mathematics used to
make atomic weapons; Monte Carlo simulation. MC worked on the Manhattan Project
because real scientists, who obeyed the laws of science when it came to using
data, were applying the mathematics to a valid data set.
Economists and Wall Street Quants threw out the data set disciplines of
science. The Quant's of Wall Street and those scientists who claimed the data
proved man made global warming share the same sin of deception. Why? For the
same reason, doing so allowed them to continue their work in the lab. They got
to continue to experiment and "do science". Science paid for by those
with a deep vested financial interest in the the false correlations proclaimed
by these soft science dogmas.
If
you take away a child's crayons and give him oil paints used by Michelangelo,
you're not going to get the Sistine Chapel. You're just going to get a bigger
mess.
If
Behavioral Finance proves anything it is how far behind the other Social
Sciences economists really are. And if the "successes" of the Social
Sciences are any indication, a lot bigger messes are waiting down the road.
Centurion
"The Standard Error of Regressions," by Deirdre N.
McCloskey and Stephen T. Ziliak, Journal of Economic Literature, 1996, pp.
97-114
THE IDEA OF statistical significance is old, as old as Cicero writing on
forecasts (Cicero, De Divinatione, I. xiii. 23). In 1773 Laplace used it to
test whether comets came from outside the solar system (Elizabeth Scott 1953,
p. 20). The first use of the very word "significance" in a
statistical context seems to be John Venn's, in 1888, speaking of differences expressed
in units of probable error,
They inform us which of
the differences in the above tables are permanent and significant, in the sense
that we may be tolerably confident that if we took another similar batch we
should find a similar difference; and which are merely transient and
insignificant, in the sense that another similar batch is about as likely as
not to reverse the conclusion we have obtained. (Venn, quoted in Lancelot
Hogben 1968, p. 325).
Statistical significance has been much used since Venn, and especially
since Ronald Fisher. The problem, and our main point, is that a difference can
be permanent (as Venn put it) without being "significant" in o ther
senses, such as for science or policy. And a difference can be significant for
science or policy and yet be insignificant statistically, ignored by the less
thoughtful researchers.
In
the 1930s Jerzy Neyman and Egon S. Pearson, and then more explicitly Abraham
Wald, argued that actual investigations should depend on substantive not merely
statistical significance. In 1933 Neyman and Pearson wrote of type I and type
II errors:
Is it more serious to
convict an innocent man or to acquit a guilty? That will depend on the
consequences of the error; is the punishment death or fine; what is the danger to
the community of released criminals; what are the current ethical views on
punishment? From the point of view of mathematical theory all that we can do is
to show how the risk of errors may be controlled and minimised. The use of
these statistical tools in any given case, in determining just how the balance
should be struck, must be left to the investigator. (Neyman and Pearson 1933,
p. 296; italics supplied)
Wald went further:
The question as to how
the form of the weight [that is, loss] function . . . should be determined, is
not a mathematical or statistical one. The statistician who wants to test
certain hypotheses must first determine the relative importance of all possible
errors, which will depend on the special purposes of his investigation. (1939,
p. 302, italics supplied)
To date no empirical studies have been undertaken measuring the use of statistical significance in economics. We here examine the alarming hypothesis that ordinary usage in economics takes statistical significance to be the same as economic significance. We compare statistical best practice against leading textbooks of recent decades and against the papers using regression analysis in the 1980s in the American Economic Review.