Home > Uncategorized > Freedman’s Rabbit Theorem

Freedman’s Rabbit Theorem

from Lars Syll

In econometrics one often gets the feeling that many of its practitioners think of it as a kind of automatic inferential machine: input data and out comes causal knowledge. This is like pulling a rabbit from a hat. Great, but as renowned statistician David Freedman had it, first you must put the rabbit in the hat. And this is where assumptions come into the picture.

The assumption of imaginary ‘superpopulations’ is one of the many dubious assumptions used in modern econometrics, and as Clint Ballinger has highlighted, this is a particularly questionable rabbit-pulling assumption:

pulling_a_rabbit_out_of_a_hat_by_candiphoenixes-d3ee5jaInferential statistics are based on taking a random sample from a larger population … and attempting to draw conclusions about a) the larger population from that data and b) the probability that the relations between measured variables are consistent or are artifacts of the sampling procedure.

However, in political science, economics, development studies and related fields the data often represents as complete an amount of data as can be measured from the real world (an ‘apparent population’). It is not the result of a random sampling from a larger population. Nevertheless, social scientists treat such data as the result of random sampling. 

Because there is no source of further cases a fiction is propagated—the data is treated as if it were from a larger population, a ‘superpopulation’ where repeated realizations of the data are imagined. Imagine there could be more worlds with more cases and the problem is fixed …

What ‘draw’ from this imaginary superpopulation does the real-world set of cases we have in hand represent? This is simply an unanswerable question. The current set of cases could be representative of the superpopulation, and it could be an extremely unrepresentative sample, a one in a million chance selection from it …

The problem is not one of statistics that need to be fixed. Rather, it is a problem of the misapplication of inferential statistics to non-inferential situations.

  1. December 21, 2022 at 11:01 pm

    When gathering data, the timing matters.
    Data from the morning after a bank crisis, shows big bank exposures to what’s very risky.
    Data from time before, (when the rabbit is put in the hat) show those exposures being built up with what’s perceived or decreed as very safe.
    https://subprimeregulations.blogspot.com/2018/08/risk-weighted-capital-requirements-for.html

  2. January 1, 2023 at 6:23 am

    TRUE. Inferential statistics can’t be taken as ‘fact’, rather it’s only a ‘lead’. As we have no time to check the entire population we take a workable sample and try to get a lead to go about further. However, the process of finding the truth is not yet over. Perhaps, parallel research on the parent population to confirm the hypothesis is always open.

  1. No trackbacks yet.

Leave a comment

This site uses Akismet to reduce spam. Learn how your comment data is processed.