Article 2N3GF Cauchy, Benford, and a problem with NHST

Cauchy, Benford, and a problem with NHST

by
John
from John D. Cook on (#2N3GF)
Introduction

Samples from a Cauchy distribution nearly follow Benford's law. I'll demonstrate this below. The more data you see, the more confident you should be of this. But with a typical statistical approach, crudely applied NHST (null hypothesis significance testing), the more data you see, the less convinced you are.

This post assumes you've read the previous post that explains what Benford's law is and looks at how well samples from a Weibull distribution follow that law.

This post has two purposes. First, we show that samples from a Cauchy distribution approximately follow Benford's law. Second, we look at problems with testing goodness of fit with NHST.

Cauchy data

We can reuse the code from the previous post to test Cauchy samples, with one modification. Cauchy samples can be negative, so we have to modify our leading_digit function to take an absolute value.

 def leading_digit(x): y = log10(abs(x)) % 1 return int(floor(10**y))

We'll also need to import cauchy from scipy.stats and change where we draw samples to use this distribution.

 samples = cauchy.rvs(0, 1, N)

Here's how a sample of 1000 Cauchy values compared to the prediction of Benford's law:

|---------------+----------+-----------|| Leading digit | Observed | Predicted ||---------------+----------+-----------|| 1 | 313 | 301 || 2 | 163 | 176 || 3 | 119 | 125 || 4 | 90 | 97 || 5 | 69 | 79 || 6 | 74 | 67 || 7 | 63 | 58 || 8 | 52 | 51 || 9 | 57 | 46 ||---------------+----------+-----------|

Here's a bar graph of the same data.cauchy_benford.png

Problems with NHST

A common way to measure goodness of fit is to use a chi-square test. The null hypothesis would be that the data follow a Benford distribution. We look at the chi-square statistic for the observed data, based on a chi-square distribution with 8 degrees of freedom (one less than the number of categories, which is 9 because of the nine digits). We compute the p-value, the probability of seeing a chi-square statistic this larger or larger, and reject our null hypothesis if this p-value is too small.

Here's how our chi-square values and p-values vary with sample size.

|-------------+------------+---------|| Sample size | chi-square | p-value ||-------------+------------+---------|| 64 | 13.542 | 0.0945 || 128 | 10.438 | 0.2356 || 256 | 13.002 | 0.1118 || 512 | 8.213 | 0.4129 || 1024 | 10.434 | 0.2358 || 2048 | 6.652 | 0.5745 || 4096 | 15.966 | 0.0429 || 8192 | 20.181 | 0.0097 || 16384 | 31.855 | 9.9e-05 || 32768 | 45.336 | 3.2e-07 ||-------------+------------+---------|

The p-values eventually get very small, but they don't decrease monotonically with sample size. This is to be expected. If the data came from a Benford distribution, i.e. if the null hypothesis were true, we'd expect the p-values to be uniformly distributed, i.e. they'd be equally likely to take on any value between 0 and 1. And not until the two largest samples do we see values that don't look consistent with uniform samples from [0, 1].

In one sense NHST has done its job. Cauchy samples do not exactly follow Benford's law, and with enough data we can show this. But we're rejecting a null hypothesis that isn't that interesting. We're showing that the data don't exactly follow Benford's law rather than showing that they do approximately follow Benford's law.

nszKU2ayUb0
External Content
Source RSS or Atom Feed
Feed Location http://feeds.feedburner.com/TheEndeavour?format=xml
Feed Title John D. Cook
Feed Link https://www.johndcook.com/blog
Reply 0 comments