# Frequentist versus Bayesian PDF for Binary Decisions Like Coin Tossing, OSHA Compliance, and Jury Trials

Requires a Wolfram Notebook System

Interact on desktop, mobile and cloud with the free Wolfram Player or other Wolfram Language products.

The probability of success in a single trial , the number of trials in an experiment , and the number of successful trials are the parameters used to illustrate both frequentist deductive reasoning and Bayesian inductive reasoning. The trials are independent and identically distributed (iid).

[more]
Contributed by: James C. Rock (July 2012)

Open content licensed under CC BY-NC-SA

## Snapshots

## Details

Discussion of Snapshots

Snapshots 1 and 2 show the results from two experiments with . The PDF is narrower for than it is for . Larger experiments lead to better precision.

Snapshots 3 and 4 show the results from two extreme experiments, with 1 of 21 trials either failing or passing, resulting in or , respectively.

Snapshots 3 and 5 show results from two experiments with 1 failure in 21 or 6 trials, resulting in or , respectively.

Things to Try with This Demonstration

Examine the shape of for and for while takes values from 1 to 20. Note that the width of the PDF narrows as increases. This concept is useful for estimating the number of samples to plan for under anticipated experimental conditions.

Examine the shape of when for various between 1 and 20. Note again that the width narrows as increases.

You have now observed the reality of experimental design: to improve experimental precision one must dramatically increase the number of trials in an iid experiment.

Discussion

A binary trial, sometimes called a Bernoulli trial, is one that results in one of two outcomes, variously called success and failure, pass and fail, heads and tails, comply and non-comply, guilty and not guilty, 1 and 0, true and false. The terms success and fail or comply and non-comply are used in occupational and environmental compliance decisions.

An experiment consists of independent, identically distributed (iid) Bernoulli trials. The probability of success in each trial is , the probability of failure in the same trial is , and is a real number in the closed interval .

Bayes's theorem is usefully expressed with likelihoods: posterior likelihood = prior likelihood × data likelihood. In this Demonstration, the prior likelihood function is the uninformative uniform distribution, so the posterior likelihood distribution = data likelihood distribution. The data likelihood for a binary experiment with iid trials is the product of the probabilities for each trial. In this Demonstration, the data likelihood is:

.

This likelihood function (LHF) is the probability of any specific outcome from an experiment with trials yielding successes. It has three variables. Two of them ( and ) are known at the end of an experiment, while the third, , is not directly observed. The frequentist assumes a value for and deduces the discrete probability distribution for the number of successes in trials. The Bayesian, through Bayes's theorem, uses the data to infer the probability distribution for the parameter .

The frequentist uses the binomial coefficient to define the number of ways successes can be arranged among trials. Each of those arrangements has the same probability, which is denoted by , , . The probability of is the product of the binomial coefficient and the probability of each individual arrangement of successes among sequential trials. In conditional probability notation (deprecated by some frequentist experts, but widely used by Bayesian experts), the resulting discrete binomial PDF is written:

.

The Bayesian uses the data ( and ) to infer a conditional PDF for and to estimate its maximum likelihood value. The area under the likelihood function, here called the likelihood (), is calculated by integrating LHF with respect to over the range of 0 to 1. This is the Eulerian Integral, and it has a well-known solution:

.

Divide the LHF by its area to find the functional form of the PDF with unit area:

.

It turns out that both of these probability density functions can be expressed in terms of well-known parametric distributions. The first is the binomial distribution with independent variable , defined by parameters and . The second is the beta distribution with independent variable defined by shape parameters derived from and , so that and. In *Mathematica*'s notation these are expressed as:

=PDF[BinomialDistribution[n, p], s] and

=PDF[BetaDistribution[s+1, n-s+1], p]

To summarize, the likelihood function for a binomial experiment consisting of Bernoulli trials is central to both frequentist deduction and Bayesian inference. To the frequentist, LHF is the probability that an experiment ends with successes arranged in one of several sequences among trials. The probability of successes in trials, irrespective of the sequence, is expressed by counting the permutations of those trials. To the Bayesian, Bayes's theorem with the uniform distribution as the uninformative prior shows that LHF has the shape of the PDF of given and , and normalizing it to unit area produces the desired PDF.

This analysis produced two desired posterior distributions: the frequentist one for given and and the Bayesian one for given and . Suitable models for both of those distributions are found in the *Mathematica* collection of distributions. Using the binomial and beta distribution formalism opens the entire *Mathematica* spectrum of statistical analysis tools for further analysis of the compliance problem or any of its analogous binary decision problems.

References

[1] D. J. Blower, *Information Processing: Boolean Algebra, Classical Logic, Cellular Automata, and Probability Manipulations*, Pensacola, FL: David Blower, Third Millennium Inferencing, 2011.

[2] P. C. Gregory, *Bayesian Logical Data Analysis for the Physical Sciences: A Comparative Approach with Mathematica® Support*, New York: Cambridge University Press, 2010.

[3] D. Sivia and J. Skilling, *Data Analysis: A Bayesian Tutorial*, 2nd ed., New York: Oxford University Press, 2006.

[4] E. T. Jaynes, *Probability Theory: The Logic of Science*, New York: Cambridge University Press, 2003.

## Permanent Citation