Distributions

Introduction

In the previous topic, we looked at the concept of a probability mass function, but there are many different forms this can take depending on the nature of the problem you’re dealing with.

Geometric distribution

The probability of kk trials required before one success. Where the kthk^{th} trial is a success.

P(X=k)=p(1p)k1\Bbb{P}(X = k) = p(1-p)^{k-1}

Where pp is the probability of success in each trail.

Binomial distribution

The probability of kk successes in nn trials.

P(X=k)=(nk)pk(1p)nk\Bbb{P}(X = k) = \dbinom{n}{k}p^k(1-p)^{n-k}

Where pp is the probability of success in each trail.

Negative binomial distribution

The probability of kk trails being required to obtain nn successes.

P(X=k)=(k1n1)pn(1p)kn\Bbb{P}(X = k) = \dbinom{k - 1}{n - 1} p^n(1 - p)^{k-n}

Where pp is the probability of success in each trail.

Poisson distribution

The Poisson distribution is an approximation of the binomial distribution, making it very similar. The key difference with Poisson is that it can take into account a rate of some kind.

P(X=k)=λkk!eλ\Bbb{P}(X = k) = \dfrac{λ^k}{k!}e^{-λ}

Where the rate is given by:

λ=npλ = np

With nn being the number of trials and pp being the probability of each trial being a success. Some questions make you model this as the rate in TT time units. In other words, two trains arriving per hour would have a rate of 2T2T where TT is the number of hours.

Some properties of Poisson distributions include:

E[X]=λE[X2]=λ2+λVar[X]=λ\begin{aligned} \Bbb{E}[X] &= λ \\\\ \Bbb{E}[X^2] &= λ^2 + λ \\\\ Var[X] &= λ \\\\ \end{aligned}

Hypergeometric distribution

The probability of kk blue balls being chosen from a bag of NN balls, where a total of nn balls are chosen and KK balls in the bag were blue.

P(X=k)=(Kk)(NKnk)(Nn)\Bbb{P}(X = k) = \dfrac{\dbinom{K}{k} \dbinom{N - K}{n - k}}{\dbinom{N}{n}}

Where pp is the probability of success in each trail.

More

More on this can be found on page 447 of the textbook.