Entropy (geometric distribution, failures before success) Entropy is a measure of uncertainty in a probability distribution. For the geometric distribution that models the number of failures before the first success, the probability mass function is: P(X = k) = (1 - p)^k p, \quad k = 0, 1, 2, \dots The entropy H(X) for this distribution is defined as: \begin{align} H(X) &= - \sum_{k=0}^\infty P(X = k) \ln P(X = k) \\ &= - \sum_{k=0}^\infty (1 - p)^k p \ln \left( (1 - p)^k p \right) \\ &= - \sum_{k=0}^\infty (1 - p)^k p \left[ k \ln(1 - p) + \ln p \right] \\ &= -\log p - \frac{1 - p}{p} \log(1 - p) \end{align} The entropy increases as the probability p decreases, reflecting greater uncertainty as success becomes rarer.
Fisher's information (geometric distribution, failures before success) Fisher information measures the amount of information that an observable random variable X carries about an unknown parameter p. For the geometric distribution (failures before the first success), the Fisher information with respect to p is given by: I(p) = \frac{1}{p^2(1 - p)}
Proof: • The
likelihood function for a geometric random variable X is: L(p; X) = (1 - p)^X p • The
log-likelihood function is: \ln L(p; X) = X \ln(1 - p) + \ln p • The score function (first derivative of the log-likelihood w.r.t. p) is: \frac{\partial}{\partial p} \ln L(p; X) = \frac{1}{p} - \frac{X}{1 - p} • The second derivative of the log-likelihood function is: \frac{\partial^2}{\partial p^2} \ln L(p; X) = -\frac{1}{p^2} - \frac{X}{(1 - p)^2} •
Fisher information is calculated as the negative expected value of the second derivative: \begin{align} I(p) &= -E\left[\frac{\partial^2}{\partial p^2} \ln L(p; X)\right] \\ &= - \left(-\frac{1}{p^2} - \frac{1 - p}{p (1 - p)^2} \right) \\ &= \frac{1}{p^2(1 - p)} \end{align} Fisher information increases as p decreases, indicating that rarer successes provide more information about the parameter p.
Entropy (geometric distribution, trials until success) For the geometric distribution modeling the number of trials until the first success, the probability mass function is: P(X = k) = (1 - p)^{k - 1} p, \quad k = 1, 2, 3, \dots The entropy H(X) for this distribution is the same as that of version modeling trials until failure, \begin{align} H(X) &= - \log p - \frac{1 - p}{p} \log(1 - p) \end{align}
Fisher's information (geometric distribution, trials until success) Fisher information for the geometric distribution modeling the number of trials until the first success is given by: I(p) = \frac{1}{p^2(1 - p)}
Proof: • The
likelihood function for a geometric random variable X is: :: L(p; X) = (1 - p)^{X - 1} p • The
log-likelihood function is: : \ln L(p; X) = (X - 1) \ln(1 - p) + \ln p • The score function (first derivative of the log-likelihood w.r.t. p) is: :: \frac{\partial}{\partial p} \ln L(p; X) = \frac{1}{p} - \frac{X - 1}{1 - p} • The second derivative of the log-likelihood function is: :: \frac{\partial^2}{\partial p^2} \ln L(p; X) = -\frac{1}{p^2} - \frac{X - 1}{(1 - p)^2} •
Fisher information is calculated as the negative expected value of the second derivative: \begin{align} I(p) &= -E\left[\frac{\partial^2}{\partial p^2} \ln L(p; X)\right] \\ &= - \left(-\frac{1}{p^2} - \frac{1 - p}{p (1 - p)^2} \right) \\ &= \frac{1}{p^2(1 - p)} \end{align}
General properties • The
probability generating functions of geometric random variables X and Y defined over \mathbb{N} and \mathbb{N}_0 are, respectively,\begin{align} \varphi_X(t) &= \frac{pe^{it}}{1-(1-p)e^{it}},\\[10pt] \varphi_Y(t) &= \frac{p}{1-(1-p)e^{it}}. \end{align} • The
entropy of a geometric distribution with parameter p is • The geometric distribution defined on \mathbb{N}_0 is
infinitely divisible, that is, for any positive integer n, there exist n independent identically distributed random variables whose sum is also geometrically distributed. This is because the negative binomial distribution can be derived from a Poisson-stopped sum of
logarithmic random variables. ==Related distributions==