Marginal probability mass function Given a known
joint distribution of two
discrete random variables, say, and , the marginal distribution of either variable – for example – is the
probability distribution of when the values of are not taken into consideration. This can be calculated by summing the
joint probability distribution over all values of . Naturally, the converse is also true: the marginal distribution can be obtained for by summing over the separate values of . p_X(x_i)=\sum_{j}p(x_i,y_j), \quad \text{and} \quad p_Y(y_j)=\sum_{i}p(x_i,y_j) A
marginal probability can always be written as an
expected value: p_X(x) = \int_y p_{X \mid Y}(x \mid y) \, p_Y(y) \, \mathrm{d}y = \operatorname{E}_{Y} [p_{X \mid Y}(x \mid Y)]\;. Intuitively, the marginal probability of
X is computed by examining the conditional probability of
X given a particular value of
Y, and then averaging this conditional probability over the distribution of all values of
Y. This follows from the definition of
expected value (after applying the
law of the unconscious statistician) \operatorname{E}_Y [f(Y)] = \int_y f(y) p_Y(y) \, \mathrm{d}y. Therefore, marginalization provides the rule for the transformation of the probability distribution of a random variable
Y and another random variable : p_X(x) = \int_y p_{X \mid Y}(x \mid y) \, p_Y(y) \, \mathrm{d}y = \int_y \delta\big(x - g(y)\big) \, p_Y(y) \, \mathrm{d}y.
Marginal probability density function Given two
continuous random variables
X and
Y whose
joint distribution is known, then the marginal
probability density function can be obtained by integrating the
joint probability density, , over
Y, and vice versa. That is \begin{align} f_X(x) = \int_c^d f(x,y) \, dy \\ f_Y(y) = \int_a^b f(x,y) \, dx \end{align} where x\in[a,b], and y\in[c,d].
Marginal cumulative distribution function Finding the marginal
cumulative distribution function from the joint cumulative distribution function is easy. Recall that: • For
discrete random variables, F(x,y) = P(X\leq x, Y\leq y) • For
continuous random variables, F(x,y) = \int_{a}^{x} \int_{c}^{y} f(x',y') \, dy' dx' If
X and
Y jointly take values on [
a,
b] × [
c,
d] then F_X(x) = F(x,d) \quad \text{and} \quad F_Y(y) = F(b,y) If
d is ∞, then this becomes a limit F_X(x) = \lim_{y \to \infty} F(x,y). Likewise for F_Y(y). == Marginal distribution vs. conditional distribution ==