A more general definition of conditional mutual information, applicable to random variables with continuous or other arbitrary distributions, will depend on the concept of
regular conditional probability. Let (\Omega, \mathcal F, \mathfrak P) be a
probability space, and let the random variables X, Y, and Z each be defined as a Borel-measurable function from \Omega to some state space endowed with a topological structure. Consider the
Borel measure (on the
σ-algebra generated by the open sets) in the state space of each
random variable defined by assigning each Borel set the \mathfrak P-measure of its preimage in \mathcal F. This is called the
pushforward measure X _* \mathfrak P = \mathfrak P\big(X^{-1}(\cdot)\big). The
support of a random variable is defined to be the
topological support of this measure, i.e. \mathrm{supp}\,X = \mathrm{supp}\,X _* \mathfrak P. Now we can formally define the
conditional probability measure given the value of one (or, via the
product topology, more) of the random variables. Let M be a measurable subset of \Omega, (i.e. M \in \mathcal F,) and let x \in \mathrm{supp}\,X. Then, using the
disintegration theorem: :\mathfrak P(M | X=x) = \lim_{U \ni x} \frac {\mathfrak P(M \cap \{X \in U\})} {\mathfrak P(\{X \in U\})} \qquad \textrm{and} \qquad \mathfrak P(M|X) = \int_M d\mathfrak P\big(\omega|X=X(\omega)\big), where the limit is taken over the open neighborhoods U of x, as they are allowed to become arbitrarily smaller with respect to
set inclusion. Finally we can define the conditional mutual information via
Lebesgue integration: :I(X;Y|Z) = \int_\Omega \log \Bigl( \frac {d \mathfrak P(\omega|X,Z)\, d\mathfrak P(\omega|Y,Z)} {d \mathfrak P(\omega|Z)\, d\mathfrak P(\omega|X,Y,Z)} \Bigr) d \mathfrak P(\omega), where the integrand is the logarithm of a
Radon–Nikodym derivative involving some of the conditional probability measures we have just defined. ==Note on notation==