Keywords: cdf | pmf | pdf | uniform distribution | bernoulli distribution | conditional | marginal |

## Summary

Remember that a Random Variable is a mapping $ X: \Omega \rightarrow \mathbb{R}$ that assigns a real number $X(\omega)$ to each outcome $\omega$ in a sample space $\Omega$. The definitions below are taken from Larry Wasserman’s All of Statistics.

## Cumulative distribution Function

The **cumulative distribution function**, or the **CDF**, is a function

,

defined by

A note on notation: $X$ is a random variable while $x$ is a particular value of the random variable.

Let $X$ be the random variable representing the number of heads in two coin tosses. Then $x$ can take on values 0, 1 and 2. The CDF for this random variable can be drawn thus (taken from All of Stats):

Notice that this function is right-continuous and defined for all $x$, even if $x $does not take real values in-between the integers.

## Probability Mass and Distribution Function

$X$ is called a **discrete random variable** if it takes countably many values ${x_1, x_2,…}$. We define the **probability function** or the **probability mass function** (**pmf**) for X by:

$f_X$ **is a probability**.

The pmf for the number of heads in two coin tosses (taken from All of Stats) looks like this:

On the other hand, a random variable is called a **continuous random variable** if there exists a function $f_X$ such that $f_X (x) \ge 0$ for all x, $\int_{-\infty}^{\infty} f_X (x) dx = 1$ and for every a ≤ b,

The function $f_X$ is called the probability density function (pdf). We have the CDF:

and $f_X (x) = \frac{d F_X (x)}{dx}$ at all points x at which $F_X$ is differentiable.

Continuous variables are confusing. Note:

- $p(X=x) = 0$ for every $x$. You
**cant think**of $f_X(x)$ as $p(X=x)$. This holds only for discretes. You can only get probabilities from a pdf by integrating, if only over a very small paty of the space. - A pdf can be bigger than 1 unlike a probability mass function, since probability masses represent actual probabilities.

### A continuous example: the Uniform Distribution

Suppose that X has pdf A random variable with this density is said to have a Uniform (0,1) distribution. This is meant to capture the idea of choosing a point at random between 0 and 1. The cdf is given by: and can be visualized as so (again from All of Stats):

### A discrete example: the Bernoulli Distribution

The **Bernoulli Distribution** represents the distribution a coin flip. Let the random variable $X$ represent such a coin flip, where $X=1$ is heads, and $X=0$ is tails. Let us further say that the probability of heads is $p$ ($p=0.5$ is a fair coin).

We then say:

which is to be read as $X$ **has distribution** $Bernoulli(p)$. The pmf or probability function associated with the Bernoulli distribution is

for p in the range 0 to 1. This pmf may be written as

for x in the set {0,1}.

$p$ is called a parameter of the Bernoulli distribution.

## Conditional and Marginal Distributions

Marginal mass functions are defined in analog to probabilities. Thus:

Similarly, marginal densities are defined using integrals:

Notice there is no interpretation of the marginal densities in the continuous case as probabilities. An example here if $f(x,y) = e^{-(x+y)}$ defined on the positive quadrant. The marginal is an exponential defined on the positive part of the line.

Conditional mass function is similarly, just a conditional probability. So:

The similar formula for continuous densities might be suspected to a bit more complex, because we are conditioning on the event $Y=y$ which strictly speaking has 0 probability. But it can be proved that the same formula holds for densities with some additional requirements and interpretation:

where we must assume that $f_Y(y) > 0$. Then we have the interpretation that for some event A:

An example of this is the uniform distribution on the unit square. Suppose then that $y=0.3$. Then the conditional density is a uniform density on the line between 0 and 1 at $y=0.3$.