20 Joint Distributions
- The joint distribution of random variables
and (defined on the same probability space) is a probability distribution on pairs. - In this context, the distribution of one of the variables alone is called a marginal distribution.
20.1 Joint probability mass functions
Example 20.1 Flip a fair coin four times and record the results in order, e.g. HHTT means two heads followed by two tails. We’re interested in the proportion of the flips which immediately follow a H that result in H. In particular, what is the expected value of this proportion? (Make a guess before proceeding.)
For example, if the sequence is HHTT then there are two flips which follow H (the second and third flips) of which 1 results in H (the second flip), so the proportion of interest is 1/2. We cannot measure this proportion if no flips follow a H, i.e. the outcome is either TTTT or TTTH; in these cases, we would discard the outcome and try again.
Let:
be the number of flips immediately following H (e.g., for HHTT) be the number of flips immediately following H that result in H (e.g., for HHTT) be the proportion of flips immediately following H that result in H (e.g. for HHTT)
Make a table of all possible outcomes and the corresponding values of
.
Make a two-way table representing the joint probability mass function of
and .
Make a table specifying the pmf of
.
Compute
. Surprised? Is equal to ?
- The joint probability mass function (pmf) of two discrete random variables
defined on a probability space with probability measure is the function defined by - Remember to specify the possible
pairs when defining a joint pmf.
Example 20.2 Let
Compute and interpret the probability that the home teams hits 2 home runs and the away team hits 1 home run.
Construct a two-way table representation of the joint pmf (you can use software or a spreadsheet).
Compute and interpret the probability that each team hits at most 3 home runs.
Compute and interpret the probability that both teams combine to hit a total of 3 home runs.
Compute and interpret the probability that the home team and the away team hit the same number of home runs.
- Recall that we can obtain marginal distributions from a joint distribution.
- Marginal pmfs are determined by the joint pmf via the law of total probability.
- If we imagine a plot with blocks whose heights represent the joint probabilities, the marginal probability of a particular value of one variable can be obtained by “stacking” all the blocks corresponding to that value.
Example 20.3 Continuing Example 20.2. Let
Compute and interpret the probability that the home team hits 2 home runs.
Find the marginal pmf of
, and identify the marginal distribution by name.
Compute and interpret the probability that the away team hits 1 home run.
Find the marginal pmf of
, and identify the marginal by name.
Use the joint pmf to compute the probability that the home team hits 2 home runs and the away team hits 1 home run. How does it relate to the marginal probabilities from the previous parts? What does this imply about the events
and ?
How does the joint pmf relate to the marginal pmfs from the previous parts? What do you think this implies about
and ?
In light of the previous part, how you could use spinners to simulate and
pair?
20.2 Joint probability density fuctions
- The joint distribution of two continuous random variables can be specified by a joint pdf, a surface specifying the density of
pairs. - The probability that the
pair of random variables lies is some region is the volume under the joint pdf surface over the region.
Example 20.4 Suppose that
has a Normal(0, 1) distribution has a Uniform(-2, 2) distribution and are generated independently .
Sketch a plot representing the joint pdf of
- The joint probability density function (pdf) of two continuous random variables
defined on a probability space with probability measure is the function which satisfies, for any region - A joint pdf is a surface with height
at . - The probability that the
pair of random variables lies in the region is the volume under the pdf surface over the region - The height of the density surface at a particular
pair is related to the probability that takes a value “close to” :
Example 20.5 Let
Is the joint pdf a function of both
and ? How?
Why is
equal to 0 if ?
Sketch a plot of the joint pdf. What does its shape say about the distribution of
and in this context?
Set up the integral to find
.
Sketch a plot of the marginal pdf of
. Be sure to specify possible values.
Find the marginal pdf of
at .
Find the marginal pdf of
. Be sure to specify possible values. (Can you identify the marginal distribution of by name?)
Compute and interpret
.
Sketch the marginal pdf of
. Be sure to specify possible values.
Find the marginal pdf of
at .
Find the marginal pdf of
. Be sure to specify possible values of .
Compute and interpret
.
Is
equal to the product of and ? Why?
- The joint distribution is a distribution on
pairs. A mathematical expression of a joint distribution is a function of both values of and values of . Pay special attention to the possible values; the possible values of one variable might be restricted by the value of the other. - The marginal distribution of
is a distribution on values only, regardless of the value of . A mathematical expression of a marginal distribution will have only values of the single variable in it; for example, an expression for the marginal distribution of will only have in it (no , not even in the possible values).
20.3 Independence of random variables
- Two random variables
and defined on a probability space with probability measure are independent if for all . That is, two random variables are independent if their joint cdf is the product of their marginal cdfs. - Random variables
and are independent if and only if the joint distribution factors into the product of the marginal distributions. The definition is in terms of cdfs, but analogous statements are true for pmfs and pdfs. - Random variables
and are independent if and only if their joint distribution can be factored into the product of a function of values of alone and a function of values of alone. That is, and are independent if and only if there exist functions and for which and are independent if and only if the joint distribution factors into a product of the marginal distributions. The above result says that you can determine if that’s true without first finding the marginal distributions.
Example 20.6 Let
Without doing any calculations, determine if
and are independent and find the marginal distributions.
Sketch a plot of the joint pdf of
and .
Without integrating, find
.