Week 7: Poisson Processes, Joint Distributions, and Covariance

0. Logistical Info

  • Section date: 11/1
  • Associated lectures: 10/24, 10/26
  • Associated pset: Pset 7, due 11/3
  • Office hours on 11/1 from 7-9pm at Quincy Dining Hall
  • Remember to fill out the attendance form

0.1 Summary + Practice Problem PDFs

Summary + Practice Problems PDF

Practice Problem Solutions PDF

1. Moment Generating Functions (MGFs)

For a random variable X, the nth moment is E(Xn).

For a random variable X, the moment generating function (MGF) is MX(t)=E(etX) for tR. If the MGF exists, then MX(0)=1,dndtnMX(t)|t=0=MX(n)(t)=E(Xn). You should sanity-check that MX(0)=1 whenever you calculate an MGF.

Useful MGF results:

  • For independent random variables, X,Y with MGFs MX,MY, then MX+Y(t)=MX(t)MY(t).
  • For random variable X and scalars a,b, Ma+bX(t)=eatMX(bt) since Ma+bX(t)=E(et(a+bX))=eatE(ebtX).

A distribution is uniquely determined by any of the following:

  • PMF (common for discrete),
  • PDF,
  • CDF (common for continuous),
  • MGF, or
  • matching to a named distribution (common).

2. Poisson Processes

Consider a problem similar to Blissville/Blotchville, where T1,T2,, represent the arrival times of busses (the amount of time from when we started waiting to when each bus arrives). Then the bus arrival process is a Poisson process with rate λ if it satisfies the following conditions:

  1. For any interval in time of length t>0, the number of arrivals in that interval is distributed Pois(λt).
  2. For any non-overlapping (disjoint) intervals of time, the number of bus arrivals are independent. This applies for any “arrival process” where T1,T2, correspond to arrival times.
Pay attention to units: λ is a rate. So if λ has units of arrivals per hour, then t should have units of hours.

Results:

  • Inter-arrival times: In a Poisson process with rate λ, the inter-arrival times (the time for the first arrival, T1, and the times between consecutive arrives T2T1,T3T2,) are each independently distributed T1,T2T1,T3T2,i.i.d.Expo(λ).
Additionally note that T2,T3,, are not exponentially distributed. In fact, they follow Gamma distributions (which we will introduce soon): TnGamma(n,λ).
  • Count-time duality: Fix a time t>0. Let Nt be the number of arrivals in the time interval [0,t], and let Tn be the arrival time of the n-th arrival. Then (Tn>t)=(Nt<n).

3. Marginal, Conditional, and Joint Distributions

Marginal, conditional, and joint distributions

Consider two random variables X,Y.

JointMarginalConditional
Distribution(X,Y)XX|Y=y
PMFP(X=x,Y=y)P(X=x)P(X=x|Y=y)
CDFP(Xx,Yy)P(Xx)P(Xx|Y=y)

For example, P(X|Y=y) is a marginal PMF. All of these apply if we flip X and Y, and PDFs follow analogously from PMFs.

  • Marginalization: If we know the joint distribution of random variables (X,Y), then we can find the marginal distribution of X (and analogously, Y) by LOTP: P(X=x)=yP(X=x,Y=y),X,Y discrete.fX(x)=fX,Y(x,y),X,Y continuous.
Note that marginal distributions of X and Y are not sufficient (not enough information) to find the joint distribution of X,Y.
  • Joint from marginal and conditional: If we know the marginal distribution of X and the conditional distributions Y|X=x for any x, then we can find the joint distribution of (X,Y) by factoring out our probability: P(X=x,Y=y)=P(X=x)P(Y=y|X=x),X,Y discrete.fX,Y(x,y)=fX(x)fY|X=x(y), X,Y continuous.
Independence of random variables: Random variables X,Y are independent if for all x and y, any of the following hold (they imply each other, if valid): FX,Y(x,y)=P(Xx,YY)=P(Xx)P(YY)=FX(x)FY(y), CDFs for any X,Y.P(X=x,Y=y)=P(X=x)P(Y=y),PMFs for discrete X,Y.fX,Y(x,y)=fX(x)fY(y),PDFs for continuous, X,Y.
  • 2D LOTUS: Let X,Y be random variables with known joint distribution. For g:support(X)×support(Y)R, LOTUS extends to 2 dimensions (or analogously for any larger dimensions) to give E(g(X,Y))={xyg(x,y)P(X=x,Y=y), X,Y discreteg(x,y)fX,Y(x,y)dxdy, X,Y continuous.

4. Covariance and Correlation

Covariance and Correlation

The covariance of random variables X,Y is Cov(X,Y)=E([XEX][YEY]) where EX is shorthand for E(X). Equivalently, Cov(X,Y)=E(XY)E(X)E(Y).

The correlation of random variables X,Y is Corr(X,Y)=Cov(X,Y)Var(X)Var(Y) =Cov(X,Y)SD(X)SD(Y), where SD(X)=Var(X) is the standard deviation of X. Equivalently, we first standardize X and Y, then find their covariance: Corr(X,Y)=Cov(XE(X)SD(X),YE(Y)SD(Y)).

X and Y are

  • positively correlated if Corr(X,Y)>0,
  • negatively correlated if Corr(X,Y)<0,
  • uncorrelated if Corr(X,Y)=0.

Since correlation and covariance have the same sign, this also applies for positive/negative/zero covariance.

Properties of covariance: see page 327 in Blitzstein & Huang for full list. Let X,Y,W,Z be random variables, as well as those of the form X1,X2,,.

  • If X,Y are independent, then Cov(X,Y)=0 (so X,Y are uncorrelated).
  • Cov(X,X)=Var(X).
  • Var(iXi)=iVar(Xi)+i<j2Cov(Xi,Xj).
    • This can be especially useful for finding the variance of a sum of indicators.
  • Cov(X+Y,W+Z)=Cov(X,W)+Cov(X,Z)+Cov(Y,W)+Cov(Y,Z).
  • Cov(aX,bY)=abCov(X,Y).

The last two properties are referred to as bilinearity.

Properties of correlation Let X,Y be random variables.

  • -If X,Y are independent, then Corr(X,Y)=0 (so X,Y are uncorrelated)
  • 1Corr(X,Y)1.
Uncorrelated does NOT imply independent: In the previous two results, we noted independent random variables have zero correlation and zero covariance. However, the converse does not apply: uncorrelated random variables are not necessarily independent.
Previous
Next