STAT110: Fall 2020
  • Welcome!
  • FAQ
  • Resources
  • Section 1
    • Questions
  • Section 2
    • Questions
  • Section 3
    • Questions
  • Section 4
    • Questions
  • Section 5
    • Questions
  • Section 6
    • Questions
  • Section 7
    • Questions
  • Section 8
    • Questions
  • Section 9
    • Questions
  • Section 10
    • Questions
Powered by GitBook
On this page
  • CDF properties
  • Universality of the Uniform
  • Multivariate LOTUS (Law of the Unconscious Statistician)
  • Poisson Process
  • Moment Generating Functions
  • MGF Properties

Was this helpful?

Section 5

Continuous Probability and Poisson Process (BH Chapter 5)

What is a Continuous Random Variable (CRV)? A continuous random variable can take on any possible value within a certain interval (for example, [0, 1]), whereas a discrete random variable can only take on variables in a list of countable values (for example, all the integers, or the values 1,12,14,181, \frac{1}{2} , \frac{1}{4} , \frac{1}{8}1,21​,41​,81​ , etc.)

PMF’s vs. PDF’s Discrete R.V’s have Probability Mass Functions, while continuous R.V.’s have Probability Density Functions. We visualize a PDF as a graph where the x axis is the support of our values

Intuitively, what do the yyy values represent? Well it doesn't make sense to say P(X=x)P(X = x)P(X=x) for a continuous r.v. XXX because P(X=x)=0P(X = x) = 0P(X=x)=0 for all xxx. Think about the yyy value in the above graph as: the relative frequency for getting a value within ϵ\epsilonϵ of the xxx value where ϵ\epsilonϵ is small

What is the Cumulative Density Function (CDF)? It is the following function of x.

F(x)=P(X≤x)F(x) = P(X \leq x)F(x)=P(X≤x)

CDF properties

1) FFF is increasing.

2) FFF is right-continuous.

3) F(x)→1F(x) \rightarrow 1F(x)→1 as x→∞x \rightarrow \inftyx→∞, F(x)→0F(x) \rightarrow 0F(x)→0 as x→−∞x \rightarrow -\inftyx→−∞

What is the Probability Density Function (PDF)? The PDF, f (x), is the derivative of the CDF.

F′(x)=f(x)F'(x) = f(x)F′(x)=f(x)
F(x)=∫−∞xf(t)dtF(x) = \int_{-\infty}^x f(t)dtF(x)=∫−∞x​f(t)dt
F(b)−F(a)=∫abf(x)dxF(b) - F(a) = \int^b_a f(x)dxF(b)−F(a)=∫ab​f(x)dx

Thus to find the probability that a CRV takes on a value in an interval, you can integrate the PDF, thus finding the area under the density curve.

Two additional properties of a PDF:

  • It must integrate to 1 (because the probability that a CRV falls in the interval [−∞,∞][-\infty, \infty][−∞,∞] is 1

  • The PDF must always be nonnegative.

∫−∞∞f(x)dx,f(x)≥0\int^\infty_{-\infty}f(x)dx, \hspace{2 cm} f(x) \geq 0∫−∞∞​f(x)dx,f(x)≥0

How do I find the expected value of a CRV? Where in discrete cases you sum over the probabilities, in continuous cases you integrate over the densities.

E(X)=∫−∞∞xf(x)dxE(X) = \int^\infty_{-\infty}xf(x)dxE(X)=∫−∞∞​xf(x)dx

Universality of the Uniform

When you plug any random variable into its own CDF, you get a Uniform[0,1] random variable. When you put a Uniform[0,1] into an inverse CDF, you get the corresponding random variable. For example, let’s say that a random variable X has a CDF

F(x)=1−e−xF(x) = 1 - e^{-x}F(x)=1−e−x

By the Universality of the the Uniform, if we plug in X into this function then we get a uniformly distributed random variable.

F(X)=1−e−X∼UF(X) = 1 - e^{-X} \sim UF(X)=1−e−X∼U

Similarly, since F(X)∼UF(X) \sim UF(X)∼U then X∼F−1(U)X \sim F^{−1}(U)X∼F−1(U). The key point is that for any continuous random variable X, we can transform it into a uniform random variable and back by using its CDF.

Multivariate LOTUS (Law of the Unconscious Statistician)

In one dimension, we have: E(g(X))=∑xg(x)P(X=x)E(g(X)) = \sum_xg(x)P(X=x)E(g(X))=∑x​g(x)P(X=x), or E(g(X))=∫−∞∞g(x)fX(x)dxE(g(X)) = \int_{-\infty}^{\infty}g(x)f_X(x) d xE(g(X))=∫−∞∞​g(x)fX​(x)dx

For discrete random variables: E(g(X,Y))=∑x∑yg(x,y)P(X=x,Y=y)E(g(X, Y)) = \sum_x\sum_yg(x, y)P(X=x, Y=y)E(g(X,Y))=∑x​∑y​g(x,y)P(X=x,Y=y) For continuous random variables: E(g(X,Y))=∫−∞∞∫−∞∞g(x,y)fX,Y(x,y)dxdyE(g(X, Y)) = \int_{-\infty}^{\infty}\int_{-\infty}^{\infty}g(x, y)f_{X,Y}(x, y) d x d yE(g(X,Y))=∫−∞∞​∫−∞∞​g(x,y)fX,Y​(x,y)dxdy

Poisson Process

The Poisson process gives a story that links the Exponential distribution with the Poisson distribution. A Poisson process with rate λ\lambdaλ has the following properties:

  • The number of arrivals that occur in an interval of length ttt is distributed Pois(λt)Pois(\lambda t)Pois(λt).

  • The number of arrivals that occur in disjoint intervals are independent of each other.

Count-Time Duality- Instead of asking how many events occur within some amount of time, we can flip the question around and ask how long it takes until some number of events occur. Let TnT_nTn​ be the amount of time it takes until the n-th event occurs and let NtN_tNt​ be the number of events that occur within time ttt. What relationship do we have between TnT_nTn​ and NtN_tNt​?

P(Tn>t)=P(Nt<n)P(T_n > t) = P(N_t < n)P(Tn​>t)=P(Nt​<n)

To reason about this in words, the event that the n-th arrival time is greater than ttt is equivalent to the event that the number of arrivals by time ttt is less than nnn.

Using count-time duality, we can discern the distribution of TnT_nTn​. Let us first look at the distribution for the first arrival.

For the first arrival, we have P(T1≤t)=1−P(T1>t)=1−P(Nt<1)=1−e−λtP(T_1 \leq t) = 1 - P(T_1 > t) = 1 - P(N_t < 1) = 1 - e^{-\lambda t}P(T1​≤t)=1−P(T1​>t)=1−P(Nt​<1)=1−e−λt This is the Expo(λ)Expo(\lambda)Expo(λ) CDF!

Moment - Moments describe the shape of a distribution. The first three moments, are related to Mean, Variance, and Skewness of a distribution. The k−thk-thk−th moment of a random variable X is

μk′=E(Xk)\mu'_k = E(X^k)μk′​=E(Xk)

Mean, Variance, and other moments (Skewness, Kurtosis, etc.) can be expressed in terms of the moments of a random variable.

  • Mean μ1′=E(X)\mu'_1 = E(X)μ1′​=E(X)

  • Variance μ2′=E(X2)=Var(X)+(μ1′)2\mu'_2 = E(X^2) = Var(X) + (\mu_1')^2μ2′​=E(X2)=Var(X)+(μ1′​)2

Moment Generating Functions

MGF For any random variable X, the moment generating function (MGF) of X is an exponential function of X if it exists for a finitely-sized interval centered around 0. The MGF is just a function of a dummy variable t.

MX(t)=E(etX)M_X(t) = E(e^{tX})MX​(t)=E(etX)

Why is it called the Moment Generating Function? Because the k−thk-thk−th derivative of the moment generating function evaluated 0 is the k−thk-thk−th moment of X!

μk′=E(Xk)=MX(k)(0)\mu_k' = E(X^k) = M_X^{(k)}(0)μk′​=E(Xk)=MX(k)​(0)

Why does this relationship hold? By differentiation under the integral sign and then plugging in t=0t=0t=0:

MX(k)(t)=dkdtkE(etX)=E(dkdtketX)=E(XketX)M_X^{(k)}(t) = \frac{d^k}{dt^k}E(e^{tX}) = E(\frac{d^k}{dt^k}e^{tX}) = E(X^ke^{tX})MX(k)​(t)=dtkdk​E(etX)=E(dtkdk​etX)=E(XketX)
MX(k)(0)=E(Xke0X)=E(Xk)=μk′M_X^{(k)}(0) = E(X^ke^{0X}) = E(X^k) = \mu_k'MX(k)​(0)=E(Xke0X)=E(Xk)=μk′​

MGF Properties

MGF of linear combination of X: MGF of linear combination of X. If we have Y=aX+cY = aX + cY=aX+c, then MY(t)=E(et(aX+c))=ectE(e(at)X)=ectMX(at)M_Y(t) = E(e^{t(aX + c)}) = e^{ct}E(e^{(at)X}) = e^{ct}M_X(at)MY​(t)=E(et(aX+c))=ectE(e(at)X)=ectMX​(at)

Uniqueness of the MGF: If it exists, the MGF uniquely defines the distribution. This means that for any two random variables X and Y, they are distributed the same (their CDFs/PDFs are equal) if and only if their MGF’s are equal. You can’t have different PDFs when you have two random variables that have the same MGF.

Summing Independent R.V.s by Multiplying MGFs:

If X and Y are independent, then the MGF of the sum of two random variables is the product of the MGFs of those two random variables.

M(X+Y)(t)=E(et(X+Y))=E(etXetY)=E(etX)E(etY)=MX(t)⋅MY(t)M_{(X+Y)}(t) = E(e^{t(X + Y)}) = E(e^{tX}e^{tY}) = E(e^{tX})E(e^{tY}) = M_X(t) \cdot M_Y(t)M(X+Y)​(t)=E(et(X+Y))=E(etXetY)=E(etX)E(etY)=MX​(t)⋅MY​(t)
M(X+Y)(t)=MX(t)⋅MY(t)M_{(X+Y)}(t) = M_X(t) \cdot M_Y(t)M(X+Y)​(t)=MX​(t)⋅MY​(t)
PreviousQuestionsNextQuestions

Last updated 4 years ago

Was this helpful?