Image-Umang Bhalla

Continuous Random Variable

Tanav Bajaj

--

Topics Covered- CDF , PDF of Continuous Random Variables along with Markov’s and Chebyshev’s inequality.

Prerequisite- Random Variables

As mentioned in my first article a random variable X with CDF Fₓ is said to be continuous random variable if Fₓ is continuous at every x. CDF has no jumps or steps .

Cumulative Distribution Function (CDF)

CDF of a random variable X , denoted Fₓ(x) is a function R to [0,1] defined as Fₓ(x)= P(X ≤ x)

Properties

  • Fₓ(b)-Fₓ(a) = P(a ≤ X ≤ b)
  • Fₓ is a non-decreasing function which takes non-negative values
  • As X -> -∞ Fₓ goes to 0
  • As X ->∞ as Fₓ goes to 1

Continuous Random Variable

A random variable with CDF Fₓ(x) is said to be continuous if Fₓ(x) is a continuous function.

  • CDF has no steps or jumps
  • So , P(X=x)=0 for all x
  • And P( a ≤ X ≤ b) = P( a< X<b)
  • Example of such distributions — Exponential , Gaussian Distributions

Probability Density Function

A continuous random variable X with CDF Fₓ(x) is said to have PDF fₓ(x) for all x₀

Fₓ(X₀) = ∫ fₓ(x) dx (this integral is from negative infinity to x₀)

CDF is integral of PDF

For a random variable X with PDF fₓ , an event A is a subset of the real line and its probability is computed as

Why do we study PDF?

Value of PDF around fₓ(x₀) is related to taking value around x such that higher the PDF , higher chance X lie there. So it helps in making probability computations easier.

∫ fₓ(x) = P(a ≤ X ≤ b) is always 1.

Side Note-

The set of possible outputs is called the support

supp(X)= {x : fₓ(x)>0}

Functions of Continuous Random Variable

Theorem: Monotonic differentiable function

Suppose X is a continuous random variable with PDF fₓ . Let g(x) be monotonic for x ∈ supp(X) with derivative g’(x) = dg(x)/dx . Then, the PDF of Y = g(X) is

Expected value of function of continuous random variable:

Let X be a continuous random variable with density fₓ(x). Let g : R → R be a
function. The expected value of g(X), denoted E[g(X)], is given by

whenever the above integral exists. The integral may diverge to ±∞ or may not exist in some cases.

Expected value (mean) of a continuous random variable:

Mean, denoted E[X] or μX or simply μ is given by

Variance of a continuous random variable:

Variance, denoted Var[X] or σ²ₓ or simply σ² is given by

Continuous Random Variables have 2 major Inequalities

Markov’s inequality:

If X is a continuous random variable with mean μ and non-negative supp(X) (i.e.P(X < 0) = 0), then

Chebyshev’s inequality:

If X is a continuous random variable with mean ( μ )and variance σ² , then

--

--

Tanav Bajaj
Tanav Bajaj

Written by Tanav Bajaj

Caffeine-fueled Prompt Engineer who can say "Hello World!" and train ML models like it's nobody's business!

No responses yet