02/17/2021
Use the left and right arrow keys to navigate the presentation forward and backward respectively. You can also use the arrows at the bottom right of the screen to navigate with a mouse.
FAIR USE ACT DISCLAIMER: This site is for educational purposes only. This website may contain copyrighted material, the use of which has not been specifically authorized by the copyright holders. The material is made available on this website as a way to advance teaching, and copyright-protected materials are used to the extent necessary to make this class function in a distance learning environment. The Fair Use Copyright Disclaimer is under section 107 of the Copyright Act of 1976, allowance is made for “fair use” for purposes such as criticism, comment, news reporting, teaching, scholarship, education and research.
The following topics will be covered in this lecture:
Courtesy of Mario Triola, Essentials of Statistics, 6th edition
Random Variable A random variable is a function that assigns a real number to each outcome in the sample space of a random experiment.
Notation A random variable is denoted by an uppercase letter such as \( X \). After an experiment is conducted, the measured value of the random variable is denoted by a lowercase letter such as \( x \)
Courtesy of Ania Panorska CC
Discrete random variable is a random variable with a finite (or countably infinite) range.
Courtesy of Ania Panorska CC
Continuous random variable is a random variable with an interval (either finite or infinite) of real numbers for its range.
The probability distribution of a random variable \( X \) is a description of the probabilities associated with the possible values of \( X \).
Probability Mass FunctionFor a discrete random variable \( X \) with possible values \( x_1, x_2,\dots, x_n \), a probability mass function is a function such that
- \( f(x_i)=P(X=x_i) \)
- \( f(x_i)\geq 0 \)
- \( \sum_{i=1}^n f(x_i)=1 \)
Courtesy of Montgomery & Runger, Applied Statistics and Probability for Engineers, 7th edition
An alternate method for describing a random variable’s probability distribution is with cumulative probabilities such as \( P(X \leq x) \).
Like a probability mass function, a cumulative distribution function provides probabilities, but always over some range of outcomes.
Consider the digital channel example from earlier:
EXAMPLE: There is a chance that a bit transmitted through a digital transmission channel is received in error.
Question: How can we compute the probability that three or fewer bits are in error?
In the last slide we saw that
\[ \begin{align} P(X=0)=0.6561&\;\;P(X=1)=0.2916\\ P(X=2)=0.0486&\;\;P(X=3)=0.0036\\ P(X=4)=0.0001 & \end{align} \]
Therefore,
\[ \begin{align} P(X\leq 3) &= P(X=0 \text{ or } X=1\text{ or }X=2\text{ or }X=3)\\ &= P(X=0)+P(X=1)+P(X=2)+P(X=3) \end{align} \] by computing the probability over the union of disjoint events.
Notice, the different outcomes for \( X \) correspond to collections of events in the sample space as,
\[ \begin{align} (X=0) \equiv\{(T,T,T,T)\} & & (X=1) \equiv \{(F,T,T,T), (T,F,T,T), (T,T,F,T), (T,T,T,F)\} , \cdots\\ \end{align} \] but where these collections of events are disjoint.
\[ P(X\leq 3) = 0.6561+0.2916+0.0486+0.0036 = 0.9999 \]
Cumulative Distribution FunctionThe cumulative distribution function of a discrete random variable \( X \), denoted as \( F(x) \), is \[ F(x) = P(X \leq x) = \sum_{x_i \leq x} f(x_i) \]
Properties of a Cumulative Distribution Function
- \( F(x)=P(X\leq x)=\sum_{x_i\leq x}f(x_i) \)
- \( 0\leq F(x) \leq 1 \)
- If \( x\leq y \), then \( F(x)\leq F(y) \)
Suppose a random variable \( X \) can assume only integer values,
Consider again the digital channel example
If now we want to find \( P(X\leq 1.5) \)
Therefore \( F(x)=0.9477 \) for all \( 1\leq x < 2 \)
Moreover, we can then partition the cumulative distribution \( F \) on intervals:
\[ \begin{align} F(x) = \begin{cases} 0 & x < 0 \\ 0.6561 & 0 \leq x < 1\\ 0.9477 & 1 \leq x < 2\\ 0.9963 & 2 \leq x < 3\\ 0.9999 & 3 \leq x < 4\\ 1 & 4\leq x \end{cases} \end{align} \]
This way we ca see that \( F(x) \) is piecewise constant between values \( x_1, x_2,\dots \)
Courtesy of Montgomery & Runger, Applied Statistics and Probability for Engineers, 7th edition
We have now constructed the fundamental tools in probability for random experiments.
The axioms of probability like the the multiplication rule, the addition rule, total probability rule, etc… are useful for constructing
However, in real examples, we usually make numeric measurements \( x \) of a random variable \( X \).
The collection of all possible outcomes \( \{x_i\} \) for the random variable \( X \) partitions the sample space into disjoint events.
Therefore, we can conveniently calculate probabilities of different measurable outcomes of the experiment through a probability mass function or a cumulative distribution function.
For a given possible outcome \( x_i \), we define the probability mass function by
\[ f(x_i) = P(X=x_i) \]
which may be associated with a collection of different measureable events in the sample space.
Similarly, for any possible value \( x \), we define the cumulative distribution function as the probability of a range of values
\[ F(x) = P(X\leq x) \]
Both of these provide different representations of the probability distribution, i.e., the complete collection of possible outcomes of \( X \) and their associated probabilities.