04/07/2021
Use the left and right arrow keys to navigate the presentation forward and backward respectively. You can also use the arrows at the bottom right of the screen to navigate with a mouse.
FAIR USE ACT DISCLAIMER: This site is for educational purposes only. This website may contain copyrighted material, the use of which has not been specifically authorized by the copyright holders. The material is made available on this website as a way to advance teaching, and copyright-protected materials are used to the extent necessary to make this class function in a distance learning environment. The Fair Use Copyright Disclaimer is under section 107 of the Copyright Act of 1976, allowance is made for “fair use” for purposes such as criticism, comment, news reporting, teaching, scholarship, education and research.
The following topics will be covered in this lecture:
Recall, any function of a random sample, i.e., any statistic, is modeled as a random variable.
If \( h \) is a general function used to compute some statistic, we thus define
\[ \hat{\Theta} = h(X_1, \cdots, X_n) \]
to be a random variable that will depend on the particular realizations of \( X_1,\cdots, X_n \).
We call the probability distribution of a statistic a sampling distribution.
Sampling Distribution
The probability distribution of a statistic is called a sampling distribution.
The sample mean
\[ \hat{\Theta} = \overline{X} = h(X_1, \cdots, X_n)= \frac{1}{n}\sum_{i=1}^n X_i \]
is now one example for which we have a model of the sampling distribution.
Specifically, the central limit theorem says that the sampling distribution of the sample mean is \( \overline{X} \sim N\left(\mu, \frac{\sigma^2}{n}\right) \) when \( X \) is normal, or if \( n \) is sufficiently large.
Point estimators
A point estimate of some population parameter \( \theta \) is a single numerical value \( \hat{\theta} \) of a statistic \( \hat{\Theta} \). This is a particular realization of the random variable \( \hat{\Theta} \), viewed as a random variable; \( \hat{\Theta} \) is called the point estimator.
We want an estimator to be “close” in some sense to the true value of the unknown parameter, but we know that it happens to be a random variable.
In this way, we need to describe how close this estimator is to the true value in a probabilistic sense.
As we have seen before, there are important parameters that describe a probability distribution or a data set:
The central limit theorem actually provided both of these (and the sampling distribution) for the sample mean:
The two above parameters thus give us a means of describing “how close” the sample mean \( \overline{X} \) tends to be to the population mean \( \mu \) in a probabilistic sense.
The notion of the “center” of the sampling distribution can be useful as a general criteria for estimators.
Formally, we say that \( \hat{\Theta} \) is an unbiased estimator of \( \theta \) if the expected value of \( \hat{\theta} \) is equal to \( \theta \).
This is equivalent to saying that the mean of the probability distribution of \( \hat{\Theta} \) (or the mean of the sampling distribution of \( \hat{\Theta} \)) is equal to \( \theta \).
Bias of an Estimator
The point estimator \( \hat{\Theta} \) is an unbiased estimator for the parameter \( \theta \) if \[ \mathbb{E}\left[\hat{\Theta}\right] = \theta \] If the estimator is not unbiased, then the difference \[ \mathbb{E}\left[\hat{\Theta}\right] - \theta \] is called the bias of the estimator \( \hat{\Theta} \). When an estimator is unbiased, the bias is zero; that is, \[ \begin{align} \mathbb{E}\left[\hat{\Theta}\right] - \theta &= \theta - \theta \\ &=0 \end{align} \]
If we consider the expected value to represent the average value over infinite replications;
A particular realization of \( \hat{\Theta} \) will generally not equal the true value \( \theta \).
However, replications of the experiment will give a good approximation of the true value \( \theta \).
Courtesy of Montgomery & Runger, Applied Statistics and Probability for Engineers, 7th edition
Minimum Variance Unbiased Estimator
If we consider all unbiased estimators of \( \theta \), the one with the smallest variance is called the minimum variance unbiased estimator (MVUE).
Courtesy of Montgomery & Runger, Applied Statistics and Probability for Engineers, 7th edition
If \( X_1, X_2 , \cdots , X_n \) is a random sample of size \( n \) from a normal distribution with mean \( \mu \) and variance \( \sigma^2 \), the sample mean \( \overline{X} \) is the MVUE for \( \mu \).
As noted before, the variance is a “natural” measure of spread mathematically for theoretical reasons, but it is in the units squared of the original units.
For this reason, when we talk about the spread of an estimator's sampling distribution, we typically discuss the standard error.
The standard error Let \( \hat{\Theta} \) be an estimator of \( \theta \). The standard error error of \( \hat{\Theta} \) is its standard deviation given by \[ \sigma_\hat{\Theta} = \sqrt{\mathrm{var}\left(\hat{\Theta}\right)}. \] If the standard error involves unknown parameters that can be estimated, substitution of those values into the equation above produces an estimated standard error denoted \( \hat{\sigma}_\hat{\Theta} \). It is also common to write the standard error as \( \mathrm{SE}\left(\hat{\Theta}\right) \).
Q: can anyone recall what the standard error is of the sample mean? That is, what is the standard deviation of the sampling distribution (for a normal sample or \( n \) large)?
\[ \overline{X}\sim N\left(\mu, \frac{\sigma^2}{n}\right). \]
\[ \sigma_{\overline{X}} = \frac{\sigma}{\sqrt{n}}. \]
As was discussed before, there are times that we may not know all the parameters that describe the standard error.
For example, suppose we draw \( X_1, \cdots, X_n \) from a normal population, for which we know neither the mean nor the variance.
Let the unknown and unobservable theoretical parameters be denoted \( \mu \) and \( \sigma \) as usual.
The sample mean has the sampling distribution,
\[ \overline{X} \sim N\left( \mu, \frac{\sigma^2}{n}\right), \]
and therefore standard error \( \sigma_{\overline{X}} = \frac{\sigma}{\sqrt{n}} \).
However, we stated that \( \sigma \) itself is unknown.
In this case, we will estimate the standard error as
\[ \hat{\sigma}_\overline{X} = \frac{s}{\sqrt{n}} \] with the sample standard deviation \( s \).
This is what is meant to estimate the standard error.
This particular example will be extremely important for confidence intervals.
An article in the Journal of Heat Transfer (Trans. ASME, Sec. C, 96, 1974, p. 59) described a new method of measuring the thermal conductivity of Armco iron.
Using a temperature of \( 100^\circ \) F and a power input of 550 watts, the following 10 measurements of thermal conductivity (in Btu/hr-ft-∘ F) were obtained:
\[ 41.60, 41.48, 42.34, 41.95, 41.86, 42.18, 41.72, 42.26, 41.81, 42.04 \]
A point estimate of the mean thermal conductivity at \( 100^\circ \) F and 550 watts is the sample mean or
\[ \overline{x} = 41.924 \]
The standard error of the sample mean is \( \sigma_\overline{X}=\frac{\sigma}{\sqrt{n}} \);
\[ \hat{\sigma}_\overline{X} = \frac{s}{\sqrt{n}}= \frac{0.284}{\sqrt{10}} \approx 0.0898 \]
Notice that the standard error is about 0.2 percent of the sample mean, implying that we have obtained a relatively precise point estimate of thermal conductivity.
Assume that thermal conductivity is normally distributed, then two times the standard error is
\[ 2\hat{\sigma}_\overline{X} = 2(0.0898) = 0.1796. \]
The empirical rule says that about 95% of realizations of the sample mean lie within two standard deviations of the true mean \( \mu \).
Therefore, we are highly confident that the true mean thermal conductivity is within the interval 41.924 ± 0.1796 or between \( [41.744 , 42.104] \).
We will now formalize this logic into confidence intervals.
Engineers are often involved in estimating parameters.
For example, there is an ASTM Standard E23 that defines a technique called the Charpy V-notch method for notched bar impact testing of metallic materials.
The impact energy is often used to determine whether the material experiences a ductile-to-brittle transition as the temperature decreases.
Suppose that we have tested a sample of \( 10 \) specimens of a particular material with this procedure. We know that we can use the sample average \( \overline{X} \) to estimate the true mean impact energy \( \mu \).
However, we also know that the true mean impact energy is unlikely to be exactly equal to your estimate due to sampling error.
Reporting the results of your test as a single number is unappealing because \( \overline{X} \) may be an accurate estimator, but it doesn't tell us how precise the estimate is.
A way to quantify our uncertainty is to report the estimate in terms of a range of plausible values called a confidence interval.
We are assuming that \( \theta \) is a deterministic but unknown constant.
\[ \begin{align} P\left(\vert \hat{\Theta} - \theta \vert \geq k \sigma_\hat{\Theta}\right) \leq \frac{1}{k^2} \end{align} \]
\[ \begin{align} P\left(\vert \hat{\Theta} - \theta \vert < 2 \sigma_\hat{\Theta}\right) > \frac{3}{4} \end{align} \]
Now we will suppose the estimate \( \hat{\Theta} \) gives us based on some data is \( \hat{\theta}=299852.4 \)
We can say that \( \theta \in (299652.4, 300 052.4) \) with confidence 75%.
Confidence intervals give us a systematic procedure as above to guarantee with a level of confidence that our plausible values for the parameter \( \theta \) include the true value.
In the remaining course, we will be concerned with dual questions:
These two ideas are known as confidence intervals and hypothesis testing respectively.
Suppose that \( X_1 , X_2, \cdots , X_n \) is a random sample from a normal distribution with unknown mean \( \mu \) and known variance \( \sigma^2 \).
By the central limit theorem, we know that the sample mean \( \overline{X} \) is distributed
\[ \overline{X} \sim N\left(\mu, \frac{\sigma^2}{n}\right) \]
We may standardize \( \overline{X} \) by subtracting its mean and dividing by its standard deviation,
\[ \begin{align} Z = \frac{\overline{X} - \mu}{\frac{\sigma}{\sqrt{n}}}. \end{align} \]
The random variable \( Z \) has a standard normal distribution.
A confidence interval estimate for \( \mu \) is an interval of the form
\[ l ≤ \mu ≤ u, \] where the end-points \( l \) and \( u \) are computed from the sample data.
Because different samples will produce different values of \( l \) and \( u \), these end-points are values of random variables \( L \) and \( U \), respectively.
Suppose that we can determine values of \( L \) and \( U \) such that the following probability statement is true:
\[ P\{L \leq \mu \leq U\} = 1 − \alpha \] where \( 0 \leq \alpha \leq 1 \).
There is a probability of \( 1 − \alpha \) of selecting a sample for which the CI will contain the true value of \( \mu \).
Once we have selected the sample, so that \( X_1 = x_1 , X_2 = x_2 , \cdots , X_n = x_n \), and computed \( l \) and \( u \), the resulting confidence interval for \( \mu \) is
\[ l \leq \mu \leq u. \]
The end-points or bounds \( l \) and \( u \) are called the lower- and upper-confidence limits (bounds), respectively, and \( 1 − \alpha \) is called the confidence coefficient (or level).
Again,
$$l \leq \mu \leq u.$$ is a fixed numerical statement with nothing random about it, this is true or untrue.
The goal then is to find the procedure that defines the random variables \( L,U \) for which the procedure will succeed \( (1-\alpha)\times 100\% \) of the time.
Courtesy of Mario Triola, Essentials of Statistics, 6th edition
Courtesy of Mario Triola, Essentials of Statistics, 6th edition
The last argument also required that the variance \( \sigma^2 \) was known.
To compute the confidence intervals as above, we need to introduce a new function, the quantile function:
qnorm(p, mean, sd)
– this is the quantile function that gives the critical value associated to the value \( \alpha=p \).qnorm(0.025)
[1] -1.959964
qnorm(0.975)
[1] 1.959964
qnorm(p)
where \( p=1-\frac{\alpha}{2} \).Suppose we know that, \( \overline{X} \) is the sample mean from a normal population with (unknown) mean \( \mu= 10 \), standard deviation \( \sigma= 2 \) and sample size \( n=16 \), then
\[ \overline{X} \sim N\left(10, \frac{4}{16}\right). \]
Notice that the standard error is thus given as \( \sigma_\overline{X} = \sqrt{\frac{1}{4}} = \frac{1}{2} \)
If \( \overline{X} \) is observed to take the value \( \overline{x}=9 \), then we can construct the \( 95\% \) confidence interval for \( \mu \) as,
se <- 0.5
z_alpha_over_2 <- qnorm(0.975)
ci <- c(9 - se*z_alpha_over_2 , 9+se*z_alpha_over_2)
ci
[1] 8.020018 9.979982
Let's suppose instead we repeat the argument but select \( \alpha=0.01 \).
This corresponds to a critical value \( z_\frac{\alpha}{2}=z_{0.005} \).
To compute the the corresponding critical value, we are looking for p
=\( 1-\frac{\alpha}{2}=0.995 \)
se <- 0.5
z_alpha_over_2 <- qnorm(0.995)
ci <- c(9 - se*z_alpha_over_2 , 9+se*z_alpha_over_2)
ci
[1] 7.712085 10.287915
Notice that this \( (1-\alpha)\times 100\% = 99\% \) confidence interval does contain the true mean.
This confidence interval is somewhat wider as we want to guarantee that the procedure will work \( 99\% \) of the time.
Therefore, we need to include a wider range of plausible values when we construct such an interval.
Courtesy of Montgomery & Runger, Applied Statistics and Probability for Engineers, 7th edition