SOC 221 • Lecture 5
Monday, July 7, 2025
Inferential statistics
Statistical procedures used
to draw conclusions
(or inferences) about
a population based on
data drawn from a sample
\(\bar{X}\) —INFERENCE–> \(\mu_x\)
Inferential statistics
Statistical procedures used
to draw conclusions
(or inferences) about
a population based on
data drawn from a sample
\(\bar{X}\) —INFERENCE–> \(\mu_x\)
Population parameter: The characteristic of the population that we are interested in knowing (i.e. the mean study time of all UW students)
Sample statistic: The characteristic of the sample that we actually observe (i.e. the mean study time of a SAMPLE of UW students)
Even with a good sample, the sample is likely to differ from the population (sample statistic is likely to be different from our population parameter, just by chance)
Assessing the risk of being wrong (being way off) in making an inference from observed sample statistics to unknown population parameters.
\(\bar{X}\) —INFERENCE–> \(\mu_x\)
Sample statistic: The characteristic of the sample that we actually observe (i.e. the mean study time of a SAMPLE of UW students)
For example: draw a random sample of 100 students and observe \(\bar{X} = 14.5\)
Population parameter: The characteristic of the population that we are interested in knowing (i.e. the mean study time of all UW students)
Our goal: Estimate the unknown population parameter \(\mu_x = ?\)
One option: Just assume that the population mean is equal to the sample mean
\[ \mu_x = \bar{X} = 14.5 \]
\[ \mu_x \neq \bar{X} \]
SAMPLING ERROR
The difference
between a sample
statistic used to
estimate a population
parameter and the
actual (but unknown)
value of the
population
parameter.
Better option: Create a CONFIDENCE INTERVAL
\[ \bar{X} \]
CONFIDENCE INTERVAL
A range of scores, centered on the sample mean, in which we think the population mean is likely to be located.
Building a margin of error around our sample mean
Better option: Create a CONFIDENCE INTERVAL
\[ \bar{X} \]
Confidence intervals are defined by the Confidence Level (our certainty that the confidence interval will contain the true population parameter)
With a 95% confidence interval, we are 95% sure that the interval we construct contains the true population parameter we are interested in.
Over the long run, there is a probability of .99 that the actual population parameter falls within a 99% confidence interval that we construct
Better option: Create a CONFIDENCE INTERVAL
\[ \bar{X} \]
CONFIDENCE INTERVAL
A range of scores, centered on the sample mean, in which we think the population mean is likely to be located.
Q: What is the ADVANTAGE of a confidence interval over a point estimate?
A: We are less likely to be wrong, and we can quantify the risk of being wrong.
Sampling Distribution
A theoretical probability distribution of all possible sample values for the statistic in which we are interested.
Allows us to think about how our one single sample result relates to all possible sample results and, by extension, the population we are interested in drawing inferences about.
Imagine drawing ONE sample from the population and calculating the sample mean
Imagine drawing ONE sample from the population and calculating the sample mean
Just by chance, our sample mean may be a little different than the true population mean.
Distance between the sample mean and the actual population mean reflects random sampling error
Now imagine drawing LOTS of independent samples and calculating the mean of X for each one.
Interesting features of the
sampling distribution
(when certain conditions are met):
• Normal distribution
• Mean of the sampling distribution is equal to the population mean
This produces a SAMPLING DISTRIBUTION of sample means
So can use what we know about the normal distribution to think about the location of sample means relative to the population mean.
68.26% of samples will be within 1 \(\sigma\) of the pop. mean
95% of sample means will be within 1.96 \(\sigma\) of the pop. mean
If all possible random samples of size \(N\) are drawn from a population with the mean \(\mu_X\) and the standard deviation \(\sigma_X\), then as \(N\) becomes larger, the sampling distribution of the sample means becomes approximately normal, with mean equal to \(\mu_X\) and standard deviation equal to the population standard deviation divided by square root of the sample size.
Check out this new symbol:
The population mean of sample means
(i.e., the average of the means you would
calculate from all possible samples)
Standard error:
The standard deviation in
the sampling distribution
(in this case, of all
possible sample means)
Just over 68% of sample means will be within ±1 standard error of the true population mean
\(95.44\%\) of sample means will be within \(\pm2\) standard errors of the true population mean
And exactly \(95\%\) will be within \(1.96\) standard errors
\(99.73\%\) of sample means will be within \(\pm3\) standard errors of the true population mean
If, 95% of the samples are within 1.96 standard errors of the true population mean . . .
If, 95% of the samples are within 1.96 standard errors of the true population mean . . .
. . . then for 95 samples out of 100, we will find the true population mean if we look within 1.96 standard errors of the sample mean.
If, 95% of the samples are within 1.96 standard errors of the true population mean . . .
. . . then for 95 samples out of 100, we will find the true population mean if we look within 1.96 standard errors of the sample mean.
If, 95% of the samples are within 1.96 standard errors of the true population mean . . .
. . . then for 95 samples out of 100, we will find the true population mean if we look within 1.96 standard errors of the sample mean.
\(\bar{X}\) —INFERENCE–> \(\mu_x\)
Sample statistic: The characteristic of the sample that we actually observe (i.e. the mean study time of a SAMPLE of UW students)
For example: draw a random sample of 100 students and observe \(\bar{X} = 14.5\)
Population parameter: The characteristic of the population that we are interested in knowing (i.e. the mean study time of all UW students)
Our goal: Estimate the unknown population parameter \(\mu_x = ?\)
\[ \bar{X} \]
Create a range of scores, centered on the sample mean, in which we think the population mean is likely to be located
Building a margin of error around our sample mean
Confidence interval:
\(\text{sample statistic} \pm \text{margin of error}\)
⬅️
All confidence intervals take this form
A researcher is studying weekly TV viewing habits of college students. The population mean is unknown, but the population standard deviation is \(\sigma = 5\) hours.
Suppose a simple random sample of \(n = 100\) students is taken, and the sample mean viewing time is \(\bar{X}\) = 12 hours.
Standard Error:
\[
SE = \frac{\sigma}{\sqrt{n}} = \frac{5}{\sqrt{100}} = 0.5 \text{ hours}
\]
Shape of Sampling Distribution:
Since \(n = 100\) is large, by the Central Limit Theorem, the sampling distribution of \(\bar{X}\) is approximately normal, regardless of population shape.
What if \(n = 25\)?
\[
SE = \frac{5}{\sqrt{25}} = 1.0
\]
So the standard error increases when sample size decreases.
Imagine a population where the distribution of individual income is strongly skewed right with a mean of $50,000 and a standard deviation of $10,000.
\(n = 5\): The sampling distribution of \(\bar{X}\) is not approximately normal — small sample from a strongly skewed population.
\(n = 50\): Sampling distribution is approximately normal — CLT applies for large enough \(n\) (typically \(n \geq 30\)).
Why is CLT useful?
It allows us to use the normal model for sample means even when the population is not normal, as long as \(n\) is large.
Two different researchers each take a simple random sample from the same population of high school students.
Standard Errors: \[
SE_A = \frac{1.8}{\sqrt{30}} \approx 0.33 \qquad SE_B = \frac{1.8}{\sqrt{90}} \approx 0.19
\]
So Researcher B has a smaller standard error.
Who is closer to \(\mu\)?
Researcher B — lower SE means their estimate is more precise.
What does this show?
Larger sample sizes → smaller SE → less variability in sample means.