How does sample size affect mean standard deviation and standard error of the mean?

Example: Mean NFL Salary

The built-in dataset "NFL Contracts (2015 in millions)" was used to construct the two sampling distributions below. In the first, a sample size of 10 was used. In the second, a sample size of 100 was used.

Sample size of 10:

200Show Data TableGenerate 100 Samples Choose samples of size n =samples = 5000mean = 2.195std. error = 0.936Generate 10 SamplesGenerate 1 SampleLeft TailTw o - TailRight TailNFL Contr acts (2015 in millions)Sampling Dotplot of MeanEdit DataUpload File Change ColuReset Plot10150 100500122.1953456789Generate 100 0 Samples

Sample size of 100:

120Show Data TableGenerate 100 SamplesChoose samples of size n =samples = 5000mean = 2.236std. error = 0.296Generate 10 SamplesGenerate 1 SampleLeft TailTwo - TailRight TailNFL Contracts (2015 in millions)Sampling Dotplot of MeanEdit D ataUpload FileChange ColuReset Plot1001008060402001.501.752.002.252.502.753.003.25Generate 1000 Samples2.236

With a sample size of 10, the standard error of the mean was 0.936. With a sample size of 100 the standard error of the mean was 0.296. When the sample size increased the standard error decreased.

Also know that the population was strongly skewed to the right. With the smaller sample size, the sampling distribution was also skewed to the right, though not as strongly skewed as the population. With the larger sample size, the sampling distribution was approximately normal.

Published on December 11, 2020 by Pritha Bhandari. Revised on July 6, 2022.

The standard error of the mean, or simply standard error, indicates how different the population mean is likely to be from a sample mean. It tells you how much the sample mean would vary if you were to repeat a study using new samples from within a single population.

The standard error of the mean (SE or SEM) is the most commonly reported type of standard error. But you can also find the standard error for other statistics, like medians or proportions. The standard error is a common measure of sampling error—the difference between a population parameter and a sample statistic.

Why standard error matters

In statistics, data from samples is used to understand larger populations. Standard error matters because it helps you estimate how well your sample data represents the whole population.

With probability sampling, where elements of a sample are randomly selected, you can collect data that is likely to be representative of the population. However, even with probability samples, some sampling error will remain. That’s because a sample will never perfectly match the population it comes from in terms of measures like means and standard deviations.

By calculating standard error, you can estimate how representative your sample is of your population and make valid conclusions.

A high standard error shows that sample means are widely spread around the population mean—your sample may not closely represent your population. A low standard error shows that sample means are closely distributed around the population mean—your sample is representative of your population.

You can decrease standard error by increasing sample size. Using a large, random sample is the best way to minimize sampling bias.

Standard error vs standard deviation

Standard error and standard deviation are both measures of variability:

  • The standard deviation describes variability within a single sample.
  • The standard error estimates the variability across multiple samples of a population.

The standard deviation is a descriptive statistic that can be calculated from sample data. In contrast, the standard error is an inferential statistic that can only be estimated (unless the real population parameter is known).

Example: Standard error vs standard deviation In a random sample of 200 students, the mean math SAT score is 550. In this case, the sample is the 200 students, while the population is all test takers in the region.

The standard deviation of the math scores is 180. This number reflects on average how much each score differs from the sample mean score of 550.

The standard error of the math scores, on the other hand, tells you how much the sample mean score of 550 differs from other sample mean scores, in samples of equal size, in the population of all test takers in the region.

Standard error formula

The standard error of the mean is calculated using the standard deviation and the sample size.

From the formula, you’ll see that the sample size is inversely proportional to the standard error. This means that the larger the sample, the smaller the standard error, because the sample statistic will be closer to approaching the population parameter.

Different formulas are used depending on whether the population standard deviation is known. These formulas work for samples with more than 20 elements (n > 20).

When population parameters are known

When the population standard deviation is known, you can use it in the below formula to calculate standard error precisely.

FormulaExplanation
  • is standard error
  • is population standard deviation
  • is the number of elements in the sample

When population parameters are unknown

When the population standard deviation is unknown, you can use the below formula to only estimate standard error. This formula takes the sample standard deviation as a point estimate for the population standard deviation.

FormulaExplanation
  • is standard error
  • is sample standard deviation
  • is the number of elements in the sample
Example: Using the standard error formulaTo estimate the standard error for math SAT scores, you follow two steps.

First, find the square root of your sample size (n).

FormulaCalculation

Next, divide the sample standard deviation by the number you found in step one.

FormulaCalculation

The standard error of math SAT scores is 12.8.

How should you report the standard error?

You can report the standard error alongside the mean or in a confidence interval to communicate the uncertainty around the mean.

Example: Reporting the mean and standard error The mean math SAT score of a random sample of test takers is 550 ± 12.8 (SE).

The best way to report the standard error is in a confidence interval because readers won’t have to do any additional math to come up with a meaningful interval.

A confidence interval is a range of values where an unknown population parameter is expected to lie most of the time, if you were to repeat your study with new random samples.

With a 95% confidence level, 95% of all sample means will be expected to lie within a confidence interval of ± 1.96 standard errors of the sample mean.

Based on random sampling, the true population parameter is also estimated to lie within this range with 95% confidence.

Example: Constructing a 95% confidence intervalYou construct a 95% confidence interval (CI) to estimate the population mean math SAT score.

For a normally distributed characteristic, like SAT scores, 95% of all sample means fall within roughly 4 standard errors of the sample mean.

Confidence interval formulaLower limitUpper limit

CI = x̄ ± (1.96 × SE)

= sample mean = 550
SE = standard error = 12.8

x̄ − (1.96 × SE)

550 (1.96 × 12.8) = 525

x̄ + (1.96 × SE)

550 + (1.96 × 12.8) = 575

With random sampling, a 95% CI [525 575] tells you that there is a 0.95 probability that the population mean math SAT score is between 525 and 575.

Other standard errors

Aside from the standard error of the mean (and other statistics), there are two other standard errors you might come across: the standard error of the estimate and the standard error of measurement.

The standard error of the estimate is related to regression analysis. This reflects the variability around the estimated regression line and the accuracy of the regression model. Using the standard error of the estimate, you can construct a confidence interval for the true regression coefficient.

The standard error of measurement is about the reliability of a measure. It indicates how variable the measurement error of a test is, and it’s often reported in standardized testing. The standard error of measurement can be used to create a confidence interval for the true score of an element or an individual.

Frequently asked questions about standard error

What is standard error?

The standard error of the mean, or simply standard error, indicates how different the population mean is likely to be from a sample mean. It tells you how much the sample mean would vary if you were to repeat a study using new samples from within a single population.

Cite this Scribbr article

If you want to cite this source, you can copy and paste the citation or click the “Cite this Scribbr article” button to automatically add the citation to our free Citation Generator.

Bhandari, P. (2022, July 06). What Is Standard Error? | How to Calculate (Guide with Examples). Scribbr. Retrieved November 9, 2022, from //www.scribbr.com/statistics/standard-error/

Is this article helpful?

You have already voted. Thanks :-) Your vote is saved :-) Processing your vote...

Does sample size affect mean and standard deviation?

Therefore, as a sample size increases, the sample mean and standard deviation will be closer in value to the population mean μ and standard deviation σ .

How does sample size effect standard error of the mean?

The standard error is also inversely proportional to the sample size; the larger the sample size, the smaller the standard error because the statistic will approach the actual value. The standard error is considered part of inferential statistics. It represents the standard deviation of the mean within a dataset.

What is the relationship between the sample standard deviation and the standard error?

Standard error and standard deviation are both measures of variability. The standard deviation reflects variability within a sample, while the standard error estimates the variability across samples of a population.

Toplist

Neuester Beitrag

Stichworte