In the output, the null and alternative hypotheses help you to verify that you entered the correct value for the hypothesized mean.
The standard deviation is the most common measure of dispersion, or how spread out the data are about the mean. The symbol σ (sigma) is often used to represent the standard deviation of a population, while s is used to represent the standard deviation of a sample. Variation that is random or natural to a process is often referred to as noise.
The standard deviation uses the same units as the data.
Use the standard deviation to determine how spread out the data are from the mean. A higher standard deviation value indicates greater spread in the data. A good rule of thumb for a normal distribution is that approximately 68% of the values fall within one standard deviation of the mean, 95% of the values fall within two standard deviations, and 99.7% of the values fall within three standard deviations.
Minitab displays two different values for the standard deviation. The known standard deviation is the standard deviation of the population that you entered before running the analysis. The standard deviation is used to calculate the confidence intervals and Z-value. StDev is the standard deviation of your sample data. Because you know the standard deviation of the population, Minitab does not use the value from the sample for any of the calculations.
If you do not know the standard deviation of the population, use 1-Sample t.
The sample size (N) is the total number of observations in the sample.
The sample size affects the confidence interval and the power of the test.
Usually, a larger sample size results in a narrower confidence interval. A larger sample size also gives the test more power to detect a difference. For more information, go to What is power?.
The mean summarizes the sample values with a single value that represents the center of the data. The mean is the average of the data, which is the sum of all the observations divided by the number of observations.
The mean of the sample data is an estimate of the population mean.
Because the mean is based on sample data and not on the entire population, it is unlikely that the sample mean equals the population mean. To better estimate the population mean, use the confidence interval.
The standard error of the mean (SE Mean) estimates the variability between sample means that you would obtain if you took repeated samples from the same population. Whereas the standard error of the mean estimates the variability between samples, the standard deviation measures the variability within a single sample.
For example, you have a mean delivery time of 3.80 days, with a standard deviation of 1.43 days, from a random sample of 312 delivery times. These numbers yield a standard error of the mean of 0.08 days (1.43 divided by the square root of 312). If you took multiple random samples of the same size, from the same population, the standard deviation of those different sample means would be around 0.08 days.
Use the standard error of the mean to determine how precisely the sample mean estimates the population mean.
A smaller value of the standard error of the mean indicates a more precise estimate of the population mean. Usually, a larger standard deviation results in a larger standard error of the mean and a less precise estimate of the population mean. A larger sample size results in a smaller standard error of the mean and a more precise estimate of the population mean.
Minitab uses the standard error of the mean to calculate the confidence interval.
The confidence interval provides a range of likely values for the population mean. Because samples are random, two samples from a population are unlikely to yield identical confidence intervals. But, if you repeated your sample many times, a certain percentage of the resulting confidence intervals or bounds would contain the unknown population mean. The percentage of these confidence intervals or bounds that contain the mean is the confidence level of the interval. For example, a 95% confidence level indicates that if you take 100 random samples from the population, you could expect approximately 95 of the samples to produce intervals that contain the population mean.
An upper bound defines a value that the population mean is likely to be less than. A lower bound defines a value that the population mean is likely to be greater than.
The confidence interval helps you assess the practical significance of your results. Use your specialized knowledge to determine whether the confidence interval includes values that have practical significance for your situation. If the interval is too wide to be useful, consider increasing your sample size. For more information, go to Ways to get a more precise confidence interval.
N | Mean | StDev | SE Mean | 95% CI for μ |
---|---|---|---|---|
20 | 16.460 | 2.258 | 0.581 | (15.321, 17.599) |
In these results, the estimate of the population mean for fat percentage is 16.46%. You can be 95% confident that the population mean is between 15.321% and 17.599%.
The Z-value is a test statistic for Z-tests that measures the difference between an observed statistic and its hypothesized population parameter in units of standard error.
You can compare the Z-value to critical values of the standard normal distribution to determine whether to reject the null hypothesis. However, using the p-value of the test to make the same determination is usually more practical and convenient.
To determine whether to reject the null hypothesis, compare the Z-value to your critical value. The critical value is Z_{1-α/2} for a two–sided test and Z_{1-α} for a one–sided test. For a two-sided test, if the absolute value of the Z-value is greater than the critical value, you reject the null hypothesis. If the absolute value of the Z-value is less than the critical value, you fail to reject the null hypothesis. You can calculate the critical value in Minitab or find the critical value from a standard normal table in most statistics books. For more information, go to Using the inverse cumulative distribution function (ICDF) and click "Use the ICDF to calculate critical values".
The p-value is a probability that measures the evidence against the null hypothesis. A smaller p-value provides stronger evidence against the null hypothesis.
Use the p-value to determine whether the population mean is statistically different from the hypothesized mean.
A histogram divides sample values into many intervals and represents the frequency of data values in each interval with a bar.
Use a histogram to assess the shape and spread of the data. Histograms are best when the sample size is greater than 20.
Examine the spread of your data to determine whether your data appear to be skewed. When data are skewed, the majority of the data are located on the high or low side of the graph. Often, skewness is easiest to detect with a histogram or boxplot.
Data that are severely skewed can affect the validity of the p-value if your sample is small (less than 20 values). If your data are severely skewed and you have a small sample, consider increasing your sample size.
Outliers, which are data values that are far away from other data values, can strongly affect the results of your analysis. Often, outliers are easiest to identify on a boxplot.
Try to identify the cause of any outliers. Correct any data–entry errors or measurement errors. Consider removing data values for abnormal, one-time events (also called special causes). Then, repeat the analysis. For more information, go to Identifying outliers.
An individual value plot displays the individual values in the sample. Each circle represents one observation. An individual value plot is especially useful when you have relatively few observations and when you also need to assess the effect of each observation.
Use an individual value plot to examine the spread of the data and to identify any potential outliers. Individual value plots are best when the sample size is less than 50.
Examine the spread of your data to determine whether your data appear to be skewed. When data are skewed, the majority of the data are located on the high or low side of the graph. Often, skewness is easiest to detect with a histogram or boxplot.
Data that are severely skewed can affect the validity of the p-value if your sample is small (less than 20 values). If your data are severely skewed and you have a small sample, consider increasing your sample size.
Outliers, which are data values that are far away from other data values, can strongly affect the results of your analysis. Often, outliers are easiest to identify on a boxplot.
Try to identify the cause of any outliers. Correct any data–entry errors or measurement errors. Consider removing data values for abnormal, one-time events (also called special causes). Then, repeat the analysis. For more information, go to Identifying outliers.
A boxplot provides a graphical summary of the distribution of a sample. The boxplot shows the shape, central tendency, and variability of the data.
Use a boxplot to examine the spread of the data and to identify any potential outliers. Boxplots are best when the sample size is greater than 20.
Examine the spread of your data to determine whether your data appear to be skewed. When data are skewed, the majority of the data are located on the high or low side of the graph. Often, skewness is easiest to detect with a histogram or boxplot.
Data that are severely skewed can affect the validity of the p-value if your sample is small (less than 20 values). If your data are severely skewed and you have a small sample, consider increasing your sample size.
Outliers, which are data values that are far away from other data values, can strongly affect the results of your analysis. Often, outliers are easiest to identify on a boxplot.
Try to identify the cause of any outliers. Correct any data–entry errors or measurement errors. Consider removing data values for abnormal, one-time events (also called special causes). Then, repeat the analysis. For more information, go to Identifying outliers.