Abstract

In a previous article, I discussed how we can practically construct a confidence interval that is likely to include the true population mean from our single sample by taking advantage of the properties of the sampling distribution. In many cases, we can make a skewed distribution symmetrical by transforming the data; this allows us to use the properties of the sampling distribution.The standard error (SE) of the sample mean is the estimated standard deviation of the sampling distribution; because we carry out the study only once, we have only 1 sample. Even though we do not observe the sampling distribution, since this would require multiple samples, we can calculate the standard deviation of the sampling distribution given the standard deviation and size of our 1 sample (σn).Again, using information from this 1 sample, along with knowledge about the sampling distribution of a mean, we can construct a confidence interval. As long as the size of our sample is reasonably large (>30), we know that the sampling distribution of a mean is normal. This allows us to infer a range of values within which the true population mean is likely to lie. This range of values is called a “confidence interval” because we can be reasonably confident that it contains the true mean.I described earlier that the confidence interval of the mean follows the general formula:mean±1.96×SE=14.5±1.96×SEIf we took thousands of samples and for each sample calculated the mean and associated 95% confidence interval, we would expect 95% of these confidence intervals to include the population mean.When we are making statistical comparisons, we state the null (Ho) and the alternative (Ha) hypotheses.The null hypothesis usually states that there is no difference when comparing groups, and we look for evidence to disprove the null hypothesis and accept the alternative. In our previous example of adolescent patients seeking orthodontic therapy, our sample's mean age was 14.5 years, and we hypothesized that our true population mean (μ) was 15.3 years. We are interested in assessing whether our assumption about the population mean holds. In other words, we would like to test whether our sample mean differs from the hypothesized true population mean.To accomplish this, we compare 2 hypotheses: (1) the null hypothesis that the population mean is equal to μ, H0:μ0 = μ; and (2) the alternative hypothesis that the population mean is any value except for μ, H1:μ1≠μ.The statistical process to assess the null hypothesis is called the statistical test. Since the alternative hypothesis is defined as 2-sided, the statistical test is a 2-sided test. The 2-sided test is applied more frequently than the “strict” 1-sided test (eg, H1:μ1>μ).Hypothesis testingThis step allows us to calculate a statistical formula known as the test statistic: a measure of how extreme our data are. The general form of the test statistic compares the observed value estimated from the sample, x¯, and the population mean under the null hypothesis, μ0 = μ. It also takes into account the population variability by using the standard error.We saw that sampling distributions are approximately normal when the sample size is large. The letter “z” is often used in association with normal distribution, so the value of the test statistic is called a “z value.” It is equal toz=x¯−μ0SEbetween the sample mean and the population mean (under the null hypothesis), and it is a measure of how far the observed mean is from the population mean under the null hypothesis, expressed in units of standard errors. Hypothesis tests that make use of a z value are generally called “z tests.” In our example, we are making inferences based on 1 sample (comparing the mean [14.5 years] from our sample of adolescent orthodontic patients with the hypothesized true population mean [15.3 years] of adolescent orthodontic patients), so we call this a 1-sample z test. If the null hypothesis is true, all the many values of z that could be obtained from different samples taken from the same population would be distributed as a standard normal distribution.In our example, the sample mean is normally distributed with a mean of μ = 14.5 and a standard error of SE=SDn=5.7974=0.67.The assumptions to implement a z test are that (1) the population is normally distributed, (2) the population's standard deviation is known, and (3) the sample size is large enough (>30).1Kirkwood B.R. Sterne J.A.C. Essential medical statistics.2nd ed. Blackwell, Oxford, United Kingdom2003: 58-70Google Scholar When assumptions 2 and 3 do not hold, then another test is used, called the “t test.” The implementation of this test dictates the estimation of the sample's standard deviation under the formula S=∑i = 1n(xi−x¯)2/(n−1), where x¯ is the sample mean and n is the sample size. Then, a t value can be calculated as follows:t=x¯−μ0S/nIf the null hypothesis is true, all the many t values that could be obtained from different samples from the same population would be distributed as a t distribution (also known as “Student t distribution”). The t distribution has a bell-like shape centered at zero (like the standard normal distribution), but the bell is smaller and broader for small samples. The shape of the t distribution depends on its degrees of freedom (df), a measure of how much of the sample size is used to determine the t distribution. In the 1-sample case, the degrees of freedom of a t distribution are equal to the sample size minus 1. The degrees of freedom depend on the sample size. As the sample size increases, the degrees of freedom increase; hence, according to the central limit theorem, the t distribution looks more and more like the standard normal distribution.So, when the sample size is small, say less than 30, the formula for a 95% confidence interval for a mean becomes the estimated mean ±t1 − 0.952,n −1×SE, where t1 − 0.952,n −1 is a multiplier known as the upper 0.025 percentile point on the t distribution with n − 1 df. In other words, when the sample size is large and we use the z distribution, the multiplier is always 1.96. However, when we use smaller sample sizes and apply the t distribution, the multiplier varies depending on the sample size and tends to reach 1.96 from above as the sample size increases.The result of the larger multiplier is that the calculated 95% confidence interval from the t distribution is larger than the confidence interval resulting from the z distribution. The wider confidence interval accounts for the extra uncertainty from the small sample. For hypothesis testing, the formula is the same, but the P value must be read from tables of the t distribution. The t distribution has critical values that are larger than the z distribution to account for the extra uncertainty because of the small sample. The t distribution approximates well the z distribution when the sample size is larger than 30.The Figure shows that the multiplier from the t distribution with 20 df is 2.09 compared with 1.96 of the z distribution.In the next article, Part 2, I will discuss how to obtain the P value and complete the statistical testing example. In a previous article, I discussed how we can practically construct a confidence interval that is likely to include the true population mean from our single sample by taking advantage of the properties of the sampling distribution. In many cases, we can make a skewed distribution symmetrical by transforming the data; this allows us to use the properties of the sampling distribution. The standard error (SE) of the sample mean is the estimated standard deviation of the sampling distribution; because we carry out the study only once, we have only 1 sample. Even though we do not observe the sampling distribution, since this would require multiple samples, we can calculate the standard deviation of the sampling distribution given the standard deviation and size of our 1 sample (σn). Again, using information from this 1 sample, along with knowledge about the sampling distribution of a mean, we can construct a confidence interval. As long as the size of our sample is reasonably large (>30), we know that the sampling distribution of a mean is normal. This allows us to infer a range of values within which the true population mean is likely to lie. This range of values is called a “confidence interval” because we can be reasonably confident that it contains the true mean. I described earlier that the confidence interval of the mean follows the general formula:mean±1.96×SE=14.5±1.96×SE If we took thousands of samples and for each sample calculated the mean and associated 95% confidence interval, we would expect 95% of these confidence intervals to include the population mean. When we are making statistical comparisons, we state the null (Ho) and the alternative (Ha) hypotheses. The null hypothesis usually states that there is no difference when comparing groups, and we look for evidence to disprove the null hypothesis and accept the alternative. In our previous example of adolescent patients seeking orthodontic therapy, our sample's mean age was 14.5 years, and we hypothesized that our true population mean (μ) was 15.3 years. We are interested in assessing whether our assumption about the population mean holds. In other words, we would like to test whether our sample mean differs from the hypothesized true population mean. To accomplish this, we compare 2 hypotheses: (1) the null hypothesis that the population mean is equal to μ, H0:μ0 = μ; and (2) the alternative hypothesis that the population mean is any value except for μ, H1:μ1≠μ. The statistical process to assess the null hypothesis is called the statistical test. Since the alternative hypothesis is defined as 2-sided, the statistical test is a 2-sided test. The 2-sided test is applied more frequently than the “strict” 1-sided test (eg, H1:μ1>μ). Hypothesis testingThis step allows us to calculate a statistical formula known as the test statistic: a measure of how extreme our data are. The general form of the test statistic compares the observed value estimated from the sample, x¯, and the population mean under the null hypothesis, μ0 = μ. It also takes into account the population variability by using the standard error.We saw that sampling distributions are approximately normal when the sample size is large. The letter “z” is often used in association with normal distribution, so the value of the test statistic is called a “z value.” It is equal toz=x¯−μ0SEbetween the sample mean and the population mean (under the null hypothesis), and it is a measure of how far the observed mean is from the population mean under the null hypothesis, expressed in units of standard errors. Hypothesis tests that make use of a z value are generally called “z tests.” In our example, we are making inferences based on 1 sample (comparing the mean [14.5 years] from our sample of adolescent orthodontic patients with the hypothesized true population mean [15.3 years] of adolescent orthodontic patients), so we call this a 1-sample z test. If the null hypothesis is true, all the many values of z that could be obtained from different samples taken from the same population would be distributed as a standard normal distribution.In our example, the sample mean is normally distributed with a mean of μ = 14.5 and a standard error of SE=SDn=5.7974=0.67.The assumptions to implement a z test are that (1) the population is normally distributed, (2) the population's standard deviation is known, and (3) the sample size is large enough (>30).1Kirkwood B.R. Sterne J.A.C. Essential medical statistics.2nd ed. Blackwell, Oxford, United Kingdom2003: 58-70Google Scholar When assumptions 2 and 3 do not hold, then another test is used, called the “t test.” The implementation of this test dictates the estimation of the sample's standard deviation under the formula S=∑i = 1n(xi−x¯)2/(n−1), where x¯ is the sample mean and n is the sample size. Then, a t value can be calculated as follows:t=x¯−μ0S/nIf the null hypothesis is true, all the many t values that could be obtained from different samples from the same population would be distributed as a t distribution (also known as “Student t distribution”). The t distribution has a bell-like shape centered at zero (like the standard normal distribution), but the bell is smaller and broader for small samples. The shape of the t distribution depends on its degrees of freedom (df), a measure of how much of the sample size is used to determine the t distribution. In the 1-sample case, the degrees of freedom of a t distribution are equal to the sample size minus 1. The degrees of freedom depend on the sample size. As the sample size increases, the degrees of freedom increase; hence, according to the central limit theorem, the t distribution looks more and more like the standard normal distribution.So, when the sample size is small, say less than 30, the formula for a 95% confidence interval for a mean becomes the estimated mean ±t1 − 0.952,n −1×SE, where t1 − 0.952,n −1 is a multiplier known as the upper 0.025 percentile point on the t distribution with n − 1 df. In other words, when the sample size is large and we use the z distribution, the multiplier is always 1.96. However, when we use smaller sample sizes and apply the t distribution, the multiplier varies depending on the sample size and tends to reach 1.96 from above as the sample size increases.The result of the larger multiplier is that the calculated 95% confidence interval from the t distribution is larger than the confidence interval resulting from the z distribution. The wider confidence interval accounts for the extra uncertainty from the small sample. For hypothesis testing, the formula is the same, but the P value must be read from tables of the t distribution. The t distribution has critical values that are larger than the z distribution to account for the extra uncertainty because of the small sample. The t distribution approximates well the z distribution when the sample size is larger than 30.The Figure shows that the multiplier from the t distribution with 20 df is 2.09 compared with 1.96 of the z distribution.In the next article, Part 2, I will discuss how to obtain the P value and complete the statistical testing example. Hypothesis testingThis step allows us to calculate a statistical formula known as the test statistic: a measure of how extreme our data are. The general form of the test statistic compares the observed value estimated from the sample, x¯, and the population mean under the null hypothesis, μ0 = μ. It also takes into account the population variability by using the standard error.We saw that sampling distributions are approximately normal when the sample size is large. The letter “z” is often used in association with normal distribution, so the value of the test statistic is called a “z value.” It is equal toz=x¯−μ0SEbetween the sample mean and the population mean (under the null hypothesis), and it is a measure of how far the observed mean is from the population mean under the null hypothesis, expressed in units of standard errors. Hypothesis tests that make use of a z value are generally called “z tests.” In our example, we are making inferences based on 1 sample (comparing the mean [14.5 years] from our sample of adolescent orthodontic patients with the hypothesized true population mean [15.3 years] of adolescent orthodontic patients), so we call this a 1-sample z test. If the null hypothesis is true, all the many values of z that could be obtained from different samples taken from the same population would be distributed as a standard normal distribution.In our example, the sample mean is normally distributed with a mean of μ = 14.5 and a standard error of SE=SDn=5.7974=0.67.The assumptions to implement a z test are that (1) the population is normally distributed, (2) the population's standard deviation is known, and (3) the sample size is large enough (>30).1Kirkwood B.R. Sterne J.A.C. Essential medical statistics.2nd ed. Blackwell, Oxford, United Kingdom2003: 58-70Google Scholar When assumptions 2 and 3 do not hold, then another test is used, called the “t test.” The implementation of this test dictates the estimation of the sample's standard deviation under the formula S=∑i = 1n(xi−x¯)2/(n−1), where x¯ is the sample mean and n is the sample size. Then, a t value can be calculated as follows:t=x¯−μ0S/nIf the null hypothesis is true, all the many t values that could be obtained from different samples from the same population would be distributed as a t distribution (also known as “Student t distribution”). The t distribution has a bell-like shape centered at zero (like the standard normal distribution), but the bell is smaller and broader for small samples. The shape of the t distribution depends on its degrees of freedom (df), a measure of how much of the sample size is used to determine the t distribution. In the 1-sample case, the degrees of freedom of a t distribution are equal to the sample size minus 1. The degrees of freedom depend on the sample size. As the sample size increases, the degrees of freedom increase; hence, according to the central limit theorem, the t distribution looks more and more like the standard normal distribution.So, when the sample size is small, say less than 30, the formula for a 95% confidence interval for a mean becomes the estimated mean ±t1 − 0.952,n −1×SE, where t1 − 0.952,n −1 is a multiplier known as the upper 0.025 percentile point on the t distribution with n − 1 df. In other words, when the sample size is large and we use the z distribution, the multiplier is always 1.96. However, when we use smaller sample sizes and apply the t distribution, the multiplier varies depending on the sample size and tends to reach 1.96 from above as the sample size increases.The result of the larger multiplier is that the calculated 95% confidence interval from the t distribution is larger than the confidence interval resulting from the z distribution. The wider confidence interval accounts for the extra uncertainty from the small sample. For hypothesis testing, the formula is the same, but the P value must be read from tables of the t distribution. The t distribution has critical values that are larger than the z distribution to account for the extra uncertainty because of the small sample. The t distribution approximates well the z distribution when the sample size is larger than 30.The Figure shows that the multiplier from the t distribution with 20 df is 2.09 compared with 1.96 of the z distribution.In the next article, Part 2, I will discuss how to obtain the P value and complete the statistical testing example. This step allows us to calculate a statistical formula known as the test statistic: a measure of how extreme our data are. The general form of the test statistic compares the observed value estimated from the sample, x¯, and the population mean under the null hypothesis, μ0 = μ. It also takes into account the population variability by using the standard error. We saw that sampling distributions are approximately normal when the sample size is large. The letter “z” is often used in association with normal distribution, so the value of the test statistic is called a “z value.” It is equal toz=x¯−μ0SEbetween the sample mean and the population mean (under the null hypothesis), and it is a measure of how far the observed mean is from the population mean under the null hypothesis, expressed in units of standard errors. Hypothesis tests that make use of a z value are generally called “z tests.” In our example, we are making inferences based on 1 sample (comparing the mean [14.5 years] from our sample of adolescent orthodontic patients with the hypothesized true population mean [15.3 years] of adolescent orthodontic patients), so we call this a 1-sample z test. If the null hypothesis is true, all the many values of z that could be obtained from different samples taken from the same population would be distributed as a standard normal distribution. In our example, the sample mean is normally distributed with a mean of μ = 14.5 and a standard error of SE=SDn=5.7974=0.67. The assumptions to implement a z test are that (1) the population is normally distributed, (2) the population's standard deviation is known, and (3) the sample size is large enough (>30).1Kirkwood B.R. Sterne J.A.C. Essential medical statistics.2nd ed. Blackwell, Oxford, United Kingdom2003: 58-70Google Scholar When assumptions 2 and 3 do not hold, then another test is used, called the “t test.” The implementation of this test dictates the estimation of the sample's standard deviation under the formula S=∑i = 1n(xi−x¯)2/(n−1), where x¯ is the sample mean and n is the sample size. Then, a t value can be calculated as follows:t=x¯−μ0S/n If the null hypothesis is true, all the many t values that could be obtained from different samples from the same population would be distributed as a t distribution (also known as “Student t distribution”). The t distribution has a bell-like shape centered at zero (like the standard normal distribution), but the bell is smaller and broader for small samples. The shape of the t distribution depends on its degrees of freedom (df), a measure of how much of the sample size is used to determine the t distribution. In the 1-sample case, the degrees of freedom of a t distribution are equal to the sample size minus 1. The degrees of freedom depend on the sample size. As the sample size increases, the degrees of freedom increase; hence, according to the central limit theorem, the t distribution looks more and more like the standard normal distribution. So, when the sample size is small, say less than 30, the formula for a 95% confidence interval for a mean becomes the estimated mean ±t1 − 0.952,n −1×SE, where t1 − 0.952,n −1 is a multiplier known as the upper 0.025 percentile point on the t distribution with n − 1 df. In other words, when the sample size is large and we use the z distribution, the multiplier is always 1.96. However, when we use smaller sample sizes and apply the t distribution, the multiplier varies depending on the sample size and tends to reach 1.96 from above as the sample size increases. The result of the larger multiplier is that the calculated 95% confidence interval from the t distribution is larger than the confidence interval resulting from the z distribution. The wider confidence interval accounts for the extra uncertainty from the small sample. For hypothesis testing, the formula is the same, but the P value must be read from tables of the t distribution. The t distribution has critical values that are larger than the z distribution to account for the extra uncertainty because of the small sample. The t distribution approximates well the z distribution when the sample size is larger than 30. The Figure shows that the multiplier from the t distribution with 20 df is 2.09 compared with 1.96 of the z distribution. In the next article, Part 2, I will discuss how to obtain the P value and complete the statistical testing example.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call