SAS Annotated Output
Proc TTest
The
ttest procedure performs t-tests for one sample, two samples and
paired observations. The single-sample t-test compares the mean of the
sample to a given number (which you supply). The dependent-sample t-test
compares the difference in the means from the two variables to a
given number (usually 0), while taking into account the fact that the scores are
not independent. The independent samples t-test compares the difference in
the means from the two groups to a given value (usually 0). In other
words, it tests whether the difference in the means is 0. In
our examples, we will use the
hsb2
data set.
Single sample t-test
For this example, we will compare the mean of the variable
write with
a pre-selected value of 50. In practice, the value against which the mean
is compared should be based on theoretical considerations and/or previous
research.
proc ttest data="D:\hsb2" H0=50;
var write;
run;
The TTEST Procedure
Statistics
Lower CL Upper CL Lower CL Upper CL
Variable N Mean Mean Mean Std Dev Std Dev Std Dev Std Err
write 200 51.453 52.775 54.097 8.6318 9.4786 10.511 0.6702
T-Tests
Variable DF t Value Pr > |t|
write 199 4.14 <.0001
Summary statistics
Statistics
Lower CL Upper CL Lower CL Upper CL
Variablea Nb Meanc Meand Meanc Std Deve Std Devf Std Deve Std Errg
write 200 51.453 52.775 54.097 8.6318 9.4786 10.511 0.6702
a.
Variable - This is the list of variables. Each variable
that was listed on the
var statement will have its own line in this part
of the output.
b.
N - This is the number of valid (i.e., non-missing)
observations used in calculating the t-test.
c.
Lower CL Mean and
Upper CL Mean - These are the lower
and upper bounds of the confidence interval for the mean. A confidence interval
for the mean specifies a range of values within which the unknown population
parameter, in this case the mean, may lie. It is given by
where s
is the sample deviation of the observations and N is the number of valid
observations. The t-value in the formula can be computed or found in any
statistics book with the degree of freedom being N-1 and the p-value being 1-alpha/2,
where alpha is the confidence level and by default is .95. If we
drew 200 random samples, then about 190 (200*.95) times, the confidence interval
would capture the parameter mean of the population.
d. Mean - This is the
mean of the variable.
e.
Lower CL Std Dev and
Upper CL Std Dev - Those are the
lower and upper bound of the confidence interval for the standard deviation. A
confidence interval for the standard deviation specifies a range of values
within which the unknown parameter, in this case, the standard deviation, may
lie. The computation of the confidence interval is based on a chi-square
distribution and is given by the following formula
where
S2 is the estimated variance of the variable and
alpha is the confidence level. If we drew 200 random samples, then about 190
(200*.95) of times, the confidence interval would capture the parameter standard
deviation of the population.
f.
Std Dev - This is the standard deviation of the variable.
g.
Std Err - This is the estimated standard deviation of the
sample mean. If we drew repeated samples of size 200, we would expect the
standard deviation of the sample means to be close to the standard error.
The standard deviation of the distribution of sample mean is estimated as the
standard deviation of the sample divided by the square root of sample size.
This provides a measure of the variability of the sample mean. The Central
Limit Theorem tells us that the sample means are approximately normally
distributed when the sample size is 30 or greater.
Test statistics
The single sample t-test tests the null hypothesis that the population mean
is equal to the given number specified using the option
H0= . The
default value in SAS for H0 is 0. It calculates the t-statistic and its
p-value for the null hypothesis under the assumption that the sample comes from
an approximately normal distribution. If the p-value associated with the t-test
is small (usually set at p < 0.05), there is evidence that the mean is different
from the hypothesized value. If the p-value associated with the t-test is
not small (p > 0.05), then the null hypothesis is not rejected, and you conclude
that the mean is not different from the hypothesized value.
In our example, the t-value for variable
write is 4.14 with 199
degrees of freedom. The corresponding p-value is .0001, which is less than
0.05. We conclude that the mean of variable
write is different from
50.
T-Tests
Variablea DFh t Valuei Pr > |t|j
write 199 4.14 <.0001
a.
Variable - This is the list of variables. Each variable
that was listed on the
var statement will have its own line in this part
of the output.
If a var
statement is not specified, proc ttest will conduct a t-test on all
numerical variables in the dataset.
h. DF - The
degrees of freedom for the single sample t-test is simply the number of valid
observations minus 1. We loose one degree of
freedom because we have estimated the mean from the sample. We have used
some of the information from the data to estimate the mean; therefore, it is not
available to use for the test and the degrees of freedom accounts for this.
i.
t Value - This is the Student t-statistic. It is the
ratio of the difference between the sample mean and the given number to the
standard error of the mean. Since that the standard error of the mean
measure the variability of the sample mean, the smaller the standard error of
the mean, the more likely that our sample mean is close to the true population
mean. This is illustrated by the following three figures.
All three cases the difference between the population means
are the same. But with large variability of sample means, two populations
overlap a great deal. Therefore, the difference may well come by chance.
On the other hand, with small variability, the difference is more clear.
The smaller the standard error of the mean, the larger the magnitude of the
t-value. Therefore, the smaller the p-value. The t-value takes into
account of this fact.
j.
Pr > |t| - The p-value is the two-tailed
probability computed using t distribution. It is the probability of
observing a
greater absolute value of t under the null hypothesis. For a one-tailed
test, halve this probability. If p-value is less than the pre-specified
alpha level (usually .05 or .01) we
will conclude that mean is statistically significantly different from zero. For example,
the p-value for
write is smaller than 0.05. So we conclude that the
mean for
write is significantly different from 50.
Dependent group t-test
A dependent group t-test is used when the observations are not independent of
one another. In the example below, the same students took both the writing
and the reading test. Hence, you would expect there to be a relationship
between the scores provided by each student. The dependent group t-test
accounts for this. In the example below, the t-value for the difference
between the variables
write and
read is 0.87 with 199
degrees of freedom, and the corresponding p-value is .3868. This is greater
than our pre-specified alpha level, 0.05. We conclude that the difference between the variables
write and
read is not statistically significantly different from 0.
In other words, the means for
write and
read are not statistically
significantly different from one another.
proc ttest data="D:\hsb2";
paired write*read;
run;
The TTEST Procedure
Statistics
Lower CL Upper CL Lower CL Upper CL
Difference N Mean Mean Mean Std Dev Std Dev Std Dev Std Err
write - read 200 -0.694 0.545 1.7841 8.0928 8.8867 9.8546 0.6284
T-Tests
Difference DF t Value Pr > |t|
write - read 199 0.87 0.3868
Summary statistics
The TTEST Procedure
Statistics
Lower CL Upper CL Lower CL Upper CL
Differencea Nb Meanc Meand Meanc Std Deve Std Devf Std Deve Std Errg
write - read 200 -0.694 0.545 1.7841 8.0928 8.8867 9.8546 0.6284
a.
Difference - This is the list of variables.
b.
N - This is the number of valid (i.e., non-missing)
observations used in calculating the t-test.
c.
Lower CL Mean and
Upper CL Mean - These are the lower
and upper bounds of the confidence interval for the mean. A confidence
interval for the mean specifies a range of values within which the unknown
population parameter, in this case the mean, may lie. It is given by
where s
is the sample deviation of the observations and N is the number of valid
observations. The t-value in the formula can be computed or found in any
statistics book with the degree of freedom being N-1 and the p-value being 1-alpha/2,
where alpha is the confidence level and by default is .95. If we
drew 200 random samples, then about 190 (200*.95) times, the confidence interval
would capture the parameter mean of the population.
d. Mean - This is the
mean of the variable.
e.
Lower CL Std Dev and
Upper CL Std Dev - Those are the
lower and upper bound of the confidence interval for the standard deviation. A
confidence interval for the standard deviation specifies a range of values
within which the unknown parameter, in this case, the standard deviation, may
lie. The computation of the confidence interval is based on a chi-square
distribution and is given by the following formula
where
S2 is the estimated variance of the variable and
alpha is the confidence level. If we drew 200 random samples, then about 190
(200*.95) of times, the confidence interval would capture the parameter standard
deviation of the population.
f.
Std Dev - This is the standard deviation of the variable.
g.
Std Err - This is the estimated standard deviation of the
sample mean. If we drew repeated samples of size 200, we would expect the
standard deviation of the sample means to be close to the standard error.
The standard deviation of the distribution of sample mean is estimated as the
standard deviation of the sample divided by the square root of sample size.
This provides a measure of the variability of the sample mean. The Central
Limit Theorem tells us that the sample means are approximately normally
distributed when the sample size is 30 or greater.
Test statistics
T-Tests
Differenceh DFi t Valuej Pr > |t|k
write - read 199 0.87 0.3868
h.
Difference - The t-test for dependent groups is to form a
single random sample of the paired difference. Therefore, essentially it is a
simple random sample test. The interpretation for t-value and p-value is the
same as for the case of simple random sample.
i.
DF - The degrees of freedom for the paired observations is
simply the number of observations minus 1. This is because the test is conducted
on the one sample of the paired differences.
j.
t Value - This is the t-statistic. It is the ratio of
the mean of the difference in means to the standard error of the difference
(.545/.6284).
k.
Pr > |t| - The p-value is the two-tailed probability computed
using t distribution. It is the probability of observing a greater absolute value of
t under the null hypothesis. For a one-tailed test, halve this
probability. If p-value is less than our pre-specified alpha level,
usually 0.05, we will conclude that
the difference is significantly from zero. For example, the p-value for
the difference between write and read is greater than 0.05, so we conclude that
the difference in means is not statistically significantly different from 0.
Independent group t-test
This t-test is designed to compare means of same variable between two groups.
In our example, we compare the mean writing score between the group of
female students and the group of male students. Ideally, these subjects are
randomly selected from a larger population of subjects. Depending on if we
assume that the variances for both populations are the same or not, the standard
error of the mean of the difference between the groups and the degree of freedom
are computed differently. That yields two possible different t-statistic and two
different p-values. When using the t-test for comparing independent groups, we
need to test the hypothesis on equal variance and this is a part of the output
that
proc ttest produces. The interpretation for p-value is the same as
in other type of t-tests.
proc ttest data="D:\hsb2";;
class female;
var write;
run;
The TTEST Procedure
Statistics
Lower CL Upper CL Lower CL Upper CL
Variable female N Mean Mean Mean Std Dev Std Dev Std Dev Std Err
write 0 91 47.975 50.121 52.267 8.9947 10.305 12.066 1.0803
write 1 109 53.447 54.991 56.535 7.1786 8.1337 9.3843 0.7791
write Diff (1-2) -7.442 -4.87 -2.298 8.3622 9.1846 10.188 1.3042
T-Tests
Variable Method Variances DF t Value Pr > |t|
write Pooled Equal 198 -3.73 0.0002
write Satterthwaite Unequal 170 -3.66 0.0003
Equality of Variances
Variable Method Num DF Den DF F Value Pr > F
write Folded F 90 108 1.61 0.0187
Summary statistics
Statistics
Lower CL Upper CL Lower CL Upper CL
Variablea femaleb Nc Meand Meane Meand Std Devf Std Devg Std Devf Std Errh
write 0 91 47.975 50.121 52.267 8.9947 10.305 12.066 1.0803
write 1 109 53.447 54.991 56.535 7.1786 8.1337 9.3843 0.7791
write Diff (1-2) -7.442 -4.87 -2.298 8.3622 9.1846 10.188 1.3042
a.
Variable - This column lists the dependent variable(s).
In our example, the dependent variable is
write.
b.
female -
This column gives
values of the class variable, in our case female. This variable is
necessary for doing the independent group t-test and is specified by class
statement.
c.
N - This is the number of valid (i.e., non-missing)
observations in each group defined by the variable listed on the
class
statement (often called the independent variable).
d.
Lower CL Mean and
Upper CL Mean - These are the lower
and upper confidence limits of the mean. By default, they are 95%
confidence limits.
e.
Mean - This is the mean of the dependent variable for each
level of the independent variable. On the last line the difference between
the means is given.
f.
Lower CL Std Dev and
Upper LC Std Dev - These are the
lower and upper 95% confidence limits for the standard deviation for the
dependent variable for each level of the independent variable.
g.
Std Dev - This is the standard deviation of the dependent
variable for each of the levels of the independent variable. On the last
line the standard deviation for the difference is given.
h.
Std Err - This is the standard error of the mean.
Test statistics
T-Tests
Variablea Methodi Variancesj DFk t Valuel Pr > |t|m
write Pooled Equal 198 -3.73 0.0002
write Satterthwaite Unequal 170 -3.66 0.0003
Equality of Variances
Variablea Methodi Num DFn Den DFn F Valueo Pr > Fp
write Folded F 90 108 1.61 0.0187
a.
Variable - This column lists the dependent variable(s). In
our example, the dependent variable is
write.
i.
Method - This column specifies the method for computing the
standard error of the difference of the means. The method of computing
this value is based on the assumption regarding the
variances of the two groups. If we assume that the two populations have the same variance,
then the first method, called pooled variance estimator, is used. Otherwise, when
the variances are not assumed to be equal, the Satterthwaite's method is used.
j.
Variances - The pooled estimator of variance is a weighted
average of the two sample variances, with more weight given to the larger sample
and is defined to be
s2 =
((n1-1)s1+(n2-1)s2)/(n1+n2-2),
where s1 and s2
are the sample variances and n1 and n2 are the sample sizes for the two groups.
the This is called pooled variance. The standard error of the mean of the
difference is the pooled variance adjusted by the sample sizes. It is defined to
be the square root of the product of pooled variance and (1/n1+1/n2). In our
example, n1=109, n2=91. The pooled variance = 108*8.1337
2+90*10.305
2/198=84.355.
It follows that the standard error of the mean of the difference =
sqrt(84.355*(1/109+1/91))=1.304. This yields our t-statistic to be
-4.87/1.304=-3.734.
Satterthwaite is an alternative to the pooled-variance t test and is used
when the assumption that the two populations have equal variances seems
unreasonable. It provides a t statistic that asymptotically (that is, as the
sample sizes become large) approaches a t distribution, allowing for an
approximate t test to be calculated when the population variances are not equal.
k.
DF - The degrees of freedom for the paired observations is
simply the number of observations minus 2. We use one degree of freedom for
estimating the mean of each group, and because there are two groups, we use two
degrees of freedom.
l.
t Value - This t-test is designed to compare means between
two groups of the same
variable such as in our example, we compare the mean writing
score between the group of female students and the group of male students.
Depending on if we assume that the variances for both
populations are the same or not, the standard error of the mean of the
difference between the groups and the degrees of freedom are computed
differently. That yields two possible different t-statistic and two
different p-values. When using the t-test for comparing independent
groups, you need to look at the variances for the two groups. As long as the two
variances are close (one is not more than two or three times the other), go with
the equal variances test. The interpretation for the p-value is the same as in
other types of t-tests.
m.
Pr > |t| - The p-value is the two-tailed probability
computed using the t distribution. It is the probability of observing a t-value of
equal or greater absolute value under the null hypothesis. For a
one-tailed test, halve this probability. If the p-value is less than our
pre-specified alpha level, usually 0.05, we will conclude that the difference is significantly different from zero.
For example, the p-value for the difference between females and males is less
than 0.05, so we conclude that the difference in means is statistically
significantly different from 0.
n.
Num DF and
Den DF - The F distribution is the ratio of
two estimates of variances. Therefore it has two parameters, the degrees of
freedom of the numerator and the degrees of freedom of the denominator. In SAS
convention, the numerator corresponds to the sample with larger variance and the
denominator corresponds to the sample with smaller variance. In our example,
the male students group (
female=0) has variance of 10.305^2 (the standard
deviation squared) and for the female
students the variance is 8.1337^2. Therefore, the degrees of freedom for the numerator is
91-1=90 and the degrees of freedom for the denominator 109-1=108.
o.
F Value - SAS labels the F statistic not F, but F', for a
specific reason. The test statistic of the two-sample F test is a ratio of
sample variances, F = s
12/s
22 where
it is completely arbitrary which sample is labeled sample 1 and which is
labeled sample 2. SAS's convention is to put the
larger sample variance in
the numerator and the smaller one in the denominator. This is called the
folded F-statistic,
F' = max(s12,s22)/min(s12,s22)
which will always be greater than 1. Consequently, the F
test rejects the null hypothesis only for large values of F'. In this case, we get 10.305^2 /
8.1337^2 = 1.605165, which SAS rounds to 1.61.
p.
Pr > F -
This is the
two-tailed significance probability. In our example, the probability is less
than 0.05. So there is evidence that the variances for the two groups, female
students and male students, are different. Therefore, we may want to use the
second method (Satterthwaite variance estimator) for our t-test.