Solutions Manual for Fundamental Statistics for the ...
Student Manual for Fundamental Statistics for the Behavioral Sciences (7th edition)
David C. Howell
The University of Vermont
Contents
Chapter 1 Introduction
Chapter 2 Basic Concepts
Chapter 3 Displaying Data
Chapter 4 Measures of Central Tendency
Chapter 5 Measures of Variability
Chapter 6 The Normal Distribution
Chapter 7 Basic Concepts of Probability
Chapter 8 Sampling Distributions and Hypothesis Testing
Chapter 9 Correlation
Chapter 10 Regression
Chapter 11 Multiple Regression
Chapter 12 Hypothesis Tests Applied to Means: One Sample
Chapter 13 Hypothesis Tests Applied to Means: Two Related Samples
Chapter 14 Hypothesis Tests Applied to Means: Two Independent Samples
Chapter 15 Power
Chapter 16 One-way Analysis of Variance
Chapter 17 Factorial Analysis of Variance
Chapter 18 Repeated-Measures Analysis of Variance
Chapter 19 Chi-Square
Chapter 20 Nonparametric and Distribution-Free Statistical Tests
Chapter 21 Choosing the Appropriate Analysis
Preface
The purpose of this manual is to provide answers to students using the accompanying text, Fundamental Statistics for the Behavioral Sciences, 7th ed. I have provided complete answers to all of the odd-numbered questions. I am often asked for answers to even-numbered exercises as well. I do not provide those because many instructors want to have exercises without answers. I am attempting to balance the two competing needs.
You may find on occasion that you do not have the same answer that I do. Much of this will depend on the degree to which you or I round off intermediate steps. Sometimes it will make a surprising difference. If your answer looks close to mine, and you did it the same way that I did, then don’t worry about small differences. It is even possible that I made an error.
I know that there will be errors in some of these answers. There always are. Even the most compulsive problem solver is bound to make errors, and it has been a long time since anyone accused me of being compulsive. I do try, honest I do, but something always slips past—sometimes they even slip past while I am correcting another error. So I maintain a page on the web listing the errors that I and other have found. If you find an error (minor and obvious typos don’t count unless they involve numbers), please check there and let me know if it is a new one. Some classes even compete to see who can find the most errors—it’s rough when you have to compete with a whole class.
The address for the main web page, is
, and the link to the Errata is there.
Important note: Due to the way hypertext links are shown by Microsoft Word, the underlining often obscures a single underline character, as in “More_Stuff.” If you see a space in an address, it is often really a “_.”
Chapter 1-Introduction
1.1 A good example is the development of tolerance to caffeine. People who do not normally drink caffeinated coffee are often startled by the effect of one or two cups of regular coffee, whereas those who normally drink regular coffee see no such effect. To test for a context effect of caffeine, you would first need to develop a dependent variable measuring the alerting effect of caffeine, which could be a vigilance task. You could test for a context effect by serving a group of users of decaffeinated coffee two cups of regular coffee every morning in their office for a month, but have them drink decaf the rest of the time. The vigilance test would be given shortly after the coffee, and tolerance would be seen by an increase in errors over days. At the end of the month, they would be tested after drinking caffeinated coffee in the same and in a different setting.
The important points here are:
1. Tolerance is shown by an increase in errors on the vigilance task.
1. To see the effect of context, subjects need to be presented with caffeine in two different contexts.
2. There needs to be a difference between the vigilance performance in the two contexts.
1.3 Contexts affects people’s response to alcohol, to off-color jokes, or to observed aggressive behavior.
1.5 The sample would be the addicts that we observe.
1.7 Not all people in the city are listed in the phone book. In particular, women and children are underrepresented. A phone book is particularly out of date as a random selection device with the increase in the use of cell phones.
Many telephone surveys really miss the general population, and instead focus on a restricted population, dominated by male adults.
1.9 In the tolerance study discussed in the text, we really do not care what the mean length of paw-lick latency is. No one would be excited to know that a mouse can stand on a surface at 105 degrees for 3.2 seconds without licking its paws. But we do very much care that the population mean of paw-lick latencies for morphine-tolerant mice is longer in one context than in another.
1.11 I would expect that your mother would continue to wander around in a daze, wondering what happened.
1.13 Three examples of measurement data: performance on a vigilance task; typing speed, blood alcohol level.
1.15 Relationship: The relationship between stress and susceptibility to disease; the relationship between driving speed and accident rate.
1.17 You could have one group of mice trained and tested in the same condition, one group trained in one condition and tested in the other, and a group given a placebo in the training context but given morphine in the testing condition.
1.19 This is an Internet search exercise without a fixed answer. The Statistics Homepage is an online statistics text. Various departments offer data sets, computing advice, and clarifying examples.
Chapter 2-Basic Concepts
2.1 Nominal: names of students in the class; Ordinal: the order in which students hand in their first exam; Interval: the student’s grade on that first exam; Ratio: the amount of time that the student spent studying for that exam.
2.3 If the rat lies down to sleep in the maze, after performing successfully for several trials, this probably says little about what the animal has learned in the task. It may say more about the animals level of motivation.
In this exercise I am trying to get the students to see that there is often quite a difference between what you and I think our variable is measuring and what it actually measures. Just because we label something as a measure of learning does not make it so. Just because the numbers increase on a ratio scale (twice as much time in the maze) doesn’t mean that what those numbers are actually measuring is ratio (twice as much learning).
2.5 We have to assume the following at the very least (and I am sure I left out some)
1. Mice are adequate models for human behavior.
1. Morphine tolerance effects in mice are like heroin tolerance effects in humans,
2. Time on a warm surface is in some way analogous to a human response to heroin.
3. A context shift for mice is analogous to a context shift for humans.
4. A drug overdose is analogous to pain tolerance.
2.7 The independent variables are the sex of the subject and the sex of the other person.
2.9 The experimenter expected to find that women would eat less in the presence of a male partner than in the presence of a female partner. Men, on the other hand, were not expected to vary the amount that they ate as a function of sex of their partner.
2.11 We would treat a discrete variable as if it were continuous if it had many different levels and were at least ordinal.
2.13 When I drew 50 numbers 3 times I obtained 29, 26, and 19 even numbers, respectively. For my third drawing only 38 percent of my numbers were even, which is probably less than I might have expected—especially if I didn’t have a fair amount of experience with similar exercises.
2.15 Eyes level condition:
a) X3 = 2.03; X5 = 1.05; X8 = 1.86
b) ∑X = 14.82
c) [pic]
2.17 Eyes level condition:
a) (∑X)2 = 14.822 = 219.6324; ∑X2 = 1.652 + ... + 1.732 = 23.22
b) ∑X/N = 14.82/10 = 1.482
c) This is the mean, a type of average.
The above answers are the variance and standard deviation of Y. You really aren’t going to do much more calculation that this.
2.19 Putting the two sets of data together:
a) Multiply pairwise
b) ∑XY = 22.27496
c) ∑X∑Y = 14.82*14.63 = 216.82
d) ΣXY ≠ ΣXΣY. They do differ, as you would expect.
e) [pic]
2.21 X 5 7 3 6 3 ∑X = 24
X + 4 9 11 7 10 7 ∑(X + 4) = 44 = (24 + 5*4)
2.23 In the text I spoke about room temperature as an ordinal scale of comfort (at least up to some point). Room temperature is a continuous measure, even though with respect to comfort it only measures at an ordinal level.
2.25 The Beth Perez story:
a) The dependent variable is the weekly allowance, measured in dollars and cents, and the independent variable is the sex of the child.
b) We are dealing with a selected sample—the children in her class.
c) The age of the students would influence the overall mean. The fact that these children are classmates could easily lead to socially appropriate responses—or what the children deem to be socially appropriate in their setting.
d) At least within her school, Beth could randomly sample by taking a student roster, assigning each student a number, and matching those up with numbers drawn from a random number table. Random assignment to Sex would obviously be impossible.
e) I don’t see negative aspects of the lack of random assignment here because that is the nature of the variable under consideration. It would be better if we could randomly assign a child to a sex and see the result, but we clearly can’t.
f) The outcome of the study could be influenced by the desire of some children to exaggerate their allowance, or to minimize it so as not to appear too different from their peers. I would suspect that boys would be likely to exaggerate.
g) The descriptive features of the study are her statements that the boys in her class received $3.18 per week in allowance, on average, while the girls received an average of $2.63. The inferential aspects are the inferences to the population of all children, concluding that “boys” get more than “girls.”
2.27 I would record the sequence number of each song that is played and then plot them on a graph. I can’t tell if they are truly random, but if I see a pattern to the points I can be quite sure that they are not random.
I think that it is important for students to become involved with the Internet early on. There is so much material out there that will be helpful, and you have to start finding it now. I find it impossible to believe that my explanations of concepts are always the best explanations that could be given and that they serve each student equally well. If one explanation doesn’t make sense, you can find others that may.
Chapter 3-Displaying Data
3.1 Katz et al (1990) No Passage Group:
[pic]
There is too little data to say very much about the shape of this distribution, but it certainly isn’t looking normally distributed.
3.3 I would use stems of 3*, 3., 4*, 4. 5*, and 5. for this display.
3.5 Compared to those who read the passages:
a) Almost everyone who read the passages did better than the best person who did not read them. Certainly knowing what you are talking about is a good thing (though not always practiced).
4 | 3* |
b) 68966 | 3. |
44343 | 4* |
6669697 | 4. |
42102 | 5* |
57557 | 5. | 5669
| 6* |
| 6. | 66
| 7* | 21232231
| 7. | 5
| HI | 91 93
Notice that I have entered the data in the order in which I encountered
them, rather than in increasing order. It makes it easier.
c) It is obvious that the two groups are very different in their performance. We would be worried if they weren’t.
d) This is an Internet exercise with no fixed answer. That source is far more advanced than the students would be at this time, but I think that they should be able to read it if they just skip over what they don’t understand.
3.7 The following is a plot (as a histogram) of reaction times collapsed across all variables.
[pic]
3.9 Histogram of GPA scores
[pic]
3.11 (1) Mexico has very many young people and very few old people, while Spain has a more even distribution. (2) The difference between males and females is more pronounced at most ages in Spain than it is in Mexico. (3) You can see the high infant mortality rate in Mexico.
3.13 The distribution of those whose attendance is poor is far more spread out than the distribution of normal attendees. This would be expected because a few very good students can score well on tests even when they don’t attend, but most of the poor attenders are generally poor students who would score badly no matter what. The difference between the average grades of these two groups is obvious.
3.15 As the degree of rotation increases, the distribution of reaction time scores appears to move from left to right—which is also an increase.
I think it is a good idea to really think through this problem, rather than to just take the answer as given. It is important to see that looking at data can lead to conclusions to scientific questions, even without formal statistical tests. Many students have a hard time seeing the relationship between data and a question they would like to ask. (Probably many older adults do as well.)
3.17 The data points are probably not independent in that data set. As time went on, there would be changes in the subject’s performance. At first he might get better with practice, but then fatigue would start to set in. Since the data are given in the order in which they were collected, at least within each condition, data nearer in time should be more similar than data farther apart in time.
3.19 The amount of shock that a subject delivers to a white participant does not vary as a function of whether or not that subject has been insulted by the experimenter. However, the black participants do suffer more shocks when the subject has been insulted.
3.21 Wikipedia gives an excellent set of data on HIV/AIDS prevalence at
3.23 There is a tremendous increase in Down’s Syndrome in children born to older mothers. This increase doesn’t really take off until mothers are in their 40s, but with parents delaying having children, this is a potential problem.
[pic]
3.25 Smoking and low birthweight:
The data are given as the percentage of births for each group that were less than 2500 grams.
[pic]
The relationship is unlikely to be a fluke because it is so consistent year after year. You can see that within each group there is very little variability.
Students often wonder why behavioral scientists care about what appears to be a public health problem. But public health problems are very often behavioral problems. Psychologists spend a great deal of time dealing with the behavioral consequences of low birthweight, and trying to find ways of lowering the rate, and with addictions such as smoking.
3.27 White females have a longer life expectancy than black females, but the difference has shrunk considerably since 1920, though recent changes have been modest.
Chapter 4-Measures of Central Tendency
4.1 Mode = 72; Median = 72, Mean = 70.18
4.3 Even without reading the passage, students are still getting about twice as many items correct as they would by chance. This suggests that the test, while testing reading comprehension, is also testing something else. I am not surprised at these results because most students can guess at better than chance levels.
I think it is worth pointing out that these data suggest that the test measures something other than reading comprehension. Most students just say “they were able to guess intelligently,” without realizing that this means that the test is somehow measuring guessing ability. This will become more obvious when we talk about correlation in Chapter 9. Any positively skewed distribution will have a mean greater than the median.
4.5 The mean falls above the median.
4.7 Rats running a straight alley maze:
[pic]
4.9 Multiplying by a constant (5):
Original data 8 3 5 5 6 2 Mean = 4.833, Mode = 5, Median = 5
Revised data 40 15 25 25 30 10 Mean = 24.17 = 5×4.833, Mode = 25,
Median = 25
4.11 Measures of central tendency for ADDSC and GPA:
ADDSC
Mode = 50
Median = 50
Mean = 4629/88 = 52.6
GPA
Mode = 3.00
Median = 2.635
Mean = 216.15/88 =2.46
4.13 The means are very nearly the same for the two conditions.
[pic]
4.15 The only measure that is acceptable for nominal data is the mode, because the mode is the only one that does not depend on the relationships among the points on the scale.
4.17 Class attendance:
Regular Attendees Mean = 276.42; Median = 276
Poor Attendees Mean = 248.33; Median = 256
The two groups were 20 points apart in terms of the medians, and about 25 points apart in terms of means. Clearly, those students who come to class do better.
Because this is not a true experiment (we don’t assign subjects to groups at random), we don’t know exactly what it means. I would like to think that students did poorly because they didn’t hear my brilliant presentations, but it could also be that poorer students in general are less likely to come to class. This is an issue of confounding, and it is a good example making the preference for random assignment apparent in a situation with which most students can identify.
4.19 This is an Internet activity in which there is no fixed answer.
4.21 a) mean = 46.57; 10% trimmed mean = 46.67.
b) mean = 28.4; 10% trimmed mean = 25.0
c) Trimming was more effective in the second example because the second distribution was quite positively skewed.
4.23 The Male Optimists had a mean of 1.016, while the Male Pessimists had a mean of 0.945. This difference is very reliable.
Chapter 5-Measures of Variability
5.1 Variability of NoPassage group:
Range = 57 – 34 = 23
St. Dev. = 6.83
Variance = 46.62
5.3 The variability of the NoPassage group is much smaller than the variability of the Passage group. If this difference turns out to be reliable, it could possibly be explained by the fact that the questions for the Passage group are asking for more than guessing and test-taking skills, and there may be greater variability due to variability in knowledge. On the other hand, it is not uncommon to find one standard deviation equal to two to three times another in small samples.
5.5 Percentages within two standard deviations in Exercise 5.2
s = 10.61
[pic] + 2(10.61) = 70.18 + 21.22 = 48.96 — 91.4
16 scores (or 94%) lie within 2 standard deviations of the mean
5.7 Multiplying or dividing by a constant:
Original 2 3 4 4 5 5 9 [pic] 1= 4.57 s1 = 2.23
X * 2 4 6 8 8 10 10 18 [pic] 2= 9.14 s2 = 4.45
X / 2 1 1.5 2 2 2.5 2.5 4.5 [pic] 3 = 2.29 s3 = 1.11
5.9 Convert revised data to mean = 0
Since adding or subtracting a constant will not change the standard deviation, but will change the mean, I can subtract 3.27 from every score for X2 in Exercise 5.8, making the mean = 0, and keeping s2 = 1.0. the new values are
X3 -0.889 0.539 -1.842 0.539 -0.413 1.016 1.016 [pic]1= 0 s1 = 1
5.11 Boxplot for Exercise 5.1:
Median location = (N + 1)/2 = 29/2 = 14.5
Median = 46
Hinge location = (median location +1)/2 = 15/2 = 7.5
Hinge = 43 and 52
H-spread = 52 – 43 = 9
Inner fences = hinges + 1.5*H-spread = hinges + 1.5*9 = hinges + 13.5 = 29.5 and 65.5
Adjacent values = 34 and 57
30 35 40 45 50 55 60
5.13 Boxplot for ADDSC:
Median location = (N + 1)/2 = 89/2 = 44.5
Median = 50
Hinge location = (median location +1)/2 = 45/2 = 22.5
Hinge = 44.5 and 60.5
H-spread = 60.5 – 44.5 = 16
Inner fences = hinges + 1.5*H-spread = hinges + 1.5*16 = hinges + 24 = 20.5 and 85.5
Adjacent values = 26 and 78
30 35 40 45 50 55 60 70 80 90 100
5.15 Variance when you add a score equal to the mean.
[pic]
Note that the new variance is (1-1/N) times the old variance.
The point that I was trying to make here is that adding scores that don’t deviate from the mean actually decrease the variance because they decrease the average deviation from the mean.
5.17 Angle of rotation:
[pic]
5.19 The following is a cut-and-paste from the JMP help screen. (I don’t expect students to make all of these distinctions from what they are given, because many of the lines overlap.)
[pic]
5.21 Treatment of anorexia:
I would hypothesize that the two treatment groups would show more of a weight gain than the control group, but I have no reason to predict which treatment group would do better. I would assume that the variability would be about the same within each group.
Complete (Before and After) data for the three groups—from which difference scores were derived:
| |Cognitive |Family |Control |
| |Behavioral |Therapy | |
|Mean | 3.01 | 7.26 | -.45 |
|Median |1.40 |9.00 |-.35 |
|St. Dev. |7.31 |7.16 |7.99 |
[pic] [pic]
[pic]
If we look at the weight gain or loss, it would appear that the Control group remained stable, but the two treatment groups gained weight. The gain is greater for the Family Therapy group.
5.23 The descriptive statistics from SPSS are given below. The variable labels should be clear.
[pic]
Notice that the Winsorized variance is considerably greater than the trimmed variance, as it should be. However, it is lower than the variance of the original data, reflecting the fact that the extreme values have been replaced. Cognitive behavior scores were positively skewed, with several quite high values and one or two low values. Trimming and Winsorizing reduced the influence of those values. This causes the Winsorized variance to be considerably smaller than the original variance. The trimmed mean is considerably smaller than the original mean, but the Winsorized mean is only slightly smaller.
Chapter 6-The Normal Distribution
6.1 Distribution of original values:
[pic]
For the first distribution the abscissa would take on the values of:
1 2 3 4 5 6 7
For the second distribution the values would be:
-3 -2 -1 0 1 2 3
For the third distribution the values would be:
-1.90 -1.27 -0.63 0 0.63 1.27 1.90
In these calculations I used the parameters as given, rather than the statistics calculated on the sample.
6.3 Psychology 1 exam grades:
[pic]
a) The percentage between 165 and 225 is the percentage between z = -1.0 and z = 1.0. This is twice the area between z = 0 and z = 1 = 2×0.3413 = .6826.
b) The percentage below 195 is just the percentage below z = 0 = .500.
c) The percentage below z = 1 is the percentage in the larger portion = .8413.
6.5 Guessing on the Psychology 1 exam:
a) We know the mean and standard deviation if the students guess; they are 75 and 7.5, respectively. We also know that a z score of 1.28 cuts off the upper 10%. We simply need to convert z = 1.28 to a raw score.
[pic]
b) For the top 25% of the students the logic is the same except that z = 0.675.
[pic]
c) For the bottom 5% the cutoff will be z = -1.645.
[pic]
d) I would conclude that students were not just guessing, and could make use of test-taking skills that they had acquired over the years.
There is a difference between Exercises 6.3 and 6.4 on the one hand, and 6.5 on the other. In the first two we are talking about performance on the test if students take it normally. There the mean is 195. In Exercise 6.5 we are talking about performance if the students just guessed purely at random without seeing the questions, but only the answers. Here the mean is 75, with a standard deviation of 7.5. These parameters are given by the binomial distribution with N = 300, p = .25, and q = .75, though the students would certainly not be expected to know this.
6.7 Reading scores for fourth and ninth grade children:
a)
[pic]
b) To do better than the average 9th grade student, the 4th grader would have to have a score of 30 or higher.
[pic]
The probability that a fourth grader would exceed a score of 30 is the probability of a z greater than 1.00 = .1587.
c) The probability of a 9th grader doing worse than the average 4th grader is the probability of a 9th grader getting a score below 25, which is the probability of being more than half a standard deviation below the mean, which is .3085.
6.9 Diagnostically meaningful cutoffs:
[pic]
A T score of 62.8 is the score that cuts off the top 10% of the distribution, and is therefore a diagnostically meaningful cutoff.
6.11 Seat belt study:
[pic]
b) We need the probability of getting a 62 if the mean is 44 with a standard deviation of 7.
[pic]
The probability of z > 2.57 = .0051. This is such a small probability that we will probably conclude that the student just made up the data, rather than collecting them honestly.
6.13 Distribution of correct responses
a) Distribution
b) The easiest way to find the cutoff for the lowest 10% is to simply take the sample data and count them, empirically finding the point with 10% of the scores below it.
6.15 Reaction time data:
[pic]
For a normal distribution we would expect 75% of the scores to be equal to or less than 2.06 seconds. In our data the 75th percentile is 1.88 seconds.
6.17 Identifying the highest 2% of Behavior Problem scores:
The upper 2% is cut off by z = 2.05
[pic]
The critical cutoff is a score of 70.5.
6.19 The statisticians were upset because, by defining “overweight” as weighing more than 95% of peers (i.e. above the 95th percentile), the article seemed to be suggesting that there were 22% of children in the top 5%. Moreover, the article says that in 1986 only 8% of children were in the top 15%. That is just silly—it is analogous to “all of the children are above average.” I assume that they meant to say that 22% (etc.) were above what the 95th percentile was some years ago, but that is a different thing. Even if that is the case, the results still look too extreme to be likely.
6.21 Histogram of combined data on emotional stability
[pic]
Notice that we have combined two normal distributions with the same mean, but the resulting distribution is not normal, as can be seen by comparing it to the superimposed normal curve. If the means were very different the distribution would become bimodal.
Chapter 7-Basic Concepts of Probability
7.1 Views of probability:
a) Analytic—If two tennis players are exactly equally skillful so that the outcome of their match is random, the probability is .50 that Player A will win the upcoming match.
b) Relative Frequency—If in past matches Player A has beaten Player B on 13 of the 17 occasions they have played, then, unless something has changed, Player A has a probability of 13/17 = .76 of winning their upcoming match.
c) Subjective—Player A’s coach feels that she has a probability of .90 of winning her upcoming match with Player B.
7.3 More raffle tickets:
a) The probability winning second prize given that you did not win first is 1/999 = .001.
b) The probability that mom comes in first and you are second = 1/1000 * 1/999 = .000001.
c) The probability of you first and mom second = 1/1000 * 1/999 = .000001
d) The probability that the two of you will take the top two prizes is .000001 + .000001 = .000002.
7.5 Part a) of Exercise 7.3 dealt with conditional probabilities.
7.7 What is the probability that you will feel better about your life given that you seek psychological counseling? The research hypothesis is that those who seek help when they need it feel better about life than those who refuse to seek help.
7.9 The mother and child are both sleeping for 11 hours, so the probabilities must be based on the remaining 13 hours.
p(mom looking) = 2/13 = .154; p(baby looking) = 3/13 = .231; p(both looking) = .154*.231 = .036.
7.11 We would expect 3.33 percent of the fliers to end up in the trash if the message and the behavior were independent. In fact, Geller et al. found 4.5 percent of those fliers in the trash. This may look like a very small difference, but given the number of fliers that were handed out, it is a reliable one. It would appear that having a message on a flier increases its probability of being disposed of properly.
7.13 A continuous variable that is routinely treated as if it were discrete is children’s learning abilities, where placement in classes often assumes that the child falls within one category or another.
7.15 If we assume that we know nothing about the applicant, the probability of their being admitted is the probability that they fall above the 80th percentile (which equals .20) times the probability that they will be admitted if they do, which is 10/100 = .10. The probability is .20*.10 = .02. Alternatively, we know that 10 out of 500 are admitted, so we could take the probability as being 10/500 = .02, which is the same thing.
7.17 ADDSC N = 88 [pic] = 52.6 s = 12.42 [calculated from data set]
[pic]
The probability associated with z = -.21 is .5832.
7.19 Dropouts with ADDSC > 60:
p(dropout|ADDSC > 60) = 7/25 = .28
7.21 Conditional and unconditional probability of dropping out:
p(dropout) = 10/88 = .11
p(dropout|ADDSC > 60) = .28
Students are much more likely to drop out of school if they scored at or above ADDSC = 60 in elementary school.
7.23 If there is no discrimination in housing, then a person’s race and whether or not they are offered a particular unit of housing are independent events. We could calculate the probability that a particular unit (or a unit in a particular section of the city) will be offered to anyone in a specific income group. We can also calculate the probability that the customer is a member of an ethnic minority. We can then calculate the probability of that person being shown the unit assuming independence and compare that answer against the actual proportion of times a member of an ethnic minority was offered such a unit.
7.25 The data again would appear to show that the U. S. Attorneys are more likely to request the death penalty when the victim was White than when the victim was Non-white. (This finding is statistically significant, though we won’t address that question until Chapter 19.)
7.27 In this situation we begin with the hypothesis that African Americans are fairly represented in the population. If so, we would expect 0.43% of the pool of 2124 people from which juries are drawn are African American. That comes out to be an expectation of 9.13 people. But the pool actually only had 4 African Americans. We would not expect exactly 9 people—we might have 7 or 8. But 4 sounds awfully small That is such an unlikely event if the pool is fair that we would probably conclude that the pool is not a fair representation of the population of Vermont. An important point here is that this is a conditional probability. If the pool is fair the probability of this event is only .05—an unlikely result.
Chapter 8-Hypothesis Testing
8.1 Last night’s hockey game:
a) Null hypothesis: The game was actually an NHL hockey game.
b) On the basis of that null hypothesis I expected that each team would earn somewhere between 0 and 6 points. I then looked at the actual points and concluded that they were way out of line with what I would expect if this were an NHL hockey game. I therefore rejected the null hypothesis. Notice that I haven’t drawn a conclusion about what type of game it actually was, because that is not what I set out to test.
8.3 A Type I error would be concluding that I was shortchanged when in fact I was not.
8.5 The rejection region is the set of outcomes for which we would reject the null hypothesis. The critical value would be the minimum amount of change below which I would reject the null. It is the border of the rejection region.
8.7 For the Mode test I would draw a very large number of samples and calculate the mode, range, and their ratio (M). I would then plot the resulting values of M.
8.9 Guessing the height of the chapel.
a) The null hypothesis is that the average of two guesses is as accurate as one guess. The alternative hypothesis is that the average guess is more accurate than the single guess.
b) A Type I error would be to reject the null hypothesis when the two kinds of guesses are equally accurate. A Type II error would be failing to reject the null hypothesis when the average guess is better than the single guess.
c) I would be tempted to use a one-tailed test simply because it is hard to image that the average guess would be less accurate, on average, than the single guess.
8.11 A sampling distribution is just a special case of a general distribution in which the thing that we are plotting is a statistic which is the result of repeated sampling.
8.13 Magen et al (2008) study
a) The null hypothesis is that the phrasing of the question will not effect the outcome—the means of the two groups are equal in the population. The alternative hypothesis is that the mean will depend on which condition the person is in.
b) I would compare the two group means.
c) If the difference is significant I would conclude that the phrasing of the choice makes a real difference in the outcome.
8.15 Rerunning Exercise 8.14 for ( = .01:
We first have to find the cutoff for ( = .01 under a normal distribution. The critical value of z = 2.33 (one-tailed), which corresponds to a raw score of 42.69 (from a population with μ = 59 and σ = 7).
We then find where 42.69 lies relative to the distribution under H1:
[pic]
From the appendix we find that .85.08% of the scores fall above this cutoff. Therefore ( = .851.
8.17 To determine whether there is a true relationship between grades and course evaluations I would find a statistic that reflected the degree of relationship between two variables. (The students will see such a statistic (r) in the next chapter.) I would then calculate the sampling distribution of that statistic in a situation in which there is no relationship between two variables. Finally, I would calculate the statistic for a representative set of students and classes and compare my sample value with the sampling distribution of that statistic.
8.19 Allowances for fourth-grade students:
a) The null hypothesis in this case would be the hypothesis that boys and girls receive the same allowance on average.
b) I would use a two-tailed test because I want to reject the null whenever there is a difference in favor of one gender over the other.
c) I would reject the null whenever the obtained difference between the average allowances were greater than I would be lead to expect if they were paid the same in the population.
d) I would increase the sample size and get something other than a self-report of allowances.
8.21 Hypothesis testing and the judicial system
The judicial system operates in ways similar to our standard logic of hypothesis testing. However, in a court we are particularly concerned with the danger of convicting an innocent person. In a trial the null hypothesis is equivalent to the assumption that the accused person is innocent. We set a very small probability of a Type I error, which is far smaller than we normally do in an experiment. Presumably the jury tries to set that probability as close to 0 as they reasonably can. By setting the probability of a Type I error so low, they knowingly allow the probability of a Type II error (releasing a guilty person) to rise, because that is thought to be the lesser evil.
Chapter 9-Correlation
9.1 Low birthweight statistics:
[pic]
The two outliers would appear to have a distorting effect on the correlation coefficient. However, if you replot the data without those points the relationship is still apparent and the correlation only drops to -.54.
9.3 With 24 degrees of freedom, and two-tailed test at α = .05 would require r > ± .388.
9.5 We can conclude that infant mortality is closely tied to both income and the availability of contraception. Infants born to people living in poverty are much more likely to die before their first birthday, and the availability of contraception significantly reduces the number of infants put at risk in the first place.
9.7 Because both income and contraception are related to mortality, we might expect that using them together would lead to a substantial increase in predictability. But note that they are correlated with each other, and therefore share some of the same variance.
9.9 Psychologists have a professional interest in infant mortality because some of the variables that contribute to infant mortality are behavioral ones, and we care about understanding, and often controlling, behavior. Psychologist have an important role to play in world health that has little to do with pills and irrigation systems.
There is a great deal of data available on these issues, and you can easily find it on the Internet. If you are interested in this question, you might also be interested in searching for similar literature on HIV/AIDS.
This question was partly intended to make students think about the fact that all sorts of things are of interest to psychologists. We don’t just run animals in a maze or inquire into people’s dirty minds. In addition, low birthweight is a risk factor for all sorts of infant outcomes.
9.11 The relationship is extremely curvilinear, even though the linear correlation is quite high. You can see that the best fitting line misses almost all of the data points at each end of the distribution.
9.13 The relationship between test scores in Katz’ study and SAT scores for application purposes is a relevant question because we would not be satisfied with a set of data that used SAT questions and yet gave answers that were not in line with SAT performance. We want to know that the tests are measuring at least roughly the same thing. In addition, by knowing the correlation between SATs and performance without seeing the questions, we get a better understanding of some of what the SAT is measuring.
9.15 Correlation for the data in Exercise 9.14:
SAT: mean = 598.57 ∑X = 16760 St. Dev. = 61.57
Test: mean = 46.21 ∑Y = 1294 St. Dev. = 6.73
[pic]
With 26 df we would need a correlation of .374 to be significant. Since our value exceeds that, we can conclude that the relationship between test scores and the SAT is reliably different from 0.
9.17 When we say that two correlations are not significantly different, we mean that they are sufficiently close that they could both have come from samples from populations with exactly the same population correlation coefficient.
9.19 The answer to this question depends on the students’ expectations.
9.21 It is sometimes appropriate to find the correlation between two variables even if you know that the relationship is slightly curvilinear. A straight line often does a remarkably good job of fitting a curved function, provided that it is not too curved.
9.23 The amount of money that a country spends on health care may have little to do with life expectancy because to change a country’s life expectancy you have to change the health of a great many individuals. Spending a great deal of money on one person, even if it were to extend her life by dozens of years, would not change the average life expectancy in any noticeable way. Often the things that make a major change in life expectancy, like inoculations, really cost very little money.
The African Red Cross estimates that there are 300-500 million cases of malaria each year, resulting in 1.5 to 2.5 million deaths. In particular, more than 90% of the deaths are in children under 5 years of age, and they occur predominantly in sub-Saharan Africa. Malaria cases could be cut by up to a third with insecticide treated bednets, which are very cheap by U.S. healthcare standards.
9.25 Extremely exaggerated data on male and female weight and height to show a negative slope within gender but a positive slope across gender:
Height 68 72 66 69 70 66 60 64 65 63
Weight 185 175 190 180 180 135 155 145 140 150
Gender Male Male Male Male Male Fem. Fem. Fem. Fem. Fem.
[pic]
What we are effectively plotting here is the relationship between Gender and Weight, more than between Height and Weight.
9.27 We have confounding effects here. If we want to claim that red wine consumption lowers the incidence of heart disease, we have a problem because the consumption of red wine is highest in those areas with the greatest solar radiation, which is another potential cause of the effect. We would have to look at the relationship between red wine and heart disease controlling for the effects of solar radiation.
9.29 This is an Internet search with no fixed answer.
Chapter 10-Regression
10.1 Regression equation predicting low birthweight from high-risk fertility.
Y = Low Birthweight Percentage
X = High-risk Fertility
[pic] = 6.70 sY = 0.698 sY2 = 0.487
[pic] = 46.00 sX = 6.289 sX2 = 39.553
covXY = 2.7245
[pic]
10.3 If the high risk fertility rate jumped to 70, we would predict that the incidence of birthweight < 2500gr would go to 8.35.
[pic]
This assumes that there is a causal relationship, which is plausible in some ways, but not proven.
It may be trivial to point this out, but here we have a real world situation where we can say something about changing trends in society and their possible effects.
10.5 I would be more comfortable speaking about the effects on Senegal because it is already at approximately the mean income level and we are not extrapolating for an extreme country.
This may have little to do with a statistics course in psychology, but there have been some noticeable improvements in infant mortality in Senegal, and one device that has made a difference is a warm table on which newborn infants can be placed. This may interest students who probably think of advances in medicine in terms of MRIs.
10.7 Prediction of Symptoms score for a Stress score of 45:
Regression equation: [pic] = 0.7831X + 73.891
If X = 45: = 0.7831*45 + 73.891
Predicted Symptoms = 109.13
10.9 Subtracting 10 points from every X or Y score would not change the correlation in the slightest. The relationship between X and Y would remain the same.
10.11 Diagram to illustrate Exercise 10.10:
[pic]
10.13 Adding a constant to Y:
[pic]
a) From this figure you can see that adding 2.5 to Y simply raised the regression line by 2.5 units.
b) The correlation would be unaffected.
10.15 Predicting GPA (Y) from ADDSC (X):
[pic]
When Hans Huessy and I first collected these data I was somewhat disheartened by how well we were doing (and to some extent I still am). We can take a measure in elementary school that is quickly filled out by the teacher, and make an excellent prediction about how the student will do in high school. That may be nice statistically, but I don’t think we like to feel that children are that locked in.
10.17 The correlation dropped to -.478 when I added and subtracted .04 from each Y value. This drop was caused by the addition of error variance.
One way to solve for the point at which they become equal is to plot a few predicted values and draw regression lines. Where the lines cross is the point at which they are equal. A more exact way of to set the two equations equal to each other and solve for X.
[pic]
10.19 Weight as a function of height for males:
[pic]
The regression solution that follows is a modification of printout from SPSS.
Equation Number 1 Dependent Variable.. WEIGHT
Variable(s) Entered on Step Number
1.. HEIGHT
Multiple R .60368
R Square .36443
Adjusted R Square .35287
Standard Error 14.99167
Analysis of Variance
DF Sum of Squares Mean Square
Regression 1 7087.79984 7087.79984
Residual 55 12361.25279 224.75005
F = 31.53637 Signif F = .0000
------------------ Variables in the Equation ------------------
Variable B SE B Beta T Sig T
HEIGHT 4.355868 .775656 .603680 5.616 .0000
(Constant) -149.933617 54.916943 -2.730 .0085
b) The intercept is given as the “constant” and is -149.93, which has no interpretable meaning with these data. The slope of 4.356 tells us that a one-unit increase in height is associated with a 4.356 increase in weight.
c) The correlation is .60, telling us that for females 36% of the variability in weight is associated with variability in height.
d) Both the correlation and the slope are significantly different from 0, as shown by an F of 31.54 and a (equivalent) t of 5.616.
10.21 Predicting my own weight, for which I use the equation from Exercise 10.19:
[pic] = 4.356*height - 149.93
[pic] = 4.356*68 - 149.93 = 146.28
a) The residual is Y - [pic] = 156 - 146.28 = 9.72. (I have gained some weight since I last used this example.)
b) If the students who supplied the data gave biased responses, then, to the degree that the data are biased, the coefficients are biased and the prediction will not apply accurately to me.
10.23 Predictions for a 5’6” male and female
For the male, [pic] = 4.356*66 - 149.93 = 137.57
For a female, [pic] = 2.578*66 - 44.859 = 125.29
Difference = 12.28 pounds
10.25 Plot of Reaction Time against Trials for only the Yes/5-stimuli trials:
[pic]
The following regression solution is a modification of SPSS printout.
Equation Number 1 Dependent Variable.. RXTIME
Variable(s) Entered on Step Number
1.. TRIAL
Multiple R .01640
R Square .00027
Adjusted R Square -.02056
Standard Error 12.76543
Analysis of Variance
DF Sum of Squares Mean Square
Regression 1 2.10363 2.10363
Residual 48 7821.89637 162.95617
F = .01291 Signif F = .9100
------------------ Variables in the Equation ------------------
Variable B SE B Beta T Sig T
TRIAL -.014214 .125100 -.016397 -.114 .9100
(Constant) 67.805186 28.267795 2.399 .0204
The slope is only -0.014, and it is not remotely significant. For this set of data we can conclude that there is not a linear trend for reaction times to change over time. From the scatterplot above we can see no hint that there is any nonlinear pattern, either.
10.27 The evils of television:
[pic]
Regression equations:
Boys [pic] = -4.821X + 283.61
Girls [pic] = -3.460X + 268.39
b) The slopes are roughly equal, given the few data points we have, with a slightly greater decrease with increased time for boys. The difference in intercepts reflects the fact that the line for the girls is about 9 points below that for boys.
c) Television can not be used as an explanation for poorer scores in girls, because we see that girls score below boys even when we control for television viewing.
10.29 Draw a scattering of 10 data points and drop your pencil on it.
b) As you move the pencil vertically you are changing the intercept.
c) As you rotate the pencil you are changing the slope.
d) You can come up with a very good line simply by rotating and raising or lowering your pencil so as to make the deviations from the lines as small as possible. (We really minimize squared deviations, but I don’t expect anyone’s eyes to be good enough to do that.)
10.31 Galton’s data
a) The correlation is .459 and the regression equation is [pic] = .646×midparent + 23.942. (Remember to weight cases by “freq”.)
b) I reran the regression requesting that SPSS save the Unstandardized prediction and residual.
c) [pic]
d) The children in the lowest quartile slightly exceed their parents mean (67.12 vs 66.66) and those in the highest quartile average slightly shorter than their parents (68.09 vs 68.31).
e) It is easiest if you force both axes to have the same range and specify that the regression line is [pic] = 1×X + 0. (If you prefer, you can use an intercept of 0.22 to equate the means of the parents and children.)
[pic]
Chapter 11-Multiple Regression
11.1 Predicting quality of life:
a) All other variables held constant, a difference of +1 degree in Temperature is associated with a difference of -.01 in perceived Quality of Life. A difference of $1000 in median income, again with all other variables held constant, is associated with a +.05 difference in perceived Quality of Life. A similar interpretation applies of b3 and b4. Since values of 0 cannot reasonably occur for all predictors, the intercept has no meaningful interpretation.
b) [pic] = 5.37 - .01(55) + .05(12) + .003(500) - .01(200) = 4.92
c) [pic] = 5.37 - .01(55) + .05(12) + .003(100) - .01(200) = 3.72
11.3 Religious Influence and religious Hope contribute significantly to the prediction, but not religious Involvement.
It is worth pointing out here that even though religion Involvement does not contribute significantly to the multiple regression, it does have a significant simple correlation with Optimism. The matrix of correlations (where N = 600) is
OPTIMISM RELINVOL RELINF RELHOPE
OPTIMISM 1.0000 .1667 .2725 .2663
P= . P= .000 P= .000 P= .000
RELINVOL .1667 1.0000 .4487 .5439
P= .000 P= . P= .000 P= .000
RELINF .2725 .4487 1.0000 .4187
P= .000 P= .000 P= . P= .000
RELHOPE .2663 .5439 .4187 1.0000
P= .000 P= .000 P= .000 P= .
11.5 I would have speculated that religious Involvement was not a significant predictor because of its overlap with the other predictors, but the tolerances kick a hole in that theory to some extent.
That’s what happens when you ask a question before you are sure of the answer. (
11.7 Adjusted R2 for 15 cases in Exercise 11.6:
[pic]
[pic]
Since a squared value cannot be negative, we will declare it undefined. This is all the more reasonable in light of the fact that we cannot reject H0:R* = 0.
11.9 The multiple correlation between the predictors and the percentage of births under 2500 grams is .855. The incidence of low birthweight increases when there are more mothers under 17, when mothers have fewer than 12 years of education, and when mothers are unmarried. All of the predictors are associated with young mothers. (As the question noted, there are too few observations for a meaningful analysis of the variables in question.)
11.11 The multiple correlation between Depression and the three predictor variables was significant, with R = .49 [F(3,131) = 14.11, p = .0000]. Thus approximately 25% of the variability in Depression can be accounted for by variability in these predictors. The results show us that depression among students who have lost a parent through death is positively associated with an elevated level of perceived vulnerability to future loss and negatively associated with the level of social support. The age at which the student lost his or her parent does not appear to play a role.
11.13 The fact that the frequency of the behavior was not a factor in reporting is an interesting finding. My first thought would be that it is highly correlated with the Offensiveness, and that Offensiveness is carrying the burden. But a look at the simple correlation shows that the two variables are correlated at less than r = .20.
11.15 Using random variables as predictors:
I drew the following data directly from the random number tables in the appendix (and I didn’t cheat).
Y X1 X2 X3 X4 X5
5 3 7 2 7 5
2 1 6 0 9 5
3 5 2 9 1 2
6 4 1 8 7 9
9 1 0 2 9 4
2 7 6 7 1 7
6 9 2 8 8 1
3 7 3 0 4 9
9 3 3 7 9 4
8 5 6 5 6 4
The multiple correlation for these data is .739, which is astonishingly high. Fortunately, the F test on the regression is not significant. Notice that we have only twice as many subjects as predictors.
This question is bound to lead to the question of how many cases we need per variable. There is no good answer to this question. Some will tell you that there should be at least 10 cases per predictor. I know of no argument in defense of such a rule. Harris (1985) has suggested a rule that says that N should exceed the number of predictors by at least 50. Cohen (1988) has argued from the point of view of power, and gives the example that a population correlation coefficient of .30 would require a sample size of 187 to have power = .80. This latter is sobering, but it is not a good argument here because we have not yet discussed power in any meaningful way.
11.17 Predicting weight:
[pic]
11.19 The weighted average is 3.68, which is very close to the regression coefficient for Height when we control for Gender.
11.21 Sex is important to include in this relationship because women tend to be smaller than men, and thus probably have smaller, though not less effective, brains, but we probably don’t want that contamination in our data. However, note that Sex was not significant in the previous answer, though the sample size (and hence power) is low.
11.23 I could argue that PctSAT is a nuisance variable because we are not particularly interested in the variable itself, but only in controlling it to allow us to have a clearer view of Expend, which is the variable in which we are interested. At the same time, it is an important contributor to the prediction of Combined, but we are led away from noticing that because of our predominant interest in Expend.
11.25 The scatterplot follows and shows that the squared correlation is .434, which is just what we found from the regression solution.
[pic]
Chapter 12—Hypothesis Tests Applied to Means: One Sample
12.1 Distribution of 100 random digits:
[pic]
12.3 The mean and standard deviation of the sample are 4.1 and 2.82, respectively, which are reasonably close to the parameters of the population for which the sample was drawn (4.5 and 2.6, respectively). The mean of the distribution of means is 4.28, which is somewhat closer to the population mean, and the standard deviation is 1.22.
a) The Central Limit theorem would predict a sampling distribution of the mean with a mean of 4.5 and a standard deviation of 2.6/(5 = 1.16.
b) These values are close to the values that we would expect.
12.5 If you had drawn 50 samples of size 15, the mean of the sampling distribution should still approximate the mean of the population, but the standard error of that distribution would now be only 2.67/(15 = 0.689.
12.7 Why doesn’t the previous question address the issue of the terrible state of North Dakota’s educational system? These students are certainly not a random sample of high school students in North Dakota or elsewhere. Moreover, they scored above the mean of 500, which would certainly not be expected if North Dakota’s system were inadequate. In addition, there is no definition of what is meant by “a terrible state,” nor any idea of whether or not the SAT measures such a concept.
12.9 Unlike the results in the two previous questions, this interval probably is a fair estimate of the confidence interval for P/T ratio across the country. It is not itself biased by the bias in the sampling of SAT scores.
12.11 Weight gain exercise:
For these data the mean weight gain was 3.01 pounds, with a standard deviation of 7.3 pounds. This gives us
[pic]
With 28 df the critical value at ( = .05, two-tailed, is 2.048, which will allow us to reject the null hypothesis and conclude that the girls gained weight at better than chance levels in this experiment.
There is an important movement within statistics right now in the direction of laying a much heavier emphasis on confidence limits than on null hypothesis tests. I think this is a very good example of a place where a behavioral scientist might make good use of a confidence interval. I didn’t ask the you to calculate these limits, but they are 0.227 and 5.787. You should think about what these limits mean and about why they are useful.
12.13 Effect size measure for data in Exercise 12.11:
One effect size measure would simply be the mean weight gain of 3.01 pounds. That statistic has real meaning to us, especially if we keep the size of a standard deviation in mind. A dubious alternative method would be to calculate an estimate of [pic]using the standard deviation of the gain scores as our base.
[pic]
If I knew the standard deviation at baseline, that would make a good denominator. Unfortunately that information is not available, and 7.3 is the standard deviation of weight gains, and it is difficult to see how that creates a reasonable metric..
12.15 I needed to solve for t in Exercise 12.14 because I did not know the population variance.
12.17 Testing the null hypothesis that children under stress report lower levels of anxiety:
[pic]
With 35 df the critical value of t at ( = .05 (two-tailed) is ±2.03. We can reject H0 and conclude that children under stress show significantly lower levels of anxiety than normal children.
Here is another situation where the data say that children report lower levels of anxiety, but it was necessary to first verify that their reports could be relied upon.
12.19 Yes, the results in Exercise 12.18 are consistent with the t test in Exercise 12.17. The t test showed that these children showed lower levels of anxiety than the normal population, and the confidence interval did not include 14.55.
Chapter 13—Hypothesis Tests Applied to Means: Two Related Samples
13.1 Sexual satisfaction of married couples. (Dependent variable = 1 for never fun and 4 for always fun.)
Husband Mean = 2.725 St. Dev. = 1.165
Wife Mean = 2.791 St. Dev. = 1.080
Difference Mean = -0.066 St. Dev. = 1.298
St. error diff = 0.136 N = 90
[pic]
With 90 df the critical value of t is approximately ±1.98, so we cannot reject the null hypothesis. We have no reason to conclude that wives are more or less satisfied, on the average, than their husbands.
This is a matched-sample t because responses came from married couples. I would hope that there is some relationship between the sexual satisfaction of one member of the couple and the satisfaction of the other—but perhaps that is hoping for too much.
13.3 Scatterplot of data from Exercise 13.1:
(The frequencies of each combination are shown above the points.)
[pic]
The correlation is .33, which is significant at ( = .05
This analysis finally addresses the degree of compatibility between couples, rather than mean differences. The correlation is significant, but it is not very large. That scatterplot is not very informative because of the discreteness of the scale and hence the overlapping of points.
13.5 The most important thing about a t test is the assumption that the mean (or difference between means) is normally distributed. Even though the individual values can only range over the integers 1 – 4, the mean of 91 subjects can take on a large number of possible values between 1 and 4. It is a continuous variable for all practical purposes, and can exhibit substantial variability.
I drew 10,000 random samples of N = 91 from treating Husband scores as a population. The distribution of means follows.
[pic]
13.7 We used a paired-t test for the data in Exercise 13.6 because the data were paired in the sense of coming from the same subject. Some subjects generally showed more beta-endorphins at any time than others, and we wanted to eliminate this subject-to-subject variability that has nothing to do with stress. In fact, there isn’t much of a relationship between the two measures, but we can’t fairly ignore it anyway. (Even though the correlation is not statistically significant, I think that we would look foolish if we did not treat these as paired data.)
13.9 If you look at the actual numbers given in Exercise 13.6, you would generally be led to expect that whatever was used to measure beta-endorphins was only accurate to the nearest half unit. Fair enough, but then where did values of 5.8 and 4.7 come from? If we can tell the difference to a tenth of a unit, why are most, but not all, of the scores reported to the nearest .5? It’s a puzzle.
13.11 You would not want to use a repeated measures design in any situation where the first measure will “tip off” or sensitize a subject to what comes next. Thus if you are going to show a subject something and then ask him to recall it, the next time you show any item the subject will expect to have to recall it. Similarly we should be careful about repeated measures in drug studies because drugs often last surprisingly long in the body.
13.13 How many subjects do we need?
First of all, in Exercise 13.6 we had 19 subjects, giving us 18 df. This means that for a one tailed test at α = .01 we will need a t of at least 2.552 to be significant. So we can substitute everything we know about the data except for the N, and solve for N.
[pic]
This exercise should be a good lead in to power, because you should be able to see the logic of this without knowing a thing about power. But in the chapter on power we are really doing the same thing but disguising it behind a bunch of Greek symbols. (Well, perhaps that’s a bit unfair.) Notice that we had to guess at N to get the critical value of t before we could calculate the needed N. Using the sample size they had is a reasonable approximation.
13.15 As the correlation between the two variables increases, the standard error of the difference will decrease, and the resulting t will increase.
13.17 First guess versus average guess
[pic]
[pic]
Notice that this is the same t as we had in Exercise 13.12. This is because there is a perfect linear relationship between first, second, and average guesses. (If you know the first guess and the average, you can compute what the second guess must have been.)
13.19 If I had subtracted the Before scores from the After scores I would simply change the sign of the mean and the sign of the t. There would be no other effect.
13.21 There is no answer I can give for this question because it asks the students to design a study.
Chapter 14—Hypothesis Tests Applied to Means: Two Independent Samples
14.1 Reanalysis of Exercise 13.1 as if the observations were independent:
Males Mean = 2.725 s = 1.165 NM = 91
Females Mean = 2.791 s = 1.080 NF = 91
[pic]
[t.05(180) = ±1.98] Do not reject the null hypothesis.
We can conclude that we have no reason to doubt the hypothesis that males and females are equal with respect to sexual satisfaction.
There was no need to pool the variances here because the sample sizes were equal. If we did pool them, the pooled variance would have been 1.262.
14.3 The difference between the t in Exercises 13.1 and 14.1 is small because the relationship between the two variables was so small.
14.5 Random assignment plays the role of assuring (as much as is possible) that there were no systematic differences between the subjects assigned to the two groups. Without random assignment it might be possible that those who signed up for the family therapy condition were more motivated, or had more serious problems, than those in the control group.
14.7 You can not use random assignment to homophobic categories for a study like the study of homophobia because the group assignment is the property of the participants themselves. The lack of random assignment here will not invalidate the findings.
14.9 In Exercise 14.8 it could well have been that there was much less variability in the schizophrenic group than in the normal group because the number of TATs showing positive parent-child relationship could have had a floor effect at 0.0. This did not happen, but it is important to check for it anyway.
14.11 Experimenter bias effect:
Expect Good Mean = 18.778 s = 3.930 N = 9
Expect Poor Mean = 17.625 s = 4.173 N = 8
[pic]
[pic]
[t.05(15) = ±2.131] Do not reject the null hypothesis.
We cannot conclude that our data show the experimenter bias effect.
14.13 Confidence limits for Exercise 14.8:
Mean difference = 1.45 standard error = 0.545 t.05(38) = 2.02
[pic]
Note that the answers to Exercises 14.11 and 14.12 are in line with the hypothesis tests, in that when we rejected the null hypothesis the confidence limits did not include 0, and when we did not reject the null, they did include 0.
14.15 Comparing GPA for those with low and high ADDSC scores:
[pic]
[t.05(86) = ±1.98] Reject H0 and conclude that people with high ADDSC scores in elementary school have lower grade point averages in ninth grade than people with lower scores.
Here I pooled the variances even though the Ns were substantially different because the variance estimates were so similar.
14.17 The answer to 14.15 tells you that ADDSC scores have significant predictability of grade point average several years later. Moreover the answer to Exercise 14.16 tells you that this difference is substantial.
This is a nice example of a situation in which it is easy to see a test of means as a test of predictability.
14.19 Anger with a reason is just fine.
[pic]
The critical value is approximately 2.00, so we will reject the null hypothesis and conclude that when given a reason for a woman’s anger, she is given more status than when no reason was given for the anger.
14.21 If the two variances are equal, they will be equal to the pooled variance.
If you have a problem seeing this, you can take any two equal variances and unequal Ns and try it for yourself. The answer becomes obvious when you do.
Chapter 15—Power
15.1 The statement on skiing is intended to point out that just because two things are different doesn’t mean that the larger (better, greater, etc.) one will always come out ahead. To take a different example, one treatment might be better than another for anorexia, but I would be very surprised if the difference was statistically significant every time, or even that its mean was always greater than the other mean. I just hope that it is significant most of the time.
15.3 Power for socially desirable responses:
Assume the population mean = 4.39 and the population standard deviation = 2.61
a) Effect size:
[pic]
b) delta:
[pic]
c) power = .22
Notice that the value of ( here is exactly the same as the value of t in that example. This is as it should be.
15.5 For Exercise 15.3 we would need ( approximately equal to 2.50, 2.80, and 3.25 for power of .70, .80, and .90, respectively.
[pic]
Notice how quickly the required sample sizes increase, and how as p increases the N required increases faster and faster.
15.7 Diagram of Exercise 15.6:
[pic]
15.9 Avoidance behavior in rabbits using a one-sample t test:
a) For power = .50 we need ( = 1.95.
[pic]
b) For power = .80 we need ( = 2.80.
[pic]
Because subjects come in whole units, we would need 16 subjects for power = .50 and 32 subjects for power = .80
15.11 Avoidance behavior in rabbits with unequal sample sizes:
[pic]
With ( = 1.46, power = .31
15.13 Cognitive development of LBW and normal babies at 1 year—modified data:
a) Power calculations
[pic]
With ( = -1.19, power = .22
b) t test:
[pic]
[t.05(38) = ±2.205] Do not reject the null hypothesis.
c) The t is numerically equal to (, although t is calculated from statistics and ( is calculated from parameters. In other words, ( is equal to the t we would get if the data came out with statistics equal to the parameters,
15.15 The significant t with the smaller N is more impressive, because that test had less power than the other, so the underlying difference is probably greater.
The fact that a significant difference with a small N is more impressive should not lead you to conclude that small sample sizes are to be preferred.
15.17 Social awareness of ex-delinquents—which subject pool would be better to use?
[pic]Normal = 38 N = 50
[pic]College = 35 N = 100 [pic]Dropout = 30 N = 25
[pic]
Assuming equal standard deviations, the H. S. dropout group of 25 would result in a higher value of (, and therefore higher power.
15.19 Total Sample Sizes Required for Power = .60, ( = .05, Two-Tailed (( = 2.20)
|Effect Size |( |One-Sample t |Two-Sample t (per group)|Two-Sample t (overall) |
|Small |0.20 |121 |242 |484 |
|Medium |0.50 |20 |39 |78 |
|Large |0.80 |8 |16 |32 |
15.21 When can power = (?
The mean under H1 should fall at the critical value under H0. The question implies a one-tailed test. Thus the mean is 1.645 standard errors above µ0, which is 100.
µ = 100 + 1.645(X
= 100 + 1.645(15)/(25
= 104.935
When µ = 104.935, power would equal (.
15.23 The power of the comparison of TATs of parents of schizophrenic and normal subjects.
[pic]
Power = .75
15.25 Aronson’s research on stereotype threat.
[pic]
[pic]
[pic]
[pic]
[pic]
From Appendix D5 the power of this experiment, if these are accurate estimates of the parameters, is .658.
Chapter 16—One-way Analysis of Variance
I am assuming that most people would prefer to see the solutions to these problems as computer printout. (I will use SPSS for consistency.)
16.1 Analysis of Eysenck’s data:
a) The analysis of variance:
- - - - - O N E W A Y - - - - -
Variable RECALL
By Variable GROUP Group Membership
Analysis of Variance
Sum of Mean F F
Source D.F. Squares Squares Ratio Prob.
Between Groups 1 266.4500 266.4500 25.2294 .0001
Within Groups 18 190.1000 10.5611
Total 19 456.5500
Standard Standard
Group Count Mean Deviation Error 95 Pct Conf Int for Mean
Grp 1 10 19.3000 2.6687 .8439 17.3909 TO 21.2091
Grp 2 10 12.0000 3.7417 1.1832 9.3234 TO 14.6766
Total 20 15.6500 4.9019 1.0961 13.3558 TO 17.9442
b) t test
t-tests for Independent Samples of GROUP Group Membership
Number
Variable of Cases Mean SD SE of Mean
----------------------------------------------------------------------
RECALL
Young 10 19.3000 2.669 .844
Older 10 12.0000 3.742 1.183
----------------------------------------------------------------------
Mean Difference = 7.3000
Levene's Test for Equality of Variances: F= .383 P= .544
t-test for Equality of Means 95%
Variances t-value df 2-Tail Sig SE of Diff CI for Diff
----------------------------------------------------------------------
Equal 5.02 18 .000 1.453 (4.247, 10.353)
Unequal 5.02 16.27 .000 1.453 (4.223, 10.377)
Notice that if you square the t value of 5.02 you obtain 25.20, which is the same as the F in the analysis of variance. Notice also that the analysis of variance procedure produces confidence limits on the means, whereas the t procedure produces confidence limits on the difference of means.
16.3 Expanding on Exercise 16.2:
a) Combine the Low groups together and the High groups together:
Variable RECALL
By Variable LOWHIGH
Analysis of Variance
Sum of Mean F F
Source D.F. Squares Squares Ratio Prob.
Between Groups 1 792.1000 792.1000 59.4505 .0000
Within Groups 38 506.3000 13.3237
Total 39 1298.4000
Standard Standard
Group Count Mean Deviation Error 95 Pct Conf Int for Mean
Grp 1 20 6.7500 1.6182 .3618 5.9927 TO 7.5073
Grp 2 20 15.6500 4.9019 1.0961 13.3558 TO 17.9442
Total 40 11.2000 5.7699 .9123 9.3547 TO 13.0453
Here we have compared recall under conditions of Low versus High processing, and can conclude that higher levels of processing lead to significantly better recall.
b) The answer is still a bit difficult to interpret because both groups contain both younger and older subjects, and it is possible that the effect holds for one age group but not for the other.
16.5 (2 and (2 for the data in Exercise 16.1:
SSgroup = 266.45
SStotal = 456.55
MSerror = 10.564
k = 2
[pic]
16.7 Foa et al. (1991) study:
|Group |n |Mean |S.D. |Total |Variance |
|SIT |14 |11.07 |3.95 |155 |15.6025 |
|PE |10 |15.40 |11.12 |154 |123.6544 |
|SC |11 |18.09 |7.13 |199 |50.8369 |
|WL |10 |19.50 |7.11 |195 |50.5521 |
|Total |45 |15.622 | |703 | |
[pic]
[pic]
From these values we can fill in the complete summary table and compute the F value.
|Source |df |SS |MS |F |
|Treatment |3 |507.840 |169.280 |3.04 |
|Error |41 |2279.067 |55.587 | |
|Total |44 |2786.907 | | |
[F.05(3,41) = 2.84] We can reject the null hypothesis and conclude that there are significant differences between groups. Some treatments are more effective than others.
b)
[pic]
c) It would appear that the more interventionist treatments lead to fewer symptoms than the less interventionist ones, although we would have to run multiple comparisons to tell exactly which groups are different from which other groups.
16.9 If the sample sizes in Exercise 16.7 were twice as large, that would double the SStreat and MStreat. However it would have no effect on MSerror, which is simply the average of the group variances. The result would be that the F value would be doubled.
16.11 Effect size for tests in Exercise 16.10.
It only makes sense to calculate an effect size for significant comparisons in this study, so we will deal with SIT vs SC.
[pic]
The SIT group is nearly a full standard deviation lower in symptoms when compared to the SC group, which is a control group.
16.13 ANOVA on GPAs for the ADDSC data:
Variable GPA
By Variable Group
Sum of Mean F F
Source D.F. Squares Squares Ratio Prob.
Between Groups 2 22.5004 11.2502 22.7362 .0000
Within Groups 85 42.0591 .4948
Total 87 64.5595
Standard Standard
Group Count Mean Deviation Error 95 Pct Conf Int for Mean
Grp 1 14 3.2536 .5209 .1392 2.9528 TO 3.5543
Grp 2 49 2.5920 .6936 .0991 2.3928 TO 2.7913
Grp 3 25 1.7436 .8020 .1604 1.4125 TO 2.0747
Total 88 2.4563 .8614 .0918 2.2737 TO 2.6388
There is a significant difference between the groups, telling us that there is a relationship between ADDSC score in elementary school and the GPA the student has in 9th grade. From the means it is clear that the GPA declines as the ADDSC score increases.
These are real data, and they tell us that a teacher in elementary school can already pick out those students who will do well and badly in high school. I have always found these results depressing and worrisome, even though psychologists are supposed to like to be able to predict. There are some things I wish weren’t so predictable.
16.15 Analysis of Darley and Latané data:
|Group |n |Mean |Total |
|1 |13 |0.87 |11.31 |
|2 |26 |0.72 |18.72 |
|3 |13 |0.51 |6.63 |
|Total |52 | |36.66 |
[pic]
From these values we can fill in the complete summary table and compute the F value.
|Source |df |SS |MS |F |
|Treatment |2 |0.854 |0.427 |8.06 |
|Error |49 |2.597 |0.053 | |
|Total |51 |3.451 | | |
[F.05(2,49) = 3.18] We can reject the null hypothesis and conclude that subjects are less likely to summon help quickly if there are other bystanders around.
16.17 Bonferroni test on data in Exercise 16.2:
Both of these comparisons will be made using t tests. The means are given in Exercise 16.15 above.
[pic]
For 36 df for error and for 2 comparisons at a familywise error rate of ( = .05, the critical value of t = 2.34. There is clearly not a significant difference between young and old subjects on tasks requiring little cognitive processing, but there is a significant difference for tasks requiring substantial cognitive processing. The probability that at least one of these statements represents a Type I error is at most .05.
It is worth pointing out that when we are using MSerror as our variance estimate, and have equal sample sizes, the computations are very simple because we only need to calculate the denominator once.
16.19 Effect size for WL versus SIT
[pic]
The two groups differ by over a standard deviation.
16.21 Spilich et al. data on a cognitive task:
Variable ERRORS
By Variable SMOKEGRP
Analysis of Variance
Sum of Mean F F
Source D.F. Squares Squares Ratio Prob.
Between Groups 2 2643.3778 1321.6889 4.7444 .0139
Within Groups 42 11700.4000 278.5810
Total 44 14343.7778
Standard Standard
Group Count Mean Deviation Error 95 Pct Conf Int for Mean
Grp 1 15 28.8667 14.6866 3.7921 20.7335 TO 36.9998
Grp 2 15 39.9333 20.1334 5.1984 28.7838 TO 51.0828
Grp 3 15 47.5333 14.6525 3.7833 39.4191 TO 55.6476
Total 45 38.7778 18.0553 2.6915 33.3534 TO 44.2022
Here we have a task that involves more cognitive involvement, and it does show a difference due to smoking condition. The non-smokers performed with fewer errors than the other two groups, although we will need to wait until the next exercise to see the multiple comparisons.
16.23 Spilich et al. data on driving simulation:
Variable ERRORS
By Variable SMOKEGRP
Analysis of Variance
Sum of Mean F F
Source D.F. Squares Squares Ratio Prob.
Between Groups 2 437.6444 218.8222 9.2584 .0005
Within Groups 42 992.6667 23.6349
Total 44 1430.3111
Standard Standard
Group Count Mean Deviation Error 95 Pct Conf Int for Mean
Grp 1 15 2.3333 2.2887 .5909 1.0659 TO 3.6008
Grp 2 15 6.8000 5.4406 1.4048 3.7871 TO 9.8129
Grp 3 15 9.9333 6.0056 1.5506 6.6076 TO 13.2591
Total 45 6.3556 5.7015 .8499 4.6426 TO 8.0685
Here we have a case in which the active smokers again performed worse than the non-smokers, and the differences are significant.
16.25 Attractiveness of faces
a) The research hypothesis would be the hypothesis that faces averaged over more photographs would be judged more attractive than faces averaged over fewer photographs.
b) Data analysis
[pic]
c) Conclusions
The group means are significantly different. From the descriptive statistics we can see that the means consistently rise as we increase the number of faces over which the composite was created.
Chapter 17—Factorial Analysis of Variance
17.1 Thomas and Wang (1996) study:
a) This design can be characterized as a 3 ( 2 factorial, with 3 levels of Strategy and 2 levels of delay.
b) I would expect that recall would be better when subjects generated their own key words, and worse when subjects were in the rote learning condition. I would also expect better recall for the shorter retention interval. (But what do I know?)
c)
Summaries of RECALL
By levels of STRATEGY
DELAY
Variable Value Label Mean Std Dev Cases
For Entire Population 11.602564 7.843170 78
STRATEGY 1.0000 9.461538 6.906407 26
DELAY 1.0000 14.923077 5.330127 13
DELAY 2.0000 4.000000 2.516611 13
STRATEGY 2.0000 11.269231 9.606488 26
DELAY 1.0000 20.538462 1.983910 13
DELAY 2.0000 2.000000 1.471960 13
STRATEGY 3.0000 14.076923 6.183352 26
DELAY 1.0000 15.384615 5.454944 13
DELAY 2.0000 12.769231 6.796492 13
17.3 Analysis of variance on data in Exercise 17.1:
RECALL by STRATEGY
DELAY
UNIQUE sums of squares
All effects entered simultaneously
Sum of Mean Sig
Source of Variation Squares DF Square F of F
Main Effects 2510.603 3 836.868 42.992 .000
STRATEGY 281.256 2 140.628 7.224 .001
DELAY 2229.346 1 2229.346 114.526 .000
2-Way Interactions 824.538 2 412.269 21.179 .000
STRATEGY DELAY 824.538 2 412.269 21.179 .000
Explained 3335.141 5 667.028 34.267 .000
Residual 1401.538 72 19.466
Total 4736.679 77 61.515
There are significant differences due to both Strategy and Delay, but, more importantly, there is a significant interaction.
This is a good example for showing all three effects. The Delay and Interaction effects are obvious, but the overall Strategy effect is harder to see. You would do well to calculate the Strategy means, which are 9.46, 11.27, and 14.08, respectively. It will help if you draw those means on the figure for Exercise 17.2.
.
17.5 Bonferroni tests to clarify simple effects for data in Exercise 17.4:
[pic]
[pic]
For 6 comparisons with 36 df, the critical value of t is 2.80.
For the 5-minute delay, the condition with the key words provided by the experimenter is significantly better than both the condition in which the subjects generate their own key words and the rote learning condition. The latter two are not different from each other.
For the 2-day delay, the rote learning condition is better than either of the other two conditions, which do not differ between themselves.
We clearly see a different pattern of differences at the two delay conditions. The most surprising result (to me) in the superiority of rote learning with a 2 day interval.
In running these Bonferroni tests, I had a choice. I could have thought of each simple effect as a family of comparisons, and obtained the critical value of t with 3 comparisons for each. Instead I chose to treat the whole set of 6 comparisons as a family and adjust the Bonferroni for 6 tests. There is no hard and fast rule here, and many might do it the other way. The results would not change regardless of what I decided.
17.7 The results in the last few exercises have suggested to me that if I were studying for a Spanish exam, I would fall back on rote learning, painful as it sounds and as much against common wisdom as it is.
17.9 In this experiment we have as many primiparous mothers as multiparous ones, which certainly does not reflect the population. Similarly, we have as many LBW infants as full-term ones, which is certainly not a reflection of reality. The mean for primiparous mothers is based on an equal number of LBW and full-term infants, which we know is not representative of the population of all primiparous births. Comparisons between groups are still legitimate, but it makes no sense to take the mean of all primiparous moms combined as a reflection of any meaningful population mean.
Many of our experiments are run this way (with equal sample sizes across groups that are not equally represented in the population), and it is important to distinguish between the legitimacy of between group comparisons and the legitimacy of combined means.
17.11 Simple effects versus t tests for Exercise 17.10.
a) If I had run a t test between those means my result would simply be the square root of the F = 1.328 that I obtained.
b) If I used MSerror for my estimated error term it would give me a t that is the square root of the F that I would have had if I had used the overall MSerror, instead of the MSerror obtained in computing the simple effect.
17.13 Analysis of variance for Spilich et al. Study:
[pic]
The main effect of Task and the interaction are significant. The main effect of Task is of no interest because there is no reason why different tasks should be equally difficult., We don’t care about the main effect of Smoking either because it is created by large effects for two levels of Task and no effect for the third. What is important is the interaction.
This is a good example of a situation in which main effects are of little interest. For example, saying that smoking harms performance is not really accurate. Smoking harms performance on some tasks, but not on others. Often main effects are still interpretable in the presence of an interaction, but not here.
17.15 Simple effects to clarify the Spilich et al. Example.
We have already seen these simple effects in Chapter 16, in Exercises 16.18, 16.19, and 16.21.
17.17 Factorial analysis of the data in Exercise 16.2:
[pic]
Here we see that we have a significant effect due to age, with younger subjects outperforming older subjects, and a significant effect due to the level of processing, with better recall of material processed at a higher level. Most importantly, we have a significant interaction, reflecting the fact that there is no important difference between younger and older subjects for the task with low levels of processing, but there is a big difference when the task calls for a high level of processing—younger subjects seem to benefit more from that processing (or do more of it).
17.19 Nurcombe et al study of maternal adaptation.
[pic]
b) The program worked as intended and there was no interaction between groups and educational level.
17.21 Effect size for Level of Processing in Exercise 17.17
[pic]
This is a very large effect size, but the data show an extreme difference between the two levels of processing.
I used the square root of MSerror here because that was in line with what I did in the text. But a good case could be made for adding Age and the interaction sums of squares back in and calculating a new error term. That would produce
[pic]
which is considerably smaller but still a very large effect.
17.23 Set of data for a 2 ( 2 design with no main effects but an interaction:
Cell means:
[pic]
8 12
12 8
17.25 Magnitude of effect for Exercise 17.1
Summary table from Exercise 17.1:
|Source |df |SS |MS |F |
|Strategy |2 |281.256 |140.628 |7.224 |
|Delay |1 |2229.346 |2229.346 |114.526 |
|S ( D |2 |824.538 |412.269 |21.179 |
|Error |72 |1401.538 |19.466 | |
|Total |77 |4736.679 | | |
[pic]
17.27 Magnitude of effect for Exercise 17.13:
Summary table from Exercise 17.13:
|Source |df |SS |MS |F |
|Task |2 |28661.526 |14330.763 |132.895 |
|SmokeGrp |2 |1813.748 |906.874 |8.41 |
|T ( S |4 |1269.452 |317.363 |2.943 |
|Error |126 |13587.200 |107.835 | |
|Total |134 |45331.926 | | |
[pic]
[pic]
[pic]
17.29 The two magnitude of effect measures ((2 and (2) will agree when the error term is small relative to the effect in question, and will disagree when there is a substantial amount of error relative to the effect. But notice that this is a comparison of MSerror and a sum of squares, and sums of squares can be large when there are many degrees of freedom for them. So to some extent, all other things equal, the two terms will be in closer agreement when there are several degrees of freedom for the treatment effect.
17.31 You should restrict the number of simple effects you examine to those in which you are particularly interested (on a priori grounds), because the familywise error rate will increase as the number of tests increases.
Although we routinely talk about familywise error rates with respect to multiple comparison procedures, they really apply whenever you run more than one test, whether you consider them tests on main effects and interactions, or tests on simple effects, or tests on multiple contrasts. A test is a test as far as the error rate is concerned.
|Source |df |SS |MS |F |
|Gender |1 |223.49 |223.49 |10.78 |
|Condition |1 |1.35 |1.35 | 46
b) We can reject the null hypothesis and conclude that first-born children are more independent than their second-born siblings.
Here is a good example of where we would use a “matched sample” test even though the same children do not perform in both conditions (nor could they). We are assuming that brothers and sisters are more similar to each other than they are to other children. Thus if the first-born is particularly independent, we would guess that the second-born has a higher than chance expectation of being more independent. They share a common environment.
20.9 Data in Exercise 20.7 plotted as a function of the first-born’s score:
[pic]
The scatterplot shows that the difference between the pairs is heavily dependent upon the score of the first-born.
20.11 The Wilcoxon matched-pairs signed-ranks test tests the null hypothesis that paired scores were drawn from identical populations or from symmetric populations with the same mean (and median). The corresponding t test tests the null hypothesis that the paired scores were drawn from populations with the same mean and assumes normality.
This is an illustration of the argument that you buy things with assumptions. By making the more stringent assumptions of a t test, we buy greater specificity in our conclusions. However if those assumptions are false, we may have used an inappropriate test.
20.13 Rejection of the null hypothesis by a t test is a more specific statement than rejection using the appropriate distribution-free test because, by making assumptions about normality and homogeneity of variance, the t test refers specifically to population means—although it is also dependent on those assumptions.
20.15 Truancy and home situation of delinquent adolescents:
Analysis using the Kruskall-Wallis one-way analysis of variance:
|Natural Home |Foster Home |Group Home |
|Score |Rank |Score |Rank |Score |Rank |
|15 |18 |16 |19 |10 |9 |
|18 |22 |14 |16 |13 |13.5 |
|19 |24.5 |20 |26 |14 |16 |
|14 |16 |22 |27 |11 |10 |
|5 |4.5 |19 |24.5 |7 |6.5 |
|8 |8 |5 |4.5 |3 |2 |
|12 |11.5 |17 |20 |4 |3 |
|13 |13.5 |18 |22 |18 |22 |
|7 |6.5 |12 |11.5 |2 |1 |
|Ri |124.5 | |170.5 | |83 |
N = 27
n = 9
[pic]
We can reject the null hypothesis and conclude the placement of these adolescents has an effect on truancy rates.
This analysis doesn’t directly answer the question the psychologist wanted answered, because he wanted to show that the group home was better than the others. He might follow this up with Mann-Whitney tests serving in the role of multiple comparison procedures, applying a Bonferroni correction (although it might be difficult to find the necessary critical values.) Alternatively, he could just run a single Mann-Whitney between the group home and the combined data of the other two placements.
20.17 The study in Exercise 20.16 has the advantage over the one in Exercise 20.15 in that it eliminates the influence of individual differences (differences in overall level of truancy from one person to another).
20.19 For the data in Exercise 20-5:
a) Analyzed by chi-square:
| |More |Fewer |Total |
|Observed |7 |3 |10 |
|Expected |5 |5 |10 |
[pic]
We cannot reject the null hypothesis.
b) Analyzed by Friedman’s test:
|Before |After |
|Score |Rank |Score |Rank |
|8 |2 |7 |1 |
|4 |1 |9 |2 |
|2 |1 |3 |2 |
|2 |1 |6 |2 |
|4 |2 |3 |1 |
|8 |1 |10 |2 |
|3 |1 |6 |2 |
|1 |1 |7 |2 |
|3 |1 |8 |2 |
|9 |2 |7 |1 |
|Totals |13 | |17 |
N = 13 k= 2
[pic]
These are exactly equivalent tests.
20.21 “The mathematics of a lady tasting tea;”
|First Cup |Second Cup |Third Cup |
|Score |Rank |Score |Rank |Score |Rank |
|8 |3 |3 |2 |2 |1 |
|15 |3 |14 |2 |4 |1 |
|16 |2 |17 |3 |14 |1 |
|7 |3 |5 |2 |4 |1 |
|9 |3 |3 |4 |6 |2 |
|8 |2 |9 |3 |4 |1 |
|10 |3 |3 |1 |4 |2 |
|12 |3 |10 |2 |2 |1 |
|Totals |22 | |16 | |10 |
N = 8 k= 3
[pic]
We can reject the null hypothesis and conclude that people don’t really like tea made with used tea bags.
Chapter 21—Choosing the Appropriate Test
[N.B. Please review the disclaimer concerning these answers at the beginning of Chapter 21. There are many different ways to think about a study.]
21.1 This test involves comparing two proportions, and the easiest way to do that is to set up a 2 × 2 contingency table with Group on one dimension and Mastery on the other.
21.3 This is a repeated measures analysis of variance with assessment periods as the repeated measure and condition as the between subject variable. If measurements were taken on several occasions I would like to see the data plotted over time, but all we currently have is the data at the end of the treatment phase.
21.5 This is a t test for two independent groups—children of divorced families and children of intact families.
21.7 This is a standard one-way analysis of variance. I would be most likely to conduct a planned comparison of the positive and negative conditions.
21.9 This is a two-way analysis of variance, with secure and insecure as one dimension and mortality vs. dental as the other. No multiple contrasts are called for because there are only two levels of each variable.
21.11 This could be treated as a two-way analysis of variance if we break the data down by race and by Afrocentric facial features. A problem with this is that we would presumably have more Afrocentric features for black inmates, which would lead to unequal sample sizes (i.e. an unbalanced design).
21.13 This is a regression problem where we are time is one variable and the difference in happiness between liberals and conservatives (by year) is the other variable.
21.15 The most important thing to do would be to plot the data over time looking for trends. A repeated measures analysis of variance would tell you if differences are significant, but it is the direction of differences, and whether they return to baseline, that is likely to be most informative. The authors further broke down the participants in terms of their preoccupation with 9/11 and looked at differences between those groups. Interestingly, even the least preoccupied group showed changes over time.
21.17 This is a difficult one, partly because it depends on what Payne wants to know. I assume that she wants to know how rankings of characteristics agree across sexes or across years. She could first find the mean rank assigned to each characteristic separately for each sex and year. Because the raw data were originally ranks, I would probably be inclined to then rank these mean values. She could then calculate Spearman’s rS between males and females for each year or between years within each sex. The correlations would be obtained for the ten pairs of scores (one per characteristic).
21.19 This is a 2 ( 4 analysis of variance with two levels of sex and 4 levels of occupation. The major emphasis is on the occupations, so multiple comparisons of those means would be appropriate.
21.21 There are two independent groups in this experiment. The authors should use a Mann-Whitney test to compare average locus of control scores.
21.23 This is a situation for a chi-square goodness-of-fit test. The conditions are Rotated versus Stationary, and the count is the number of subjects choosing that condition as giving stronger contours. The expected values would be 37/2 = 18.5. The data are sufficiently extreme that a test in superfluous.
21.25 This is another complex repeated-measures analysis of variance. The comparison of recall of the two lists (one learned before administration of the drug and the other learned after) is a repeated measurement because the same subjects are involved. The comparison of the Drug versus Saline groups is a between-subjects effect because the groups involve different subjects.
21.27 This is basically a correlational study, where we separately correlate the two dependent variables with amount of alcohol consumed. Given the 14 year gap, and all of the other factors that affect development, we should not expect very strong correlations even under the best of conditions.
-----------------------
α
β
2.987 = 3.87-2.03*2.61/√36
2.75
3.87
................
................
In order to avoid copyright disputes, this page is only a partial summary.
To fulfill the demand for quickly locating and searching documents.
It is intelligent file search solution for home and business.
Related download
- to the student
- density curves math
- statistics 1601 university of minnesota
- multivariate analyses introduction examples where
- introduction to statistics
- chapter 2 modeling distributions of data
- practice of statistics texas edition answer key
- mathematics sample test grade 5 2010 2013
- chapter 11 inference for distributions of categorical data
- moore 5th edition chapter 4 section 5