Name of Test: - University of Alberta



Test Review: Gray Oral Reading Test-4 (GORT-4)

|Name of Test: Gray Oral Reading Test-4 (GORT-4) |

|Author(s): Wiederholt, J. Lee and Bryant, Brian R. |

|Publisher/Year: Pro-Ed 1967, 1986, 1992, 2001 |

|Forms: Form A and Form B |

|Age Range: 6 years, 0 months to 18 years, 11 months |

|Norming Sample: |

| |

|GORT-4 used a test administrator recruitment procedure common to other Pro-Ed tests in which the customer database was used to invite potential examiners to participate. At the four major test |

|sites (one in each geographic region), site coordinators supervised the administration of tests to students. At these sites, the authors state that “several hundred children in Grades 1 through |

|12” (Wiederholt & Bryant, 2001, p. 49) participated in Fall, 1999 and Fall, 2000. |

| |

|Authors report that students with disabilities were those who attended regular rather than special education classes. The sample was stratified by ages and region, gender, race, ethnicity, |

|parental income, and parental education. This data was presented in table form (Wiederholt & Bryant, 2001, Table 4.2, pp. 53-55). The percentage of the sample compares favourably to the national|

|school age percentages taken from the 1997 Statistical Abstract of the United States: The National Data Book by the U.S. Bureau of Census. |

| |

|Total Number: 1 677 students |

|Number and Age: 6 years (n=87) to 18 years (n-72) with 12 month intervals for groups |

|Location: 28 states (and the District of Columbia in four major geographic regions) |

|Demographics: Age, race (White, Black, Other), ethnicity (Native American, African American, Hispanic, Asian, European American and Other), gender, urban vs. rural, geographic region (NW, |

|Midwest, West, and South), family income (under 15,000 to over 75, 000) and educational attainment of parents (less than bachelor’s degree, bachelor’s degree, and master’s/professional/doctoral |

|degrees) and disability status (no disability, learning disability, speech/language disorder, mental retardation, and other) were reported. The authors stratified the sample on these |

|characteristics. |

|Rural/Urban: yes |

|SES: by family income |

|Other: no |

|Summary Prepared By: Eleanor Stewart 18 October 2007 |

|Test Description/Overview: |

| |

|The test kit consists of the examiner’s manual, the student workbook containing both Forms A and B, and Form A and Form B record forms. The kit is contained in a sturdy cardboard box. The |

|examiner needs a stopwatch. |

| |

|The preface provides an interesting history of Dr. William Gray and his career in reading assessment and research. His influence is acknowledged with the test name, Gray Oral Reading Test, a |

|testament to his original work in developing the present day edition. Each edition of the GORT is described in order for the reader to appreciate the progression toward the revisions made for |

|the fourth edition. |

|Comment from the Buros reviewer: “The GORT-3 was well received, but criticism of the technical aspects of the test was made. These areas of criticism involved documentation of the equivalency of|

|Forms A and B, need for more extensive validity data for assessing student progress in reading, socioeconomic description of students in the standardization sample, and the normative data being |

|15 years old” (Crumpton, & Miller Whitehead, 2003, p. 418). |

| |

|Improvements included in the 2001 edition of the GORT-4, are listed on page x (Wiederholt & Bryant, 2001). Changes include: new normative data comparable to the U.S. population, normative sample|

|stratification, bias is addressed, additional reliability is calculated for subsamples, validity studies addressing new normative sample as well as subgroups are included, and clear evidence of |

|form equivalence. |

| |

|In the introductory chapter, the authors describe the nature of reading and the purpose of the GORT-4. They begin with the statement, “The nature of reading is complex and not completely |

|understood” (Wiederholt & Bryant, 2001, p. 1). They outline a brief history of how reading has been conceptualized first as “thought getting” by Huey in 1908 followed by Gray’s proposition that |

|“reading assumes not only fluent, accurate recognition of words, but in addition, a clear grasp of the meanings implied by the symbols used” (Quote in text, p. 1). The authors state that |

|theorists hold three broad views of how reading develops. These views are described as: bottom-up, top-down, and interactive theories. They briefly outline the central tenets of each theory. |

|They then introduce the view that certain skills may be more important at certain stages of reading development with a brief comment on phonemic skills. Finally, they re-state Gray’s view as |

|follows, “Regardless of how one learns to read, the ultimate goal of reading instruction is the same-that is, the fluent recognition of words and a clear grasp of the meaning implied by the |

|symbols used. The GORT-4 is a measure of this ultimate goal” (p. 2). |

| |

|Comment: I was surprised at how short this introduction was at only 1.5 pages. Also, the brief statement above is the only reference to the authors’ views. Supposedly this constitutes their |

|rationale. Perhaps earlier editions contained more detailed information and rationale leading the authors to assume that readers are familiar enough with the test and its history to make their |

|own judgments about rationale and theory. I just wonder about this because whenever I use a test, I recognize that by doing so I am operating within the authors’ theoretical view of the |

|phenomenon of interest. |

| |

|The two equivalent forms have 14 stories each. Each story is read by the student. A series of five multiple choice questions follow that aim to assess the student’s comprehension. The results |

|are scored as follows (in quotations from the text): |

|“Rate- the amount of time taken by a student to read a story |

|Accuracy- the student’s ability to pronounce each word in the story correctly |

|Fluency- the student’s rate and accuracy scores combined |

|Comprehension- the appropriateness of the student’s responses to questions about the content of each story read |

|Overall reading ability- a combination of a student’s Fluency (i.e., Rate and Accuracy) Comprehension Scores” (Wiederholt & Bryant, 2001, p. 3). |

| |

|The Flesch-Kincaid Readability Formula was used to determine the reading levels of the stories. The results indicated that Story 1 had a 0.0 readability index and ranges were exhibited for |

|subsequent stories with the final five (Stories 9-14) demonstrating 12.0. |

| |

|The authors describe the equating procedures used to align Forms A and B. They state that they followed Kolen and Brennan’s (1995) recommendation for correcting differences in difficulty that |

|might exist between the two forms (Wiederholt & Bryant, 2001, p. 50). They describe the linear equating procedure chosen including the equation used. |

| |

|The student workbook contains both Form A and Form B stories. The type print appears to be 12 point for stories 1-7, 10 point for stories 8-14 for both forms. Each story is presented as a |

|paragraph printed in the middle of the page with the corresponding questions printed on the reverse. |

| |

|Several appendices (A-G) are included. A to D present normative score tables, standard scores, and age and grade equivalents. Information about the samples used in validity studies is presented |

|in Appendix E. The miscue worksheet is found in F and two examples of completed sample performances are found in Appendix G. |

| |

|Purpose of Test: The purpose of this test is to assess oral reading rate, accuracy, fluency, and comprehension, identify strengths, document progress, and use in research. The test is intended |

|to help identify students who perform below their peers in reading proficiency, to identify strengths and weaknesses, to document progress in intervention, and to use in research on reading. |

|Each of these purposes is briefly described in the manual (Wiederholt & Bryant, 2001, pp. 4-5). |

| |

|Areas Tested: |

|Reading Comprehension |

|Other: fluency, accuracy, rate |

| |

|Who can Administer: A range of individuals such as teachers, special educators, and speech pathologists are listed as potential testers. The authors state that whoever uses the GORT-4 should |

|have formal training in test administration and interpretation and suggest that such training is available through colleges and “local agencies or private consultants” (Wiederholt & Bryant, |

|2001, p. 7) however, they note that testers should consult schools and professional organizations if the purpose of testing involves the diagnosis of disabilities or qualification for special |

|programs. The authors provide five general rules to make test administration more reliable. These are familiar injunctions about familiarizing oneself with the test prior to administering it, |

|creating an appropriate test environment, etc. |

| |

|Administration Time: The manual indicates 15 to 45 minutes (Wiederholt & Bryant, 2001, p. 10). Though testing can be completed in one session, two sessions are also allowed if, for example, the |

|student becomes fatigued. |

|Test Administration (General and Subtests): |

| |

|Chapter 2 presents the administration and scoring information. Entry points are identified by grade and starting point story with the caveat that testing should begin at the child’s grade level |

|unless the examiner has prior knowledge that the child may be a poor reader in which case testing may begin at a lower level. For the Comprehension subtests, basal and ceiling rules are |

|identified with three examples provided to assist the examiner. If a basal is not initially established, the examiner is instructed to proceed to ceiling and then return to establish five |

|correct responses to comprehension questions. Form A and B use the same basal and ceiling rules. The examiner scores the student’s response as correct (1 point) or incorrect (0 point) and sums |

|the correct responses for each story which is transferred to Section II on the record form. A completed example of the form is presented in Figure 2.2 (Wiederholt & Bryant, 2001, p. 12). For the|

|Fluency score, the examiner is concerned with timing the reader as the passage is read to completion. Then the time taken (in seconds) is recorded in the record booklet. Time intervals are |

|converted to values on a 0 to 5 point Rate score. Similarly, accuracy is recorded as the number of deviations from the print and converted to a 0 to 5 point Accuracy Score. The Fluency score is |

|the sum of the Rate and Accuracy Scores. The Fluency score determines the basal as a score of 9 or 10 for a story and ceiling as a score of 2 or less. Figure 2.3 illustrates the procedure for |

|establishing basal and ceiling scores in this manner (Wiederholt & Bryant, 2001, p. 14). |

| |

|Chapter 2 continues with specific administration and scoring instructions (Wiederholt & Bryant, 2001, pp. 15-26) in which prompting and timing, noting deviations, determining scores, and |

|identifying and calculating miscues is discussed. The examiner’s instructions to the student are printed in red type in this section. Examples of each marking system are provided in Figures 2.5 |

|and 2.6 (i.e., slash marking and miscue marking). The authors explain that they chose to present a simple marking system though they are aware that some examiners may choose to separate graphic |

|and phonemic miscues. They state that the examiner must analyze at least two different stories with a minimum of 25 miscues across the two stories. The student’s miscues are classified according|

|to the outline in Section V of the record booklet. An example is illustrated in Figure 2.7 (p. 22). Five categories of miscues are used as follows: meaning similarity, function similarity, |

|graphic/phonemic similarity, multiple sources, and self-correction (each is described in Chapter 1; and specific examples are provided in this chapter for purpose of assisting the reader to |

|learn the system). The student’s miscues per category are then totaled and a percentage is calculated for all miscue categories. The authors allow that other types of miscues may be produced. |

|These miscues should be noted along with general impressions and observations of the student’s behaviour and performance during the reading tasks. Space is provided on the record booklet for |

|this purpose. |

| |

|Comment: The miscue schema is easy to follow and use. I think that it may be easier for someone who works in linguistics or related area. Novice examiners may have some difficulty with this |

|test. Also, I wonder if subjective judgment enters into scoring, especially in assessing meaning elements in miscuing. I think that examiners should be proficient English language users to use |

|this system accurately and fairly. |

|Test Interpretation: |

| |

|Interpreting the results of the standardized scores is explained beginning with the Total Scores which the authors state have no clinical value in themselves but are used as the basis for |

|calculating the standard scores, percentiles, and age and grade equivalents. The subtest standard scores are based on 100 +/-3. The Oral Reading Quotient has a base of 100 with standard |

|deviation of 15. Percentiles are explained and follow the norm for this type of score. Age and grade equivalents are based on the average performance of the normative group. The authors state |

|that these scores were developed “through interpolation, extrapolation, and smoothing…” (Wiederholt & Bryant, 2001, p. 36). They acknowledge the well-known APA concerns but note that examiners |

|will have to align their results with employer requirements and legislative requirements. Therefore, the GORT-4 authors deliver these scores “reluctantly”. They encourage examiners to read the |

|cautions regarding these scores and cite references to Aiken (2000), Anastasi and Urbina (1997), Linn and Gronlund (1995), and others. |

| |

|In the following section, the authors turn their attention to Discrepancy Analyses. They state that the GORT-4 can be used to compare results with other measures of reading and to “measure the |

|effects of instruction over time” (Wiederholt & Bryant, 2001, p. 37). They note that only the ORQ should be used for discrepancy analysis. They provide an example of how to conduct the analysis |

|with a student’s results using Anastasi and Urbina’s (1997) formula to calculate how large a difference is needed to be statistically significant. They conclude by stating, “This analysis is a |

|relatively unsophisticated way of examining differences between an individual’s pretest and posttest ORQs. It is not recommended for examining differences between groups of individuals’ pretest |

|and posttest ORQs” (Wiederholt & Bryant, 2001, p. 38). For group differences, the authors recommend using an experimental design appropriate for comparisons. |

| |

|In terms of comparing with other measures, analysis can be made between standard scores to determine significance of difference. The authors note that the same procedure is followed for |

|statistical significance but that clinical usefulness is otherwise determined. Following Reynolds (1990), the authors outline the formula to be used in order to identify a severe discrepancy |

|that is clinically useful. The authors provide Table 3.4 to illustrate the severe discrepancy analyses results for comparisons between the GORT-4 and language and intelligence tests. Also |

|provided is an example of a comparison between the GORT-4 ORQ and the Silent Reading Quotient (SRQ) of the Gray Silent Reading Tests (p< .05). Continuing the student example discussed earlier, |

|the authors present the example illustrating the 30 point criterion cut-off (i.e., as they state, “Any Difference Score that meets or exceeds 30 points is severely discrepant”, (Wiederholt & |

|Bryant, 2001, p. 39). |

| |

|In the section, “Analyzing Oral Reading Miscues”, the authors describe how to use the optional miscue classification system to make descriptive judgments about the student’s deviations from |

|print. Furthering their example of the student, they provide analysis in terms of percentage miscues before and after instruction to show how the examiner can use this information. |

| |

|The GORT-4 is one of three tests, along with the Gray Silent Reading Tests and the Gray Diagnostic Reading Tests-Second Edition, that can be used together to comprehensively examine aspects of |

|reading ability. They briefly outline the historical antecedent, the original Gray Silent Reading Test, and acknowledge that the relationship between oral and silent reading remains |

|controversial. The authors review the content of each test with a view to how to select the appropriate test to address specific questions. They state that the GORT-4 serves to assess oral |

|reading ability alone whereas administration of the GSRT will yield a comparison of the student’s oral and silent reading ability. The GDRT-2 is used “when the reader’s ability is so poor that |

|he or she cannot handle the demands of passage reading, or when you want to look at a variety of reading skills” (Wiederholt & Bryant, 2001, p. 44). In the last section in the chapter on |

|interpreting results of the GORT-4, the authors discuss additional assessment and instruction. |

|Standardization: Age equivalent scores Grade equivalent scores Percentiles Standard scores (mean 10 , SD 3) Stanines |

|Other: Composite (from standard scores) and Conversion Scores (for rate and accuracy) |

|These conversions are made when the time taken to read the passage and the number of deviations are converted into number values from 0 to 5 that then are added to lend themselves to the overall|

|Fluency Score which ranges from 0 to 10. The Fluency Score is said to equally emphasize both rate and accuracy aspects. The procedure for developing the conversion scores from a sample of the |

|normative data is presented on page 57. Basically, the authors divided the performance into quarters with the top 25% representing the best readers receiving a rating of 5, followed by the next |

|group and so on down to the bottom values that represented the performance of the students 6% who attained a value converted to 0. |

| |

|The authors comment about age and grade equivalent scores and the concerns raised about their use in the chapter on interpreting scores (Wiederholt & Bryant, 2001, p. 37). They state their |

|position as follows, “We prefer you use GORT-4 standard scores, quotients, and percentiles rather than age and grade equivalents when reporting results to parents and other professionals” (p. |

|37). They cite several references including Aiken (2000), and Anastasi and Urbina (1997). In addition to the use of norm-referenced scores, the authors present a procedure for conducting a |

|discrepancy analysis. |

|Reliability |

| |

|Internal consistency of items: Cronbach’s coefficient alphas were computed for the entire norming sample by age groups. |

|Large value alphas were reported for the subtests and composite (ORQ) ranging from .91 to .97. |

| |

|Test-retest: Forty-nine students in elementary (n=30), middle school (n=10), and high school (n=9) participated in retesting. The students, 22 males and 27 females, ranged in age from 6 years to|

|18 years. The majority, 78% were European Americans with participation from 16% African American, and 2% Other. The interval between tests was “approximately two weeks” (Wiederholt & Bryant, |

|2001, p. 66). Correlation coefficients were corrected for restricted sample range where needed. The results indicated a range from .85 to .95 for the four subtests and ORQ. |

| |

|Inter-rater: Thirty randomly selected completed test records were independently scored by two Pro-Ed staff. The results indicated inter-rater reliability from .94 to .99. |

| |

|Other: none |

|Validity: The extensively detailed chapter on validity is not surprising given the long history and research related to the test. |

| |

|Content: The authors note that using stories with corresponding comprehension questions was the format originally chosen by Gray. The story format continues to be a practical and logical way to |

|assess reading abilities as they are used extensively in the students’ classroom experiences. Other reading tests also use the story format. Therefore, face validity is established. The |

|composition of the stories reflected general interest, controlled syntactic structure, and age and grade appropriate vocabulary. The authors describe the specific procedures and materials they |

|consulted in this process. They note that the stories for all but the youngest children remain unchanged from earlier editions. The only new story is the one provided for the entry level where |

|emerging oral reading ability is assessed. |

| |

|Comment: This is a credible process, a number of sources were consulted for vocabulary in particular. |

| |

|In terms of developing the comprehension questions, the authors describe both the types of questions used and the research to support their choice of comprehension questions (i.e., literal, |

|inferential, critical, and affective). Also, vocabulary was controlled so that “the vocabulary would not be more difficult than that used in the stories” (Wiederholt & Bryant, 2001, p. 75). |

| |

|Reading fluency has received substantial research attention according the authors. In particular, they note that “poor fluent word recognition is the most frequently cited factor in reading |

|disorders” (Wiederholt & Bryant, 2001, p. 75). The GORT-4 combines rate and accuracy to reflect the evidence indicating that good readers are proficient at both of these aspects. |

| |

|The sequence of the stories is arranged in ascending order of difficulty. The controversy over readability formulas is stated with references to the literature. The authors, who wished to |

|provide some indication of readability, chose the Flesch-Kincaid Readability Formula to do so. Table 6.1 (Wiederholt & Bryant, 2001) provides the results for the stories. However, the authors |

|also used the data from the normative sample to determine the final order of stories. |

| |

|In terms of deviations from print, the term preferred by the authors, a synopsis of the research history on reading errors and classifications is given. On the basis of the research, the authors|

|decided to use their system of five types of miscues described elsewhere in this review. Their system, they comment, is similar to others currently in use (references are given). |

| |

|Conventional item analysis provided quantitative evidence for content validity. Although the stories, save for one, are the same as previous editions, the analysis yielded satisfactory item |

|discrimination coefficients, corrected for part-whole effect, and item difficulties. Tables 6.2 and 6.3 display the data (Wiederholt & Bryant, 2001). |

| |

|Criterion Prediction Validity: The authors chose to refer to studies that used the GORT-3 and GORT-R for evidence of aspects of criterion prediction validity noting that this is an acceptable |

|practice where the items, administration, and response format remain unchanged in the new edition. Therefore, their discussion focused on the comparison of the new edition to the earlier ones |

|mentioned and to other measures of reading, including the GDRT-2 and GSRT. The results indicate high coefficients (with median correlations from .63 to .75) except for Comprehension which is |

|moderate (.45). Table 6.5 displays the correlation data for comparisons with other reading measures. |

| |

|Construct Identification Validity: Throughout this section, the authors provide background statements to highlight the relationships that they seek to demonstrate. For example, they state in the|

|section on measures of spoken language, “Authorities have long agreed that reading is related to spoken language to some meaningful degree. In fact, the correlation between the two variables is |

|often offered as evidence for construct-identification validity for most tests of spoken language and reading…” (Wiederholt & Bryant, 2001, p. 87). |

| |

|Age differentiation: Mean scores and standard deviations are displayed in table form demonstrating an age-related progression in scores (i.e., scores become larger as children age). Coefficients|

|for both Forms A and B were analyzed with the following coefficients reported: Rate (.74), Accuracy (.73), Fluency (.66), and Comprehension (.74). Authors note that these coefficients are, |

|according to their reference, (MacEachron, 1982), to be high. |

|Intercorrelation: Correlations between each of the subtest measures were shown to be statistically significant at the .01 level which “represents a moderate to very high degree of relationship” |

|(Wiederholt & Bryant, 2001, p. 87). |

|Relationship to measures of spoken language: Results of two studies are stated: Cornwall (1992) and Catts (1993) found median coefficients of .49 for Fluency and .57 for Comprehension. Table 6.7|

|displays the data. |

|Relationship to measures of written language: The authors summarize the results of four separate studies that demonstrated relationship to written language (three studies used the GORT-3, one |

|the GORT-4, Cornwall, 1992). Results are displayed in Table 6.8 with the authors pointing out that “the best evidence for validity is the median coefficient for the ORQ because it represents the|

|overall picture regarding the relationship between reading and writing” (Wiederholt & Bryant, 2001, p. 89). The authors reported the median coefficient as.59 for ORQ. |

|Relationship to tests of intelligence or aptitude: The GORT-4 was correlated as follows: |

|WISC-III (individual scales and full scale) ranged from .40 and .67, |

|Scholastic Aptitude Scale ranged from .34 to .83, |

|Short Form Test of Academic Aptitude ranged from .26-.50, |

|WISC-R (individual scales and full scale) ranged from .43 to .80, and |

|Median correlations for all studies ranged from .45 to .62. |

|Relationship to RAN: Three studies report the relationship of RAN to GORT-4. These studies found median correlation coefficients to be .49 for Fluency and .48 for Comprehension. The authors |

|state that these results compare favourably with the .41 reported in the RAN literature. |

|Group differentiation Using the mean standard scores for the normative sample and subgroups, the predicted outcomes were demonstrated in regard to gender, race/ethnicity (European Americans, |

|African Americans, and Hispanic Americans), and disability (learning disability, serious emotional disturbance, and attention-deficit/hyperactivity disorder). (N.B. The sample for serious |

|emotional disturbance was small n=15). The learning disability group performed Below Average whereas the gifted children were Above Average as expected. All results are presented in Table 6.11. |

|Additionally, other studies reported significant differences between the groups of students studied (Wiederholt & Bryant, 2001, pp. 94-95). Each study is briefly described in terms of the |

|sample, sample size, and the comparisons made. |

|Change over time: Intervention studies reported the ability of GORT-4 to detect change due to intervention. These studies are summarized both in the text and in Table 6.13. The authors state |

|that, “all three studies reported in this section support this use of the GORT-4 and clearly provide convincing evidence of the GORT-4’s construct identification validity” (Wiederholt & Bryant,|

|2001, p. 96). |

|Item validity: The discriminating powers previously reported in Table 6.2 provide evidence of construct-identification validity. The authors note that these results are replications of those |

|found in earlier studies using the GORT-R. |

| |

|Differential Item Functioning: Swaminathan and Rogers’s (1990) logistic regression procedure was used to detect DIF. The authors explain their rationale and procedures undertaken. The results of|

|560 comparisons found that nine were significant at the .001 level. Effect size was investigated and none of the nine comparisons indicated moderate or large effect sizes. On the basis of this |

|evidence the authors conclude that “the test is unbiased in regard to gender, race, and ethnicity” (Wiederholt & Bryant, 2001, p. 82). |

| |

|Other: none |

|Summary/Conclusions/Observations: This is a comprehensive and well-researched test, attesting to its continued relevance and use. |

|Clinical/Diagnostic Usefulness: |

|On the basis of this review, I would use this test for its intended purposes trusting that it would yield important and clinically useful information. I was particularly interested in the miscue|

|analysis as I think that clinicians would use this information as a beginning point for exploring the individual student’s approach to reading. |

References

Aiken, L. R. (2000). Psychological testing and assessment (8th ed.). Needham Heights, MA: MIT Press.

Anastasi, A., & Urbina, S. (1997). Psychological testing (7th ed.). Upper SaddleRiver, NK: Prentice-Hall.

Catts, H. (1993). The relationship between speech-language impairments and reading disabilities. Journal of Speech and Hearing Research, 36, 948-958.

Cornwall, A. (1992). The relationship of phonological awareness, rapid naming, and verbal memory to severe reading and spelling disability. Journal of Learning Disabilities, 25(8), 532-538.

Crumpton, N.L., & Miller Whitehead, M. (2003). Test review of the Gray Oral Reading Test, Fourth Edition. In B. S. Plake, J.C. Impara, and R.A. Spies (Eds.), The fifteenth mental measurement yearbook (pp. 417-421). Lincoln, NE: Buros Institute of Mental Measurements.

Kolen, M. J., & Brennan, R. L. (1995). Test equating methods and practices. New York: Springer.

Linn, R. L., & Gronlund, N. E. (1995). Measurement and evaluation in teaching (7th ed.). Upper Saddle River, NJ: Merrill.

MacEachron, A. E. (1982). Basic statistics in the human sciences. Austin, TX: Pro-Ed.

Swiminathan, H., & Rogers, H. J. (1990). Detecting differential item functioning using logistic regressions procedures. Journal of Educational Measurement, 26, 55-66.

U.S. Bureau of the Census. (1997). Statistical abstract of the United States: The national data book. Washington, DC: U.S. Department of Commerce.

Wiederholt, J. L., & Bryant, B. R. (2001). GORT 4 Gray Oral Reading Tests Examiner’s Manual. Austin, TX: Pro-Ed.

To cite this document:

Hayward, D. V., Stewart, G. E., Phillips, L. M., Norris, S. P., & Lovell, M. A. (2008). Test review: Gray oral reading test-4 (GORT-4). Language, Phonological Awareness, and Reading Test Directory (pp. 1-12). Edmonton, AB: Canadian Centre for Research on Literacy. Retrieved [insert date] from .

................
................

In order to avoid copyright disputes, this page is only a partial summary.

Google Online Preview   Download