CARS Projects And Rearch

 

The Research Base of Open Court and its Translation into Instructional Policy in California

Children Reading

by Jerry Treadway, Professor of Education
San Diego State University

Summary

In a recent paper that has been broadly disseminated via the Internet, Moustafa and Land have reached erroneous conclusions about Open Court Publishing's reading series, Collections for Young Scholars. Chief among them are: (1) that Open Court's reading program, Collections for Young Scholars, was adopted by the California State School Board despite recommendations to the contrary by the state's textbook adoption committee; (2) that, more generally, California's recent policy and legislation on reading instruction were unduly influenced to favor the Open Court curriculum by the results of a single study conducted in Houston and sponsored by NICHD (Foorman et al. 1998); and (3) that the Open Court curriculum produces inferior reading outcomes as compared with a variety of other commercial programs.

In this paper, all three allegations are refuted. Contrary to Moustafa and Land's claims: (1) Open Court was approved at every stage in California's 1996 textbook adoption process; (2) the NICHD Houston study post-dated, and therefore had no causal effect on California's reading initiative, and (3) when several serious flaws in Moustafa and Land's analysis are corrected, the Open Court schools are shown to significantly outperform otherwise comparable non-Open Court schools.

Introduction

Moustafa and Land's paper, "The Research Base of Open Court and its Translation into Instructional Policy," recently disseminated over the Internet has received wide-spread but undeserved attention. In their article, Moustafa and Land have made serious allegations about the legitimacy and ethics of the process through which Open Court's reading program, Collections for Young Scholars was adopted in California and about the program's instructional adequacy.

As background, during the 1990s, the public was repeatedly confronted with hard data from the National Assessment of Educational Progress (NAEP 1994, 1995) indicating that many children in U.S. schools could not read well enough to comprehend or learn from grade-level materials. Over the same period of time, however, reading researchers formally published an array of scientifically validated evidence that, given proper instruction, virtually every healthy child can learn to read. It is within this context that the Open Court program became topical. Through both rigorous scientific studies (e.g., Foorman, et al. 1998) and the separate experiments of a number of school districts around the country (e.g., Sacramento, CA; Baltimore, MD; Ft. Worth, TX), implementation of the Open Court program has been shown to produce marked improvement in students' reading achievement. Moreover, in both the prestigious and carefully vetted reports of the National Research Council (1998) and the National Reading Panel (2000) and in curriculum evaluations of a number of other authors and organizations (e.g., AFT), the Open Court curriculum has gained special recognition for its coverage of those instructional components that research has found most critical to children's reading success. Arguably as a result of such reports, the Open Court program is increasingly recommended to schools seeking to improve their reading outcomes.

In short, if Moustafa and land were correct in their allegations, their paper would be well warranted, Yet, the evidence shows that Moustafa and Land are wront in their message and allegations, and that their electronic broadcast of this paper amounts to a grievous disservice to all who are earnestly concerned with how best to help children learn to read. In the present paper it will be shown that Moustafa and Land's thesis is seriously flawed in both information and argument.

The remainder of this paper will present excerpts from Moustafa and Land's paper (in Bold, with page numbers corresponding to those in the pdf file on Moustafa's website), followed by discussion of the errors embodied in each. Because Moustafa and Land's paper is so thoroughly riddled with misinformation, it was possible to respond only to selected points. The reader is therefore warned not to assume the correctness of any porton of Moustafa and Land's text that are not excerpted or addressed herein.

(1) Moustafa & Land, p. 1:
In 1996, California's Reading/Language Arts Textbook Adoption Committee, composed of qualified reading/language arts teacher specialists, recommended that Open Court's Collections for Young Scholars not [emphasis theirs] be placed on the state's textbook adoption list.

The process through which publisher materials were reviewed for California's 1996 Reading/Language Arts adoption included two evaluation committees and one adoption board, namely, the State School Board. The first evaluation committee, the Instructional Materials Evaluation Panel (IMEP), was made up of the "qualified reading/language arts teacher specialists" to whom Moustafa and Land allude. On the basis of their evaluation, the IMEP recommended, in the affirmative, that Open Court be adopted. The second committee to evaluate the materials was the Curriculum Development and Supplemental Materials Commission (Curriculum Commission). The Curriculum Commission also recommended, in the affirmative, that Open Court be adopted. Following is the summary evaluation of Open Court that the IMEP and the Curriculum Commission presented to the State School Board, as documented by the California Department of Education:

Collections for Young Scholars develops independent and self-motivated critical thinkers who reflect on, discuss, and write about their reactions to rich literature. The program is integrated, sequentially developed, and increasingly complex and demanding. Students regularly are asked to confer with peers, develop questions, respond in a variety of ways, and make connections across selections and other curricular areas. Support is provided across all grade levels for skill development, including development of phonemic awareness, and phonetic skills at the early grades. Little evidence was noted of the availability of literature in the five most common languages in California other than English. A strength of this program lies in the activities, materials, and strategies that treat students as capable learners and in those integrated assessment activities that are closely related to classroom instruction. (Resource Evaluation: Reading/Language Arts and English as a Second Language Adoption Report: Review of Instructional Resources: Kindergarten Through Grade Eight, 1997, p. 39). Note that, though not published for wide distribution by the state printing house until 1997, this report was submitted to the Board of Education during July, 1996.)

In other words, and contrary to Moustafa and Land's claim, in voting to include Open Court on California's list of adopted programs, the State School board was accepting the recommendation of both of the committees, the IMEP and the Curriculum Commission, that were responsible for materials evaluation for the State of California. The foregoing shows that Moustafa and Land's first major conclusion that Open Court was not recommended by the IMEP is proven false.

(2) Moustafa & Land, p. 1:
This study...compares the SAT9 reading scores of schools using Open Court against comparable schools using non-scripted programs.

Moustafa and Land's description of their comparison schools as "non-scripted" is misleading in that it invites readers to think "scripted" each time they encounter "Open Court." Open Court is not a "scripted" program. Scripted materials, by definition, are those that provide teachers with the exact words through which to explain concepts and conduct activities. For example, directions to be given to students in the administration of standardized tests are "scripted." In fact, there are a few reading programs on the market that are scripted. However, Open Court is not one of them.

Although the activities and materials provided in the Open Court manual are carefully described and sequenced, the program makes no attempt to limit communication between teachers and students. To the contrary, much of the program is built on the premise that open inquiry and reflection by students are critical both to learning and, as they allow teachers to determine how best to clarify and enrich concepts, to instruction.

In view of this, more appropriate labels for the schools in contrast in Moustafa and Land's study would be "Open Court" versus "non-Open Court" schools or, more specifically, "Open Court" schools versus schools using "other comprehensive commercial reading programs."

(3) Moustafa & Land, p. 1:
Open Court's Collection for Young Scholars, a reading instruction program written by an electrical engineer in the 1960s...

The "electrical engineer" in reference, Blouke Carus (Ph.D, Cal. Tech.), owned Open Court Publishing when Collections for Young Scholars was authored. He continues to publish children's periodicals including Cricket, Lady Bug, Baby Bug, Spider, Cicada (mostly fiction) and Muse (the official children's periodical of the Smithsonian Institute). However, Dr. Carus did not write the Collection For Young Scholars, nor does he author any of the other Open Court works he publishes. In the early 1970's, he decided to develop and publish Open Court in response to what he perceived to be the poor quality of his children's instructional materials. In March 1996, the Carus family sold Open Court's classroom reading series to McGraw-Hill.

Over the past thirty years, the authorship of Open Court has changed somewhat across editions. The senior authors of the 1995 edition, Collections for Young Scholars, were: Marilyn Jager Adams, Valerie Anderson, Carl Bereiter, Ann Brown, Joe Campione, Robbie Case, Jan Hirshberg, Anne McKeough, Michael Pressley, and Marlene Scardamalia. This information is available on the title page of the Open Court teacher guides or student anthologies. In addition, brief biographies of the authors are provided in the front matter of the teacher guides.

(4)Moustafa & Land, p. 1:
The story of Open Court ... begins [with the study conducted by Foorman et al. (1998) in Houston]. There were many problems with the [Foorman et al., 1998] research.

There are many errors associated with Moustafa and Land's description of the design and results of the Foorman et al. (1998) study. For example, in asserting that the researchers collected data on only "3 to 8 children in each of the four types of classrooms," the reader might conclude that the study was too small to permit thorough analysis. Closer examination of the study shows, however, that the results are based on data collected from 53 classroom teachers, 28 Title 1 teachers, and 285 children in 19 different elementary schools. Further, a glance at the comprehension data in Table 5 in Foorman et al.'s (1998, p. 50) article shows that the Open Court students outperformed every comparison group on every comprehension measure. As possible source of Moustafa and Land's claims to the contrary, one cell in the table shows a higher standard score (83.1 versus 81.8) for one of the groups receiving "contemporary reading instruction" over the Open Court group. However, the table also indicates that 25% of this "contemporary" group read too poorly to participate in this test.

Among the Open Court children, only 14% were so excused. Moreover, Foorman et al. (1998, p. 51) found that, overall, "The direct instruction [Open Court] group approached national average on decoding (43rd percentile) and passage comprehension (45th percentile) compared with the IC-R ['contemporary'] group's means of 29th and 35th percentile, respectively." Not only did those children whom Moustafa and Land term the "contemporary" group score more poorly on average but further, as reported by Foorman et al. (1998, p. 51), they were "much more likely" to score at levels conventionally accepted as indicative of reading disability.

Readers interested in the Houston researchers' own responses to such criticisms are referred to an article (Foorman et al., 2000) that was recently invited by the Educational Researcher (which is the association journal of the American Educational Research Association) and also to the Houston research team's website (cars.uth.tmc.edu).

(5)Moustafa & Land, p. 1:
For one, the study received significant financial and personnel support from McGraw-Hill, the publisher of Open Court (Taylor, 1998, pp. 79 & 298).

All instructional materials in the Foorman, et. al, (1998) study, including the Open Court materials, were provided, free of charge, to the research project. The Open Court materials were provided by Open Court Publishing, not McGraw-Hill Publishing which didn't own the Open Court Reading Program in 1994. As reported by Foorman et al. (1998, p. 40), "Because [the direct code] group used basal materials that were new to the teachers, a representative from the publisher spent one day orienting the teachers to the materials." This level of inservice support is routinely provided by publishers to schools newly adopting the Open Court curriculum and did not exceed the inservice opportunities enjoyed by teachers in other conditions in the study.

The insinuation of "significant financial and personnel support" is that the publisher provided special incentives-financial and perhaps otherwise-to researchers involved in the study. This is among the unfounded allegations made in Taylor's (1998) book that is cited by Moustafa and Land. Such charges of scientific compromise and influence peddling are extremely serious and have been widely repeated. Yet, there exists not one shred of evidence that points to any impropriety, financial or otherwise, on the part of the research team or the publishers of Open Court in the conduct of this study. Indeed, the excerpt below is from a note that the Houston researchers wrote to Taylor before she published her book and which she chose not to include.

To reiterate previous responses, the study was not biased in favor of a particular group, nor is there any relationship between our group and the current or previous owners of Open Court. In fact, we approached Open Court at the recommendation of the school district in which we were conducting the research. We proposed to use DISTAR, but it was not acceptable to the District. District officials suggested Open Court because the District used Open Court Math. When we approached Open Court, which at that time was not owned by SRA/McGraw-Hill, we learned that there was a new edition of Open Court Reading. We did not have sufficient funds to purchase Open Court because it was not budgeted in the original proposal. Open Court generously provided us with the prepublication version of the curriculum, and some assistance with implementation, but teachers who used Open Court did not receive more intensive training (Response to Taylor, 1996).

Open Court Publishing assisted with this project at their own expense, and at considerable risk of exposure for, after all, the study was conducted in schools that previously had poor reading outcomes.

It must also be appreciated that what is at stake here extends well beyond any particulars of the Foorman et al. (1998) study. "Unfortunately," as argued in the National Research Council's report, "the efficacy of commercial basal programs is rarely evaluated.... Given the programs' potential for supporting teachers, as well as teachers' widespread use and even dependence on these programs in the classroom, such evaluation should be a priority for public policy" (1998, p. 194). Finally, it needs to be reiterated that the Open Court program was not acquired by McGraw-Hill until spring of 1996. In 1994, when the study began, McGraw-Hill's only connection was as owner of the rejected program, DISTAR. In other words, the suggestion that the contribution of materials and training to the Houston project was part of a methodical and foresighted conspiracy by McGraw-Hill to influence the outcome of the study and to use the results to enhance market share is simply untrue.

(6) Moustafa & Land, p. 2:
How did the NICHD Houston research come to effect [sic] instructional policy in California?

The Houston study (Foorman, et al. (1998) postdated and thus had no causal effect on either key policy or legislation. By May, 1996, California's two major policy documents on reading were completed, or nearly so. Every Child a Reader: The Report of the California Reading Task Force (1995), had been published. A follow up reading advisory called Teaching Reading: A Balanced, Comprehensive Approach to Teaching Reading in Prekindergarten Through Grade Three, Program Reading Advisory (1997), was largely finished by early May 1996. (As was the case with the adoption report, the State of California printing house took over six month to publish the document for wide dissemination). The reading advisory was developed to provide more information about the components of an exemplary reading program. When completed, it was adopted by the California Department of Education, the State School Board, and the Commission on Teacher Credentialing, and was sent to every primary grade teacher and administrator in California. In part it stated:

To be complete and balanced and to meet the literacy needs of all students, including English language learners and students with special needs, any early reading program must include the following instructional components: phonemic awareness; letter names and shapes; systematic, explicit phonics; spelling; vocabulary development; comprehension; and higher-order thinking; and appropriate instructional materials (page 4).

The California legislature's decision to require explicit, systematic phonics in instructional materials and their decision to require training in explicit systematic phonics for teachers also had been established prior to May 1996. The following bills (including chapter numbers-the last two numbers are the year the bill became law) became law or had finished the legislative process and were waiting funding prior to May 1996.

Assembly Bill 1504, Burton, (Ch. 764-95) among other things, required the systematic instruction of spelling.

Assembly Bill 170 Alpert, Burton, Conroy, (Ch. 765-95) (Also known as the "ABC" bill), required, among other things, that reading and mathematics, "…include, but not be limited to, systematic, explicit phonics, spelling, and basic computational skills…"

Assembly Bill 3482, Davis, Johnston (Ch. 196-96), and the Goals 2000 federal legislation that was tied to AB 3482, provided funds for the professional development of teachers by approved providers. The requirements, known as "A to K" were to include, but not be limited to, the following:

A. phonemic awareness
B. systematic, explicit phonics instruction (including sound-symbol relationships, decoding, and word attack skills)
C. spelling instruction
D. diagnosis of reading deficiencies
E. research on how children learn to read
F. research on how proficient readers read
G. structure of the English language
H. relationships between reading, writing, and spelling
I. planning and delivery of appropriate reading instruction based on assessment and evaluation
J. means of improving reading comprehension, and
K. pupil independent reading of good books and the relationship of that activity to improved reading performance.

A widely stated complaint is that the legislation privileged phonics to the exclusion of all else. As generally overlooked by detractors, however, these skills were but several on a list of requirements aimed at ensuring complete coverage of the issues and challenges of early literacy instruction, where that instruction includes language development, comprehension instruction, and wide reading.

Again, as is clearly seen, instructional policy in reading in California could not have been initiated and/or developed as a result of the Foorman et al. (1998) study.

(7) Moustafa & Land, p. 2
The NICHD Houston research was presented to legislators and other policy makers as "reliable, replicable" research. May, 1996,...Foorman and the Director of NICHD, Reid Lyon, presented it to the California Senate Education Committee. Researchers and educators with views consistent with four decades of replicated research were not allowed to testify.

In May 1996, the California Assembly Education Committee (and not the "Senate Education Committee"] asked NICHD to provide a review of their reading research program. Reid Lyon, who is NICHD Branch Chief for Child Development and Behavior (and not "Director of NICHD"), complied. In doing so, he asked Barbara Foorman to present data on the Houston study as an example of how intervention research is designed and conducted. For a legislative committee to invite expert testimony on a topic before them is not unusual; it is normal procedure.

Beyond that, Moustafa and Land seem to be suggesting either that the Houston study (Foorman, et al., 1998) does not qualify as reliable replicable research or that its results conflict with those of other studies that do so qualify. Yet this is not the case. Based on an exhaustive search of the last thirty years' of refereed journals, the National Reading Panel (2000) statistically confirmed that, across soundly designed experimental studies, and regardless of the type of less methodical alphabetic instruction received by the comparison groups, systematic phonics instruction is of significant and inarguable benefit to young readers. Furthermore, though the Houston study was unusual in its large-scale, real-world implementation, its results were not. Across the many substantial and formal investigations of the issue, the conclusion is always the same (Adams, 1990; Anderson, R. C. 1985; Bond & Dykstra, 1967; Chall, 1967; National Reading Panel, 2000 National Research Council, 1998). Literacy growth depends on learning to recognize written words easily and accurately, and learning to recognize words easily and accurately is more quickly and securely achieved where students receive explicit, systematic instruction in the logic and conventions of the writing system.

In summary, and contrary to Moustafa and Land's claim, the Foorman et al. (1998) study did not have an impact on reading policy in California.

(8) Moustafa & Land, p. 2
In December 1996, the California Board of Education overruled the reading/language arts teacher specialists on its Reading/Lanuage Arts Committee and added Open Court to California's textbook adoption list. In 1997, the Board of Education of Sacramento City Unified School District adopted Open Court for 55 of its 60 elementary school, K-6.

Again, as reviewed in point (1) above, the first sentence of this excerpt is simply not true. Although the second is correct, it carries an insinuation that warrants address. In particular, the Sacramento City School Board's decision to adopt the Open Court curriculum was both considered and participatory, involving more than a dozen public hearings. In addition, the Board's selection of the Open Court program was supported by both the district's administration and the Sacramento City Teachers' Association. It enjoyed widespread support within the district, support that has been justified by subsequent increases in reading scores.

(9) Moustafa & Land, p.3:
In their analyses of Open Court's success in Sacramento, Helfand and Ellerbee compared Sacramento's scores in one grade in 1998 to scores in the same grade in 1999. In doing so, they were analyzing teaching, not learning.

As background, the issue here is that the SAT9 scores of Sacramento City Unified School District rose dramatically from 1998 to 1999. In the above excerpt, Moustafa and Land are challenging Helfand and Ellerbee for publicly suggesting that the test score rise was related to the district's contemporaneous shift to the Open Court reading program. The problem is that Moustafa and Land's criticism makes no sense. First, their notion that Helfand and Ellerbee's interpretation of the SAT9 scores pertained to teaching and not learning is untenable. Student performance is a variable that depends on both teaching and learning. Second, since the question at hand is whether the instructional program had a measurable impact on student growth, differences in teaching would seem of primary interest anyhow.

In any case, Helfand and Ellerbee's logic is wholly sound. That is, it is a safe assumption that, across the 60 elementary schools of the Sacramento City Unified District, neither the population of teachers nor the population of children changed significantly from 1998 to 1999. By implication, then, the dramatic change in the district's test outcomes must have been due to some other factor. Given that the only unique and significant district-wide change from 1998 to 1999 was the adoption of the Open Court program, the conclusion that the new instructional program was a causal factor in students' improved outcomes is well-warranted.

(10) Moustafa & Land, p. 4
We investigated the effectiveness of Open Court schools against non-scripted programs in schools serving similar economic groups in the Los Angeles Unified School District.

Schools differ in many ways beside classroom instruction and many of these differences influence the impact of classroom instruction. Yet, Moustafa and Land chose instead to gauge school comparability on the basis of a single characteristic: The percentage of children receiving free or reduced lunch. This is surprising because the State of California has already developed a multiple criteria system for establishing school comparability.

The State of California, as provided by the Public Schools Accountability Act (PSAA) (Ch. 3/1999) has developed both a "State Rank" and a "Similar Schools Rank." The Statewide Rank provides a ranking of schools by deciles by grade level of instruction. To develop the Similar Schools Rank category, the State of California generates separately, for each of its 8,563 schools, a list of 100 other schools with similar extra-instructional characteristics. These extra-instructional factors are:

  • Pupil mobility

  • Pupil ethnicity

  • Pupil socioeconomic status

  • Percentage of teachers who are fully credentialed

  • Percentage of teachers who hold emergency credentials

  • Percentage of pupils who are English language learners

  • Average class size per grade level

  • Whether the schools operate multitrack year-round educational programs"

When placed together, a school's reading scores and their State Schools Rank and Similar Schools Rank provide a reliable measure of its academic achievement.

Given all the numbers that Moustafa and Land pulled from the API website, it is hard to understand how they overlooked these characteristics and scores. Yet, they chose instead to gauge school comparability on the basis of a single characteristic: the percentage of children receiving free or reduced lunch. Regrettably, even on this single variable, Moustafa and Land failed to constitute their Open Court and non-Open Court samples fairly, the percentage of children receiving free or reduced lunch across the Open Court schools was 89% while for the non-Open Court schools, it was only 79%.

(11) Moustafa & Land, p. 4
We limited our study to schools that either (1) used Open Court, not in combination with another program or (2) one of the four non-scripted programs in the study. Of these 155 schools, 19 used Open Court. Nine Open Court schools were long-term Open Court schools.

Because Moustafa and Land give no indication of how they ascertained a school's use of Open Court, all 19 were contacted and conversations regarding their reading programs took place. Of Moustafa and Land's 9 "long-term Open Court schools," 7 began using Open Court on a limited basis in 1987 as part of a program called the "Ten Schools Program." The other two schools (116th St. and 118th St.) in Moustafa and Land's "long-term" Open Court sample did not join the "Ten Schools Program" (now called "Ten plus Two") and begin using Open Court until the 1998-9 school year. Noting that Moustafa and Land's analysis is based on Spring 1999, SAT9 scores, these last two schools do not qualify as "long term" Open Court users.

By contacting the schools, it was confirmed that Open Court has been used selectively, if at all, in many of the nineteen schools mentioned by Moustafa and Land. Only schools in the original Ten Schools Project have consistently used Open Court and then usually only in grades 1 and 2. In grades 3-5, several of the Ten Schools said they used other approaches, the most common called "Core Literature." Further, Open Court did not produce a comprehensive kindergarten program until 1995. Since that program was not adopted in California until the end of 1996, it was not used in any of the schools until 1997.

(12) Moustafa & Land, p. 4
We compared Open Court against the non-scripted programs in two ways: (1) average second to fifth grade SAT 9 reading increases/decreases at each school...In the second to fifth grade comparisons, we sorted the schools into one of three groups: those with increases of 10 or more percentile points between second and fifth grade, and those with decreases of 10 or more percentile points between second and fifth grade.

For purposes of evaluating the relative educational benefits of one school versus another, no informational value whatsoever is gained by comparing the numbers obtained by subtracting second from fifth grade 1999 SAT9 scores.

In dismissing Helfand and Ellerbee's interpretation of Sacramento's SAT9 scores, Moustafa and Land argued that proper assessment of learning requires one "To compare the same-or nearly the same-children," and to do so, they continued, "we need to compare scores from one grade, one year, to the next grade, the next year" (p. 3). Presumably, the approach they adopt here, of assessing learning by subtracting Grade 2 from Grade 5 SAT9 scores, is an effort to turn their argument to practice. But wait. All of the test scores used in their analysis are from the 1999 SAT9. The differences on which they have built their analysis were obtained by subtracting the SAT 9 scores for students completing Grade 2 in 1999 from those for students completing Grade 5 in 1999. There is no sense in which Moustafa and Land are comparing the same children across different grades, as they insisted was the proper procedure. Instead, they are comparing different children in different grades. Additionally, the children whose scores are subtracted one from the other are separated by four school years and come from schools with notoriously high mobility rates. Moreover, they come from schools where the past four years have witnessed a dramatic transition in early reading pedagogy.

Even so, Moustafa and Land's analysis is negated by a more fundamental flaw. In particular, their use of the difference between the Grade 2 and Grade 5 SAT9 scores fails to take into account the schools' overall educational performance. By their logic, for example, a school that performed in the 99th percentile in both second and fifth grade would be categorized as showing no sign of educational quality. To clarify the problem, consider the actual school comparison in Table 1.

Table 1

Moustafa and Land categorize the educational performance of the Sheridan School as outstanding and the Kelso School as subnormal

School %
Free
Lunch*
%
Cred.
Teachers*
%
Students
Tested*
SAT9
Reading
%-ile
Grade 2**
SAT9
Reading
%-ile
Grade 5**
SAT9
Change:
Grades
2-5
State
Rank*
Similar
School
Rank*
Sheridan(non-Open Court) 95 71 89 17 32 +15 1 3
Kelso(Open Court) 89 72 89 72 61 -11 10 10


 

Three sets of mean scores are provided here. The 2nd to 5th grade reading scores, the Statewide Rank, the Similar Schools Rank. The combination of the reading scores and the State and Similar Schools Rank gives a comprehensive view of a school's academic health. The first set of mean scores is for the original ten schools. The second set of mean scores is for the combined Ten Schools, now called the "10+2" schools. The third set of mean scores is for the original ten comparison schools.

Statistical analyses of these data confirm that, even including the two schools that did not join the Ten Schools Project until 1998-9 school year, the Open Court (OC) schools significantly outperformed the comparison schools as measured by their respective Statewide School Rankings, t(17) = 2.24, p = .04, or by their Similar Schools Ranking, t(17) = 3.30, p = .004 (all p values are for two-tailed comparisons). Further, a 2x4 (Project x Grades) analysis of variance affirmed a strong effect of project participation across all grades, F(1,20) = 279.57, p < .0001, as well as an effect of grades, F(3,60) = 5.62, p<.01, but no interaction between them, F(3,60) = 0.68, p=.57. The performance differences are even more pronounced when the original Ten Schools alone are compared to the original comparison schools. The point of these analyses is simply that given an appropriately constituted and a priori comparison sample, sensible measures of educational impact, and sound statistical methods, evidence exists that the Open Court schools outperformed the non-Open Court schools. In sum, Moustafa and Land's third major conclusion that the non-Open Court schools in the Los Angeles Unified School district outperformed comparable Open Court schools is erroneous.

Conclusion

Moustafa and Land's attempt to discredit Open Court is deeply flawed in both scholarship and argument. The facts are wrong and the analyses are misleading. The three major premises made in the paper are proven wrong. The record reveals that Open Court was recommended by all evaluation and adoption boards, that reading policy in California relied on a wide variety of research, not just one study, and that, when comparable schools were analyzed, the Open Court Schools outperformed the non-Open Court schools.

Bibliography

Adams, M. J, (1990). Beginning to read: Thinking and learning about print, Cambridge, MA: MIT Press.

American Fedration of Teachers, (http://www.aft.org/Edissues/whatworks/seven/oc/index.htm).

Anderson, R. C., Hiebert, E. H., Scott, J. A., and Wilkinson, I. A. G. (1985). Becoming a nation of readers: The report of the commission on reading. Champaign IL: Center for the Study of Reading.

Bond, G. L., and Dykstra, R. (1967). The cooperative research program in first-grade reading instruction. Reading Research Quarterly, 2, 5-142.

Resource Evaluation: Reading/Language Arts and English as a Second Language Adoption Report: Review of Instructional Resources: Kindergarten Through Grade Eight. (1997, p. 39). Sacramento, CA: California Department of Education, (The evaluations were finished and the report written by July 1996. It wasn't published in final form by the State of California printing house until 1997.)

Chall, J. S. (1967). Learning to read: The Great Debate. New York: McGraw-Hill. Collections for Young Scholars. (1995). Peru, IL: Open Court Publishing Company. Every child a reader: The report of the California Reading Task Force. (1995). Sacramento: California Department of Education.

Foorman, B.R., Francis, D.J., Fletcher, J.M., Schatschneider, C., and Mehta, P. (1998). The role of instruction in learning to read: Preventing reading failure in at-risk children, Journal of Educational Psychology, 90 (1), 37-55.

Moustafa, M, & Land R. (2000). The research base of Open Court and its translation into instructional policy in California, found on the web at (http://curriculum.calstatela.edu/faculty/mmousta).

Moustafa, M. (2000). The research base of Open Court and Its Translation into Instructional Policy in California -- The Data. (http://curriculum.calstatela.edu/faculty/mmousta/Data)

National Assessment of Educational Progress, (1994). The NAEP 1992 Technical Report. Princeton, NJ: Educational Testing Service

National Assessment of Educational Progress, (1995). NAEP 1994 Reading: A First Look-Finding from the National Assessment of Educational Progress (Revised Edition). Washington, DC: U.S. Government Printing Office

National Reading Panel, (2000). Teaching children to read: An evidence-based assessment of the scientific research literature on reading and its implication for reading instruction. Rockville, MD: National Institute of Child Health and Human Development, National Institutes of Health.

National Research Council. (1998). Preventing reading difficulties in young children. Washington, D.C.: National Academy Press.

Office of Student Integration, (2000). 1999 Academic Performance Index, (API) report TSP/original comparison schools. Los Angeles, CA: Los Angeles Unified School District.

Public Schools Accountability Act, (1999). (www.cde.ca.gov/psaa/api/base/fachsheet.pdf)

Star Data, (1999). (http://STAR.cde.ca.gov/)

Taylor. D. (1998). Beginning to Read and the Spin Doctors of Science. Urbana, Ill: National Council of Teachers of English.

Teaching Reading: A balanced comprehensive approach to teaching reading in preK-grade 3. (1997). Sacramento, CA: California State Board of Education

Author Note

The author wishes to acknowledge and thank the following colleagues who have provided information and/or editorial support: Marilyn Jager Adams, Barbara Foorman, and Reid Lyon. Hollis Scarborough also is gratefully acknowledged for providing statistical support.




back to top