CARS Projects And Rearch

 

CARS' Response to Coles' Misreading Reading

Children Reading

May 3, 2000

We would like to correct several errors in Coles’ (2000) account of our 1998 article published in the Journal of Educational Psychology.

  1. Coles claims that students at the control school were not tutored and that we dropped the control school in the JEP article because the school was too high poverty and the students were "tough." (p. 33) This claim is incorrect. Students at this unseen control were tutored. The school was never dropped, but, in fact, appears in the JEP article. Critics, including Coles and D. Taylor have misrepresented the purpose of this group in our study. In fact, its role in the design was to serve as a control on the professional development that was provided to teachers who delivered the three study curricula. That is, teachers at this control school participated in the District’s usual staff development rather than the professional development provided by our research staff.

  2. Coles criticizes our use of "a less-stringent standard" (p. 33, bottom) to point out that group differences on the Woodcock-Johnson Passage Comprehension (PC) favoring the Direct Code group were meaningful. Coles is referring to our use of effect sizes - the most widely used statistic for reporting the strength of statistical effects. The effect size on PC for the direct code vs. whole-language contrast was .76. Effect sizes of .20 are typically considered weak; .50 moderate; .80 and larger are considered strong. Thus, the effect size of .76 is fairly large and is of practical significance. Coles’ statement implies that we were misrepresenting the data.

  3. Coles’ recalculation of our Woodcock-Johnson data: In the spirit of professional collegialism, we sent Coles a disk of the data for the 1998 JEP study when he requested our data. The way in which he seemed to manipulate the PC means in his commentary to Education Week (January 27, 1999) is referred to in our response published in Education Week (April 21, 1999) and detailed on our website (http://cars.uth.tmc.edu). A copy of this response is attached. At that time, Dr. Coles’ would not share his calculations with us. Now that we have access to his calculations from the book, we see that he did not delete data based on classroom means, but rather deleted data from all classrooms in a given curriculum in a particular grade at a particular school based on the school-level mean. In particular, Coles’ removed the highest scoring Direct Code school in Grade 1 (2 classrooms with 5 students in each) and the lowest scoring school in the Implicit Code instructional group in Grade 2 (6 classrooms with 1, 2 or 3 students in each). He then "eyeballs" the remaining school means without regard for the number of classrooms involved in computing those means, or the number of students involved in computing the classroom level means that make up the school means. Coles gives all school means equal weights and, in so doing, dismisses the low scores from the Implicit Code school with the most students.

    Coles further claims (p. 35) that the Direct Code group’s higher PC scores came from two classrooms in a school in which the scores of the students were "uncharacteristically uniform, unlike the considerably varied student scores in all other classrooms and schools." This statement is not correct. The two smallest classroom standard deviations were 1.0 and 4.9 for Implicit Code, as compared to 2.6 and 4.8 for Embedded Code, and 4.8 and 5.8 for Direct Code. If anything, the Implicit Code classroom with a smaller standard deviation is more "uncharacteristically uniform." We disagree with his emphasis on school means over classroom means. But given his desire to focus on school means, we note his failure to emphasize the differences in outcomes in the one school where both Direct Code and Implicit Code were used. Specifically, in that one school, the two Direct Code classrooms had a combined average of 100.4 in Grade 1, whereas the two Implicit Code classrooms had a combined average of 93.4. In Grade 2, the combined average for the two Implicit Code classrooms was 95.8 as compared to 93.3 for the two Direct Code classrooms. These are clear examples of Coles’ jaundiced and biased "re-analysis" of the data in the 1998 paper.

  4. Coles misrepresents the word reading differences as levels of attainment rather than rate. Thus, it’s not that 38% of the whole-language students could read 2.5 words or fewer out of 50. The correct statement is that 38% of the whole-language students improved in word reading at a rate of 2.5 words or less per year. This rate estimate has been miscontrued by Coles and by others. We have emphasized that one cannot interpret this measure in an absolute sense, but only as it relates to the experimental word list used in the study. It is, however, perfectly legitimate to compare this rate to the mean rates of growth observed in the other instructional groups. The results speak for themselves in showing that many children in the Implicit Code and Embedded Code groups do not change over the year.

  5. Coles misrepresents the pre-publication history of the 1998 JEP article. For an accurate accounting, see our response to B. Taylor et al. on our website.

  6. Our 1998 study is only one study out of 30 years of studies that have pointed to the importance of explicit alphabetic instruction for children at-risk for reading difficulties (see Fletcher & Lyon, 1999, for a review). We clearly state 9 specific limitations of the 1998 study in the conclusions. Coles ignores our statement of limitations and ignores the context in which this study resides.

  7. Personal attack on Foorman and Adams: Marilyn Adams is not lying, as Coles suggests, when she said that my request to use Open Court "came out of the blue." (p. 44) Our 1993 grant proposal had indicated that we would use DISTAR. The school district at the last minute said that we could not use DISTAR and that if we wanted a DI approach that they recommended Open Court Reading because they used Open Court Math. So the district directed us to Open Court and Foorman was not even aware that Open Court had been rewritten when she made the request. For that matter, she and Adams did not have any formal or colleagial relationship at that time, knowing each other only through reputation. A relationship did develop, which was stimulated by their appointments to the NRC committee that produced the report "Preventing reading difficulties in young children." Why Coles needs to attribute motive and malign both Adams and Foorman is mysterious.

Coles is also incorrect in his description of how Phonemic awareness in young children (Adams, Foorman, Lundberg, & Beeler, 1998) came about. Foorman had written the Lundberg, Frost, and Petersen (1988) phonological awareness activities into our proposal to NICHD in 1992-93. Ingvar Lundberg told Foorman about an English translation of the activities by Adams and Huggins. Foorman contacted Marilyn Adams to ask for a copy for research purposes. Adams graciously agreed. We used her translation during the 1993-94 and 1994-95 school years, with teachers giving us feedback that we incorporated into the activities. People started contacting us for copies of the revised curriculum. Thus, after this kindergarten research was complete in 1996-97, Adams and Foorman began to prepare the curriculum for publication. So as to avoid any perception of conflict of interest, Foorman divested herself of royalties from the sales of the publication. When Coles contacted Foorman about the publication date for Phonemic awareness in young children she explained that it was 1998, long after our kindergarten studies had ended, and that she had divested herself of royalties. Nevertheless, Coles depicts Foorman as highly unethical (p. 50): "...I am not certain where the royalties will go, but direct financial benefit from the book’s publication is not my concern. My concern is about interlocking research and collaboration that can create conflicts of interest, which in turn can impair objectivity. Capital in the form of professional capital can be accrued in various ways, such as through enhancing one’s professional reputation, status, and influence and by obtaining research and program grants. One way of earning these kinds of professional gains is through an educational program that is highly regarded and widely used." Again, what motivates Coles to malign researchers with whom he simply disagrees?

When the Coles book first came out we disregarded it as so preposterous and personal that no one would take it seriously. Since then it has shown up, along with Denny Taylor’s (1998) Spin Doctors of Science, on the desk of officials in school districts where we conduct research studies. In one district, officials suddenly began reconsidering their approval for our study because of the suggestions in these books that we were unethical. The subtitle of Coles book is particularly worrisome from a school district perspective: "The bad science that hurts children." No matter that Coles never tells us what the good science is that helps children. The title does sell books and scares non-researchers. However, in this book (Misreading Reading) the personal attacks are inaccurate, unprofessional, and unforgivable, and harmful to reading research and the children it serves.

Sincerely,

Barbara R. Foorman, Ph.D. and Jack Fletcher, Ph.D.




back to top