Jim Hill High School

Empirical Research-Assessments
Home
INTASC Standards Reflection (Behavior Management)
INTASC Standards Reflection (Reading)
Model Standards for Beginning Teacher Licensing, Assessment and Development
Interstate New Teacher Assessment and Support Consortium (INTASC)
Position Paper: Classroom Management
Position Paper: The Foundation of Every State is its Youth
PowerPoint of a Lesson
April Logs
Empirical Research- Problem Based Learning/ Instruction
March Logs
Research Paper- Methods and Strategies of Teaching
February Logs
Two Week Lesson Plan (April 21 & April 28) Integrated Theme
Two Week Lesson Plan (March 31) 10 Week Thematic Unit
Two Week Lesson Plan (April 21 & April 28) Block Schedule
Lesson Plan (Reteaching)
Lesson Plan (Diversity)
Lesson Plan (Technology)
My Journey Through Student Teaching
January Logs
Case Study Reflective Journals
Empirical Research-Assessments
Empirical Research- At-Risk Students
Empirical Research- Teacher Education
Empirical Research- Early Intervention
Position Paper: Rationales for Discrepancies between Abilities and Achievement
Position Paper: The Advantages and Needs of INTASC Standards for Pre-Service Teachers
How INTASC Standards are Applied in my Lesson Plans
Position Paper: Parent Involvement
Behavior Management Mini-Portfolio
What is an Effective Teacher
Resume
No Child Left Behind Act
Two Week Lesson Plan (Jan. 28 & Feb. 11) Block Schedule
Two Week Lesson Plan (March 10 & March 25) Block Schedule
Two Week Lesson Plan (Feb. 18 & March 3) Block Schedule
Basic Philosophies of Education
Why I Want to be a Teacher
Philosophy of Education

Enter subhead content here

Empirical Article- Assessments 

 

Reference:

Burns, M. K., & Senesac, B. V. (2005). Comparison of dual discrepancy criteria to assess response to intervention. Journal of School Psychology, 43, 393-406.

 

Purpose:

The study furthers previous research concerning dual discrepancy (level and growth) by comparing the validity and diagnostic implications of four dual discrepancy (DD) models. There were two main questions that were addressed. They were the following: Which approach of DD best differentiates reading skills as measured by a norm-referenced test of reading achievement? What prevalence rates are associated with the four DD models?

 

Population/Sample:

There were 146 students in this study from first (14.6%), second (37.7%), and third (47.7%) grades that were nominated by their teacher as having reading difficulties. All of these individuals scored at or below the 25th percentile on a district-administered group test of reading. There was an equal distribution of females (50%) and males (50%).  Among these students were 27.8% Caucasians, 2.8% African Americans, and 2.6% Hispanics. 57.71% of these students were eligible for the federal free of reduced lunch program.

 

Methods:

-There were two interventions that were used in this study. The students participated in remedial program entitled, Help One Student to Succeed (HOSTS), or they received Title I support. HOSTS is a comprehensive literacy program that supplements classroom reading instruction by trained volunteer tutors. Individual reading and computer-based assessments are conducted with children having reading difficulties. This data is used to make personalized interventions for each student. This includes daily and weekly lesson plans and tutoring for four thirty-minute sessions each week. Tutoring was available Monday through Friday and students may have had different tutors in these days. The HOSTS teacher is present during tutoring sessions to monitor the instruction and provide feedback to the tutors.

- Because the intervention can vary, Title I was selected. This included weekly individual reading instruction from the Title I consultant or small group instruction in class or in a pull-out model. Five of the nine schools used the HOSTS program and four used Title I services. Also, five schools were randomly selected from all the schools that used the HOSTS program. Five elementary schools not using the HOSTS program were selected using student enrollment, percentage of students receiving free or reduced lunch, average student to teacher ratio, and percentage of children who passed the fourth grade tests.

- Data from the Dynamic Indicators of Basic Early Literacy Skills (DIBELS) test were collected mid-year and end-of-year. DIBELS is an assessment developed from the curriculum-based measurement model to assess early literacy development (phonological awareness) for kindergarten and first graders.

-The Gray Oral Reading Test (GORT-4) was used to study level or reading proficiency.

-Students were seen individually in January to take the tests, and again in May of the same school year.

- The DIBELS median fluency rate for the end-year assessment was the post-intervention reading level score. The mid-year score was subtracted from the end-year score and this was the growth score. This score was then rank ordered within three grades to create responsiveness groups using normative criteria. Student response was examined in which the mean growth was computed. Those students whose post-intervention or end-year DIBELS score fell within the “at- risk” criterion and whose fluency change score fell at or below the non-responsiveness criterion were considered dually discrepant.

 

Findings:

-All three percentile-rank models greatly differentiated the reading scores of DD from other students, but the one standard deviation approach did not. Reading scores were compared according to two groups of student response. The three percentile rank criteria largely differentiated the groups’ reading scores, and the one standard deviation criterion did not. A chi squared statistic was computed for those with DD and non-DD. The 50th percentile led to a significant effect for grade, but this may have been because of an error.

-There were 23.3% for the 25th percentile, 28.1% for the 33rd percentile, 41.8% for the 50th percentile, and 12.3% for the one standard deviation criterion. There were 74 students who received the intervention. Eight were DD for the 25th percentile, 11 for the 33rd percentile, 22 for the 50th percentile, and 3 for the one standard deviation group. The prevalence rate for children in 1st, 2nd, and 3rd grades was 3.7%, for 25th percentile, 4.4% for the 33rd percentile, 6.6% for the 50th percentile, and 1.9% for the one standard deviation below the mean.

 

  

Implications for Education

 

This study showed that dual discrepancy (DD) was most effective for the 25th and 33rd percentile rank groupings than for the 50th percentile and one standard deviation criteria. According to the author, because there was consistency in the prevalence rates, the DD method seemed to be reliable. The author states that this data should be compared to national prevalence data for learning disabilities to thoroughly understand the utility of this model. He also states that further research needs to be done to find best practices in the diagnosing of learning disabilities within a response to intervention model.

 I think that perhaps the results may have had to do with the fact that the Title I program allowed for group instruction, whereas the HOSTS program did not. Also, the results may have not been as reliable based on the fact that in the HOSTS program, the students may have received a different tutor (with a different teaching method) each time he or she went to tutoring. In addition, I would want to know the difference in the results, for example, if an Informal Reading Inventory was administered as well.

Furthermore, I believe that no one or two tests can determine everything a student knows or does not know. There are many factors that can play a role in the results of tests scores. For example, a child may be an intelligent individual, but may be having a bad day or isn’t feeling well. That is a reason why teachers should administer ongoing assessments. By doing this, they will have more reliable results.

Although the author stated that his study was reliable, I am not too sure that it was. In this study, he stated that there may have been an error in the scores. In addition to this, he stated that further research needed to be done to refine DD and that additional comparisons of definitions for non-responsiveness of growth data was needed. More research should have been done in this study to truly make it reliable.

 

 

Enter supporting content here