Inclusive Classroom Profile (ICP): Findings from the first US Demonstration Study Paper Presented at the AERA Annual Meeting, San Francisco, CA April, 27 – May 1, 2013 Elena P. Soukakou, Tracey A. West, Pam J. Winton, John H. Sideris Frank Porter Graham Child Development Institute, University of North Carolina at Chapel Hill Lia M. Rucker, North Carolina Rated License Assessment Project 1 Inclusive Classroom Profile (ICP): Findings from the first US Demonstration Study Objectives The Inclusive Classroom Profile (ICP) (Soukakou, 2012) is a classroom observation measure designed specifically to measure the quality of inclusive classroom practices. It is based on research evidence on the effectiveness of specialized instructional strategies for meeting the individual needs of preschool children with disabilities in inclusive settings (Odom, 2004; National Professional Development Center on Inclusion, 2011; Buysse and Hollingsworth, 2009). This session will share findings from the first demonstration study in the US conducted by FPG Child Development Institute, University of North Carolina. The study investigated the acceptability, reliability and validity of the ICP measure in collaboration with the NC Rated License Assessment Project, the Division of Child Development and Early Education, and the NC Department of Public Instruction, Exceptional Children. This study involved field testing the ICP in 51 settings in North Carolina with the following objectives: (1) Examine the acceptability of the ICP measure for program evaluation within an accountability structure such as quality rating systems. (2) Assess the effectiveness of training procedures for ensuring accurate administration. (3) Assess the psychometric properties of the ICP, including the measure’s inter-rater reliability, factor structure and construct validity. (4) Explore the relationship between inclusive quality and various child and classroom characteristics (e.g. relationship between quality and teachers’ special education qualifications, number of children with disabilities and severity of disability). Background/Theoretical Framework Never has so much attention been paid to improving the quality of early childhood programs serving high - need children. Through the Race to the Top –Early Learning Challenge Program and other reform efforts, states are being challenged to develop systems for rating, monitoring and improving early learning and development programs. States need reliable, valid tools for measuring classroom quality that are sensitive to and inclusive of each and every child (Odom, Buysse, & Soukakou, 2011). The ICP was developed in response to a lack of validated instruments designed specifically to measure the quality of inclusive practices (Odom & Bailey, 2001). It was designed to complement existing program quality measures with new quality indicators that assess specific, 2 classroom practices that support the developmental progress of children with disabilities in early childhood settings. In the form of a structured observation 7- point rating scale, the ICP includes 12 items that are assessed predominately through observation. A few quality indicators are assessed through teacher interview. Examples of the ICP items include the following: adaptation of space and materials, adult involvement in peer interactions, adaptation of group activities, support for communication, and family-professional partnerships. The development of the ICP was founded on a recognized premise; that is, what constitutes quality in programs where diverse learners such as children with disabilities need additional supports is possibly more multifaceted than what current quality assessment systems and tools measure (Gallagher & Lambert, 2006; Odom & Bailey, 2001; Spiker et al., 2011). An important assumption guiding the conceptualization of the ICP is that specialized instructional strategies that have research support are more likely to meet the needs of young children with disabilities in early childhood settings. For example, naturalistic instruction, which includes a set of practices such as incidental teaching methods, embedded strategies, scaffolding strategies, and peermediated interventions have empirical support and are generally encouraged in inclusive classrooms (Odom, 2004; National Professional Development Center on Inclusion, 2011). The ICP has been used in a pilot study in the United Kingdom with promising results for the measure’s reliability and validity (Soukakou, 2012). This is the first US demonstration study leading to validation. Methods The ICP was field tested in 51 inclusive preschool classrooms. Criteria for program participation in the study included: center-based, served preschool age children (three-five years of age), and served at least one child and no more than 50% with an Individualized Education Plan (IEP). To field-test the ICP measure, each inclusive classroom had an ICP assessment conducted by a trained assessor over a four month period. In addition, Early Childhood Environment Rating Scale assessments (ECERS-R; Harms, Clifford and Cryer, 2005) were gathered in all classrooms during the same data collection period for comparison to the ICP. Each classroom observation lasted approximately three hours. Classroom characteristics Of the 51 settings, 39% were child care programs, 26% were developmental day programs, 26% were Head Start programs and 9% were public preschools. In total, classrooms included 150 children with an identified disability (child with an IEP). The number of children with a disability per classroom ranged from one to eight children, with each classroom serving in average 2.94 children with an IEP. Teacher education level ranged from a high school diploma through a Master’s degree, with 51% of the teachers reporting having received a Bachelor’s degree. 3 Area and severity of disability For each child with a disability (N=150) teachers were asked to identify children’s area and severity of disability1. Children’s disabilities were identified in the following areas: behavior/social (67%), gross motor (27%), fine motor coordination (45%), sensory integration (27%); and intentional communication (90%). With regard to severity of disability, 59% of the classrooms had a least one child with a disability identified at the ‘severe’ level (1-4 point scale rating possible/mild/moderate/severe level), while 88% of classrooms had at least one child with a moderate or severe level of disability. Data sources Four assessors from the North Carolina Rated License Assessment Project (NCRLAP) collected data using the following measures: Inclusive Classroom Profile (ICP) (Soukakou, 2012), the ICP child and classroom demographic form, and the Early Childhood Rating Scales (ECERS-R) (Harms, Clifford, & Cryer, 2005). The ECERS-R was selected for use in this study because it was one of the main measures that served as a foundation for developing the ICP. Given that the ECERS-R also measures provisions for children with disabilities, correlations between the two measures could provide important information about the construct validity of the ICP and the contribution of the ICP items above and beyond those included in the ECERS-R. The NCRLAP has the contract in NC to collect ECERS-R data for NC’s Quality Rating and Improvement System (QRIS). For this study, the assessors were formally trained in using the ICP measure and classroom demographic form developed by the study researchers. The demographic information was gathered through teacher interview. Information on the acceptability and usability of the ICP was gathered through a social validity survey completed by the assessors. Assessors also participated in a full-day focus-group to provide feedback on the measure and its application. Results and Conclusions Objective 1: Examine the acceptability of the ICP measure for program evaluation. Results from the social validity survey (1-5 point scale), indicated that assessors rated the importance of the constructs measured by the ICP very highly (m= 5) and that they would highly recommend the ICP measure to others (m=5). With regard to the usability of the ICP, the assessors found the measure easy to use (m= 4). Finally, assessors rated the ‘Overview’ part of the training (information on administration and scoring, including video clips for practice purposes) as useful (m= 3.75), and all four assessors reported that they felt well prepared after the reliability training observations (m=4). Findings from the focus-group feedback meeting confirmed the high level of acceptability of the ICP measure by the assessors who used the ICP 1 Teachers could identify one or more areas of need for each child with an IEP. 4 in the study, and provided specific recommendations for improving the administration, scoring and training procedures. Objective 2: Examine the effectiveness of training procedures for accurate and reliable administration. Assessors received a three-hour training session on ICP administration and scoring followed by four reliability observations in inclusive classrooms to assess accuracy of administration and reliability proficiency against the ICP author standard. Each assessor (n=4) met a reliabilityproficiency criterion of 85% agreement within one scale point, maintained for three consecutive reliability observations against the ICP author’s rating. Mean inter-rater agreement across assessors was 98% with a range of 91-100%. Objective 3: Assess the psychometric properties of the ICP, including the measure’s inter-rater reliability, factor structure and construct validity. Inter-rater reliability was assessed in 9 reliability observations (18% of the sample), distributed over the four month data collection period. The mean inter-rater agreement across reliability observations was 87% (within one point scale). Reliability at the item level was assessed via an intra-class correlation coefficient (ICC) ranging from zero to one where one indicates perfect agreement. For most items agreement between raters was in an acceptable range (.51-.99) with a mean of .71. Only one item (Item 3) presented with low agreement (.11) indicating the need for further clarification of the item in the scoring protocol. Accuracy of ICP administration by assessors against author standards was assessed in 9 classroom observations. The mean agreement between assessor’s scores and ICP author standards (consensus score) was 94% (within one) and 75% (exact agreement) indicating that assessors maintained accurate administration of the ICP throughout the study. Internal consistency was assessed through Cronbach’s alpha analysis for the scale’s 12 items. This analysis suggested that the scale’s items were highly consistent (α = .85). Structural validity was assessed through exploratory factor analysis of the ICP items. A scree plot examination (see figure 1) of the variance explained by the number of factors, clearly showed a one-factor structure with good factor loadings (.34 - .84). Construct validity was examined through correlation analysis between the ratings on the ICP in and ratings on the Early Childhood Environment Rating Scales (ECERS-R; Harms, Clifford and Cryer, 2005). A moderate high correlation was found between the total score of the ICP and the ECERS-R (rho =.48) suggesting that the two instruments are correlated but not measuring identical constructs. A pattern of weaker correlations between the ICP and certain ECERS subscales, such as the ‘health and personal care routines’ that involve practices not measured by 5 the ICP (rho =.21) and higher correlations with ECERS subscales that measure constructs more closely related to the ICP ( ‘interactions’; rho = .38) provide preliminary evidence for the construct validity of the ICP. Table 1 presents all correlations between the ICP and ECERS-R total and item/subscale scores. Objective 4: Examine the relationship between the quality of observed inclusive practices and child and classroom characteristics. Multiple regression analyses were conducted to explore the association between observed quality of inclusive practices (as measured by the ICP) and the following classroom characteristics that were predicted to be associated with higher quality on the ICP: teacher education level, teacher special education training (course hours in special education), number of children in the classroom with an IEP, ratio between children and adults, and program type. Teacher education level was correlated at r = .41, p =.003; number of course hours in special education at r = .36, p =.014; number of children with an IEP at r = .37, p =.007; and ratio between children and adults in the classroom was correlated r = -.57, p <.001. Simple mean differences between program types were tested with one-way ANOVA. It was expected that developmental day programs and perhaps Head Start programs would have higher ICP scores because of their histories in serving young children with disabilities. This model testing was significant, F (3,47) = 13.77, p < .05 and accounted for a large amount of the variance, R2 = .47, indicating that differences in program type account for nearly half of the variance in ICP scores. The model partially confirmed the expected mean differences. Developmental day programs did indeed score highest and were significantly higher than child care programs but not significantly higher than the other two program types. Given the high correlations of the ICP with the predicted variables described above and the differences between program types, an exploratory analysis was conducted to further test the relationship between the ICP and those predictors. The following predictors of ICP scores were tested in a hierarchical series of five models with a predictor variable added at each step: teacher education level, number of course hours in special education, number of children in the classroom with an IEP, ratio between children and adults, and program type. ECERS-R total score was also entered as a predictor. The initial model tested for differences in ICP scores based on program type. Adding the predictor variables in sequence enabled us to test not only the unique contribution of each of the predictor variables, but also whether the presence of one affects the apparent impact of the others. Of particular interest were the expected program type differences and whether those differences were collinear with the other predictors. The significant difference that was found for the child care programs compared to all other program types was maintained upon controlling for teacher education, ECERS-R total score, course hours in special education, number of children with an IEP, and adults to children ratio. No other predictors were significant. 6 Significance This demonstration study provided further evidence for the reliability, validity, and usability of the ICP. Findings from the present study have implications for use of the ICP in research, quality rating systems and professional development. For instance, as a research instrument, the ICP could be used to assess quality across inclusive programs as well as to evaluate the relationship between the quality of inclusive practices and outcomes for children with disabilities and their families. As one component of a quality ratings system, the ICP could be used to evaluate the quality of inclusive practices. Finally, as a quality improvement tool, the information gathered on ICP could inform professional development models and enhance program quality improvement activities. 7 References Buysse, V., & Hollingsworth, H. L. (2009). Research synthesis points on early childhood inclusion: What every practitioner and all families should know. Young Exceptional Children, 11, 18-30. Gallagher, P. A., & Lambert, R. G. (2006). Classroom quality, concentration of children with special needs and child outcomes in Head Start. Exceptional Children, 73(1), 31–53. Harms, T., Clifford, R. M., & Cryer, D. (1998/2005). Early Childhood Environment Rating Scale-Revised Edition. New York, NY: Teacher’s College Press. National Professional Development Center on Inclusion. (2011). Research synthesis points on quality inclusive practices. Chapel Hill: University of North Carolina, FPG Child Development Institute, Author. Retrieved from http://community.fpg.unc.edu/npdci Odom, S. L., & Bailey, D. B. (2001). Inclusive preschool programs: Classroom ecology and child outcomes. In M. Guralnick (Ed.), Focus on change (pp. 253–276). Baltimore, MD: Brookes. Odom, S. L., Buysse, V., & Soukakou, E. (2011). Inclusion for young children with disabilities: A quarter century of research perspectives. Journal of Early Intervention, 33, 344-356. Odom, S. L., Vitztum, J., Wolery, R., Lieber, J., Sandall, S., Hanson, M. J., & Horn, E. (2004). Preschool inclusion in the United States: A review of research from an ecological systems perspective. Journal of Research in Special Educational Needs, 4, 17-49. Soukakou, E. P. (2012). Measuring quality in inclusive preschool classrooms: Development and validation of the Inclusive Classroom Profile (ICP). Early Childhood Research Quarterly, 27,478-488. Spiker, D., Hebbeler, K. M., & Barton, L. R. (2011). Measuring quality of ECE programs for children with disabilities. In M. Zaslow, I. Martinez-Beck, K. Tout, & T. Halle (Eds.), Quality measurement in early childhood settings (pp. 229–256). Baltimore, MD: Brookes. 8 Table 1 Rank-Order Correlations Between ICP and ECERS-R ECERS-R Scale Space and Furnishings Personal Care Language and Reasoning Program Structure Activities Interactions Parent and Staff ICP Total 0.48*** 0.21** 0.47*** 0.29* 0.30* 0.38** 0.38** ECERS Total Score 0.48*** Note: *p<.05, **p<.01, ***p<.001 Figure1. Scree plot displaying the amount of variance explained by the number of factors entered in ICP exploratory factor analysis. 9