Massachusetts Part B Annual Performance Report for FFY 2006 Submitted to the Office of Special Education Programs: February 1, 2008 Revised April 14, 2008 Massachusetts Department of Elementary and Secondary Education 350 Main Street, Malden, MA 02148 Phone 781-338-3000 TTY: N.E.T. Relay 800-439-2370 www.doe.mass.edu Part B Massachusetts Annual Performance Report (MA APR) for FFY 2006 Table of Contents Cover Letter / Overview of MA APR Development 3 Monitoring Priority: FAPE in the LRE Indicator #1: Graduation Rates 4 Indicator #2: Drop-Out Rates 6 Indicator #3: Assessment 11 Indicator #4: Suspension/Expulsion 16 Indicator #5: School Age LRE 19 Indicator #8: Parent Involvement 24 Monitoring Priority: Disproportionality Indicator #9: Disproportionality – Child with a Disability 27 Indicator #10: Disproportionality – Eligibility Category 31 Effective General Supervision / Child Find Indicator #11: Initial Evaluation Timelines 34 Effective General Supervision / Effective Transition Indicator #12: Early Childhood Transition 36 Indicator #13: Secondary Transition Effective General Supervision / General Supervision Indicator #15: Identification and Correction of Noncompliance 40 43 Indicator #16: Complaint Timelines 48 Indicator #17: Due Process Timelines 51 Indicator #18: Hearing Requests Resolved by Resolution Sessions 53 Indicator #19: Mediation Agreements 55 Indicator #20: State Reported Data 57 Appendices Appendix A: MA Parent Survey for Indicator #8 59 Appendix B: Indicator #15 Worksheet 61 Appendix C: Web-Based Monitoring System 65 Appendix D: Table 7 - Report of Dispute Resolution for FFY 2006 66 Appendix E: Indicator #20 Scoring Rubric 67 Note: Further information on the Indicators listed above and complete information for Indicators 6, 7, and 14 can be found in the Massachusetts State Performance Plan (MA SPP) found at http://www.doe.mass.edu/sped/spp/. 2 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008; Revised April 14, 2008 U.S. Department of Education ATTN: Janet Scire / Mail Stop 2600 7100 Old Landover Road Landover, MD, 20785-1506 Dear Ms. Scire: Enclosed is the Massachusetts Annual Performance Report for FFY 2006 (MA APR). The MA APR responds directly to the indicators identified by the Office of Special Education Programs (OSEP) in Information Collection 1820-0624, Part B State Performance Plan (SPP) and Annual Performance Report (APR). The MA APR provides information on Indicators 1, 2, 3, 4a, 5, 8, 9, 10, 11, 12, 13, 15, 16, 17, 18, 19, and 20, and addresses any areas identified in OSEP’s letter responding to the MA SPP/APR submitted February 1, 2007. The Massachusetts Department of Education (MASSDE) has engaged in a variety of activities to obtain broad input from stakeholders on the development of the MA APR. MASSDE convened the Massachusetts Statewide Special Education Steering Committee – which consists of state special education advisory council members, key MASSDE personnel, local education officials, parents, advocates, and representatives from higher education, charter schools, approved private special education schools, and adult service agencies – to review data, measure progress against the targets, examine methodologies, and identify key activities as appropriate for each of these indicators. Additionally, MASSDE has formed targeted interest groups focused on each indicator. These workgroups incorporate a wide variety of stakeholders who communicate throughout the year to help guide Massachusetts’ work in each area. Regarding public dissemination, the completed MA APR will be made widely available for public discussion. This will be accomplished by broad discussion in interest groups (as previously mentioned) and at the Statewide Advisory Council meeting and other conference and group discussion opportunities. MASSDE will post the MA APR on the MASSDE website at http://www.doe.mass.edu/sped/spp/, and distribute hard copies of the report to key constituents and the media. MASSDE has developed a resource titled “Special Education Results” that has been added to each district’s profile on the MASSDE website to meet the requirement that district indicator data be publicly reported (http://profiles.doe.mass.edu/special_ed.aspx). Data are currently presented in table format for Indicators 1, 2, 3, 4a, 5, and 12, and are presented through thematic maps for Indicators 1, 2, and 5. MASSDE is currently working to determine the best way to publicly report data for Indicators 8, 9, 10, 11, 13, and 14, and anticipates that data for these indicators will be available on the MASSDE website by the February 1, 2009 submission of the MA APR. If questions or additional clarification is needed regarding the MA APR, please contact me at 781.338.3388 or mmmittnacht@doe.mass.edu. Sincerely, Marcia Mittnacht State Director of Special Education Special Education Planning and Policy Development Office Massachusetts Department of Elementary and Secondary Education Cc: Jeffrey Nellhaus, Acting Commissioner of Education, Massachusetts Department of Elementary and Secondary Education Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Monitoring Priority: FAPE in the LRE Indicator 1: Percent of youth with IEPs graduating from high school with a regular diploma compared to percent of all youth in the State graduating with a regular diploma. (20 U.S.C. 1416 (a)(3)(A)) Measurement: The statewide Graduation Rate is the number of students in a cohort who graduate in four years or less, divided by the number of first-time entering 9th graders in that cohort. The denominator is adjusted so that students who transfer into Massachusetts’ public schools are added to the original cohort and students who transfer out, or who are now deceased, are subtracted from the original cohort. The quotient is multiplied by 100 to express the Graduation Rate as a percentage. The measurement for youth with an IEP is the same as the measurement for youth without an IEP. FFY 2006 (2006-2007) Measurable and Rigorous Target Students with IEPs Graduation Rate: 61.7% Actual Target Data for FFY 2006 (2006-2007): IEP Non-IEP All Students # of Students in 2006-07 cohort # of Students in 2006-07 cohort who graduated in four years or less. 2006-07 Graduation Rate 13,594 62,318 75,912 8,538 52,880 61,418 62.8% 84.9% 80.9% Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): From FFY 2005 to FFY 2006, the Graduation Rate for students with disabilities in Massachusetts increased by 1.1 percentage points, from 61.7% to 62.8%. Likewise, the Graduation Rate for students without disabilities increased from 83.9% to 84.9%, and the overall state Graduation Rate increased from 79.9% to 80.9%. The improvement in the Graduation Rate of students with disabilities in Massachusetts may be related in part to a number of improvement activities that were completed in FFY 2006. These include further development of public reporting of Special Education data, a variety of professional development initiatives for educators, and certain targeted academic support programs and grants. Over this past year, MASSDE further developed its public reporting of Graduation Rates and other Special Education data. MASSDE worked with stakeholder groups and staff from a number of school districts to determine the best way to publicly report graduation data at an LEA level. As a result of these efforts, MASSDE has developed a resource titled “Special Education Results” that has been added to each district’s profile on the MASSDE website to meet the requirement that district indicator data be publicly reported. Data for Indicator 1 are reported through tables, and thematic maps have been Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 developed to display the graduation rate for students with disabilities for each district: http://profiles.doe.mass.edu/special_ed.aspx. MASSDE also provided a number of professional development activities in FFY 2006 to support educators and administrators in helping students with disabilities reach graduation and attain positive post-secondary outcomes, including: Graduation Rate Summit - March 5, 2007: MASSDE co-sponsored this statewide conference for decision-makers and leaders in education, which included a variety of presentations and breakout sessions around issues related to graduation and dropout. Special Education Summer Institutes – Summers 2006 and 2007: MASSDE sponsored a variety of Institutes with topics such as “Effective Evaluation of Special Education Programs”, “Assistive Technology and Universal Design in the Classroom”, “Managing Behavior in an Inclusive Classroom”, “Assessing English Language Learners with Disabilities”, and the “Special Education Administrative Leadership Academy”. Individuals in each Institute participated in at least 45 contact hours and created a final product related to the topic of the institute. Project FOCUS Academy (PFA) – 2005-2008: This pilot distance-learning program, funded through a federal State Improvement Grant, included courses for educators in Universal Design for Learning, Transition/Post-School Outcomes, and Positive Behavioral Interventions and Supports. Participating schools then received grants to fund school-improvement initiatives and systems reform that benefit students with disabilities. Massachusetts FOCUS Academy (MFA) – 2007-2012: MASSDE is developing this online professional development tool based on the PFA pilot model. MFA, which is funded through a federal State Personnel Development Grant, will ultimately increase professional development opportunities for educators across the state. Special Education Program Improvement Grants – yearly: MASSDE administers non-competitive grants to school districts to support professional development around special education-related topics and induction/mentoring programs for new special educators. Finally, MASSDE administers several academic support programs each year that can help districts increase the percentage of students with disabilities who graduate from high school, including: Academic Support Services Grants: A number of grants fund school- and district-level programs that provide additional instruction and tutoring for students who have not met the Competency Determination requirement for graduation. Alternative Education: These programs are offered by districts for “at-risk” students, often in nontraditional settings. Secondary School Improvement initiatives: These MASSDE programs promote college- and careerreadiness, and include grants and conferences for high schools. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): Since the FFY 2005 APR submission, Massachusetts has revised the SPP baseline data and targets for the statewide Graduation Rate of students with disabilities. Due to data corrections in the state’s Student Information Management System (SIMS) since February 2007, the baseline Graduation Rate for the 2005-06 cohort was revised from 61.6% to 61.7% for students with disabilities, from 83.8% to 83.9% for non-disabled students, and from 79.8% to 79.9% for all students. Additionally, Five-Year Graduation Rates for the 2005-06 cohort were calculated in the fall of 2007, and so these rates were recently added to Massachusetts’ baseline data. The targets have been revised to align with MASSDE’s goal of closing the gap between the Graduation Rate of student with disabilities and the Graduation Rate of students without disabilities over the next ten years. As such, our Measurable and Rigorous Targets now reflect MASSDE’s goal of bringing the Graduation Rate for students with disabilities to 83.9% or higher by FFY 2015, with targets increasing by 4.5% every two years between FFY2005 and FFY 2015. 5 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Monitoring Priority: FAPE in the LRE Indicator 2: Percent of youth with IEPs dropping out of high school compared to the percent of all youth in the State dropping out of high school. (20 U.S.C. 1416 (a)(3)(A)) Measurement: The dropout rate is the number of students in grades 9-12 who drop out over a one-year period, from July 1 to June 30, who do not return to school by October 1 st of the next school year, divided by the total enrollment of students, times 100. The measurement for youth with an IEP is the same as the measurement for youth without an IEP. FFY Measurable and Rigorous Target 2005 (2005-2006) Students with IEPs Dropout Rate: 5.6% 2006 (2006-2007) Maintain Students with IEPs Dropout Rate of 5.6% Actual Target Data for FFY 2005 (2005-2006) and FFY 2006 (2006-2007): Because MASSDE was unable to report dropout rate data for FFY 2005 in last year’s report, OSEP has required that MASSDE report it this year. Those data are included in this section along with the FFY 2006 Dropout Rates as required by OSEP. FFY 2005 (2005-2006) # of Students Enrolled (Grades 9-12) # of Dropouts (Grades 9-12) FFY 2005 Dropout Rate IEP 43,508 2,237 5.1% Non-IEP 253,003 7,673 3.0% All Students 296,511 9,910 3.3% FFY 2006 (2006-2007) # of Students Enrolled (Grades 9-12) # of Dropouts (Grades 9-12) FFY 2006 Dropout Rate IEP 44,257 2,550 5.8% Non-IEP 253,776 8,891 3.5% All Students 298,033 11,441 3.8% 6 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): Discussion of Dropout Rates: Dropout Rates in Massachusetts decreased between FFY 2004 and FFY 2005, and then increased between FFY 2005 and FFY 2006. From FFY 2004 (Baseline Year) to FFY 2005, statewide Dropout Rates declined by half of one percentage point for students with IEPs and students without IEPs. The Dropout Rate declined from 5.6% in FFY 2004 to 5.1% in FFY 2005 for students with IEPs. In that same period, Dropout Rates declined from 3.5% to 3.0% for non-IEP students, and from 3.8% to 3.3% for all students. From FFY 2005 to FFY 2006, statewide Dropout Rates increased from 5.1% to 5.8% for students with IEPs, from 3.0% to 3.5% for non-IEP students, and from 3.3% to 3.8% for all students. The fluctuation in Dropout Rates over the past two years may reflect two significant modifications to Massachusetts’ data collection and coding procedures in the Student Information Management System (SIMS): Starting in FFY 2005, the state now cross-references SIMS data with the General Educational Development (GED) Testing Service database. Previously, MASSDE relied on districts to report students who had earned a GED. This previous method was somewhat unreliable because districts rarely have any way to track former students and determine when those students earn a GED. As a result of this new cross-referencing databases capacity, MASSDE can more accurately track students who drop out of high school and then earn a GED, decreasing the number of students who had previously been counted as dropouts. This change is probably part of the reason Dropout Rates decreased between FFY 2004 and FFY 2005. Starting in FFY 2006, MASSDE has implemented a wider range of SIMS codes and data verification procedures for reporting of students’ enrollment status. In the past, students who dropped out without notifying their district were often incorrectly coded by districts as transfer students. Districts must now indicate the enrollment status of transfer students as “Transferred – In state public”, “Transferred – In state private”, “Transferred – Out-of-State (public or private)”, “Transferred – Home-school”, or “Transferred – Adult diploma program, leading to MA diploma”. Based on these new codes, MASSDE can confirm whether or not students who were coded as transfers to other in-state districts actually showed up elsewhere in SIMS. This change probably contributed to the increase in Dropout Rates from FFY 2005 to FFY 2006 because fewer dropouts were incorrectly identified as having transferred to a different district. Given that the data collection and verification procedures used in calculating Dropout Rates have changed slightly each of the last two years, it is not possible to make meaningful year-to-year comparisons over this period of time. It should also be noted that the FFY 2006 Dropout Rates were completed only a few days prior to the February 1 APR submission deadline, and may be adjusted as MASSDE makes further corrections to the data set. However, it is clear that the gap between the Dropout Rates of students with IEPs and general education students appears to be relatively stable. Evidence of this persistent disparity in Dropout between students with disabilities and their non-disabled peers is a major factor in MASSDE’s decision to change the way in which Measurable and Rigorous Targets are set, as described in the Revisions section below. The following chart illustrates the gap over time: 7 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Annual Dropout Rate (Percent of Students Grades 9-12) Annual Dropout Rates for IEP and Non-IEP Students Over Time 7.0% 6.0% 5.8% 5.6% 5.1% 5.0% 4.0% 3.5% 3.5% 3.0% IEP Non-IEP 3.0% 2.0% 1.0% 0.0% 2004-2005 2005-2006 2006-2007 Academic Year Discussion of Improvement Activities: In an effort to reduce the Dropout Rate in Massachusetts, and to close the gap between students with disabilities and students without disabilities, a number of improvement activities were completed in FFY 2006. These include further development of public reporting of Special Education data, a variety of professional development initiatives for educators, and certain targeted academic support programs and grants. Over this past year, MASSDE further developed its public reporting of Dropout Rates and other Special Education data. Two major improvements in public reporting are: Special Education Data and Mapping: MASSDE worked with stakeholder groups and staff from a number of school districts to determine the best way to publicly report graduation data at an LEA level. As a result of these efforts, MASSDE has developed a resource titled “Special Education Results” that has been added to each district’s profile on the MASSDE website to meet the requirement that district indicator data be publicly reported. Data for Indicator 2 are reported through tables, and thematic maps have been developed to display the dropout rate for students with disabilities for each district: http://profiles.doe.mass.edu/special_ed.aspx. Public Reporting of “Cohort” Dropout Rates: MASSDE’s yearly Graduation Reports include the percentage of non-graduates in each cohort who dropped out of high school. In the 2005-2006 cohort, 19.5% of students with IEPs and 11.7% of all students dropped out. In the 2006-2007 cohort, 16.1% of students with IEPs and 9.4% of all students dropped out of high school. The availability of “cohort” Dropout Rates on the MASSDE website gives the public a more complete understanding of attainment and outcomes for a given group of students and further illustrates the gap between students with disabilities and the general population. 8 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 MASSDE also provided a number of professional development activities in FFY 2006 that may help reduce dropout by enhancing educators’ and administrators’ understanding of factors leading to dropout, and by promoting policies and practices that support students with disabilities in high school, including: Graduation Rate Summit - March 5, 2007: MASSDE co-sponsored this statewide conference for decision-makers and leaders in education, which included a variety of presentations and breakout sessions around issues related to graduation and dropout. Special Education Summer Institutes – Summers 2006 and 2007: MASSDE sponsored a variety of Institutes with topics such as “Effective Evaluation of Special Education Programs”, “Assistive Technology and Universal Design in the Classroom” and “Managing Behavior in an Inclusive Classroom”. Individuals in each Institute participated in at least 45 contact hours and created a final product related to the topic of the institute. Project FOCUS Academy (PFA) – 2005-2008: This pilot distance-learning program, funded through a federal State Improvement Grant, included courses for educators in Universal Design for Learning, Transition/Post-School Outcomes, and Positive Behavioral Interventions and Supports. Participating schools then received grants to fund school-improvement initiatives and systems reform that benefit students with disabilities. Massachusetts FOCUS Academy (MFA) – 2007-2012: MASSDE is developing this online professional development tool based on the PFA pilot model. MFA, which is funded through a federal State Personnel Development Grant, will ultimately increase professional development opportunities for educators across the state. Special Education Program Improvement Grants – yearly: MASSDE administers non-competitive grants to school districts to support professional development opportunities that increase educators competencies in Meeting the Behavioral and Social Needs of a Diverse Student Population, and in Curriculum Development, Instruction and Classroom Assessment. Positive Behavioral Interventions and Supports Grant Program – 2006-2012: This competitive grant program provides selected districts with training, resources and technical assistance to support the implementation of Positive Behavioral Interventions and Supports. Finally, MASSDE administers several academic support programs each year that can help districts reduce dropout of students with disabilities, including: Academic Support Services Grants: A number of grants fund school- and district-level programs that provide additional instruction and tutoring for students who have not met the Competency Determination requirement for graduation. Alternative Education: These programs are offered by districts for “at-risk” students, often in nontraditional settings. MASSDE actively promotes alternative education programs as a method of reducing dropout, and hosted a statewide conference on Alternative Education in October of 2007. Dropout Prevention Liaisons: Starting this year, MASSDE is requiring all school districts to designate a person at the district level to serve as the Dropout Prevention Liaison. This person is responsible for communicating with MASSDE on issues regarding dropout prevention and recovery, and for coordinating district-level initiatives. Secondary School Improvement initiatives: These MASSDE programs promote college- and careerreadiness, and include grants and conferences for high schools. Trauma-Sensitive Schools Program: MASSDE provides grants and resources to districts to support comprehensive staff trainings and school-based initiatives that help schools meet the needs of at-risk students and victims of trauma. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): Since the FFY 2005 APR submission, Massachusetts has revised the SPP targets for the statewide Dropout Rate of students with disabilities. The targets have been revised to align with MASSDE’s goal of closing the gap between the Dropout Rate of student with disabilities and the Dropout Rate of students without disabilities over the next ten years. As such, our Measurable and Rigorous Targets now reflect 9 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 MASSDE’s goal of reducing the Dropout Rate for students with disabilities to 3.5% or lower by FFY 2015, with targets decreasing in increments of approximately half of one percentage point every two years between FFY 2005 and FFY 2015. 10 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Monitoring Priority: FAPE in the LRE Indicator 3: Participation and performance of children with disabilities on statewide assessments: A. Percent of districts meeting the State’s AYP objectives for progress for disability subgroup. B. Participation rate for children with IEPs in a regular assessment with no accommodations; regular assessment with accommodations; alternate assessment against grade level standards; alternate assessment against alternate achievement standards. C. Proficiency rate for children with IEPs against grade level standards and alternate achievement standards. (20 U.S.C. 1416 (a)(3)(A)) Measurement: A. Percent = # of districts meeting the State’s AYP objectives for progress for the disability subgroup (children with IEPs) divided by the total # of districts in the State times 100. B. Participation rate = a. # of children with IEPs in grades assessed; b. # of children with IEPs in regular assessment with no accommodations (percent = b divided by a times 100); c. # of children with IEPs in regular assessment with accommodations (percent = c divided by a times 100); d. # of children with IEPs in alternate assessment against grade level standards (percent = d divided by a times 100); and e. # of children with IEPs in alternate assessment against alternate achievement standards (percent = e divided by a times 100). Account for any children included in a but not included in b, c, d, or e above Overall Percent = b + c + d + e divided by a. C. Proficiency rate = a. # of children with IEPs in grades assessed; b. # of children with IEPs in grades assessed who are proficient or above as measured by the regular assessment with no accommodations (percent = b divided by a times 100); c. # of children with IEPs in grades assessed who are proficient or above as measured by the regular assessment with accommodations (percent = c divided by a times 100); d. # of children with IEPs in grades assessed who are proficient or above as measured by the alternate assessment against grade level standards (percent = d divided by a times 100); and e. # of children with IEPs in grades assessed who are proficient or above as measured against alternate achievement standards (percent = e divided by a times 100). Overall Percent = b + c + d + e divided by a. 11 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 FFY Submitted February 1, 2008 Measurable and Rigorous Targets % Districts Meeting AYP for Disability Subgroup (3A) Participation Rate for Students with IEPs (3B) Proficiency Rate (CPI) for Students with IEPs (3C) ELA MATH ELA MATH ELA MATH Targets for FFY 2006 (2006-2007) 45% 37% 99% 99% 85.4% 76.5% Actual Target Data for FFY 2006 (2006-2007) 41.4% 43.6% 97.9% 97.6% 66.7% 54.0% Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): In order to meet the goal of students with disabilities achieving academic success, MASSDE works with districts and schools to analyze student assessment data and implement effective improvement plans as outlined in the MA SPP submitted in December 1, 2005. During the FFY 2006, several improvement activities were completed: School districts were provided with their AYP results detailing the outcomes for each subgroup. Detailed MCAS files were provided to the districts so that schools can create item-analysis charts to assist educators in identifying weakness and relevant relationships across student subgroups and subject areas, and to inform staff professional development. The Massachusetts Statewide Special Education Steering Committee met in December 2007 to review current data and activities and give input on methods of data analysis. The Student Achievement Interest Group focused on this indicator was given the opportunity to review the data collected and gave input on the analysis, the possible reasons for slippage and progress. Superintendents and principals previewed preliminary 2007 Adequate Yearly Progress (AYP) data for schools and reported discrepancies. Districts and schools making substantial gains in students’ achievement were identified (Compass awards) and best practices were disseminated. Analysis of the assessment data for students with disabilities, including testing accommodations and the impact of key factors (e.g. disability type, educational environment, level of need) on student performance is ongoing School Panel reviews were conducted in order to determine districts that could be identified as under-performing and in need of assistance. Training and technical assistance, including analysis of student assessment data and development of school improvement plans, were provided to districts and schools identified as under-performing or in need of improvement. Targeted workgroups focused on this indicator met to review improvement activities and results, to revise TA as needed and to consider dissemination activities. 12 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Analysis for Indicator 3A: Change in Distrcts Making AYP Percentage Points 50% 40% 30% 20% 10% 0% Baseline-Target 2005 & 2006 ELA MATH 45% 37% Actual FFY 2005 25% 19% Actual FFY 2006 41.4% 43.6% The targets set for FFY 2006 maintained the baseline level of performance obtained for FFY 2004. The data for FFY 2006 shows a slight slippage from the baseline data for English Language Arts (ELA). However, there is a significant increase in the percentage of districts making AYP for special education subgroups in ELA and Mathematics from the FFY 2005 data: In ELA the percentage increased from 25% to 41.4%; and For Mathematics, the percentage increased from 19% to 43.6%. This increase exceeded the 37%, the target set for FFY 2006. Massachusetts is proud to report the increase in the percentage of districts making AYP for students with disabilities in both ELA and Mathematics. MASSDE will continue to work with the focused indicator workgroup to identify the performance gaps in our schools and districts that require attention so that we will continue to meet the AYP targets. We will continue to work with educators to take the information from the accountability reporting and use it to improve the performance of students with disabilities overall. Analysis for Indicator 3B: Participation Rate for Students with IEPs Percentage Points 100% 95% 90% ELA MATH 99% 99% Actual FFY 2005 97.6% 97.7% Actual FFY 2006 97.9% 97.7% Baseline-Target 2005 & 2006 The baseline data (FFY 2004) indicated the participation rate for students with IEPs on statewide assessments was better than 99%, which MASSDE characterized as full participation. The data for FFY 2006 reflects a minor slippage in the participation rate for students with IEPs on statewide assessment. 13 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 The participation rate was 97.9% and 97.7% on the ELA and Mathematics assessment respectively. However, although MASSDE acknowledges that numerically this does represent minor slippage in comparison to the targets set, MASSDE is satisfied that the participation rate continues to be excellent for students with disabilities in state assessment programs. Analysis for Indicator 3C: The FFY 2006 Composite Performance Index (CPI) for students with disabilities is 66.7 in ELA and 54.0 in Math, which are below the targets of 85.4 for ELA and 76.5 for Mathematics set for this year. However, as presented in the charts below, the data reflect an improvement in both ELA and Math from the previous two years. The CPI in ELA was 65.0 in the baseline year of FFY 2004, was 64.8 in FFY 2005, and has now risen to 66.7 in FFY 2006. In Mathematics, the CPI was 49.5 in the baseline year of FFY 2004, and increased to 51.5 in FFY 2005 and on to 54.0 in FFY 2006. This demonstrates that the state is moving towards the targets set for FFY 2010. MCAS Proficiency Rates CPI: English Language Arts 100 80 60 40 20 0 80.5 65 FFY 2004 85.4 66.7 80.5 64.8 FFY 2005 ELA-Target ELA-Actual FFY 2006 MCAS Proficiency Rates CPI: Mathematics 100 80 60 68.7 68.7 49.5 51.5 76.5 54 Math-Target Math-Actual 40 20 0 FFY 2004 FFY 2005 FFY 2006 14 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Public Reporting: MASSDE has developed a resource titled “Special Education Results” that has been added to each district’s profile on the MASSDE website to meet the requirement that district indicator data be publicly reported. While data for Indicator 3 continue to be reported through tables, thematic maps have been developed for some of the indicators and will be developed for more as Massachusetts continues its implementation of activities for the MA SPP: http://profiles.doe.mass.edu/special_ed.aspx. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): Massachusetts has revised its data and targets for Indicator 3C to reflect the state performance targets identified in No Child Left Behind (NCLB) because we are committed to having the same high expectations for students with disabilities as we have for all students. Our methods of calculating our performance are fully articulated in the SPP. Massachusetts has submitted a revised State Performance Plan with the new baseline data and targets for Indicator 3C. 15 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Monitoring Priority: FAPE in the LRE Indicator 4: Rates of suspension and expulsion: A. Percent of districts identified by the State as having a significant discrepancy in the rates of suspensions and expulsions of children with disabilities for greater than 10 days in a school year; and B. RMASSDERVED. ((20 U.S.C. 1416(a)(3)(A); 1412(a)(22)) Measurement: A. Percent = [(# of districts identified by the State as having significant discrepancies in the rates of suspensions and expulsions of children with disabilities for greater than 10 days in a school year) divided by the (# of districts in the State)] times 100. B. RMASSDERVED. Include State’s definition of “significant discrepancy.” Indicator 4A FFY 2006 (2006-2007) Measurable and Rigorous Target The % of districts with a significant discrepancy finding will be 0% Actual Target Data for FFY 2006 (2006-2007): 2006-07 Special Education Enrollment 2006-07 State Suspension/ Expulsion Rate % of districts with suspension/ expulsion rate that is five times State Rate* % of districts with a finding of “significant discrepancy”** 1.1% 0.29% (4 districts) (1 district) *Districts with fewer than 30 students in special education were removed from this part of the analysis. The total number of districts included in the analysis was 347. **The calculation is (1/347)*100. 163,396 1.0% Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): Explanation of Progress or Slippage: The data presented in the table above is based on the state’s definition of ‘significant discrepancy’ -- a district having a suspension rate of five times the state rate for two consecutive years. One district (0.29% of all districts) met the threshold of exceeding the state rate by five times for two years in a row and was Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 found to have a ‘significant discrepancy’ in the rate of suspensions and expulsions of children with disabilities for greater than 10 days in a school year. This exceeds our rigorous target of 0% for each year. MASSDE will review the district’s policies and procedures relating to the development and implementation of IEPs, the use of positive behavioral interventions and supports, and procedural safeguards, and will require the LEA to revise it’s policies, practices, and procedures as appropriate. The data presented above also indicate the emergence of two trends in our rates of suspension and expulsion for students with disabilities. The first trend is a potential cause for concern – over the past three years, the state suspension/expulsion rate for students with disabilities has increased. The FFY 2004 rate was 0.514%, the FFY 2005 rate was 0.916%, and the FFY 2006 rate is 1.0%. As an initial step in investigating this increase, MASSDE continues to upgrade our activities in regards to ensuring that the data is appropriately reported. MASSDE continues to consider the main effort in this area to rest with appropriate procedures to ensure good reporting and an effective review of policies and procedures. The second trend identified is more positive. Despite the increase in the suspension/expulsion rate of students with disabilities, the number of districts found each year to have five times the state suspension/expulsion rate has decreased. In FFY 2004, 6 districts exceeded this threshold. The number of districts has since decreased to 4 districts in both FFY 2005 and FFY 2006. Discussion of Improvement Activities Completed: In order to meet the goal of students with disabilities achieving academic success, MASSDE continued its work with districts and schools to analyze School Safety and Discipline Report (SSDR) data and implement effective improvement plans outlined in the State Performance Plan submitted in December 1, 2005. During the FFY 2006, several improvement activities were completed: Met with Massachusetts Statewide Special Education Steering Committee to review current data and identify potential improvement activities. Identified the districts that have a suspension/expulsion rate for students with an IEP that is five times greater than the state rate. Reviewed the flagged districts’ policies, practices, and data reporting procedures for a finding of significant discrepancy. Met with the targeted workgroup focused on this indicator. Discussion included how to best ensure that districts are using the same definition of “suspension” when reporting their data. Met with Massachusetts Statewide Special Education Steering Committee to review current data disaggregated by number of suspensions. Discussion included determining a threshold for a district. Reviewed and revised the procedures used for policy and practice review and data verification review. Provided technical assistance to district data coordinators on calculating “suspensions” and correct reporting through the SSDR. Pulled a sample of districts with a low or no suspension and some with a high suspension rate, Invited these districts to be part of this forum. Met with selected districts to begin development of a ‘School District Self-Review Monitoring Protocol’ for suspension of students with disabilities Additionally, over this past year, MASSDE further developed its public reporting of suspension/expulsion data and other Special Education data. The department worked with stakeholder groups and staff from a number of school districts to determine the best way to publicly report these data at an LEA level. As a result of these efforts, MASSDE has developed a resource titled “Special Education Results” that has been added to each district’s profile on the MASSDE website to meet the requirement that district indicator data be publicly reported. Data for Indicator 4a are reported through tables, and thematic maps have been developed to display data for some of the other indicators: http://profiles.doe.mass.edu/special_ed.aspx. 17 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): In the Part B FFY 2005 SPP/APR Response Table, OSEP stated that the state was not complying with 34 CFR § 300.170 in its definition of ‘significant discrepancy’. Massachusetts has revised its definition of ‘significant discrepancy’ to be consistent with Federal requirements. Significant discrepancy is defined as having a suspension rate of five times the state rate for two consecutive years. MASSDE is submitting a revised State Performance Plan submission for this indicator to reflect this change. Additionally, in past years, Massachusetts reported that it identified several districts for a review of policies and procedures for having a suspension/expulsion rate five times the state rate: (1) five districts in FFY 2003; (2) six districts in FFY 2004; and (3) four districts in FFY 2005. In the Part B FFY 2005 SPP/APR Response Table, OSEP requested that the state report on our review of these districts. MASSDE has reviewed all of the districts’ policies, practices and procedures relating to the development and implementation of IEPs, the use of positive behavioral interventions and supports, and procedural safeguards. Based on the reviews, MASSDE did not find cause to require any of the districts to revise their policies or procedures. 18 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Monitoring Priority: FAPE in the LRE Indicator 5: Percent of children with IEPs aged 6 through 21: A. Removed from regular class less than 21% of the day; B. Removed from regular class greater than 60% of the day; or C. Served in public or private separate schools, residential placements, or homebound or hospital placements. (20 U.S.C. 1416(a)(3)(A)) Measurement: A. Percent = [(# of children with IEPs removed from regular class less than 21% of the day) divided by the (total # of students aged 6 through 21 with IEPs)] times 100. B. Percent = [(# of children with IEPs removed from regular class greater than 60% of the day) divided by the (total # of students aged 6 through 21 with IEPs)] times 100. C. Percent = [(# of children with IEPs served in public or private separate schools, residential placements, or homebound or hospital placements) divided by the (total # of students aged 6 through 21 with IEPs)] times 100. Measurable and Rigorous Target FFY % of children with IEPs aged 6 through 21 in full inclusion (Indicator 5A) % of children with IEPs aged 6 through 21 in substantially separate placements (Indicator 5B) % of children with IEPs aged 6 through 21 in outof-district placements (Indicator 5C) 43.4% 16.2% 6.8% 2006 (2006-2007) Actual Target Data for FFY 2006 (2006-2007): % of children with IEPs aged 6 through 21 in full inclusion (Indicator 5A) % of children with IEPs aged 6 through 21 in substantially separate placements (Indicator 5B) % of children with IEPs aged 6 through 21 in outof-district placements (Indicator 5C) 53.0% 15.3% 6.7% Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): In FFY 2005, the total number of students with disabilities ages 6-21 was 146,807. In FFY 2006, the total number of students with disabilities ages 6-21 was 149,211. This constitutes a 1.6% increase in students receiving special education services in the state. Despite this increase in the number of special education students, in FFY 2006 Massachusetts has exceeded its target for full inclusion. MASSDE is very pleased with the progress that has been made in this area. One reason for the increase in full inclusion could be due to the fact that districts are doing a better job in reporting data, since a change was made in 2002 in the definition of full inclusion. In addition, it is believed that districts are implementing practices that promote full inclusion of all students. The data support this by showing an increase in full inclusion close in size to the decrease in partial inclusion. As a state, the three disability categories that have increased the most in full inclusion from 2003-04 to 2006-07 are: Physical (49.2% in 2003-04; 71.9% in 2006-07); Health (45.9% in 2003-04; 67.5% in 2006-07); and SLD (38.9% in 2003-04; 60.3% in 2006-07). The chart below shows the change in full inclusion from 2003-04 (baseline data) to 2006-07 for all disability categories. Change in Full Inclusion from 2003-2007 percentage point change 25 22.7 21.6 20 16.5 21.4 17.2 14.8 15 11.3 10 7.4 16.0 13.3 6.4 4.1 5 2.1 ie s is ab i lit H ea om l th m un ic Se at io ns n or y V is io n Em ot io Se na ns l or Ph y Sp ys H ec ic ar al d i fi c of Le H ar ea ni rin ng g Di sa bi l it ie s D ea fb D l in ev d el op m en ta N l eu ro lo gi ca l In te l le ct ua l C M ul tip le D Au t is m 0 Substantially separate and out-of-district placements stayed virtually the same from one year to the next and continued efforts need to be made to move these settings closer to the long-term goals that have been set. There have, however, been some changes in the use of substantially separate settings by students with different disability types. All but two disability types decreased their use of substantially separate placements between 2003 and 2007. Students with intellectual disabilities and neurological disabilities increased their use of substantially separate placements, although the change in substantially separate placements for students with neurological disabilities is nominal (.03 percentage points). 20 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 For students with intellectual disabilities, partial inclusion is decreasing, but those students are relatively evenly split between full inclusion placements and substantially separate placements. Difference from 2003-2007 Submitted February 1, 2008 Full Inclusion Partial Inclusion +4.1% -8.4% Substantially Separate Out of District +4.2% 0.1% For students with neurological disabilities, the increase in substantially separate settings is nominal (increase of .03 percentage points). However, there is a significant movement from partial inclusion to full inclusion (16 percentage points). Difference from 2003-2007 Full Inclusion Partial Inclusion +16.0% -16.5% Substantially Separate Out of District +0.03% +0.4% In the category of out-of-district placements, students who are Deaf/Blind have increased the most in their use with 6.6 percentage points from 2003-2007. Students with autism have also increased use of out-ofdistricts placements over the past three years by 0.6 percentage points. There are also students in other disability categories that have decreased use of out-of-district placements. These include: Emotional Disabilities (decrease of 2.5 percentage points); Health (decrease of 1.1 percentage points); and Multiple Disabilities (decrease of 0.5 percentage points). Changes in Out of District Placements from 2003-2007 Autism 8 Multiple Disabilites 6.61 Health Percentage Points 6 Communication Sensory Vision 4 Emotional 2 Pyhsical 0.59 0.27 0.04 0 0 -2 -0.52 -1.05 -0.3 0.09 0.42 0.14 Sensory Hard of Hearing Specific Learning Disabilities Deafblind Developmental -2.49 -4 0.14 Neurological Intellectual These changes indicate that while students who are Deafblind or with autism are increasing in out-ofdistrict placements, students in most disability categories are holding steady or increasing their inclusive practices. 21 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Several activities have occurred around LRE that have contributed to the state’s increase in full inclusion. The Special Education Planning and Policy Development office holds Special Education Summer Institutes each year, which include topics such as: Managing Behavior in an Inclusive Classroom; Teaching Strategies for Students with Vision Impairments, Blindness or Deafblindness; and Teaching Strategies for Students with Autism Spectrum Disorders in the General Education Classroom. In addition, MASSDE offers annual Content Institutes over the summer that are designed to support teacher learning and development of skills for standards-based instruction. In 2006-07, institutes were offered in Science and Technology, Math, Reading and Writing, and the Arts. In addition to these institutes, the Special Education Planning and Policy Development office funded over 400 Program Improvement grants to districts across the state. This grant includes five priority areas: Enhancing Induction and Mentoring Programs; Serving Students with Autism Spectrum Disorders in Inclusive Settings; Curriculum Development, Instruction and Classroom Assessment; Meeting the Behavioral and Social Needs of a Diverse Student Population; and Recruitment and/or Additional Professional Development Needs as Identified by the District or Educational Collaborative. Over this past year, MASSDE further developed its public reporting of LRE data and other Special Education data. The department worked with stakeholder groups and staff from a number of school districts to determine the best way to publicly report these data at an LEA level. As a result of these efforts, MASSDE has developed a resource titled “Special Education Results” that has been added to each district’s profile on the MASSDE website to meet the requirement that district indicator data be publicly reported. Data for Indicator 5 are reported through tables, and thematic maps have been developed to display the full inclusion rate for students with disabilities for each district: http://profiles.doe.mass.edu/special_ed.aspx. Finally, MASSDE staff is working to analyze data both at the district and student level to determine appropriate technical assistance to ensure students are educated in the most appropriate, least restrictive environment for their educational needs. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): Based on the progress that has been made in all three placement areas, MASSDE considers that it has met or exceeded its FFY 2006 targets. As a result, MASSDE is submitting a revised State Performance Plan to revise the targets for this indicator. Technical assistance activities will continue to ensure accurate placements are made for students and there is an increase in full inclusion. In the original SPP, MASSDE anticipated that with the decrease in out-of-district placements, there would be an increase in substantially separate placements. However, based on the past two years of data as noted above, most of the substantially separate placements have decreased as well. Therefore, new targets are being set to decrease the percentage of students in substantially separate settings. MASSDE enjoys a strong relationship with the 144 Special Education Approved Private Schools (APS) across the state. Each of these schools self identify the population they serve by disability, age, and gender and the MASSDE Program Quality Assurance office monitors their compliance. Because there are such varied and rich options available to students with disabilities within the state, MASSDE is hesitant to reduce the targets for out-of-district placements at this time. Based on the progress that has been made in all three target areas, the improvement activities, timelines and resources for FFY 2006 remain appropriate. 22 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Revised targets: Measurable and Rigorous Target FFY % of children with IEPs aged 6 through 21 in full inclusion (Indicator 5A) % of children with IEPs aged 6 through 21 in substantially separate placements (Indicator 5B) % of children with IEPs aged 6 through 21 in outof-district placements (Indicator 5C) 2005 (2005-2006) 43.4% 16.2% 6.8% 2006 (2006-2007) 43.4% 16.2% 6.8% 2007 (2007-2008) 54.3% 15.1% 6.2% 2008 (2008-2009) 55.5% 14.9% 6.2% 2009 (2009-2010) 56.8% 14.7% 5.9% 2010 (2010-2011) 58% 14.5% 5.5% 23 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Monitoring Priority: FAPE in the LRE Indicator 8: Percent of parents with a child receiving special education services who report that schools facilitated parent involvement as a means of improving services and results for children with disabilities. (20 U.S.C. 1416(a)(3)(A)) Measurement: Percent = [(# of respondent parents who report schools facilitated parent involvement as a means of improving services and results for children with disabilities) divided by the (total # of respondent parents of children with disabilities)] times 100. FFY Measurable and Rigorous Target 2006 (2006-2007) 76% Actual Target Data for FFY 2006 (2006-2007): 77% [Calculation: (5,282 surveys that met the standard / 6,872 returned surveys) * 100] Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): In order to collect data for this indicator, MASSDE uses the “School’s Efforts to Partner with Parent scale from the Part B survey instrument developed by the National Center for Special Education Accountability Monitoring (NCSEAM). This 25-item scale (survey) addresses Indicator # 8 (see Appendix A for the survey). Massachusetts’ Indicator # 8 stakeholder group adapted the survey via the “Item Bank” provided by NCSEAM. Additional technical information on this survey instrument is available on the NCSEAM website (http://www.monitoringcenter.lsuhsc.edu/parent_family_involvement.htm). To ease the data collection activity burden for districts, MASSDE divided its 387 LEAs into four cohorts that are representative of the state as a whole, and is collecting and reporting data on this indicator based on a four-year cycle (2005-06 through 2008-09). This cohort model and plan for data collection was approved by OSEP. Over a four-year period, every district in the state will have participated in the data collection activities for this indicator. Because it has an average daily membership of over 50,000 students, Boston participates in all activities every year. Further information on the cohort model can be found at: http://www.doe.mass.edu/sped/2006/0522idea.html. To date, one half of Massachusetts’ districts (or two of four cohorts – 194 LEAs) have participated in the parent survey activity. Survey rounds were conducted in fall 2006 and spring 2007, each with a different group of districts. Surveys and cover letters were distributed to parents in three languages: English, Spanish, and Portuguese. In order to calculate the percentage to report on this indicator, MASSDE considered the responses of parents for each survey item where “very strongly agree”, “strongly agree”, or “agree” was the response. Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 The measure adopted for Massachusetts’ SPP to show “that schools facilitated parent involvement as a means of improving services and results for children with disabilities” requires agreement on at least 50% of the survey items (13 of 25). The percentage for FFY 2005 (survey conducted in fall 2006) as initially reported was 76%. This was reported in the FFY 2005 SPP submission in February 2007. Late submissions of surveys increased the positive results (77.2%). However, MASSDE will continue to base our reporting on the 76% achieved at the time of the required FFY 2005 APR submission. The FFY 2006 percentage (survey conducted in spring 2007) is as listed above: 77%. These results include an all-online pilot that was conducted in three districts. 40,476 parents were surveyed and 6,872 responded, resulting in an overall return rate of 17%. The language breakdown is as follows: English: 37,746 sent and 6,664 returned. Return rate: 17.7%. Spanish: 2,408 sent and 195 returned. Return rate: 8.1%. Portuguese: 322 sent and 13 returned. Return rate: 4.0%. Both the FFY 2005 and FFY 2006 results (76% and 77% respectively) are statistically consistent with one another. In both of the survey administrations conducted, the item reliability has met or exceeded NCSEAM’s standard of .90 (.93 for both administrations of the survey). This indicator is an area around which there has been a great deal of activity. The stakeholder interest group has met several times, both in-person and virtually, via conference call. Calls were conducted in April 2007 to review the fall 2006 parent survey results. Later, feedback was solicited from particular districts on how to best format and present them with their individual district results. This feedback was incorporated into the final report template, impacting the way in which those results were presented to districts. When they were mailed the results for their particular district, LEAs also received an accompanying technical assistance letter. This letter offered suggestions on how to read and understand their report, as well as how to work with their results in their community. Once all 194 districts were mailed their results (from fall 2006 and spring 2007 survey rounds) and technical assistance letter, they were invited to participate in conference calls to discuss their district’s results. There are several projects related to parent involvement in which MASSDE is currently engaged. Our Parent Advisory Council (PAC) Guidance Document is under revision and soon to be re-issued/updated. This document will offer suggestions to special education parent advisory councils across Massachusetts on how to succeed in achieving their mission. In addition, each year, MASSDE offers Special Education Summer Professional Development institutes. These are offered at no cost to the participating educator, where he/she is provided with 67.5 hours toward re-licensure and the option of taking the course(s) for graduate credit. Two of the institutes are particularly relevant to the topic of parent involvement: the Special Education Administrative Leadership Academy and IEP Team Facilitation Skills. For detailed information on the institutes, please see http://www.doe.mass.edu/sped/2007/institutes.pdf. Finally, MASSDE is fortunate to have been funded for both a State Improvement Grant (SIG) and a State Personnel Development Grant (SPDG). These grants have provided Massachusetts with funding for Project FOCUS Academy and Massachusetts FOCUS Academy, respectively. Project FOCUS Academy is in nine high schools across Massachusetts, where educators and other interested parties, including parents, were offered online courses in three areas: Universal Design for Learning, Positive Behavioral Interventions and Supports, and Secondary Transition. Massachusetts FOCUS Academy’s goal is to spread such opportunities statewide, creating an “online university”, including courses that are designed to improve and facilitate parental involvement. Among our partners on these grants has been the Federation for Children with Special Needs (FCSN), our statewide parent organization in Massachusetts. FCSN has conducted parent presentations on the three course topics in Project FOCUS Academy, and has created FAQ sheets and PowerPoint presentations on each area as well. These materials will be made available to parents in hard copy form, and will be posted online, along with course materials. In the coming months, we will continue to work with our stakeholder interest group around matters such as increasing survey return rates, raising the overall percentage that is reported to OSEP, and public 25 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 reporting. We are also working on the compilation of a technical assistance document / resource list for districts, containing information on best practices around parent involvement and partnership with families. We look forward to continuing and expanding upon our collaboration/partnership with the Federation for Children with Special Needs. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): MASSDE believes that the proposed targets in this indicator area remain appropriate at this time. Massachusetts will consider revising targets once all districts have been surveyed and the entire Massachusetts data set has been obtained and analyzed. MASSDE recently renewed its contract with Macro International and Ashton Associates for project completion of survey mailings and data collection/processing/analysis. Improvement activities, timelines, and resources as previously outlined remain appropriate at this time. 26 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Monitoring Priority: Disproportionality Indicator 9: Percent of districts with disproportionate representation of racial and ethnic groups in special education and related services that is the result of inappropriate identification. (20 U.S.C. 1416(a)(3)(C)) Measurement: Percent = [(# of districts with disproportionate representation of racial and ethnic groups in special education and related services that is the result of inappropriate identification) divided by (# of districts in the State)] times 100. State’s definition of “disproportionate representation”: Massachusetts defines “disproportionate representation” using a calculation of weighted or alternate risk ratio and a review of the appropriateness of a district’s policies and procedures for identifying students as disabled. MASSDE calculates a weighted or alternate risk ratio for each school district using the techniques described in detail in Westat’s “Methods for Assessing Racial/Ethnic Disproportionality in Special Education” (http://www.nichcy.org/training/B-resources.pdf). The state uses a minimum cell size of 20 for each race/ethnic group in every district. Cells less than 20 are individually reviewed to see if data irregularities for specific racial and ethnic groups in these districts would suggest disproportionate representation due to inappropriate identification. Once the calculation is made for each district, the weighted risk ratios are compared to the two previous years’ weighted risk ratios. Districts are flagged if for 3 consecutive years, they exhibit a weighted risk ratio of 3.0 or greater for possible over-representation, and of .25 or less for possible under-representation. All districts identified by way of quantitative analysis are then subject to a review of the appropriateness of their policies, practices, and procedures (PPPs) for special education eligibility determination and disability identification. Description of determination that disproportionate representation is due to inappropriate identification: Any district identified using the measurement techniques described above submitted their current PPPs to MASSDE where they were reviewed by a committee of policy analysts and compliance specialists. If the committee concluded that the PPPs were inappropriate or otherwise inconsistent with federal and state regulations, and concluded that the PPPs likely caused the disproportionate representation at least to some degree, then a district was identified as having disproportionate representation due to inappropriate identification. Regarding the FFY 2005 measurement: In the FFY 2005 SPP/APR Response Table, MASSDE was cited for not having provided FFY 2005 baseline data. As a result, OSEP required that MASSDE “must provide, in its FFY 2006 APR, baseline data from FFY 2005 on the percent of districts identified with disproportionate representation of racial and ethnic groups in special education and related services that was the result of inappropriate identification, and describe how the State made that determination…” MASSDE used the same numerical definition for the FFY 2005 and FFY 2006 definitions of disproportionate representation. However, in determining that the disproportionate representation is due to inappropriate identification practices, the process differed slightly: rather than asking districts what their PPPs were 2 years ago, MASSDE used existing PPP information from recent 27 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Coordinated Program Reviews and Mid-Cycle Coordinated Program Reviews. FFY Measurable and Rigorous Target 2005 (2005-2006) 0% 2006 (2006-2007) 0% Actual Target Data for FFY 2005 (2005-2006) and FFY 2006 (2006-2007): Because MASSDE did not report the baseline data for FFY 2005 in last year’s report, OSEP has required that MASSDE report it this year. Calculating the weighted risk ratios for FFY 2005 over-representation produced zero districts that met the criteria of three consecutive years of a WRR of 3.0 or higher. In terms of FFY 2005 under-representation, the calculation yielded zero districts with a WRR of .25 or lower for three consecutive years. Because zero districts met the criteria for the determination of overrepresentation and under-representation in FFY 2005, MASSDE met its measurable and rigorous target of 0%. In the FFY 2006 analysis, the calculations yielded similar results. For over-representation, the calculation produced zero districts that met the criteria of three consecutive years of a WRR of 3.0 or higher. For under-representation, the calculation yielded zero districts with a WRR of .25 or lower for three consecutive years. Because zero districts met the criteria for the determination of over-representation and under-representation in FFY 2006, MASSDE met its measurable and rigorous target of 0%. Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): As noted in the MA SPP, MASSDE planned to improve its disproportionality monitoring process by shifting oversight from the Coordinated Program Review process, which only targeted approximately one sixth of the districts in the state each year, to the Office of Strategic Planning, Research, and Evaluation (OSPRE). This shift in responsibilities constitutes a new improvement in MASSDE’s capacity to address disproportionality in all districts. OSPRE conducts the data analysis using various statistical packages, uses that data to determine the “cut points,” and then reviews policies, practices and procedures in those districts. In another new improvement, OSPRE conducts intervention procedures for all districts in the state that were identified as having disproportionate representation, and does so on a yearly basis. MASSDE’s annual review of district policies, practices, and procedures is a new and important step toward verifying that districts are identifying special education students in an appropriate manner. Additionally, MASSDE has made progress in its calculation and definition of “significant disproportionality” based on race and ethnicity. Prior to this year and as noted in OSEP’s response to the MA SPP submitted for this indicator in FFY 2005, MASSDE had not fully accounted for the difference between “disproportionate representation” and “significant disproportionality.” As of this year and annually hereafter, MASSDE is defining “significant disproportionality” as “any district with a weighted risk ratio or alternate risk ratio for special education identification, placement, and disciplinary actions, exceeding 5.0 for four consecutive years, and whose weighted risk ratio or alternate risk ratio is growing more disproportionate (i.e., becoming more and more over-represented) in each of those four years.” MASSDE 28 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 has made its determination of districts this year based on this definition. If MASSDE identifies a district with “significant disproportionality,” it (1) directs the district to review (and, if appropriate) revision of policies, procedures, and practices; (2) requires the district to reserve the maximum amount of funds to be used for early intervening services; and (3) requires the district to publicly report on the revision of policies, procedures, and practices. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): Massachusetts has met the rigorous, proposed target of 0% districts with disproportionate representation of racial and ethnic groups receiving special education or related services due to inappropriate identification. In addition to this accomplishment, MASSDE has identified a number of improvement activities that will strengthen the disproportionality monitoring process. Improvement Activity: ”Develop a self-assessment tool for districts that are ‘flagged’ for potential disproportionality. Districts use the self-assessment tool to examine their own policies and procedures regarding special education eligibility and disability definition.” The disproportionality self-assessment has been a success in many other states, and it is MASSDE’s plan to have a similar self-assessment tool prepared for use in districts next year. The tool will supplement the existing intervention procedure (the desk review of a district’s policies, practices and procedures). Timelines: Winter/Spring 2008 for its development; Summer 2008 for its use. Resources: State of New Jersey self-assessment, State of New York self-assessment, NCCRESt Improvement Activity: ”Develop broader and more robust communications with state disproportionality stakeholders, and create more discrete responsibilities, more frequent opportunities for interaction, and specific annual goals for the SPP disproportionality workgroup.” To this point, the SPP disproportionality workgroup has maintained sporadic membership, communicated infrequently, and at least to some extent has not been utilized to its potential. MASSDE depends on these stakeholders to provide guidance and external accountability, and in its current state, the workgroup has not been used adequately to those ends. It is MASSDE’s goal to increase communications, improve the quality of those communications, increase stakeholder accountability, and regularly consult with the workgroup regarding formulae, thresholds, and district intervention practices. Timelines: Ongoing. Resources: Institutions of Higher Education (Harvard University, etc.), OSPRE Research and Evaluation Advisory Group, District Administration, MASSDE website Improvement Activity: ”Broaden public awareness of disproportionality by way of an online resource for all things relating to disproportionality in Massachusetts.” A web-based resource devoted to providing the latest data and research on the topic of disproportionality would increase the profile and the transparency of this issue. Included in this online resource will be a research report that MASSDE will soon publish on the state of disproportionality in the commonwealth. The research report will examine national trends, examine Massachusetts’ trends, and provide the MASSDE with a contextual framework through which it can improve its assistance to districts and students. 29 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Timelines: Spring/Summer 2008. Resources: Research organizations (Westat, NCCRESt, RAND, etc.), Institutions of Higher Education, MASSDE website 30 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Monitoring Priority: Disproportionality Indicator 10: Percent of districts with disproportionate representation of racial and ethnic groups in specific disability categories that is the result of inappropriate identification. (20 U.S.C. 1416(a)(3)(C)) Measurement: Percent = [(# of districts with disproportionate representation of racial and ethnic groups in specific disability categories that is the result of inappropriate identification) divided by (# of districts in the State)] times 100. State’s definition of “disproportionate representation”: Massachusetts defines “disproportionate representation” using a calculation of weighted or alternate risk ratio and a review of the appropriateness of a district’s policies and procedures for identifying students as disabled. MASSDE calculates a weighted or alternate risk ratio for every school district in each of the six required disability categories (intellectual impairment, specific learning disabilities, emotional disturbance, speech/language impairments, other health impairments, autism) using the techniques described in detail in Westat’s “Methods for Assessing Racial/Ethnic Disproportionality in Special Education” (http://www.nichcy.org/training/B-resources.pdf). The state uses a minimum cell size of 10 for each racial/ethnic disability group in every district. Cells less than 10 are individually reviewed to see if data irregularities for specific racial and ethnic groups in these districts would suggest disproportionate representation due to inappropriate identification. Once the calculation is made for each district, the weighted risk ratios are compared to the two previous years’ weighted risk ratios. Districts are flagged if for 3 consecutive years, they exhibit a weighted risk ratio of 4.0 or greater for possible over-representation, and of .20 or less for possible under-representation. All districts identified by way of this quantitative analysis are then subject to a review of the appropriateness of their policies, practices, and procedures (PPPs) for special education eligibility determination and disability identification. Description of determination that disproportionate representation is due to inappropriate identification: Districts identified using the measurement techniques described above submitted their current PPPs to the MASSDE (or MASSDE verified recent compliance information/reviews) where the PPPs were reviewed by a committee of policy analysts and compliance specialists. If the committee concluded that the PPPs were inappropriate or otherwise inconsistent with federal and state regulations, and concluded that the PPPs likely caused the disproportionate representation at least to some degree, then a district was identified as having disproportionate representation due to inappropriate identification. Regarding the FFY 2005 measurement: In the FFY 2005 SPP/APR Response Table, MASSDE was cited for not having provided FFY 2005 baseline data. As a result, OSEP required that MASSDE “must provide, in its FFY 2006 APR, baseline data from FFY 2005 on the percent of districts identified with disproportionate representation of racial and ethnic groups in special education and related services that was the result of inappropriate identification, and describe how the State made that determination…” MASSDE used the same numerical definition for the FFY 2005 and FFY 2006 definitions of 31 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 disproportionate representation. However, in determining that the disproportionate representation is due to inappropriate identification practices, the process differed slightly: rather than asking districts what their PPPs were 2 years ago, MASSDE used existing PPP information from recent Coordinated Program Reviews and Mid-Cycle Coordinated Program Reviews. FFY Measurable and Rigorous Target 2005 (2005-2006) 0% 2006 (2006-2007) 0% Actual Target Data for FFY 2005 (2005-2006) and FFY 2006 (2006-2007): Because MASSDE did not report the baseline data for FFY 2005 in last year’s report, OSEP has required that MASSDE report it this year. Calculating the weighted risk ratios for FFY 2005 over-representation produced seven districts that met the criteria of three consecutive years of a WRR of 4.0 or higher. Using the existing PPP information from recent Coordinated Program Reviews and Mid-Cycle Coordinated Program Reviews, MASSDE determined that none of these districts had inappropriate special education identification procedures. In terms of FFY 2005 under-representation, the calculation yielded 11 districts with a WRR of .20 or lower for three consecutive years. Using the existing PPP information from recent Coordinated Program Reviews and Mid-Cycle Coordinated Program Reviews, MASSDE determined that none of these districts had inappropriate special education identification procedures. Because zero districts met the criteria for the determination of over-representation and under-representation in FFY 2005, MASSDE met its measurable and rigorous target of 0%. In the FFY 2006 analysis, the calculations yielded similar results. For over-representation, the calculation produced eleven districts that met the criteria of three consecutive years of a WRR of 4.0 or higher. MASSDE then conducted a review of these two districts’ policies, practices, and procedures, and found that the PPPs were sound, and that the disproportionality was not due to inappropriate PPPs. For under-representation in FFY 2006, the calculation yielded 12 districts with a WRR of .20 or lower for three consecutive years. MASSDE then conducted a review of these districts’ policies, practices, and procedures, and found that the PPPs were sound, and that the disproportionality was not due to inappropriate PPPs. Because zero districts met the criteria for the determination of over-representation and under-representation in FFY 2006, MASSDE met its measurable and rigorous target of 0%. Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): MASSDE’s improvement activities and explanation of progress for this indicator are the same as for Indicator 9. 32 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): MASSDE revisions and improvement activities for this indicator are the same as for Indicator 9. 33 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Monitoring Priority: Effective General Supervision Part B / Child Find Indicator 11: Percent of children with parental consent to evaluate, who were evaluated within 60 days (or State established timeline). (20 U.S.C. 1416(a)(3)(B)) Measurement: a. # of children for whom parental consent to evaluate was received. b. # determined not eligible whose evaluations were completed within State established timeline. c. # determined eligible whose evaluations were completed within State established timeline. Account for children included in a but not included in b or c. Indicate the range of days beyond the timeline when the evaluation was completed and any reasons for the delays. Percent = [(b + c) divided by (a)] times 100. FFY Measurable and Rigorous Target 2007 (2007-2008) 100% Actual Target Data for FFY 2006 (2006-2007): Based on 85 districts reporting on initial evaluations conducted in the months of October, November, and December of 2006: a. 2,825 children for whom parental consent to evaluate was received b. 1,015 children whose initial evaluations were completed within the State established timeline and found not eligible for special education services c. 1,636 children whose initial evaluations were completed within the State established timeline and found eligible for special education services % of children with parental consent to evaluate, who were evaluated within the State established timeline: 93.8% (1015+1636)/2825 Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): The current figure of 93.8% includes cases where the evaluation was completed within the State established timeline, and also includes cases in which the district had an acceptable reason for not meeting the timeline. These are reasons that were beyond the district’s control, including school cancellation due to weather, parent scheduling needs, and significant student absenteeism. In order to obtain student level data, MASSDE required each district selected for this monitoring activity to collect information and report data on all initial evaluations conducted during October, November, and 34 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 December of 2006. For additional information on the data collection process as approved by OSEP, please refer to Indicator 11 of the MA SPP. In examining the 174 cases in which a district did not meet the timeline and did not have an acceptable reason, the most common reason was due to district scheduling conflicts (16.9% of the missed timelines). Insufficient staff availability and/or availability of outside evaluators was the second most common reason (10.0%), and “lack of qualified staff to complete the evaluation on time” was the other most common reason (5.3%). Of the cases that missed the timeline, the average number of days beyond the 45-day timeline for evaluation and eligibility determination was 15 school working days. This is an improvement of 2.5 working days from the previous data collection effort that was 17.5 working days. Compared with the first cohort’s data collection effort (88.5% in compliance) reported in FFY 2005, MASSDE has seen a rise of 5.3 percentage points in the number of students for whom initial evaluations are conducted within the State established timeline. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): In the FFY 2005 SPP/APR Response Table, OSEP noted that “data demonstrating correction of noncompliance identified in FFY 2005” must be provided in the FFY 2006 APR submission. Through the data collection and analysis process described above, 14 districts were identified as not in compliance with this indicator in FFY 2005. These districts were notified and completed corrective action reports that were submitted to MASSDE within one year of the date of the finding. The corrective action report required identified districts to detail the steps taken to remedy any noncompliance by describing the policies and procedures that have been implemented to ensure that the district is able to meet the required evaluation timelines. All 14 identified districts (100%) completed their corrective actions and have successfully remedied any noncompliance within one year of the date of the finding. The next cohort of districts will participate in data collection for this indicator in Spring 2008. MASSDE will refine it’s data collection instrument to collect more detailed information from districts regarding barriers to meeting State established timelines, and will use this information to assist districts in their efforts to achieve 100% compliance. All improvement activities, timelines and resources for FFY 2006 remain appropriate. 35 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Monitoring Priority: Effective General Supervision Part B / Effective Transition Indicator 12: Percent of children referred by Part C prior to age 3, who are found eligible for Part B, and who have an IEP developed and implemented by their third birthdays. (20 U.S.C. 1416(a)(3)(B)) Measurement: a. # of children who have been served in Part C and referred to Part B for eligibility determination. b. # of those referred determined to be NOT eligible and whose eligibilities were determined prior to their third birthdays. c. # of those found eligible who have an IEP developed and implemented by their third birthdays. d. # of children for whom parent refusal to provide consent caused delays in evaluation or initial services. Account for children included in a) but not included in b, c or d. Indicate the range of days beyond the third birthday when eligibility was determined and the IEP developed and the reasons for the delays. Percent = [(c) divided by (a – b – d)] times 100. FFY Measurable and Rigorous Target 2006 (2006-2007) 100% Actual Target Data for FFY 2006 (2006-2007): a. 5,515 children served in Part C and referred to Part B for eligibility determination. b. 866 of those referred determined to be NOT eligible and whose eligibilities were determined prior to their third birthdays. c. 3099 of those found eligible who have an IEP developed and implemented by their third birthdays. d. 789 children for whom parent refusal to provide consent caused delays in evaluation or initial services. [(3099) divided by (5515 – 866 – 789) x 100] = 80.3% This shows an increase of 1% from the FFY 2004 baseline (77%) and a 6% increase from the FFY 2005 APR submission (72% OSEP revised). Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): In FFY 2006, the Massachusetts Department of Early Education and Care (EEC) revised its data collection to align with the Measurement Criteria in Indicator 12 so we can more accurately report on LEAs’ compliance performance. This year, we asked LEAs to report on the number of children whose 36 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 eligibility could not be determined due to parent refusal to provide consent, continue through the eligibility determination, or return a signed IEP. Districts reported 789 (14%) of parents refused consent, refused to follow through with the eligibility determination or to return a proposed IEP. In addition, this year we asked districts to report on the number of eligible children served after they turned three, and how many had proposed IEPs prior to turning three. Districts originally reported of the 705, and with revised data corrections yielded 697, children served after age 3, 439 (62%) of them had an IEP proposed to the family by the child’s third birthday. While we can only speculate why these proposed IEPs were not implemented, we know that districts are not consistently allowing 30 calendar days for parents to review, sign and return an IEP. Of the 697, 62 were served one to two days following their third birthdays; 140 were served 3-5 days following their third birthdays; and 100 were served 6-10 days after turning three; 136 were served 2-4 weeks after turning three; and finally 259 were served 1 month or more after turning three. We also collected data on the reasons for delays in serving eligible children by age three. Of the 705 children served after three, districts reported reasons on 265 children who did not have an IEP proposed and presented to the parent prior to the child’s third birthday. Unfortunately, we can put little faith in the reliability of these data. For example, Indicator 12 speaks specifically to children coming from Part C, found eligible and served. Many of the reasons for delays speak to children referred from Head Start after turning three; parents who refused to attend meetings or changed appointments; late referrals that did not allow sufficient time to complete the process. All of these reasons fall outside the scope of what is being evaluated as part of Indicator 12 Part C to Part B transitions. In addition, many districts misinterpreted the definition of an implemented IEP. Districts applied a very literal interpretation of served by age three and counted those children with summer birthdays, or birthdays on weekends or school vacations as served after turning three, but they had signed IEPs from the families prior to the child’s third birthday. Next year, EEC intends to provide districts with an online data collection tool as part of the early childhood special education grant, with cell protections and drop down menus with reasons. The districts will select from a standard menu so that one district’s parent refusal is not another district’s parent who postpones meetings; or allow reports of Head Start referrals. In other words, this online system will create consistency in data reporting across all of our LEAs. In addition, all data will automatically transfer into a database, eliminating data entry errors that can exist under our current reporting system. We also asked districts to report of the number of referrals from Part C/Early Intervention, how many were received at least 45 school working days prior to the child’s third birthday (the state’s eligibility timeline from consent to IEP). Districts reported of the 6,368 children referred, 5,515 (86.6%) referrals were received at least 45 school working days prior to age three. EEC was interested in knowing whether referral timing would affect data so we analyzed districts in substantial compliance (95% and above). Forty-eight and a half percent (48.5%) of districts are in substantial compliance (118 of the 243 reporting) Data show that in cases where the district is in substantial compliance (95% to 100%), 91.3% of the referrals from Part C were received at least 45 working days in advance of the child’s third birthday. In contrast, for those districts not in substantial compliance, 84% of the referrals from Part C were received within 45 working days in advance of the child’s 3rd birthday. In theory, this demonstrates that the ability to complete a child’s eligibility determination and smoothly transition him or her from Part C to Part B depends largely on the collaboration between EI programs and LEAs in making referrals, communicating and exchanging information in a timely manner. EEC has addressed these transition issues with several initiatives. In April 2007, EEC, DPH (Part C) and DOE collaborated to conduct two statewide trainings with both LEA Early Childhood staff/administrators and EI program staff/directors to review each entity’s requirements/regulations, our common SPP Indicator on transition, and brainstorm how to improve referrals, eligibility determination and ultimately reach compliance. Attendance was encouraging, with the majority of LEAs and EI programs present. We reiterated in the training that districts have to allow 30 calendar days for parents to return an IEP so services can commence. As a result, we expect a significant decrease next year in the number of children whose IEP was not in place by their 3rd birthday as a result of the district not allowing 30 days for the parents to sign and return the IEP. A copy of the Power Point is available at http://www.eec.state.ma.us/docs/Transition%20Workshop%204-07.pdf In addition, EEC is collaborating with DPH as they adapt an online training module developed by the Central Regional Resource Center. It will be a required training to meet EI competencies. EEC is currently 37 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 in discussion on how we will disseminate the training information beyond making it available to LEAs on our website. From a Part C perspective, the objective of the training is to improve transition practices. Part C’s SPP/APR from FFY04 demonstrates only 61.5% of children discharged from EI and potentially eligible for Part B had a “yes” answer to “was the LEA invited?” on the Transition Survey on the Annual Report/Self-Assessment. And 85.2% of children discharged from EI and potentially eligible for Part B had a “yes” answer to “Did a transition conference occur for this child?” In working together across agencies, we expect improvement to be inevitable. As articulated in the FFY 2005 APR, DPH and EEC completed and distributed the Best Practices in Early Childhood Transition: Guide for Families. Every EI program received five hard copies; each LEA received a hard copy. The guide is available on our website at http://www.eec.state.ma.us/docs/EECTransitionGuideForWeb(3).pdf EEC and MASSDE are currently updating an advisory of frequently asked questions relative to young children, ages 3 to 5, with disabilities. We expect to disseminate that document to all EI, LEA, Head Start and community child care programs. While OSEP is collecting data on only those children served by Part C and referred and found (in)eligible by age 3, we know that not all preschool children served in early childhood special education were served in Part C. We collect referral numbers from child care, Head Start, parents and other. In FFY 2006 (the 2006-07 school year), 8,231 children were referred from entities other than EI. EEC is hiring a Transition Specialist who will work with districts/programs and oversee all aspects of transition for young children and their families. It is our goal to improve transition practices from wherever children are coming from. Massachusetts was selected by OSEP as one of three states this year involved in Expanding Opportunities, a multi-agency team whose mission it is that all children and families have universal access to child-centered, family-friendly, inclusive settings and a seamless system of coordinated supports and services. As a team we decided to also apply to become a Special Quest Birth-Five State Leadership Team to assist us in developing a statewide plan for professional development to improve inclusive practices, including seamless transitions at age three and at age five. Our Special Quest team members are from EI, DOE, EEC, Center for Excellence, higher ed, child care, Head Start and parents. The application was submitted January 14, 2008 and the award was received in February. EEC and MASSDE have developed a consistent approach to work with districts not in substantial compliance on each of the compliance indicators. For Indicator 12, letters were sent to districts below substantial compliance for FFY 2005 (96 districts) and FFY 2006 (91 districts) and included the transition data that they reported on their early childhood special education grants. LEAs are required to verify their data and submit corrections, and review and, if applicable, modify the district’s policies and procedures around transition in order to bring the district into substantial compliance. Each district must submit a Corrective Action Report to detail the steps taken to remedy any noncompliance by describing the policies and procedures that have been implemented to ensure compliance. New policies and procedures will be shared with districts as samples. Of the 96 districts in FFY2005 in non-compliance, 13 LEAs submitted revised/corrected existing data and came into compliance, and 26 corrected the noncompliance within one year of identification through a Corrective Action Plan for an overall timely correction rate of 41%. The remaining 57 LEAs are currently engaged in completing their corrective actions and the Department will continue to work with them until the corrective actions are completed. In addition, EEC is working with DOE’s Program Quality Assurance Unit (Indicator 15: General supervision system, including monitoring, complaints, hearings, etc.) to identify and correct noncompliance as soon as possible but in no case later than one year from identification. Of the districts scheduled for a CPR review that operate preschool programs (some are regional or vocational schools and K-12 charter schools), the districts found not in substantial compliance with their early childhood transition practices from Part C to Part B, will submit their corrective action report to PQA on SE 17 (see table below) as part of the compliance monitoring process. 38 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Criterion Number SE 17 Revised April 14, 2008 Initiation of services at age three and Early Intervention transition procedures Student Records SPED Administrator 1. Special Education Program Plan Statement Early Childhood Coordinator 2. The school district encourages referrals from the Department of Public Health, other agencies, and individuals for young children when or before the child turns two-and-one-half years old in order to ensure continuity of services and to ensure the development and implementation of an IEP for eligible children by the date of the child's third birthday in accordance with federal requirements. The district implements procedures to ensure the effective transition of young children with disabilities from Early Intervention Programs through participation in transition planning conferences arranged by such programs. Team Chairperson Parents State and Federal Requirements 603 CMR 28.06(7)(b) 34 CFR 300.101(b); 300.124; 300.323(b) Part 1 of SE 17 is related to State Performance Plan Indicator 12. (See http://www.doe.mass.edu/sped/spp/.) Source: http://www.doe.mass.edu/pqa/review/cpr/instrument/sped.doc EEC participates with MASSDE/PQA on CPR reviews and will look closely at the district’s practices, interview early childhood staff and provide the necessary technical assistance. EEC will also look at districts in substantial compliance consecutive years to learn best practices to disseminate to other districts. EEC’s Inclusion Specialist will host a series of regional meetings to convene early childhood special education personnel from LEAs to share best practices in “Communities of Practice”. Topics will include transitions at age three. As reported in the FFY 2005 APR, the cross agency team convened by the Head Start Quality Initiative continues to work together on a transition MOU. We have had several drafts that have been under multiple revisions. We expect to complete it this school year. Each district’s compliance for each of the compliance indicators is publicly reported on DOE’s website at https://profiles.doe.mass.edu/special_ed.aspx Massachusetts has exceeded its activities planned for this year and will continue to collaborate with its partners DPH and DOE to strive for 100% substantial compliance across all Part C programs and LEAs. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): The targets, improvement activities, timelines and resources for FFY 2006 remain appropriate. 39 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Monitoring Priority: Effective General Supervision Part B / Effective Transition Indicator 13: Percent of youth aged 16 and above with transition planning that includes coordinated annual goals and transition services that will reasonably enable the student to meet his/her postsecondary goals in the identified areas. (20 U.S.C. 1416(a)(3)(B)) Measurement: Percent = [(# of youth with disabilities aged 16 and above with an IEP that includes coordinated, measurable, annual IEP goals and transition services that will reasonably enable the student to meet the post-secondary goals) divided by the (# of youth with an IEP age 16 and above)] times 100. FFY Measurable and Rigorous Target 2006 (2006-2007) 100% Actual Target Data for FFY 2006 (2006-2007): # of Student Records reviewed 2,111 # of Student Records with transition planning that included coordinated annual goals and transition services 2,085 Percentage of student records in compliance 98.8% Of the 2,111 student records reviewed during 2006-07, almost 99% include transition planning that includes coordinated annual goals and transition services that will reasonably enable the student to meet his/her post-secondary goals in the identified areas. In order to obtain student level data, MASSDE required each district selected for this monitoring activity to collect information on a representative sample of students aged 16-21 with IEPs. The sample student files were reviewed for evidence of full transition planning discussions. For additional information on the data collection process as approved by OSEP, please refer to Indicator 13 of the MA SPP. The data review process as conducted by districts was revised for FFY 2006 with the addition of the mandated Massachusetts Transition Planning Form (TPF). As was done with the initial data collection, a review sheet allowed districts to assess a student record for evidence of appropriate transition planning. Evidence of transition planning that includes coordinated annual goals and transition services that will reasonably enable the student to meet his/her post-secondary goals in the identified areas resulted in a finding of appropriate transition planning for the student record being reviewed. This data collection allowed districts to use the mandated Transition Planning Form (documenting full transition planning discussion) as appropriate evidence of transition planning or a completed MASSDE Transition Planning Chart (documenting full transition planning discussion), or a record review (with appropriate IEP documentation) indicating an appropriate transition planning discussion. If such documentation was not found in the student record or IEP, then the student was not considered to have received appropriate 40 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 transition planning. Districts were encouraged to provide optional comments detailing any aspect of the student’s transition plan. Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): The vast improvement over the 83.8% reported in FFY 2005 is in large part due to the implementation of the Massachusetts Transition Planning Form (TPF), a mandated form maintained with the IEP, issued in February 2007. Along with information on the MASSDE website that included completed samples, trainings on the use of the TPF were offered as part of The Comprehensive System of Personnel Development (CSPD) Training Project as part of a module titled: The Massachusetts Transition Planning Form and Effective Transition Planning. Other training opportunities included a Summer Content Institute on “Secondary School IEP Measurable Annual Goals -Objectives/Benchmark and Transition Planning Workshop” allowing participants to develop the following skills: become familiar with the laws and regulations that govern transition planning, better understand the components of transition planning, and facilitate discussions that will help students with disabilities work towards their post-secondary goals. Additional trainings and workshops have been offered (examples: Massachusetts Association of Approved Private Schools (MAAPS) conference, Perkins School For The Blind Discover conference) Additionally, courses and technical assistance were offered as part of MASSDE’s federal State Improvement Grant, Project FOCUS Academy. These focused on creating professional development programs to help students with disabilities build sound career goals and learn skills to ensure successful post-secondary outcomes. The content covered in this program (Positive Behavioral Interventions and Supports, Universal Design for Learning, post-secondary planning, and family engagement) could have a long-term impact on transition planning. Implementation activities are currently underway in participating districts. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): In the FFY 2005 SPP/APR Response Table, OSEP noted that “data demonstrating correction of noncompliance identified in FFY 2005” must be provided in the FFY 2006 APR submission. Through the data collection and analysis process described above, 19 districts were identified as not in compliance with this indicator in FFY 2005. These districts were notified and completed corrective action reports that were submitted to MASSDE within one year of the date of the finding. The corrective action report required identified districts to detail the steps taken to remedy any noncompliance by describing the policies and procedures that have been implemented to ensure that appropriate transition planning is provided to all students with disabilities ages 16 and older. Districts were also required to demonstrate that steps had been taken to ensure that transition planning has since taken place for any students reported in FFY 2005 as not having evidence of transition planning. All 19 identified districts (100%) completed their corrective actions and have successfully remedied any noncompliance within one year of the date of the finding. Regarding improvement activities, MASSDE was awarded a federal State Personnel Development Grant to continue the development of a system of free online professional development for general and special educators throughout the Commonwealth. This initiative, called Massachusetts FOCUS Academy, builds on Project FOCUS Academy (described above) and offers online courses in a variety of key areas, including positive behavioral supports, Universal Design for Learning, family engagement, and postsecondary planning. The first courses will be offered during the 2006-07 schools year, and include the first two courses in a four-course series on Post-Secondary Transition Planning. The target audience for the first course, Youth Development and Leadership, is middle and high school educators and guidance counselors working with students with disabilities. The course provides participants with a framework for 41 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 understanding how youth development - including adult mentoring, leadership, decision-making, problem solving, self-initiating, citizenship, and community service - prepares youth to meet the challenges of adolescence and adulthood. All other improvement activities, timelines and resources for FFY 2006 remain appropriate. 42 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Monitoring Priority: Effective General Supervision Part B / General Supervision Indicator 15: General supervision system (including monitoring, complaints, hearings, etc.) identifies and corrects noncompliance as soon as possible but in no case later than one year from identification. Measurement: Percent of noncompliance corrected within one year of identification: a. b. # of findings of noncompliance. # of corrections completed as soon as possible but in no case later than one year from identification. Percent = [(b) divided by (a)] times 100. For any noncompliance not corrected within one year of identification, describe what actions, including technical assistance and enforcement actions that the State has taken. FFY Measurable and Rigorous Target 2006 (2006-2007) 100% Actual Target Data for FFY 2006 (2006-2007): o 211 findings of special education noncompliance were made through the Problem Resolution System (PRS) between July 1, 2005 and June 30, 2006. 100% of these findings were corrected within one year of identification. o 844 findings of special education noncompliance were made through the Coordinated Program Review (CPR) reports published between July 1, 2005 and June 30, 2006. 53% (451) of these findings were corrected within one year of the date of the final CPR report. 87% (733) of these findings have been corrected to date. o 70 new findings of special education noncompliance were made through Mid-Cycle Review (MCR) reports published between July 1, 2005 and June 30, 2006. 64% (45) of these findings were corrected within one year of the date of the MCR report. 77% (54) of these findings have been corrected to date. o No findings of special education noncompliance were made through MASSDE’s Bureau of Special Education Appeals (dispute resolution system) between July 1, 2005 and June 30, 2006. a. 1125 findings of noncompliance were made through the PRS, CPR, or MCR system between July 1, 2005 and June 30, 2006. b. 707 findings of noncompliance were corrected within one year of the finding, for an overall oneyear rate of correction of 63%. (To date, 998 of the 1125 findings of noncompliance have been corrected, for a rate of correction to date of 89%.) 43 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 See the Appendix B: Indicator #15 Worksheet for the disaggregation of findings made from July 1, 2005 to June 30, 2006. Follow-up on Data for FFY 2005 (2005-2006): In the FFY 2005 SPP/APR Response Table, OSEP noted that “data on the correction of outstanding noncompliance identified in FFY 2004 (2004-2005)” must be provided in the FFY 2006 APR submission. As reported in the FFY 2005 APR, 62% of the 921 findings of special education noncompliance made in FFY 2004 through the CPR system had been corrected by the date of that report, February 1, 2007. As of the date of this report, 82% of the findings of special education noncompliance made in FFY 2004 have been corrected. (Note: In preparing the data for this report it was discovered that findings in some CPR reports were counted twice for last year’s report due to a data entry error. The more accurate figure for the number of findings made in CPR reports between July 1, 2004, and June 30, 2005 is 813, of which 667, or 82%, have been corrected to date.) Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): A. Discussion of Improvement Activities Completed: Year 2 (FFY 2006: 2006-07): MASSDE staff from Program Quality Assurance Services (PQA) made changes to PQA’s MCR based on the implications of the requirement that all corrective action be completed as soon as possible, but in no case later than one year from identification. Materials for the 2007-08 MCR cycle were revised to make clear that the purposes of the MCR no longer include checking on the implementation of corrective action approved or ordered by PQA after the CPR three years before, as that implementation will have been long before completed. Purposes of the MCR for the current (2007-08) monitoring year include monitoring special education issues that have been raised by complaints from a district; at the discretion of PQA staff, monitoring special education standards (“criteria”) with which noncompliance was found in the previous CPR to ensure the continuing effectiveness of the corrective action previously completed; and monitoring special education criteria which have been created or substantially changed, in response to IDEA-2004 and its regulations, since a district’s CPR. B. Explanation of Progress or Slippage: 100% of the 211 findings of noncompliance made through the Problem Resolution System for complaints between July 1, 2005 and June 30, 2006 were corrected within one year. This is the third reporting period for which MASSDE has maintained 100% compliance in this area (as reported in the MA SPP for 2004-05 and the APR for 2005-06, 100% of noncompliance from complaints was corrected within one year). Of the 844 findings of special education noncompliance that were made through the CPR reports published between July 1, 2005 and June 30, 2006, 53% (451) of these were corrected within one year of the date of the final CPR report. ( 87% (733) of these findings have been corrected to date.) Massachusetts fell markedly short of its target of 100% correction within one year. In the FFY 2005 APR for Indicator 15 submitted February 1, 2007, MASSDE reported that 62% of the findings of special education made through CPR reports published between July 1, 2004 and June 30, 2005, had been corrected “to-date”. MASSDE reported the data in this manner because the standard of correction within one year was not the standard originally applied to noncompliance found in 2004-05, and so the standard could not be fully applied to this time period. Thus the 53% reported for the 2005-06 CPR findings corrected within one year does not constitute slippage from the 62% reported as corrected to date for the 2004-05 CPR findings; in fact, the percentage of 2005-06 CPR findings corrected to date is significantly greater, 87%. 44 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Of the 70 new findings of special education noncompliance that were made through MCR reports published between July 1, 2005 and June 30, 2006, 64% (45) of these findings were corrected within one year of the date of the MCR report. 77% (54) of these findings have been corrected to date. Again, Massachusetts fell considerably short of its target of 100% correction within one year. It is impossible to tell whether 64% constitutes progress or slippage from the year before, as PQA staff only realized this year that the relatively small number of new findings in MCR reports ought to be counted for the purposes of Indicator 15. The reason for the greater percentage of MCR findings that were corrected within a year (64% versus 53% for CPR findings) is presumably largely that PQA staff orders corrective action itself as part of the MCR report, rather than giving the district a chance to develop its own corrective action to submit to PQA for approval, as is done with CPR findings. Thus time within the year is not lost waiting for the district’s corrective action plan and responding to it. Of the 1125 total findings of noncompliance made (through the PRS, CPR, or MCR system) during 2005-06, 707 findings of noncompliance were corrected within one year of the finding, for an overall one-year rate of correction of 63%. (89% (998) of these findings have been corrected to date.) Again, though the overall one-year correction rate for findings made during 2005-06 (63%) is less than the overall correction rate reported last year for findings made during 2004-05 (69%), the 63% does not constitute slippage from the 69%, as that figure was for correction to the date of the FFY 2005 APR. (The overall rate of correction to date for the findings made through the PRS, CPR, or MCR system in 2005-06 (89%) is considerably higher than the overall rate of correction reported last year for findings made through the PRS or CPR systems (69%).) C. Further Improvement Steps: MASSDE is fully aware of how far from 100% correction within one year it is for findings made through the CPR and MCR systems. Clearly, making the new one-year requirement clear to PQA staff and revising existing monitoring materials so as to feature prominently a statement of the one-year requirement have not been enough (see paragraph 2 of “Discussion of Improvement Activities Completed” for Year 1 (FFY 2005) in the FFY 2005 APR). The following are steps that MASSDE is taking in order to approach the target of 100% correction of noncompliance within one year. 1. (See paragraph 5 under “Discussion of Improvement Activities Completed” for Year 1 in the FFY 2005 APR.) In April, 2007, MASSDE hired Kyran Research Associates, Inc., of Rhode Island to develop a web-based monitoring system for PQA, described in Attachment C. PQA has begun piloting this new system during FFY 2007, a year ahead of the timeline described in the FFY 2005 APR. Its implementation will be expanded in FFY 2008 and successive years until the whole monitoring system has been switched over to the web-based model. This more efficient system, with its emphasis on self-assessment by the local agencies rather than full onsite visits by PQA staff, will, it is expected, result in quicker identification and correction of noncompliance. More responsibility on the part of school districts is expected to result in a certain amount of noncompliance being corrected immediately without MASSDE even having to notify districts of a finding. And once the web-based monitoring system is fully implemented, it will be easier for PQA to tailor its onsite investigations to monitoring priority areas or areas of exceptional need, perhaps reducing the number of CPR findings of noncompliance from the more than 800 currently made per year, thus facilitating the correction within one year of those findings. MASSDE has in the past employed a soup to nuts kind of compliance monitoring, but it is clear that MASSDE’s monitoring must be better targeted. It is also expected that the ability of the web-based system to generate compliance profiles of individual districts and for individual standards and indicators as well as aggregate compliance reporting (see Attachment C) will aid PQA in making sure that noncompliance is being corrected within one year. At the current time, PQA has no ongoing system of tracking compliance criterion by criterion. 2. Furthermore, Remedy’s Action Request software system, used to track the dates of publication of CPR and MCR reports, receipt of corrective action plans (CAPs), reviews of CAPs, receipt of progress reports, and reviews of progress reports, is unwieldy, difficult to use, and 45 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 unreliable, particularly in producing reports. It is therefore difficult now to produce a report that will give PQA administrators an accurate idea of the status of correction of noncompliance either across the board or district by district. Over the next year or two MASSDE will be phasing in new software, one of whose functions will the tracking of the correction of noncompliance. It is expected that the new software, like the web-based monitoring system, will make it easier to monitor the correction of noncompliance within one year. 3. (See paragraphs 1 and 2 of “Further Improvement Steps” in the FY 2005 APR.) PQA has begun piloting the electronic CAP/progress reporting system described in the FFY 2005 APR. The pilot did not begin until the beginning of FFY 2007; however, having been delayed by technical difficulties. It was discovered that having one electronic document containing all the relevant forms—the summary of required corrective action, a progress report status sheet, and one copy of the CAP form, CAP approval section, progress report form, and progress report approval section for each criterion found in noncompliance—was technically unfeasible. Once filled in, such a large document would be liable to crash or be impossible to open. New, separate forms and new instructions for using the forms therefore had to be developed. Because of this delay use of the electronic CAP/progress report system has not moved beyond the pilot, and feedback on the pilot has not yet been collected. If the pilot is successful, however, plans remain in place to use this system more widely until the web-based system described above is fully instituted. As previously stated, the expectation is that an electronic system will lead to speedier completion and verification of corrective action, along with generally improved communication between districts and PQA. 4. As an immediate step toward our target of 100% correction within one year, PQA will continue and intensify the training given to supervisors and staff with respect to the one-year correction requirement. It is clear that in spite of efforts over the last two and a half years many staff are still operating with assumptions and attitudes appropriate for PQA’s former system, accepted at that time by OSEP, in which the CPR CAP and progress reports were spread over a three-year period. (See description of this former system in the MA SPP.) PQA must make greater efforts to change the practice over to the one-year system of correction now required. Staff will be trained in guiding districts in structuring corrective action activities and progress reports on a timeline that leaves enough time to complete those activities and to make sure that they have been effective before the year elapses. Also, PQA is planning to hire a staff trainer; the position is going through the intra-agency approval process and is expected to be posted in the near future. The staff trainer hired will work to make consistent the approaches of the six different public school monitoring teams (for six different areas of Massachusetts); one of the trainer’s duties will be training staff on all six teams in managing the timeline issues involved in correcting all noncompliance as soon as possible but in no case later than a year from identification. D. Actions taken when CPR and MCR findings of noncompliance were not corrected within one year of identification (as noted above, all findings of noncompliance stemming from complaints were corrected within one year): General description: Under the new one-year correction mandate instituted at the beginning of the 2005-06 school year, corrective action must be completed within a year from the date of the CPR final report or MCR report. When a progress report from a district shows that corrective action, whether for a finding made in 2004-05 or a finding made in 2005-06, has not been taken by the time required or in the manner required or has not been effective in remedying noncompliance, the PQA follow-up liaison for that district makes a new order of corrective action. When a progress report is overdue from a district, the liaison will contact district personnel to ask for it. In either of these cases the liaison may well, in the course of communicating with the district on this matter, offer technical assistance on the corrective action necessary. Enforcement action is an option that is available when corrective action has not been completed as required, but it is seldom used unless the failure to complete the corrective action is egregious. 46 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Almost every district whose CPR was conducted in 2004-05 will be undergoing an MCR during 200708. During the MCR any noncompliance outstanding from the CPR will, of course, be investigated thoroughly, technical assistance rendered, and enforcement action taken if necessary. As described under “Further Improvement Steps” above (see C. 1 (end) and C. 2), PQA does not now have an adequate data system for tracking the timely correction of noncompliance found during CPRs and MCRs. Both the acquisition of new software and the institution of a new web-based monitoring system will remedy this lack, enabling action to be taken earlier when correction of noncompliance is taking too long and enabling PQA supervisors and administrators easily to identify cases of overdue correction. Program-specific follow-up activities related to uncorrected noncompliance: Findings of special education noncompliance made through CPR and MCR reports published between July 1, 2005 and June 30, 2006: Nineteen school districts have not yet corrected instances of noncompliance identified through the FFY 2005 (2005-2006) CPR and MCR reports. (One school district has had the correction of its four instances of noncompliance verified by an onsite visit since February 1, 2008, the date of the original submission of the FFY 2006 APR.) Of the 19 districts with outstanding noncompliance identified in FFY 2005, program-specific activities include: Four districts have had onsite visits conducted. Seven districts have been scheduled for onsite visits in the last few months of this school year; the last of these visits will begin on June 4, 2008. Two districts have onsite visits scheduled for the next school year (2008-2009). One district is implementing a plan to conduct an evaluation of special education programs and services, to correct the last issue remaining; the success of this implementation will be verified during an onsite visit scheduled for the next school year (2008-2009). Two districts have had overdue letters sent, and additional follow-up has been conducted. One district has been called by the PQA liaison to that district concerning an issue of FAPE (making counseling available in a student’s native language), and this one remaining issue will be followed up during an onsite visit scheduled for the next school year (2008-2009). One district has been notified concerning overdue progress reports. One district has been ordered to submit requested documentation (verifying implementation of a parent advisory council for special education—the last issue remaining) by April 15, 2008. Findings of special education noncompliance made through CPR reports published between July 1, 2004 and June 30, 2005: All districts with outstanding instances of noncompliance that were identified through CPR reports published between July 1, 2004 and June 30, 2005, have either received an onsite visit or are scheduled to receive one by June 2, 2008. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): The targets, improvement activities, timelines and resources for FFY 2006 remain appropriate. 47 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Monitoring Priority: Effective General Supervision Part B / General Supervision Indicator 16: Percent of signed written complaints with reports issued that were resolved within 60-day timeline or a timeline extended for exceptional circumstances with respect to a particular complaint. (20 U.S.C. 1416(a)(3)(B)) Measurement (data from Appendix D: Table 7): Percent = [(1.1(b) + 1.1(c)) divided by 1.1] times 100. Measurable and Rigorous Target FFY 2006 (2006-2007) 100% Actual Target Data for FFY 2006 (2006-2007): [(179 reports within timeline + 36 reports within extended timeline) / 236 complaints with reports issued] x 100 = 91% Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): Discussion of Improvement Activities Completed: Status of Improvement Activities Previously Completed: 1. The system of more intensive monitoring developed in FFY 2005 (first improvement activity for FFY 2005 in the Massachusetts SPP) continues to be implemented, particularly for complaints where the timelines have been extended. Darlene Lynch, Director of PQA, periodically prints out logs for every staff person showing complaints assigned and timelines met or missed, then disseminates these logs to the appropriate supervisor with comments on compliance or on needed improvement. 2. The modification made to Remedy’s Action Request System software in FFY 2005 (second improvement activity for FFY 2005) remains in effect, so that PQA staff continue to receive more frequent reminders from the system that action with respect to a complaint is required. 3. (See fourth improvement activity for FFY 2005 in the Massachusetts SPP.) PQA procedures, as described in the Problem Resolution System (PRS) Implementation Guide for PQA staff, continue to provide for an independent DOE investigation of the complaint where the local agency indicates that it does not intend to provide a local report in response to the complaint. (This kind of independent investigation is becoming an increasingly useful tool for making sure that complaints are handled in a timely manner.) Procedures also continue to provide that where the local agency indicates that it will not respond to the complaint within the required timelines, PQA staff may, with the approval of a supervisor, issue a letter of finding based on the complainant’s documentation. 4. (See Explanation of Progress or Slippage in the Massachusetts APR for FFY 2005.) Supervisors continue to make clear to PQA staff that where timelines are extended for exceptional circumstances with respect to a particular complaint, the extension should allow not only enough Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 time for the school or district to submit additional information but also enough time for the PQA staff person to review that information, obtain feedback on that information from the complainant, consult with the PQA supervisor, and write and issue a letter of finding or letter of closure (taking into account which days are holidays or weekends and when the staff person will be out of the office on Coordinated Program Reviews or Mid-cycle Reviews). 5. (See explanation of Progress or Slippage in APR for FFY 2005.) The PRS Implementation Guide for PQA Staff continues to include guidance on the duration of extensions. Improvement Activities Recently Completed: 1. (See first and second improvement activities for FFY 2007 in the Massachusetts SPP.) PQA has prepared a statistical report on complaint resolution in connection with preparing this APR, and has analyzed the reasons for noncompliance and barriers to timely compliance (see Explanation of Progress or Slippage below). 2. (See third improvement activity for FFY 2007 in the Massachusetts SPP.) PQA has not yet implemented any new modifications to the PRS as a result of this analysis, but it is planning one, as well as working on a modification to another system which impacts the PRS. a. Over the next year or two MASSDE will be phasing in new software one of whose functions will be the tracking of complaints and complaint resolution. This software will replace Remedy’s Action Request System, which has proven unwieldy, difficult to use, and unreliable. It is unreliable in producing reports and is inadequate when complaints fall into more than one category, have more than one closure code, or cross programs. (The same system is used to track special education complaints as is used to track complaints in a variety of other areas.) Also, this software system's categories do not match the lines used in Table 7. This year, these deficiencies resulted in a need to submit an amended Table 7. The amended Table 7 was submitted in January 2008. It is hoped that the new software will make it easier to monitor PQA staff’s adherence to complaint timelines more frequently and efficiently. b. Also, PQA has begun implementation of a web-based system for monitoring school districts and charter schools. Its implementation will be expanded in FFY 2008 (2008-09) and successive years until the whole monitoring system has been switched over to the web-based model. It is hoped that this more efficient system, with its emphasis on selfassessment by the local agencies rather than full onsite visits by PQA staff, will give PQA staff more time to spend on complaint management, thus making it easier to meet the 60day timeline. An added benefit of the new system’s emphasis on self-assessment may be that local agencies will be more proactive in identifying their own problems, thus reducing the need for complaints. Explanation of Progress or Slippage: Although MASSDE did not meet its target for FFY 2006 of 100%, it made progress on the target of 100% compliance. It increased the percent of signed written special education complaints with reports issued that were resolved within 60 days, or within a timeline extended for exceptional circumstances with respect to a particular complaint, from 81% in FFY 2005 to 91% in FFY 2006. This progress was due both to the continuing, intensive monitoring of the management of complaints in general (see paragraph 1 under “Status of Improvement Activities Previously Completed” above) and to the special emphasis placed on complaints where the timelines are extended (see paragraphs 1, 4, and 5 of that same section). Where the timeline was not extended, the percent of signed written complaints that were resolved within 60 days was 96% (179 out of 186); where the timeline was extended, the percent of signed written complaints that were resolved within the extension was 72% (36 out of 50). Though the percent of complaints resolved in a timely manner increased from 89% to 96% where timelines were not extended, 49 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 an even bigger gain came in the group of complaints where timelines were extended: the percent of complaints resolved in timely manner increased for this group from 52% in FFY 2005 to 72% in FFY 2006, a gain of 20 points. It appears that the ongoing improvement activities listed above (see paragraphs 1 – 5 under “Status of Improvement Activities Previously Completed”) are paying off. PQA will continue to press on with its more intensive system of monitoring for the management of complaints, as well as with its special emphasis on complaints where the timelines are extended. PQA is also planning improvements to software and to the monitoring of districts (described above in paragraph 2 of “Improvement Activities Recently Completed”) that should help improve the percentage of complaints resolved in a timely manner. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): Given the progress made over the past year, the targets, improvement activities, timelines and resources for FFY 2006 remain appropriate at this time. 50 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Monitoring Priority: Effective General Supervision Part B / General Supervision Indicator 17: Percent of fully adjudicated due process hearing requests that were fully adjudicated within the 45 day timeline or a timeline properly extended by the hearing officer at the request of either party. (20 U.S.C. 1416(a)(3)(B)) Measurement (data from Appendix D: Table 7): Percent= [3.2(a) and 3.2 (b) divided by 3.2] times 100 Measurable and Rigorous Target FFY 2006 (2006-2007) 100% Actual Target Data for FFY 2006 (2006-2007): # of Hearings (fully adjudicated) 26 Decisions issued within 45-day timeline or a timeline that is properly extended by the hearing officer at the request of either party. 25 % of fully adjudicated due process hearing requests that were fully adjudicated within the 45-day timeline or a timeline that is properly extended by the hearing officer at the request of either party. 96.2% Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): This represents an improvement over the percentage reported in FFY 2005 (88.8%) and is reflective of only one decision that was not issued within the 45-day timeline or timeline properly extended. The following explanation addresses the reason said decision was not issued within the timeline. First, it should be noted that the hearing officer who authored this decision is no longer employed by the BSEA. Second, the decision in question was the result of an LEA filed (as opposed to a parent filed) hearing request. In an effort to ensure uniformity in scheduling parent and LEA hearing requests, initial hearing dates for both were set by the BSEA 35 calendar days subsequent to receipt of the hearing request. This was done to accommodate the 30-day resolution period mandated for parental hearing requests (despite 51 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 the fact that such resolution period was not mandated for LEA hearing requests). As a result, hearing officers had a 40 calendar day period to issue decisions in parent requested hearings (as only five days of the 45 days allocated post-resolution period had been utilized in scheduling the initial 35 day date); however, given that in LEA filed hearing requests the 45 day timeline begins to run on the date the hearing request is filed, the hearing officer in such instance was only left with a period of 10 calendar days to issue the decision. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): The BSEA is rectifying this scheduling anomaly by revising the BSEA Hearing Rules to reflect an initial hearing date in LEA filed requests which is 20 calendar days from the date of the filing of the request, thus allowing the hearing officer 25 calendar days to issue the decision. While the BSEA had hoped to launch the revised rules during this past fiscal year, promulgation was delayed, and said revised rules will be issued effective February 1, 2008. 52 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Monitoring Priority: Effective General Supervision Part B / General Supervision Indicator 18: Percent of hearing requests that went to resolution sessions that were resolved through resolution session settlement agreements. (20 U.S.C. 1416(a)(3)(B)) Measurement (data from Appendix D: Table 7): Percent= [3.1(a) divided by 3.1] times 100 FFY Measurable and Rigorous Target 2006 (2006-2007) 48% Actual Target Data for FFY 2006 (2006-2007): (25/400) x 100 = 6.25% Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): The 6.25% reported for FFY 2006 is significantly below the FFY 2005 baseline data of 48%, and reflects the difficulties encountered when trying to collect data for this indicator. As noted in the MA SPP, the baseline data submitted for last year were not deemed reliable, and also utilized a different tool for measurement. Thus, there is no basis for comparison between data reported in FFY 2005 and FFY 2006, and all data for this indicator should be viewed with caution. Regarding the number of resolution sessions that were resolved through resolution settlement agreements, there is a large discrepancy between the number reported in FFY 2006 (25) and the number reported in FFY 2005 (212). As detailed in the MA SPP, the 212 cases reported in FFY 2005 represented 50% of all cases involving parental requests for hearing that were not resolved through mediation or substantive hearing officer decision on merits. While MASSDE felt that this method was appropriate for reporting in FFY 2005, we have since implemented procedures to improve the reliability of these data. In FFY 2006, the Bureau of Special Education Appeals (BSEA) utilized a form that was to be submitted by a party who is withdrawing a hearing request as the result of a settlement agreement reached at a resolution session. The 25 settlement agreements reported for FFY 2006 represent the number of such forms the BSEA received during the reporting period. However, there is no way to ensure that each party whose case was resolved via a settlement agreement reached at a resolution session in fact submitted such a form. In order to further improved the reliability of these data, BSEA has launched a system in which parties who had filed hearing requests but did not proceed to hearing were called by a BSEA staff person to inquire as to the means by which the matter had been resolved. However, this method has not proven as successful as anticipated, in that some parties did not respond and of those that did there was apparent confusion regarding terminology; thus it is not clear that the meetings reported were all actually resolution sessions nor that the settlements reported were all as result of resolution 53 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 sessions. As a result, the 25 settlement agreements reported for FFY 2006 should be viewed with caution. MASSDE and BSEA will continue to work toward improving this data collection system in order to ensure more accurate data. Regarding the number of resolution sessions held, the 400 reported for FFY 2006 represent a slight decrease from the 442 reported in FFY 2005. However, as described in the MA SPP, these data should be interpreted with caution for the following reasons: o there may be cases in which both a resolution session and a mediation were held; o there are likely cases in which both parties waived the resolution session and did not opt for mediation; and o there are likely cases in which the LEA failed to timely convene a resolution meeting within the 15 days and therefore it was constructively waived. As previously noted, MASSDE and the BSEA will continue to work to improve the data collection for this indicator. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): The targets, improvement activities, timelines and resources for FFY 2006 remain as originally proposed. MASSDE intends to consider revision only if a good data collection method is available. Until such time, a comparison of years will continue as a challenge for this indicator. 54 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Monitoring Priority: Effective General Supervision Part B / General Supervision Indicator 19: Percent of mediations held that resulted in mediation agreements. (20 U.S.C. 1416(a)(3)(B)) Measurement (data from Appendix D: Table 7): Percent= [2.1(a)(i) and 2.1(b)(i) divided by 2.1] times 100 Measurable and Rigorous Target FFY 2006 (2006-2007) 86% Actual Target Data for FFY 2006 (2006-2007): # of Mediations 841 # of Mediation Agreements 704 % of Mediations Held that Resulted in Mediation Agreements 83.7% Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): Improvement Activities Completed: All improvement activities identified in the MA SPP for FFY 2006 have been completed, and FFY 2007 improvement activities identified to-date in the MA SPP have been completed. Explanation of Progress or Slippage: The MASSDE mediation program is managed by the Bureau of Special Education Appeals, Mediation Office (BSEA-Mediation) and is nationally recognized as providing highly effective mediation services. The MA SPP identified our target setting for this indicator to be a maintenance target as it was strongly felt that although tracking mediation agreements was important, it would be inappropriate to suggest that we seek to “compel” parties in mediation to reach agreement. Therefore, while the 83.7% reported for FFY 2006 represents only slight improvement over the FFY 2005 percentage (83.4%) and is slightly lower than our FFY 2004 baseline year of 85.9%, we believe that 83.7% appropriately meets our essential goal 55 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 of maintaining a high level of mediation agreements. It should also be noted that the raw number of mediations conducted this fiscal year increased by 8.8% over the prior year. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): The targets, improvement activities, timelines and resources for FFY 2006 remain appropriate at this time. 56 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 Monitoring Priority: Effective General Supervision Part B / General Supervision Indicator 20: State reported data (618 data and Massachusetts State Performance Plan and Annual Performance Report) are timely and accurate. (20 U.S.C. 1416(a)(3)(B)) Measurement: State reported data, including 618 data and annual performance reports, are assessed according to the following standards: APR data are scored to ensure that data are: a. Considered valid and reliable; b. Analyzed using an OSEP-approved calculation; and c. Supplied in a manner consistent with the instructions given. 618 data are scored to ensure that data are: a. b. c. d. Timely; Complete; Able to pass edit check; and Explained when necessary through data notes. Please see Appendix E for scoring rubric used to calculate this measurement. FFY Measurable and Rigorous Target 2006 (2006-2007) 100% Actual Target Data for FFY 2006 (2006-2007): 96.6% on-time submissions, 100% of data submitted Indicator #20 Calculation A. APR Grand Total = 63 B. 618 Grand Total = 50 C. APR Grand Total (A) + 618 Grand Total (B) = 113 D. Subtotal (C divided by 119)* = E. Indicator Score (Subtotal (D) x 100) = 0.950 95.0% 57 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 The target data for FFY 2006 were based on MASSDE’s ability to produce its required data submissions for FFY 2006 (618 data: Tables 1-7, and the MA SPP/APR submission) in a manner consistent with OSEP’s data submission requirements. The percent compliance indicates the percentage of data submissions that were successfully submitted for FFY 2006. For further explanation on how this calculation was derived please refer to the scoring rubric located in Appendix E. Discussion of Improvement Activities Completed and Explanation of Progress or Slippage that occurred for FFY 2006 (2006-2007): In FFY 2005, MASSDE achieved a score of 91.6% on its submission of required state data. For FFY 2006, MASSDE showed improvement with this Indicator by achieving a score of 96.6%. However, due to unforeseen personnel changes, and the time needed for cleaning and checking of the data, delays continued to occur this year in some of our data submissions. This year, in addition to districts reporting their data through our District and School Staff Report (DSSR), 16 pilot districts reported their personnel data through MASSDE’s new Education Personnel Information Management System (EPIMS). While the DSSR collects data at an aggregate level for each school and district, EPIMS is designed to collect personnel data at the individual level. It is planned that all districts will be able to report their personnel data through EPIMS in FFY 2007 so that we may continue to provide timely and accurate personnel data for Table 2 and through the EDEN system. The Table 5 Discipline data continues to be on a cycle where we close the data submission window in October of the following academic year. This allows districts the time to submit accurate and complete data but makes it difficult for us to submit the data by the November 1 deadline. We anticipate continuing to work with our data timelines to meet the compliance deadline by FFY 2010. Revisions, with Justification, to Proposed Targets / Improvement Activities / Timelines / Resources for FFY 2006 (2006-2007): As evidenced through the measurement description and data provided above, MASSDE began using the OSEP-recommended Indicator 20 scoring rubric in FFY 2006 (see Appendix E). As a result, data for the previous years has changed, and MASSDE is submitting a revised State Performance Plan submission for this indicator. The baseline data (FFY 2004) is now 52.9%, and data for FFY 2005 is 91.6%. Given the progress made over the past year, the targets, improvement activities, timelines and resources for FFY 2006 remain appropriate. 58 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 APPENDIX A -- Massachusetts Parent Survey for Special Education for Indicator #8 This is a survey for parents of students receiving special education services. Your responses will help guide efforts to improve services and results for children and families. (Note: If you have more than one child currently receiving special education services, you may choose to submit one or more surveys, based upon your experiences as related to your children.) For each statement below, please select one of the following response choices: very strongly disagree (VSD), strongly disagree (SD), disagree (D), agree (A), strongly agree (SA), very strongly agree (VSA). You may skip any item that you feel does not apply to you or your child. Schools' Efforts to Partner with Parents 1. I am considered an equal partner with teachers and other professionals in planning my child's program. 2. I was offered special assistance (such as child care) so that I could participate in the Individualized Educational Program (IEP) meeting. 3. At the IEP meeting, we discussed how my child would participate in statewide assessments. 4. My child’s teachers give me enough time and opportunities to discuss my child’s needs and progress. 5. All of my concerns and recommendations were documented on the IEP. 6. Teachers and administrators at my child’s school invite me to share my knowledge and experience with school personnel. 7. I was given information about organizations that offer support for parents of students with disabilities. 8. I have been asked for my opinion about how well special education services are meeting my child's needs. 9. 10. 12. Teachers treat me as a team member. VSA VSD A SA VSA I feel I can disagree with my child’s special education program or services without negative consequences for me or my child. Teachers and administrators: 15. Agree SA IEP meetings are scheduled at a time and place that are convenient for me. Teachers are available to speak with me. 14. A My child's evaluation report is written in terms I understand. 11. 13. Disagree VSD SD D - seek out parent input. - show sensitivity to the needs of students with disabilities and their families. SD D Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 16. - encourage me to participate in the decision-making process. 17. - respect my cultural heritage. The school: Submitted February 1, 2008 VSD SD D A SA 18. - has a person on staff who is available to answer parents' questions. 19. - communicates regularly with me regarding my child's progress on IEP goals. 20. - gives me choices with regard to services that address my child's needs. 21. - offers parents training about special education issues. 22. - offers parents a variety of ways to communicate with teachers. 23. - gives parents the help they may need to play an active role in their child's education. 24. - provides information on agencies that can assist my child in the transition from school. 25. - explains what options parents have if they disagree with a decision of the school. Demographic Information 26. Number of years child has received special education services Less than 1 year 1-3 years 4-7 years More than 7 years 27. Child’s race/ethnicity White Black or African-American Hispanic or Latino Asian or Pacific Islander American Indian or Alaskan Native Multi-racial 28. Language spoken in the home English Spanish Portuguese Chinese Creole/Haitian Vietnamese Other _________________ 29. Child’s school level Elementary School Middle School High School 30. Child’s Primary Disability (check one) Autism Communication Impairment Deaf-Blind Impairment Developmental Delay Emotional Impairment Health Impairment Hearing Impairment Intellectual Impairment Multiple Disabilities Neurological Impairment Physical Impairment Specific Learning Disability Vision Impairment 60 VSA Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 APPENDIX B -- Indicator #15 Worksheet (Note: Indicator 4A has been added to this worksheet. Also, the description of what monitoring consists of has been revised to be appropriate to MASSDE, and the general supervision system categories “Dispute Resolution” and “Other” have been removed, as there were no findings made in these categories during this time period.) Indicator 1. Percent of youth with IEPs graduating from high school with a regular diploma. 2. Percent of youth with IEPs dropping out of high school. General Supervision System Components Monitoring: Coordinated Program Reviews, Mid-Cycle Reviews, complaints 13. Percent of youth aged 16 and above with IEP that includes coordinated, measurable, annual IEP goals and transition services that will reasonably enable student to meet the postsecondary goals. 14. Percent of youth who had IEPs, are no longer in secondary school and who have been competitively employed, enrolled in some type of postsecondary school, or both, within one year of leaving high school. 3. Participation and performance of children with disabilities on statewide assessments. 7. Percent of preschool children with IEPs who demonstrated improved outcomes. 4A. Percent of districts with a significant discrepancy in the rates of suspension/expulsion of students with disabilities # of Programs Monitored (a) # of Findings of noncompliance identified in FFY 2005 (7/1/05 – 6/30/06) (b) # of Findings from (a) for which correction was verified no later than one year from identification 81 47 240 126 57 36 For: CPRs: 51 MCRs: 31 Complaints: 2 Monitoring: Coordinated Program Reviews, Mid-Cycle Reviews, complaints For: CPRs: 51 Monitoring: Coordinated Program Reviews, Mid-Cycle Reviews, complaints For: CPRs: 51 MCRs: 31 Complaints: 0 MCRs: 31 Complaints: 3 61 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Indicator 5. Percent of children with IEPs aged 6 through 21 educational placements. 6. Percent of preschool children aged 3 through 5 – early childhood placement. General Supervision System Components Monitoring: Coordinated Program Reviews, Mid-Cycle Reviews, complaints 8. Percent of parents with a child receiving special education services who report that schools facilitated parent involvement as a means of improving services and results for children with disabilities. Monitoring: Coordinated Program Reviews, Mid-Cycle Reviews, complaints 9. Percent of districts with disproportionate representation of racial and ethnic groups in special education that is the result of inappropriate identification. Monitoring: Coordinated Program Reviews, Mid-Cycle Reviews, complaints 10. Percent of districts with disproportionate representation of racial and ethnic groups in specific disability categories that is the result of inappropriate identification. 11. Percent of children who were evaluated within 60 days of receiving parental consent for initial evaluation or, if the State establishes a timeframe within which the evaluation must be conducted, within that timeframe. Monitoring: Coordinated Program Reviews, MidCycle Reviews, complaints # of Programs Monitored Submitted February 1, 2008 (a) # of Findings of noncompliance identified in FFY 2005 (7/1/05 – 6/30/06) (b) # of Findings from (a) for which correction was verified no later than one year from identification 123 61 207 117 63 32 77 52 For: CPRs: 51 MCRs: 31 Complaints: 1 For: CPRs: 51 MCRs: 31 Complaints: 2 For: CPRs: 51 MCRs: 31 Complaints: 0 For: CPRs: 51 MCRs: 31 Complaints: 20 62 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Indicator General Supervision System Components 12. Percent of children referred by Part C prior to age 3, who are found eligible for Part B, and who have an IEP developed and implemented by their third birthdays. Monitoring: Coordinated Program Reviews, MidCycle Reviews, complaints # of Programs Monitored Submitted February 1, 2008 (a) # of Findings of noncompliance identified in FFY 2005 (7/1/05 – 6/30/06) (b) # of Findings from (a) for which correction was verified no later than one year from identification 22 12 94 46 9 6 973 535 For: CPRs: 51 MCRs: 31 Complaints: 1 Other Topical Areas Faculty, Staff, and Administration Record Keeping Monitoring: Coordinated Program Reviews, MidCycle Reviews, complaints Monitoring: Coordinated Program Reviews, MidCycle Reviews, complaints For: CPRs: 51 MCRs: 31 Complaints: 1 For: CPRs: 51 MCRs: 31 Complaints: 1 Sum the numbers down Column a and Column b Percent of noncompliance corrected within one year of identification = (column (b) sum divided by column (a) sum) times 100: 535/973 x 100 = 55% Notes: 1. The numbers of CPR findings of noncompliance counted into the number of findings of noncompliance for each indicator/indicator group or other topical area, when added up, do not equal the number of findings of noncompliance made through CPRs reported in the APR for Indicator 15. Likewise, the numbers of MCR findings of noncompliance counted into the number of findings of noncompliance for each indicator/indicator group or other topical area, when added up, do not equal the number of findings of noncompliance made through MCRs reported in the APR for Indicator 15. The reason for this is that the criteria under which Program Quality Assurance Services makes findings often have several parts. In disaggregating findings in this worksheet, where a finding was related to two indicators/indicator groups or other topical areas, it was usually treated as two findings. 2. Because MCR findings are often less detailed than CPR findings, the small number of MCR findings for criteria that were split for disaggregation purposes between two indicators/indicator groups or other 63 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 topical areas were split between those two indicators/indicator groups or other topical areas in the same proportions as the CPR findings. 3. The sum of column a, 973, differs from the total number of findings of noncompliance, 1125, reported in the APR for Indicator 15, and the sum of column b, 535, differs from the total number of findings of noncompliance corrected within one year, 707, reported in the APR for Indicator 15: a. for the reason described in Note 1 above, and also b. because only 35 of the 211 findings of noncompliance made through the PRS (complaint system) during 2005 -2006 were included in this worksheet. The remainder of the findings did not fall into any of the indicator/indicator groups or the two other topical areas included in the worksheet. The bulk of them (101) had to do with non-implementation or partial implementation of IEPs. 4. The “number of programs monitored” for CPRs is the number of districts for which a CPR final report was issued between July 1, 2005, and June 30, 2006. The “number of programs monitored” for MCRs is the number of districts for which an MCR report was issued between July 1, 2005, and June 30, 2006. The “number of programs monitored” for complaints for each indicator/indicator group or other topical area is the number of districts about which findings of noncompliance related to that indicator/indicator group or other topical area were made between July 1, 2005, and June 30, 2006. 64 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 APPENDIX C -- Web-Based Monitoring System The goal of the Web-Based Monitoring System is to directly involve school district leadership in a careful and methodical assessment of their current practices, procedures, and policies and their compliance with federal and state special education requirements. Once the districts have completed this selfassessment, they will be required to report their findings to PQA staff for evaluation. As part of the selfassessment the district will be expected to complete a variety of activities, including special education student record review, interviews, and procedure and policy reviews, with additional activities to be later defined. The districts will have access to a virtual library of forms and guidance documents, as well as a link to the most recent available data reports resulting from the Student Information Management System (SIMS) collection, Assessments/MCAS, and staffing information via MASSDE’s new Education Personnel Information Management System (EPIMS), all designed to facilitate the various self-assessment activities as specified by PQA. The district will report the results of its self-assessment by responding through MASSDE’s secure portal to a series of questions relating to a series of standards and indicators. These responses will be entered using check boxes and limited text fields for short responses, with the capacity to add word processing and spreadsheet attachments as necessary. Once MASSDE receives this electronic report, liaisons in PQA will review and analyze the results of the district’s self-assessment. MASSDE’s analysis will include a summary report that describes the results of the desk audit for all of the standards and/or indicators reviewed. The summary report will also use check boxes and limited text fields for short responses, with the capacity to add word processing and spreadsheet attachments when required. Beyond special education self-assessment reporting and MASSDE’s evaluation of it as described above, the Web-Based Monitoring System will allow districts to submit special education corrective action plans and subsequent special education progress reports in electronic templates. These e-documents will be substantially narrative in nature. The corrective action plan will describe the problem, the proposed corrective action, the person(s) responsible for implementation, and the projected date of completion, as well as explaining how the effectiveness of the corrective action will be measured. The electronic templates for the corrective action plans and progress reports will also have the capacity to add word processing and spreadsheet attachments as may be required for MASSDE’s review and approval (format to be determined). All of the information collected will be stored electronically according to the monitoring year for on-call retrieval by MASSDE for analysis and the fulfillment of reporting requirements The following overall project objectives are identified below: Improve the monitoring and follow-up processes while reducing staffing needs and paperwork, thereby lowering MASSDE and school districts’ ultimate costs; Increase district level compliance with federal and state laws and regulations; Increase process and information access and transparency to PQA and district staff; Provide an extensible, scalable architecture for monitoring education programs on a platform for the future that uses industry standards; and Complete a successful monitoring and follow-up system implementation (defined by project and functional metrics). More specifically, it is MASSDE’s expectation that it will have the capacity to generate a variety of special education reports based on data from the Web-Based Monitoring System from both PQA and the districts. These special education reports will include, but not be limited to: Compliance profiles of each individual district as well as the aggregate compliance reporting of all of the districts involved; Item analysis reports for specific standards and indicators; Review of patterns of compliance and comparisons of district compliance to statewide compliance; and The analysis of compliance rates for particular standards and indicators within a particular fiscal year with the capacity, over time, to look at district and state trend data. Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Submitted February 1, 2008 APPENDIX D -- Table 7: Report of Dispute Resolution for FFY 2006 U.S. DEPARTMENT OF EDUCATION OFFICE OF SPECIAL EDUCATION AND REHABILITATIVE SERVICES OFFICE OF SPECIAL EDUCATION PROGRAMS TABLE 7 REPORT OF DISPUTE RESOLUTION UNDER PART B, OF THE INDIVIDUALS WITH DISABILITIES EDUCATION ACT 2006-07 PAGE 1 OF 1 OMB NO.: 1820-0677 FORM EXPIRES: 08/31/2009 STATE: Massachusetts SECTION A: Written, signed complaints 383 (1) Written, signed complaints total (1.1) Complaints with reports issued 236 (a) Reports with findings 170 (b) Reports within timeline 179 (c) Reports within extended timelines (1.2) Complaints withdrawn or dismissed 36 107 40 (1.3) Complaints pending (a) Complaint pending a due process hearing 21 SECTION B: Mediation requests (2) Mediation requests total 841 (2.1) Mediations 841 (a) Mediations related to due process 36 (i) Mediation agreements 27 (b) Mediations not related to due process (i) Mediation agreements (2.2) Mediations not held (including pending) 805 677 0 SECTION C: Hearing requests (3) Hearing requests total (3.1) Resolution sessions (a) Settlement agreements (3.2) Hearings (fully adjudicated) (a) Decisions within timeline (b) Decisions within extended timeline (3.3) Resolved without a hearing 592 400 25 26 3 22 566 SECTION D: Expedited hearing requests (related to disciplinary decision) (4) Expedited hearing requests total 31 (4.1) Resolution sessions 24 (a) Settlement agreements 1 (4.2) Expedited hearings (fully adjudicated) 2 (a) Change of placement ordered 1 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 APPENDIX E -- Indicator #20 Scoring Rubric APR Data APR Indicator 1 2 3A 3B 3C 4A 5 7 8 9 10 11 12 13 14 15 16 17 18 19 Valid and Reliable Data 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 APR Score Calculation Correct Calculation 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 Followed Instructions 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 Subtotal Timely Submission Points - If the FFY 2006 APR was submitted ontime, place the number 5 in the cell on the right. Grand Total (Sum of subtotal and Timely Submission Points) = Total 2 2 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 58 5 63 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Table Table 1– Child Count Due Date: 2/1/07 Table 2Personnel Revised April 14, 2008 618 State-Reported Data Responded Passed to Data Note Complete Edit Requests Timely Data Check Total 1 1 0 1 3 1 1 1 1 4 1 1 1 1 4 1 1 0 1 3 0 1 1 1 3 1 1 1 1 4 1 1 1 1 4 Due Date: 11/1/07 Table 3– Ed. Environments Due Date: 2/1/07 Table 4– Exiting Due Date: 11/1/07 Table 5Discipline Due Date: 11/1/07 Table 6- State Assessment Due Date: 2/1/08 Table 7Dispute Resolution Due Date: 11/1/07 Subtotal 618 Score Calculation Grand Total (Subtotal x 2) = 25 50 68 Massachusetts Part B Annual Performance Report (MA APR) for FFY 2006 Revised April 14, 2008 Indicator #20 Calculation A. APR Grand Total = 63 B. 618 Grand Total = 50 C. APR Grand Total (A) + 618 Grand Total (B) = D. Subtotal (C divided by 119)* = E. Indicator Score (Subtotal (D) x 100) = 113 0.950 95.0% * Note: Any cells mark with a N/A will decrease the denominator by 1 for APR and 2 for 618 data. Definitions Timely – All data for the APR are submitted on or before February 1, 2008. Data for tables for 618 are submitted on or before each tables’ due date. NO extensions. Valid and Reliable Data - Data provided are from the correct time period, are consistent with 618 (when appropriate) and the measurement, and are consistent with previous indicator data (unless explained). Correct Calculation - Result produced follows the required calculation in the instructions for the indicator. Instructions Followed - APR provides information required in the instructions for the indicator. For example, when required, explanation provided, raw data and/or definitions given, or response provided to previous OSEP APR analysis. Complete Data – No missing sections. No placeholder data. Data submitted from all districts or agencies. For example, when the instructions for an indicator require data broken down into subparts, data for all subparts are provided. Passed Edit Check - Tables submitted to Westat do not have missing cells or internal inconsistencies. (See https://www.ideadata.org/TAMaterial.asp regarding Westat edit checks.) Responded to Data Note Requested – Provided written explanation to Westat in response to data note requests. 69