Shaping the Community Scorecard Action Research Project Final Report Foreword In spring 2010, the UK Commission for Employment and Skills and LSIS came together to explore – through a series of sector-led action research projects – the implications for colleges and providers of the UK Commission’s proposal for ‘community scorecards’ to be part of an emerging suite of public information to inform choice. The UK Commission’s 2009 report, ‘Towards Ambition 2020: skills, jobs, growth’1, proposed that informed customers should increasingly drive supply, performance and quality, and that securing this shift would involve a transfer of trust and greater autonomy for providers. The report argued that information should be provided to inform learner choice, but that a wider range of information and measures should be developed to reflect providers’ responsiveness to their local communities. LSIS has a long-standing interest in and experience of supporting and encouraging colleges and providers to play a strategic role in supporting their local communities and so were keen to work with the UK Commission on the exploration of a ‘community scorecard’ to strengthen and demonstrate providers’ contribution and accountability to their communities. The work was timely and relevant in the context of the ‘Big Society’ – founded on the high-level principles of freedom, fairness and responsibility – resonating with the intention to reorientate the accountability of public services away from government and towards customers and citizens. In this context, the National Improvement Partnership Board (NIPB) was charged with the development of early proposals for a new framework of public information for the sector, and agreed to take an overview of the ‘community scorecard’ project as well. Eleven providers took part in the action research and put a considerable amount of time and effort into exploring what a ‘community scorecard’ might look like for their institutions. The group reflected the diversity in the range and mix of institutional and administrative arrangements in different localities, and the different socio-economic profiles of the communities the sector serves. It also deliberately included providers at different stages of development in their thinking on local accountability to gain a realistic and wide perspective on developmental needs. The diversity has resulted in a wide-ranging and rich report that outlines areas of good practice but also some of the challenges experienced. We hope that the learning shared in this report may help to inform colleagues in the sector who are thinking about how to articulate as fully and transparently as possible their responsiveness and contribution to their communities within new frameworks of accountability. In LSIS, we are also very keen to understand what more we could do to support the sector to develop its thinking in this important new area of work. Caroline Mager Executive Director Policy, Research and Communications, LSIS Ian Kinder Deputy Director UK Commission for Employment and Skills UK Commission for Employment and Skills report, ‘Towards Ambition 2020: skills, jobs, growth’ www.ukces.org.uk/assets/bispartners/ukces/docs/publications/towards-ambition-2020-skills-jobs- growth.pdf 1 2 Shaping the Community Scorecard Action Research Project Contact details Enquiries about this report should be addressed, in the first instance, to: Jenny Williams Head of Policy LSIS Email: jenny.williams@lsis.org.uk or Alison Morris Programme Manager UK Commission for Employment and Skills Email: Alison.morris@ukces.org.uk Action Research Project Shaping the Community Scorecard 3 Contents Foreword02 Contents04 Executive summary05 1. Introduction10 2. Policy context12 3. Methodology15 4. Research conclusions18 5. Emerging implications and six proposals for policy development and implementation 4 Shaping the Community Scorecard Action Research Project 28 Executive summary In ‘Towards Ambition 2020: skills, jobs, growth’ (October 2009), the UK Commission made a series of proposals for “building a more strategic, agile and labour market led employment and skills system”. At the heart of these proposals was an intention to “increase trust in, and authority to, learning providers”. Providing more public information on the performance of providers, and increasing their accountability to their communities was seen as good ways to do this. LSIS has an interest in and experience of supporting and encouraging colleges and providers to play a strategic role in supporting their local communities, and agreed to work with on the development of a ‘community scorecard’ to strengthen and demonstrate providers’ contribution and accountability to their communities. At the time the project began in summer 2010, the concept of the ‘Big Society’ was emerging as an overarching framework for the Coalition Government’s high-level principles of freedom, fairness and responsibility. The concept of a ‘community scorecard’ seemed to resonate with the intention to reorientate the accountability of public services away from government and towards customers and citizens. The approach adopted was to involve the sector in shaping the development of the community scorecard from the beginning and willing providers were invited – from FE colleges, the work-based and adult and community learning sectors – to join a small action research group. The group reflected the diversity in the range and mix of institutional and administrative arrangements in different localities and the different socio-economic profiles of the communities the sector serves. It also deliberately included providers at different stages of development in their thinking about the community scorecard to gain a realistic and wide perspective on developmental needs. Each provider received a small grant towards their project and all were supported by RCU Research and Consultancy2, who coordinated the research project. The diversity of the sector, reflected in the action research group, underlined the necessity to avoid any prescriptive design for the community scorecard. Rather, the idea was to provide a spectrum of opportunities from which providers could select the most appropriate mix for their specific context. The project ran between September and December 2010, in parallel with the emergence of the new government’s approach to public service reform and its implications for FE and skills. These have included a stronger emphasis on the sector developing a more comprehensive range of public information through the National Improvement Partnership Board (NIPB), a focus on localism, and in the context of greater freedoms and flexibilities, a strong steer to the sector to develop its responsiveness to its communities and its networks of partners. During this period a set of national indicators and measures has also emerged, creating new drivers from the centre of the system. Some of the opportunities and tensions between these various developments are reflected in the challenges the providers faced in establishing their community scorecards in a rapidly evolving context. In particular, the determination of the Coalition Government to devolve greater responsibility to frontline providers and 2 http://www.rcu.co.uk/ Action Research Project Shaping the Community Scorecard 5 to emphasise the need for accountability to their communities and localities became apparent late in the action research. The action research was not, therefore, able to explore fully the implications of this shift. This report outlines the experiences of the eleven providers involved in the action research and identifies some overall research conclusions from the project. It also considers the emerging implications of these conclusions for future policy development and implementation and makes six proposals. Research conclusions The intention of the action research project was to give a group of providers the space to explore what information could be used to engage with, and to demonstrate their responsiveness and contribution to their community or communities. The diversity of the group of providers, both in relation to how far these issues had previously been considered as well as differences in the types of institution and their geographical location, has meant that the question has been explored from a number of different angles. Purpose of a community scorecard Although the outline purpose of the community scorecard had been defined – to strengthen and demonstrate providers’ contribution and accountability to their communities – the eleven providers’ rationales for taking part in the project covered an array of reasons that reflected their diversity and range. They all, however, shared the ambition of delivering services to meet community needs and to establish performance measures. Providing evidence to support public relations campaigns, funding decisions, influencing the strategic direction of the organisation, and commercial business expansion were also drivers for engagement in the project. A small group of the providers explicitly saw the development of their community scorecard as a way of strengthening their strategic role in the locality. Overall, the experience of the providers suggests that if an organisation is to develop a community scorecard, then it needs to be part of a long-term strategy and a real desire for public accountability. The community scorecard has the potential to be an invaluable continuous improvement tool for providers who genuinely seek and act upon community feedback, as well as serving a variety of other uses, from supporting community engagement to PR. Defining the community Defining the community was essential at the outset to enable providers to target their research and seek meaningful views to inform their scorecard. Some providers felt they served one diverse community, whereas others thought they served several distinct communities with differing needs. Some communities were locality-based, others reflected the interests of groups of employers, for example the land-based sector. Some providers felt that it would be difficult to develop a single community scorecard, even one just for their organisation, and chose to focus their community scorecard on one or two distinct communities that they served, whilst two provider networks explored the possibility of putting in place a single scorecard for the whole network. Securing legitimacy for the community scorecard Each provider’s journey in developing the scorecard was unique according to the internal and external environments in which they operated. Although securing legitimacy was an area of consideration for all, it was explored in more detail amongst those providers at a more advanced stage in their developmental journey. One provider used the action research project to ‘soft launch’ its community scorecard with the community, but this was the culmination of two years of development within the organisation, including with governors. Individuals engaged in the development of a community scorecard felt that top-level commitment was essential, not least to ensure that sufficient resource was available to develop and maintain the scorecard. 6 Shaping the Community Scorecard Action Research Project Approaches of individual providers The providers had different starting points and ambitions for their projects. The research was most effective when the provider had a clear view at the outset of their community and their research methodology. The majority of providers conducted primary research to inform the development of their community scorecard measures using new data sources. However, there was recognition that this was time-consuming to collect and would require significant on-going resource commitment. Others considered an approach that involved existing data sources to some degree. The level of research expertise differed amongst providers, for instance, in the design of research instruments to provide robust and verifiable results. The amount of resource available also varied widely. Emerging measures There were differing views from providers as to what type of measures to include in a community scorecard with the intelligence gathered being wide ranging and encompassing perceptions, customer satisfaction, stakeholder views and community needs. Where measures emerged beyond the existing national indicators, they were around measures of community engagement – for example numbers of students volunteering in the community, the number / range of events open to the community; the extent to which patterns of participation could be mapped to areas of disadvantage; and the extent to which the provider’s curriculum offer could be seen to be meeting the needs of the community. One provider also suggested a measure of ‘innovative and visionary leadership’. Providers recognised the value of having a national set of consistent core measures and spent some time exploring how to capture indicators of learner destinations as part of this. Equally, however, there was recognition of the value of local measures and the suggestion that guidance would be helpful for providers on a spectrum of measures they could consider, linked to the overall goals of their institution. Publication of results A number of providers considered how they might present their scorecard to the community with a web-based approach being particularly popular. Some providers also considered the extent to which the scorecard could become an interactive tool that invited user feedback. There was also consideration given as to how the scorecard would relate to existing published data. Some providers expressed concern about striking an appropriate balance between transparency and positive publicity in the presentation of their scorecards, particularly where they were operating in competitive environments. There was a general consensus that to ensure a scorecard is utilised by both the organisation and its community, it needs to be simple to use, easy to update and relevant. Support for providers It is clear from this project that providers looking to embark on the development of a community scorecard, or an approach to demonstrating their contribution to their communities, would benefit from advice and support, which at the very least includes dissemination of the findings from this research to enable providers to learn from one another’s experiences. Other support could include: mapping, analysis and dissemination of other community scorecard development activity, in the FE and skills sector and more widely across other public services; establishing of good practice forums; and development of a community scorecard toolkit. Further areas for possible support are identified in the following section on emerging implications. Action Research Project Shaping the Community Scorecard 7 Emerging implications and six proposals for policy development and implementation This section builds on the research conclusions and discusses some of the emerging implications for policy development and implementation. Overall, some concerns were raised about the term ‘community scorecard’. There are options for how this work could be described going forward. Although the action research was overseen by the NIPB, alongside its work on FE public information and the ‘Informing Choice Framework’; the shift towards greater freedoms that has emerged in the last year has created greater space and necessity for providers to develop strategic plans that are responsive to their communities. Within this context, new ways of thinking about accountability could assist providers to articulate as fully and transparently as possible their contribution to civic and community wellbeing. Audiences Ideas about the purpose of a ‘community scorecard’ varied between the providers and a range of different audiences were identified. The motivations for demonstrating accountability also vary depending on the audience, for example, the reasons for providing information to potential learners will not be the same as those for a provider wanting to make information available to their local authority. There seem to be two main groups that are external to the provider. The first broad group are ‘customers’. This group includes existing and potential customers (learners, potential learners, employers), and what could be described as ‘the wider learning community’. The wider learning community encompasses organisations and groups that support people to get involved in learning (including community groups, support groups, career guidance professionals). The second group are ‘strategic partners and civil society’. This group is broad and has increased in salience as the Coalition Government’s policy has emerged, and was therefore not a major focus of the action research. It includes local authorities and other public sector partners such as health and the police; and also private and third sector partners such as local enterprise partnerships, local chambers of commerce, or voluntary, social enterprise sector bodies. Given its core purpose of human capacity development, the role of the FE and skills sector has relevance for a wide range of stakeholders and strategic players in the locality. It will be beneficial for the sector to develop awareness of its contribution with this wide-ranging audience. Further development of the scorecard concept should therefore explore the scope and priorities of strategic partners and civil society in order to position the sector more centrally in relation to the strategic development of communities Proposal 1: It is important to recognise the wide range of audiences for information on FE and skills. The sector needs to develop its awareness of the priorities of strategic partners and civil society in order to shape its response to the emerging ‘Big Society’ concept. Presentation of information Different audiences have different information requirements. For customers, the NIPB’s work to develop a “more comprehensive range of public information”, building on the Framework for Excellence3, is important in terms of securing a nationally consistent, comparable set of information on which learners and employers can take decisions about the types of provision and providers in which they wish to invest. Other types of information for customers will be particular to different providers. The action research suggests that individual provider surveys of their communities are also helpful in highlighting the sort of local information that different audiences would find useful, for example: indicators on community 3 Skills Funding Agency Framework for Excellence http://ffe.skillsfundingagency.bis.gov.uk/ 8 Shaping the Community Scorecard Action Research Project engagement; how well participation reflects the socio-economic make-up of an area; or how effectively a provider’s curriculum offer responds to the priorities of the communities it serves. Our research suggests that securing accountability to communities needs to be part of a long-term institutional strategy and plan. Measures of community accountability should reflect the overall goals of the institution, and be negotiated with and responsive to their communities, offering a framework in which providers can be publicly accountable for their contribution to social and economic life. A key challenge for providers will be how to analyse, interpret, use and present an increasingly wide range of ‘public information’ on FE and skills in ways that are authentic and accessible. This challenge also extends to the provision of information for strategic partners. Proposal 2: It would be helpful to develop a spectrum of measures as examples for providers, but ultimately the measures providers select must enable them, with their communities, to assess progress towards and achievement of their goals. Proposal 3: Sector leaders should consider the new challenges and opportunities of working with an increasingly wide range of public information, and making sense of it with and for different audiences / stakeholders. LSIS could support the sector to explore how information and communication strategies should underpin the shift towards greater community responsiveness. Proposal 4: The implications of emerging thinking about social productivity4 could be further explored as a possible guiding framework for generating more meaningful metrics to account for providers’ strategic engagement with and contribution to their communities and localities. Dynamic engagement with communities The devolution of responsibility and greater freedoms at organisational level, together with the emphasis on responsiveness to communities presents challenges to providers to develop more dynamic engagement with their communities, and empower customers and partners to interact with them to co-design services and programmes. Many of the providers involved in the community scorecard research envisaged presenting their information in a web-based format. The potential of technology to implement providers’ information and communications strategies in highly visible and accessible ways is, we know, being explored across the sector. Such approaches could become part of a provider’s overall narrative to both customers and strategic partners, clearly connecting it to the concerns and priorities of the communities it serves. Proposal 5: Sector providers should explore innovative approaches to developing dynamic engagement with their communities, including how to harness the potential of technology, and the possible economies of scale of working across the sector. Strategic direction and ownership There was variation in the level of senior leadership involvement across the providers involved in the action research. Where there was senior leadership the work was more likely to be used to support the strategic planning of the institution. Proposal 6: Sector leaders should consider the implications of the shift to greater community responsiveness and public information for their approaches to accounting for their contribution to the economic and social life of their communities. 4 2020 Public Service Hub at the RSA; ‘The further education and skills sector in 2020: a social productivity approach’ http://www.lsis.org.uk/2020PSHQ Action Research Project Shaping the Community Scorecard 9 1. Introduction In ‘Towards Ambition 2020: skills, jobs, growth’ (October 2009), the UK Commission for Employment and Skills made a series of proposals for “building a more strategic, agile and labour market led employment and skills system”. At the heart of these proposals was an intention to “increase trust in, and authority to, learning providers”. Providing more public information on the performance of providers and increasing their accountability to their communities was seen as one way to do this. The aim was for citizens, communities, employers and members of local public service boards to be empowered to become far more directly involved in shaping outcomes and improving the performance of the employment and skills system and to be better informed to make choices. LSIS has an interest in, and experience of, supporting and encouraging colleges and providers to play a strategic role in supporting their local communities and agreed to work with the UK Commission to explore options for the development of community scorecards to strengthen and demonstrate providers’ contribution and local accountability to their communities. Alongside this project in 2010-11, LSIS has also supported twenty provider-led Peer Review and Development (PRD) groups focused on the theme of community developments5 and with the sector - is developing a strategic framework for effective community development. At the same time as the community scorecard project began, in summer 2010, the concept of the ‘Big Society’ was emerging as an overarching framework for the Coalition Government’s high-level principles of freedom, fairness and responsibility. The idea of a ‘community scorecard’ seemed to resonate with the intention to reorientate the accountability of public services away from government and towards customers and citizens. The rationale for the research was to explore some of the ways in which providers could demonstrate accountability to their communities in order to better understand the journey from ‘vertical’to ‘horizontal’ accountability. The approach adopted was to involve the sector in shaping the development of the community scorecard from the beginning. Participants were invited from FE colleges, work-based, and adult and community learning providers, to join a small action research group. The group of eleven reflected the diversity of institutional and administrative arrangements in different localities, together with the different socioeconomic profiles of the communities the sector serves. It also deliberately included providers at different stages of development in their thinking on the community scorecard to gain a realistic and wide perspective on developmental needs. Each provider received a small grant towards their project and all were supported by RCU Research and Consultancy, who coordinated the research project. The diversity of the sector, reflected in the action research group, underlined the necessity to avoid any prescriptive design for the community scorecard. The purpose of the project was not to develop a blueprint for what the community scorecard should look like, but through an action research approach to explore some of the issues involved in developing ways for providers to provide more information to their communities in order to strengthen and demonstrate their local accountability – “to secure legitimacy for their strategic role and to play their part in local public service coalitions”6. The idea was to provide a spectrum of opportunities from which providers could select the most appropriate mix for their specific context. 5 See the evaluation of the community development peer review initiative at: http://www.excellencegateway.org.uk/page.aspx?o=323173 6 See Annex 1 for the initial LSIS/UKCES outline paper on the community scorecard 10 Shaping the Community Scorecard Action Research Project At the outset7 LSIS and the UK Commission suggested that the purposes of the community scorecard could encompass: • establishing the fit between the mission of the organisation and local strategic priorities; • raising awareness of the wider contribution that providers of further education make to securing social and economic well-being in their localities; • communicating the economic and social returns to learning beyond learning outcomes; • enhancing the accountability of providers to citizens and communities; • providing a focus for developing leaders of learning and skills as strategic partners in local public consultation; and • providing a stimulus to develop partnerships for better integrated delivery, efficiencies, and improved outcomes for learners. The project ran between September and December 2010, in parallel with the emergence of the new government’s approach to public service reform and its implications for FE and skills. These have included a stronger emphasis on the sector developing a more comprehensive range of public information through the National Improvement Partnership Board (NIPB), a focus on localism, and in the context of greater freedoms and flexibilities a strong steer to the sector to develop its responsiveness to its communities and its networks of partners. During this period a set of national indicators and measures has also emerged, creating new drivers from the centre of the system8; and towards the end of the project the Framework for Excellence data was published9. Some of the opportunities and tensions between these various developments are reflected in the challenges the providers faced in establishing their community scorecards in a rapidly evolving context. In particular, the determination of the Coalition Government to devolve greater responsibility to frontline providers and to emphasise the need for accountability to their communities and localities became apparent late in the action research. The action research was not, therefore, able to explore fully the implications of this shift. This report outlines the experiences of the eleven action research providers and identifies some overall research conclusions from the project. It also considers the emerging implications of these conclusions for policy development and implementation. 7 See Annex 1 for the initial LSIS/UKCES outline paper on the community scorecard 8 BIS Structural Reform Plan - Impact Measures (October 2010) and SFA Guidance Note 6 (December 2010) 9 http://readingroom.lsc.gov.uk/SFA/FfE_Summary_Statistics_2010.pdf Action Research Project Shaping the Community Scorecard 11 2. Policy context The community scorecard idea originated as part of the UK Commission for Employment and Skills report ‘Towards Ambition 2020: skills, jobs, growth’ in October 2009 which recommended the need for a more strategic and agile labour market led employment and skills system which would encourage increased trust in learning providers. By developing a community scorecard, providers would give the public relevant information on their performance and this would in turn increase their accountability to the communities they serve. The report called for publicly-funded institutions to be more accountable to individuals, employers and communities through a system of quality labelling for public services. It went on to propose that this should include outcomes / destinations, customer satisfaction and quality indicators “balanced against a profile of the economic, social and labour market characteristics of the area in which a provider operates”. It anticipated the Coalition Government’s policy shifts, which have emphasised the importance for providers of ensuring that citizens’, communities’ and employers’ voices shape what they deliver and how they deliver it, empowering providers to offer responsive provision that fits with the priorities and patterns of social and economic life. At a recent LSIS seminar, Maggie Galliers CBE, principal of Leicester College and an early adopter of the community scorecard ideas, explained it like this: “In order for people to better understand the contribution FE can make, they need to understand what it does... and so FE needs to be more transparent and accountable to the public. That should then become the basis for a dialogue about what society values and what it wants FE to do – recognising that it can’t do everything; enabling FE to make an effective contribution to the Big Society.” The emerging Big Society concept has provided a backdrop to the action research on the community scorecard. Although still coming into focus and not without controversy, it is acting as a framework for the Coalition Government’s high-level principles of freedom, fairness and responsibility, principles that are driving reform in wider public service, as well as FE and skills. For the FE and skills sector, the skills strategy published in November 2010 made clear the government’s intention to reorientate the accountability of the sector away from government and towards customers, citizens and employers: “Colleges and training organisations must respond to the real demands demonstrated by employers and learners. “[Providers] should engage with local partners – consulting them on their business plans so that providers can set out how their offer...contributes to local economic priorities. “Development of a more comprehensive range of public information will be published in January 2012.”10 The development of a “more comprehensive range of public information” has been led by the sector, through the National Improvement Partnership Board, which has also taken an overview of the progress and outcomes of the community scorecard action research project. It builds on the Framework for Excellence data, published by the Skills Funding Agency towards the end of the community scorecard action Strategy document: Further Education New Horizon Investing in skills for sustainable growth (November 2010) Department for Business Innovation and Skills http://www.bis.gov.uk/assets/biscore/further-education-skills/docs/s/10-1272-strategy-investing-in-skills- for-sustainable-growth.pdf 10 12 Shaping the Community Scorecard Action Research Project research project in December 2010, the intention of which is that “giving learners and employers better information will drive a ‘step change’ in quality improvement”. The SFA describes the shift in the emphasis of accountability as being “founded on the relationship between provider and customer rather than provider and government”. The government has from the start set its face against centralised performance management systems for holding the sector to account. There has been clarification that the Ofsted inspection regime will be simplified and less frequent for successful providers, and, working with the sector, the Department for Business, Innovation and Skills is engaged in further work aimed at streamlining the performance architecture of the FE and skills system. Strengthening providers’ capacity to understand their communities’ needs, to effectively respond to those needs – even where they may be in tension with national policy priorities – and to be able to account transparently to their communities for their contributions, will become key skills in the new operating environment. Indeed, Lifelong Learning UK’s 2010 review of workforce skills11 suggests providers are already recognising this, citing “partnership working including engagement with communities” as a skills priority for the sector. Defining ‘community’ There is a strong emphasis in recent announcements from BIS that ‘community’ is being used in its widest sense, to include the sectoral interests of localities and employers. These interests come together around the agendas for Local Enterprise Partnerships in which, so far, skills appear to be a high priority but engagement with the FE and skills sector less so. With the changes to sub-national economic planning structures and the policy emphasis being placed on the sector’s contribution to stimulating and supporting economic growth, providers’ ability to show their responsiveness to the emerging priorities for Local Enterprise Partnerships will be an important dimension of the sector’s strategic role. Other aspects of the Coalition Government’s localism agenda also resonate with the intentions of the community scorecard to enco urage greater accountability to communities – particularly localities. Some elements of this programme continue the trajectory of the previous government – a less top-down approach, more citizen-driven services and an emphasis on the use of markets and competition. However, the current emphasis appears to be on decentralising power down to the lowest possible level, and this move to accelerate the role played by local communities is now being translated into law through a wide-ranging Bill12. In addition to the provisions to create local enterprise partnerships, other measures in the Bill strengthen the framework for local accountability by: • giving councils a general power of competence through which they will be able to set up businesses and innovate in new ways including developing shared services, employee mutuals, outsourcing and widespread commissioning of functions; • empowering local people through giving residents the power to instigate local referendums on any local issue, and voluntary and community groups the right to challenge local authorities over their services; • allowing the Secretary of State to put in place shadow elected mayors in 12 city councils and to initiate a mayoral referendum; and • abolishing the existing ethical governance framework, replacing this with a duty to promote and maintain high standards of conduct and a power to adopt a voluntary code of conduct. 11 http://www.lluk.org/wp-content/uploads/2011/01/Final-SSA-2010-UK.pdf Decentralisation and the Localism Bill: From Big Government to Big Society. A guide to the Bill is available at http://www.communities.gov.uk/documents/localgovernment/pdf/1793908.pdf 12 Action Research Project Shaping the Community Scorecard 13 LSIS has an interest in and experience of supporting and encouraging colleges and providers to play a strategic role in supporting their (local) communities and agreed to work with the UK Commission on the development of a community scorecard to strengthen and demonstrate providers’ contribution and local accountability to their communities. The initial scoping of the project13 was based on a review of the learning from previous policy research at LSIS, including: • the Total Place pilots – LSIS commissioned and worked closely with the Association of Colleges (AoC) to explore the involvement of colleges in the pilots; • the new duty on colleges to promote economic and social wellbeing in their localities – LSIS was commissioned by BIS to develop guidance for the sector; • opening up college premises to the community to support informal learning – LSIS was commissioned by BIS to produce a good practice guide; • tackling and preventing gun, knife and gang-related problems in the sector - LSIS has produced a good practice website to support the sector; and • LSIS’s support programme for equalities, diversity and community development. 13 See Annex 1 14 Shaping the Community Scorecard Action Research Project 3. Methodology LSIS invited ‘pilots of the willing’ in summer 2010 through its standard open invitation process to sector providers. A number of responses were received and reviewed by LSIS and the UK Commission. The decisions took into account the need to have a range of different types of provider involved in the project from different parts of the country, and at different stages of development with their thinking about the community scorecard to enable a spectrum of developmental issues to be considered. The successful providers were: • Leicester College (East Midlands) • Petroc (South West) • Stockton Riverside College (North East) • Barton Peveril College (Sixth Form) (South East) • Otley College (Land-based) (East) • In Touch Care Limited (Yorkshire and the Humber) • GDR Solutions (UK) Limited (South West) • Black Country Training Group (West Midlands) • The Sellafield Ltd Unionlearn Centre (North West) • Wandsworth Council Lifelong Learning (London) • The Co-operative College (North West) RCU was commissioned by LSIS and the UK Commission to provide informed and experienced consultancy support to help the eleven providers deliver an action research project that explored and helped to establish the purpose of a community scorecard, what it might look like, and how it might be used. The research took place between September and December 2010, with providers reporting in January 2011 and RCU final summative reporting in February 2011. The research methodology was a five stage approach as follows: Stage 1 – First seminar event: an opportunity for the participating providers to hear first-hand the policy context from both the UK Commission and LSIS on the community scorecard concept. Leicester College also shared their community scorecard journey to date. Providers’ initial discussions explored their different definitions of community, existing measures for evaluating impact, and their initial thoughts on the purpose and scope of their individual research projects. Stage 2 – Developing providers’ project plans: RCU mentors visited each provider to help them develop a project plan including aims, objectives, senior in-house sponsor and proposed research approach, with timescales and milestones. The purpose of the project plan was to provide a framework within which providers could undertake their action research and also to form the basis for progress review. Stage 3 – Mentoring, reporting and keeping in touch: there was regular contact between RCU mentors and the providers during the project. Providers supplied fortnightly reports to RCU detailing their progress. The reporting approach was flexible but provided an important mechanism to capture key learning, which was later reflected in providers’ final reports. A workroom was also set up on the Excellence Gateway, into which provider project plans, reports and project communications were posted. Action Research Project Shaping the Community Scorecard 15 Stage 4 – Second site visits and seminar event: RCU mentors visited providers prior to the second seminar event to review research outcomes, support providers in the analysis of results and identify any potential measures that could be included with a community scorecard. At the second seminar, attended by the UK Commission, LSIS, RCU, the participating providers and representatives from the Skills Funding Agency, each provider delivered a presentation on their project which included: the purpose of their community scorecard and key research activities undertaken; lessons learned; and next steps. This was followed by action learning sets which explored: • how well providers’ projects progressed against the original project plan; • what worked well and which particular factors / levers were most useful / successful; • what had not worked well, why, and what had been the impact of these issues; • what providers had achieved to date and what were their emerging findings; • what more needed to be done; • whether providers had been able to identify some suitable community scorecard measures; • whether providers had been able to test the relevance of any community scorecard measures with the community; • whether individual providers’ projects would continue beyond the period of the research; • how providers had communicated their research internally and externally; • whether Governors or Trustees had been involved and in what capacity; • whether providers would choose to develop a community scorecard based on what they had learnt through the research; • what factors should be considered for any extension of the community scorecard concept; and • whether the development of a community scorecard would be considered a priority issue for the management team in their organisation. 16 Shaping the Community Scorecard Action Research Project Critical reflection was also undertaken on the wider implications of the community scorecard based on the following: Clarity • From what you have learnt so far, to what extent is a community scorecard an effective way to communicate the vision and mission of the provider? • How effective is a community scorecard in clearly articulating how a provider gives value to a community? • Are there any essential elements to be included or considered? Scope • How useful is a community scorecard for your particular organisation? • What are its strengths and what are its limitations in your circumstances? • Does the community scorecard need to include a range of measures / indicators, or is it more suitable for it to focus on a particular issue / aspect? Measurability • To what extent does / will your community scorecard include quantifiable measures that allow stakeholders to make a judgement about you and your provision? Verifiability • How straightforward was it to develop your community scorecard and can it easily be developed to meet changing needs or circumstances? Transferability • How useful do you feel the community scorecard is to other organisations similar to yours? • Are there certain types, size and location of provider where a community scorecard is likely to be more suitable? Utility • To what extent do you think the community scorecard is of interest to other key stakeholders such as local authorities, public bodies, community and voluntary groups and employers? Stage 5 – Provider reports and summative report: Following the discussions at the seminar, providers were asked to submit final reports building on their presentations and the action learning / critical reflection undertaken. This final report is based on information from those reports. Action Research Project Shaping the Community Scorecard 17 4. Research conclusions Introduction This section discusses the research conclusions from the project overall, based on the individual reports from the eleven providers involved in the action research. For ease of reference, the providers involved are listed again below: • Leicester College (East Midlands) • Petroc (South West) • Stockton Riverside College (North East) • Barton Peveril College (Sixth Form) (South East) • Otley College (land-based) (East) • In Touch Care Limited (Yorkshire and the Humber) • GDR Solutions (UK) Limited (South West) • Black Country Training Group (West Midlands) • The Sellafield Ltd Unionlearn Centre (North West) • Wandsworth Council Lifelong Learning (London) • The Co-operative College (North West) Each provider’s report included: • Details of their own experiences of developing a community scorecard: rationale, purpose, project aims, methodology, definition of community, legitimacy of the scorecard, project outcomes, lessons learned and critical success factors, communication of the scorecard, and next steps; • Critical reflections on the wider implications of a community scorecard for a range of provider types and delivery settings; and • Factors to be considered by other providers looking to develop a community scorecard together with any further support that might be required to assist them. It is worth noting that as a result of this action research, all eleven providers started to identify and test the critical factors that need to be considered when developing a community scorecard. However, the extent to which each provider reached the stage of identifying measures and producing a community scorecard within the short project timescale varied, depending upon. The key findings from each project have been distilled into an overall set of eight research conclusions that cover: • the range of starting points for a diverse sector; • the purpose of a community scorecard; • defining the community; • securing legitimacy for the community scorecard; • approaches taken by individual providers; 18 Shaping the Community Scorecard Action Research Project • emerging measures; • publication of results; and • support for providers. The range of starting points for a diverse sector The diversity of the sector, reflected in the providers involved, is a constant refrain through these research conclusions. Within the group of providers engaged in the action research, there were three general FE colleges serving a multitude of community groups with diverse and wide-ranging needs. These colleges particularly wanted to explore how a community scorecard could adequately reflect the impact they were having on their communities and to identify areas where there was scope for improvement. A number of the participating providers served predominantly rural areas, including one specialist land-based college, a general FE college and a work-based learning provider. These providers wanted to ascertain how they could develop a community scorecard to reflect the particular needs of their learners, employers and community groups spread across a wide and often remote geographic area. These providers set out to develop some additional (not replacement) performance measures that would reflect the needs of isolated rural groups. For these providers, the community scorecard was viewed as potentially offering a more appropriate measure of their progress towards delivering their missions, which generally reflected ambitions to achieve learning excellence within the rural communities (including business communities) that they serve, rather than centralised performance measures. Other providers’ were keen to test the suitability of a community scorecard in verifying the impact they were having within their own organisational structures. For example, the Co-operative College wanted to explore how a community scorecard could be adopted to reflect the added value of Co-operative schools by using comparator measures such as attendance, attainment and a reduction in vandalism to review performance against levels prior to the school becoming a co-operative or against similar schools in their location. The Black Country Training Group – a network of work-based learning providers – sought to establish whether they could develop a single community scorecard that could reflect the combined impact of all the group’s members, comprising over forty training providers and voluntary groups across the West Midlands. Other providers saw the community scorecard as a means of raising awareness of services offered and as a way to gather intelligence from potential users to inform future services and support. These providers also hoped to evidence positive impact on their service users. Their aim in developing a community scorecard was not only to measure their impact on the communities they serve, but also to give evidence to their funding bodies or senior management to justify continuation of service delivery. One provider used the community scorecard research as a means of gathering information on their local community impact to support a new business strategy. The early adopter, Leicester College, focused on a ‘soft launch’ of their already developed community scorecard, and on gathering community feedback. The purpose of a community scorecard Although the outline purpose of the community scorecard had been defined – to strengthen and demonstrate providers’ contribution and local accountability to their community – the eleven providers had a range of reasons for taking part in the project. They all, however, shared the ambition of delivering services to meet community needs and to establish performance measures for this work. Providing evidence to support public relations campaigns, funding Action Research Project Shaping the Community Scorecard 19 decisions, influencing the strategic direction of their organisation and commercial business expansion was also a driver for engagement in the project. A small number of the providers saw the development of their community scorecard explicitly as a way of strengthening their strategic role in the locality. There was no single consensus about the purpose of a community scorecard by the providers. Whilst the freedom to shape the concept was warmly welcomed and seen as exciting at the beginning of the project, it actually caused some delays in executing the primary research because a large number of the providers found it difficult to establish a clear purpose for their community scorecard. Indeed, even by the end of the action research phase there were a number of providers who had defined only an outline purpose. There were also a number that shifted their purpose during the course of the action research. Not surprisingly, the actual purposes of the providers’ scorecards mirrored their rationale for being involved in the action research. These were: • to gauge how effective the community scorecard is as a tool for measuring an institution’s impact on a wide-ranging and diverse set of communities and locations; • to establish the suitability of a community scorecard in offering an alternative set of performance measures that better reflect the operating circumstances of the provider; • to measure current public opinion of the organisation, establishing whether stakeholders feel suitably informed about the organisation’s contribution to the community and ascertaining whether differences in opinion and knowledge exist across diverse community groups; • to encourage non-learners to participate in learning by better understanding their needs and those factors that influence their decision to become engaged; • to develop an evidence-gathering tool that acts as a two-way communication system between the general public and the organisation; • to showcase what the provider can offer and gather feedback on whether this is what the community wants; • to provide a framework for Co-operative schools to measure the impact of those aspects of their school organisation that are influenced / driven by co-operative values and principles; and • to show how a college serves its local community and demonstrates its accountability to that local community by publishing key information about its performance. Some providers sought the views of internal staff to inform the purpose, whereas others found it difficult to secure buy-in from within their organisation, including at a senior level. Others consulted with groups in their community to shape the development of their scorecards, with mixed results. However, the clear message from this group was that to be of any value, the scorecard must be meaningful to the communities they serve and not be perceived as a national imposition. Providers also grappled with how the community scorecard would complement existing public information sources such as Framework for Excellence, Ofsted inspection reports and success rates. The shifts in policy around these existing information sources, and the publication of the Framework for Excellence data the day before the second project seminar in December all highlighted the tensions in this important but still emerging area of work. The result of these deliberations was that some projects were small-scale in nature, potentially making it difficult to influence the culture change required for their organisations to secure true accountability to their communities. Overall, the experience of providers suggests that if an organisation is to develop a community scorecard, then it needs to be part of a long-term strategy and based on a real desire for 20 Shaping the Community Scorecard Action Research Project public accountability. It has the potential to be an invaluable continuous improvement tool for providers who genuinely seek and act upon community feedback, as well as serving a variety of other uses, from supporting community engagement to PR. Defining the community Defining the community was essential at the outset to enable providers to target their research and seek meaningful views to inform their scorecard. Some providers felt they served a diverse community, whereas others thought they served several distinct communities with differing needs. Some providers decided to define community by geography; others felt customers constituted their community. A small number took a broader approach to encompass ‘users’, ‘non-users’ and ‘stakeholders’. However, for most the key community focus for this project was service users rather than wider stakeholders and other public bodies. This may have been due in part to the time constraints of the project, or the general turbulence in the public sector landscape in autumn 2010. Some providers changed their perspectives on the scope of their definition of community in relation to this project as their action research progressed, particularly as they realised the scale of what they were planning to undertake within the timescale of the project. This resulted in some providers defining one or two distinct communities within their overall community to research as part of the project. This was often driven by a key strategy that the provider wished to pursue, for instance, increasing adult learner recruitment or employer engagement, engaging with consortium partners, developing a new venture or to gauge perceptions of the provider following a merger. On the other hand, two provider networks explored the possibility of putting in place a single scorecard for the whole network. Securing legitimacy for the community scorecard Each provider’s journey in developing the scorecard was unique according to the internal and external environment in which they operated. Although legitimacy was an area of consideration for all, it was explored in more detail amongst those providers further along the developmental journey. Leicester College, for example, used the action research project to ‘soft launch’ its scorecard with the community, but this was the culmination of two years’ development within the organisation, including with governors. Leicester College The College has developed its scorecard through the involvement of a number of key stakeholders over a two-year development period, including with student liaison committee and attendees at the College’s Annual General Meeting, and with the College governors being closely involved throughout and providing the impetus for development of the scorecard. The focus of their action research project was to build on work already undertaken, by developing a web-based version of the scorecard and conducting a ‘soft launch’ through dissemination by day-to-day contact and a direct mailing of key partners. The reason for the ‘soft launch’ was to provide an opportunity to refine and further develop the scorecard prior to re-launch in 2011. The College monitored website activity, recording 860+ hits, with 420 unique visits during the project. The College received “overwhelmingly positive comments” from the email web-link sent to key partners; however, feedback invited using a feedback button on the website elicited no response. The College subsequently identified that it still had more to do to engage the local community with the scorecard and is looking to embed it within the College’s wider communication channels. Action Research Project Shaping the Community Scorecard 21 Support for the development of a scorecard varied between providers with some having top-level governor and senior management team endorsement. Others restricted their community scorecard development to their areas of influence, almost working in isolation from the rest of the organisation, success with the project being as a result of the commitment and hard work of small teams (often 1 or 2 individuals) rather than a strategic imperative. However, the individuals engaged in the development of a community scorecard felt that top-level commitment was essential, not least to ensure that sufficient resource was available to develop and maintain the scorecard. Reasons for generally low levels of senior buy-in are unclear, although the lack of clarity around providers’ purposes in pursuing the scorecard could be a factor. Also, there was uncertainty regarding the relationship with existing datasets and how the community scorecard would sit alongside other public information such as Framework for Excellence and Ofsted reports. The project was conducted in autumn 2010 at a time when the implications of the Coalition Government’s approach to public service reform were only beginning to emerge. There was uncertainty about the new balance between national and local accountability and the relationship to future funding, and perhaps a reluctance to move too far from a focus on known performance measures. The tensions between transparency and the importance of good publicity were felt by some to be a factor, particularly in areas of competition. There was also some anxiety about potential risks to the community profile of institutions raising public expectations through community engagement / consultation during a period of austerity. Approaches by individual providers As has already been said, the providers all had different starting points and ambitions for their projects. The research was most effective when the provider had a clear view at the outset of their community and their research methodology and the project was managed within the context of measuring the contribution that the provider made to the community. Some projects ‘piggy-backed’ on other initiatives although this was generally found not to be an effective approach to developing a scorecard. Many providers had originally set a broad definition of community and as a result the initial research methodologies were wide-ranging and perhaps on reflection over-ambitious. Once the project commenced and the scale of the challenge was mapped out, it became clear to some that it would be impossible to include all communities within the scope and timescale of the project. Therefore some providers, such as Stockton Riverside College, identified a number of priority community groups, for example workless people and those in need of employability skills, as the focus for their research. Once this more precise definition had been agreed, often the scale of the task became much more manageable. Other providers took a much more pragmatic approach. Before embarking on gathering hard evidence of their community contribution, they spent the research phase holding internal meetings to explain the purpose of the community scorecard and secure the support and buy-in of internal management teaching staff, and followed this up with internal focus groups to define the community and agree which community groups were to be the audience for the scorecard. Others focused on gathering the views of their member organisations, key external stakeholders and local opinion formers. The level of research expertise differed amongst providers, for example in the design of research instruments to provide robust and verifiable results. The amount of resource available also varied widely. Broadly, large general FE colleges were more ambitious in terms of their aims and objectives and tended to define their community more broadly for research purposes. These providers planned to gather views across their communities and then disaggregate the views through their proposed research methodologies. Two of the general FE colleges had research specialists involved in the project and these organisations were experienced in gathering and sharing information. Leicester College, for example, had segmented its community into different user groups and used existing and publicly available information to gauge its contribution to the local community. 22 Shaping the Community Scorecard Action Research Project Petroc The action research project was timely for Petroc College as they were about to embark on a major strategic plan consultation exercise and opted to add a very short survey asking people to rank their perceptions of how the College contributed to the community as part of that process. The consultation was web-based and the College sent a direct email to key stakeholders, college staff, recent and current learners, schools, local authorities and elected members, totalling 20,000 individuals. In addition, the local paper ran a front page story covering the College’s strategic vision and included the web address where individuals could become involved in the consultation. Unfortunately, the level of response was low – but that in itself was an important learning outcome for the college. Some of the smaller providers, particularly from the vocational learning sector, found the action research approach more challenging, partly due to inexperience in this area and a lack of knowledge of robust research methodologies. Interestingly, the majority of the providers’ methodologies focused on conducting research to gather primary data to inform their scorecard measures, rather than utilising existing data sources. This could be attributed to the fact that providers were keen to develop a community scorecard with a local flavour and wished to take the opportunity afforded through this project to test out new approaches. It may also reflect the timing of the project: most of the research was completed before the Framework for Excellence data was published in December 2010. There was recognition, however, that conducting primary research was time-consuming and could require significant on-going resource commitment. A small number of providers considered an approach that involved using existing data sources. The learner, or potential learner, was rarely the focus of the primary research – where research was conducted with ‘users’, the responses were higher than research approaches that sought to involve ‘non-users’. This could be a reflection of the lead-in time required to engage these groups in research and the learning that providers would have to undertake before they were able to do this effectively. It also reflects that an organisation needs to feel confident, when inviting community views, that it has a clear strategy in place which: • has a clear purpose; • will not raise expectations that cannot be met; • asks relevant questions; • stimulates a valid and robust response; and • does not place the organisation in a difficult situation whereby it may have to publish or respond openly to negative perceptions and views. Many of the providers had only begun to undertake preliminary research during the duration of the project and would require further research if they were to develop a full scorecard. Emerging measures There were differing views from providers as to what type of measures to include in a community scorecard, with the intelligence gathered being wide-ranging and encompassing perceptions, customer satisfaction, stakeholder views and community needs. Some also perceived difficulties in establishing robust measures that could be replicated. Action Research Project Shaping the Community Scorecard 23 Not all providers had time to define in detail the measures that would constitute their scorecards. However, a number of providers did identify some top-level measures that could be further refined including: reputation; sustainability; community engagement (such as volunteering at events); quality; customer feedback / satisfaction; learner support; value for money; learners recruited from disadvantaged groups / areas; and curriculum meeting community needs. One provider also suggested a measure of ‘innovative and visionary leadership’. Some of these measures reflect existing national indicators, but others did go beyond that, in particular community engagement (for example, numbers of students volunteering in the community, or the number / range of events open to the community); the extent to which patterns of participation could be mapped to areas of disadvantage; and the extent to which the provider’s curriculum offer could be seen to be meeting the needs of the community. Some providers started to reflect on how ‘soft outcomes’ – for instance user feedback – could be measured and form part of the scorecard, but this was felt to be more challenging and there was insufficient time to consider this fully within this project. Some providers adopted a hybrid approach using a combination of new and existing data, thereby enabling the scorecard to be individualised, whilst at the same time complementing existing public information such as Framework for Excellence. Three provider case studies illustrate the types of measures that might be included within a community scorecard. Barton Peveril College Barton Peveril College set out to consult with delivery staff, learners and stakeholders on what a community scorecard might look like, how it could be used to the advantage of the college and how measures could be developed. The focus was adult and community learning. Staff were initially consulted and as a result proposed measures identified. The measures were then incorporated into a questionnaire that was used to consult with the community. Measures identified were: • childcare facilities • community involvement • cost of course • exam results • learner support • location and access • environmental friendliness • Ofsted • refreshments • reputation The research indicated that the measures of cost of course, exam results, location and access, and reputation were particularly important to stakeholders. 24 Shaping the Community Scorecard Action Research Project Wandsworth Council Lifelong Learning Wandsworth Council set out to review its community development frameworks and to establish whether it was collating the correct evidence to judge the effectiveness of its community development provision and if it was judging the impact of its community development provision from a sufficiently wide perspective. The research was mainly undertaken through existing groupings. Wandsworth Council was also involved in the LSIS PRD community development project. The Council’s mission and strategic priorities, together with its defined priority groups, showed how identifying one community was impossible as the Council served many and varied communities. The focus for this project was taken to be stakeholders in adult and community learning, with learners at the centre and the provider, the community and employers also represented. Indicators were considered for each of the key stakeholder groups. The learner: • percentage of learners achieving their learning outcome • percentage of learners progressing onto a higher level of learning • percentage of learners gaining a job (previously unemployed) • percentage of learners gaining a promotion / changing jobs with their existing employer • percentage of learners starting a new business / developing their existing business • percentage of learners who gained voluntary work experience • percentage of learners being more involved in their community The provider: • percentage of learners aged 60+ • percentage of learners from disadvantaged areas • percentage of learners with below level 2 qualifications • percentage of learners from BME groups • percentage of male learners • percentage of learners with disabilities / learning difficulties • percentage of learners studying Skills for Life (literacy, numeracy and ESOL) • percentage of learners retained on the course. The community: • percentage of learners volunteering with a local community group • percentage of learners becoming involved with their children’s school • percentage of learners who have gained voluntary work experience • percentage of learners being more involved in their community (e.g. joining coffee mornings, meeting neighbours in groups, joining the neighbourhood watch) • percentage of families becoming more involved in the community / school. The employer: • employees who have increased responsibility in the workplace as a result of their learning • employees who have gained promotion as a result of their learning • employees who have gained skills through acting as a workplace supervisor. Action Research Project Shaping the Community Scorecard 25 Black Country Training Group (BCTG) BCTG set out to develop an overarching community scorecard on behalf of all its training provider members that would show what end-users think of the service provided. Partners were consulted on a number of proposed measures that included: • quality – service quality standards and quality reports; • learner voice – success stories, learner feedback, surveys and questionnaires; • support – value for money, pastoral care, and information, advice and guidance; • employer feedback. Providers were generally keen to develop a community scorecard that was relevant to their organisation and the community it serves, feeling it was important that it was seen to be owned and driven by themselves and their community rather than as a national imposition. Certainly at the outset of the project, the invitation to providers to determine their own scorecard utilising whatever measures they felt appropriate was a key motivating factor, although one of the key learning points arising from the project was that this can be both challenging and time-consuming. The publication of Framework for Excellence data just prior to the December seminar served to prompt providers to consider whether a community scorecard framework with some guidance might provide a useful tool for other providers, perhaps extending so far as to incorporate some core measures, as well as the flexibility to design their own measures. Publication of results A number of providers considered how they might present their scorecard to the community, with a web-based approach being particularly popular. Some providers also considered the extent to which the scorecard could become an interactive tool that invited user feedback, but it was unclear from this project how such feedback would be managed, particularly negative messages. There was also consideration given as to how the scorecard would relate to existing published data and whether, for instance, it should become a central data dashboard with signposting to other sources of information. Others had yet to decide whether they wished to communicate their research findings at all, and a small number were concerned about how to manage potentially conflicting messages, for example, where public information seemed at odds with community views. 26 Shaping the Community Scorecard Action Research Project Support for providers Providers participating in the project generally made significant steps towards either the development, or enhancement, of their scorecard approach. However, it was found to be overly ambitious for providers to conduct primary research and develop a community scorecard from scratch within the project timescale; further time and resource would be required by some to ensure that the scorecard continued to be implemented. A number of key stages were identified in the journey to develop a community scorecard: • securing senior management commitment; • integrating within strategic plans; • defining the community; • seeking community views, • developing measures; • testing measures; • publishing the scorecard in a user-friendly format; and • continuous improvement and review. Although none of the participating providers completed all these stages, they all made progress on their individual community scorecard journeys. The development of a scorecard that is relevant both internally and externally requires time and effort; and similarly, maintaining a scorecard also requires resource commitment. It is clear from this project that providers looking to embark on the development of a community scorecard, or more generally to devise approaches to demonstrating their contribution to their communities, would benefit from advice and support which at the very least includes dissemination of the findings from this research to enable providers to learn from one another’s experiences. Other support could include: • mapping, analysis and dissemination of other community scorecard development activity – both in the FE and skills sector, and more widely across other public services; • establishing of good practice forums where providers can share their experiences, perhaps through the Excellence Gateway; and • development of a community scorecard toolkit that could provide guidance on topics such as research techniques and the relative merits of utilising existing and new data sources. Action Research Project Shaping the Community Scorecard 27 5. Emerging implications and six proposals for policy development and implementation Introduction The intention of the action research project was to give a group of providers the space to explore what information could be used to engage with, and demonstrate their responsiveness and contribution to their community or communities. The diversity of the group of providers, both in relation to how far these issues had previously been considered as well as differences in the types of institution and their geographical location, has meant that the question has been explored from a number of different angles. The research conclusions outlined in the previous section highlight the richness of the learning that this approach gave. This section builds on the research conclusions and draws out some of the emerging implications for policy development and implementation and makes six proposals. Overall, some concerns were raised about the term ‘community scorecard’. There are options for how this work could be described going forward – although the action research was overseen by the NIPB, alongside its work on FE public information and the ‘Informing Choice Framework’; the shift towards greater freedoms that has emerged in the last year has created greater space and necessity for providers to develop strategic plans that are responsive to their communities. Within this context, new ways of thinking about accountability could assist providers to articulate as fully and transparently as possible their contribution to civic and community wellbeing. Audiences Ideas about the purpose of a community scorecard varied between the providers and a range of different audiences were identified. The motivations for engaging with a community scorecard and the benefits of doing so also vary depending on the audience – for example, the reasons for providing information to potential learners will not be the same as those for a provider wanting to make information available to their local authority. There seem to be two main groups that are external to the provider. The first group are ‘customers’. This group includes existing and potential customers (learners, potential learners, employers), and what could be described as ‘the wider learning community’. The wider learning community encompasses organisations and groups that support people to get involved in learning (including community groups, support groups, career guidance professionals). Individuals and groups in this category need information to use to make decisions or to help other people make decisions. Informing choice by making information publicly available will be increasingly important as public service reform seeks to empower citizens and communities, and as the balance of responsibility for paying for provision changes. If people are asked to invest financially in their learning it seems logical to assume that they will demand more information on which to base their decisions. The benefits for providers of offering more information to customers include increased recruitment as more people understand what the provider can offer, and improved retention and success as people use information to make better decisions about which courses are right for them. The second group are ‘strategic partners’. This group is broad. It includes local authorities and other public sector partners such as health and the police, and also private and third sector partners such as Local Enterprise Partnerships, local chambers of commerce, or voluntary, social enterprise sector bodies. Given its core purpose of human capacity development, the role of the FE and skills sector has relevance for a wide range of stakeholders and strategic players in the locality. It will be beneficial for the sector to develop 28 Shaping the Community Scorecard Action Research Project awareness of its contribution to this wide-ranging audience. As more responsibility is devolved, it will be important that strategic partners in a local area understand and recognise each others’ roles to enable them to plan and work effectively in partnership. In a time of limited funding, this will be particularly important as partners look for ways to secure the maximum value from the collective investment in an area. Further development of the scorecard concept should therefore explore the scope and priorities of strategic partners and civil society in order to position the sector more centrally in relation to the strategic development of communities. Some of the providers also highlighted the importance of internal audiences and the value of the scorecard in informing their strategic planning process. There was not a single internal audience, so a community scorecard could support: • the principal / chief executive / head of service / governors / trustees to make decisions about their organisation’s strategic direction; • curriculum heads in making decisions about the mix of provision that can best meet the needs of the local community; • quality improvement managers reviewing the overall responsiveness and performance of the organisation; and • marketing departments in recruitment and other campaigns. Proposal 1: It is important to recognise the wide range of audiences for information on FE and skills. The sector needs to develop its awareness of the priorities of strategic partners and civil society in order to shape its response to the emerging ‘Big Society’ concept. Presentation of information Different audiences have different information requirements. For customers, the NIPB’s work to develop a “more comprehensive range of public information”, building on the Framework for Excellence, is important in terms of securing a nationally consistent, comparable set of information on which learners and employers can take decisions about the types of provision and providers in which they wish to invest. Other types of information for customers will be particular to different providers: for example, arrangements for the new discretionary support funds; details of facilities, resources and materials available; information about fees (including fee loans in future); and promotional offers. York Consulting, for the NIPB, has been looking at the different types of information valued by learners and employers and this will be a good source of evidence for providers wishing to develop their own provision of information. In addition, we note the new measures included in the Structural Reform Plan for BIS which may in turn become part of a wider set of public information: in particular, the proportion of college leavers who go into employment or further training and are still there after 6 months; the proportion of young people from disadvantaged backgrounds who go on to higher education by age 19; and the number of apprenticeship places. Our research suggests that securing accountability to communities needs to be part of a long-term institutional strategy and plan. Measures of community accountability should reflect the overall goals of the institution, negotiated with and responsive to its customers, community(ies) and strategic partners. Experience from the community scorecard projects suggests that individual provider surveys of their communities may be helpful in highlighting the sort of local information that customers want. Strategic partners will be interested in the consistent, comparable information available through the framework of public information, but often through a local, community-wide, or sectoral lens. So a local authority and its partners might want to know, not only that a provider performs well on the provision Action Research Project Shaping the Community Scorecard 29 of long vocational courses, but that these courses serve the key sectoral or occupational priorities of the local economy; or that a provider is investing in engagement provision in a community with poor health outcomes or a history of community tension14. The narrative will be different for different providers serving different communities. A key challenge for providers will be how to analyse, interpret, use and present an increasingly wide range of public information and evidence on FE and skills in ways that are authentic and accessible. The growing emphasis on being responsive to customers and communities places a premium on providers’ information and stakeholder communication strategies – marking a shift from previous arrangements which tended to see the primary audience for ‘performance information’ as the funding agencies. Ofsted and government. There is a wide range of data already available to assist providers, but our research suggests some organisations would welcome support in thinking about how to use and present this information in ways that will help to demonstrate their contribution to their community and strengthen their engagement with economic and social partners. The 2020 Public Services Hub’s recent report15 on a ‘social productivity’ approach to FE and skills argues for the importance of all public services being able to ‘incubate’ and express clearly their social value within communities, and this should be an area for further exploration. Proposal 2: It would be helpful to develop a spectrum of measures as examples for providers, but ultimately the measures providers select must enable them, with their communities, to assess progress towards and achievement of their goals. Proposal 3: Sector leaders should consider the new challenges and opportunities of working with an increasingly wide range of public information, and making sense of it with and for different audiences / stakeholders. LSIS could support the sector to explore how information and communication strategies should underpin the shift towards greater community responsiveness. Proposal 4: The implications of emerging thinking about social productivity could be further explored as a possible guiding framework for generating more meaningful metrics to account for providers’ strategic engagement with and contribution to their communities and localities. Dynamic engagement with communities In the action research, although some providers focussed on presenting existing data in an accessible format, it was interesting to note that a number of providers took the approach of conducting primary research, and the approach that was most frequently taken was to ask the community for their views of the provider. The providers that took this approach encountered a number of issues including: • The resources involved in carrying out primary research are not insignificant. This relates to the specialist skills required by staff as well as the time this takes and the associated cost. • At the time the surveys were conducted (autumn 2010) there was a genuine concern amongst providers about raising the expectations of the community when reduced funding meant providers were not sure they would be able to respond to increased demand. • Not all feedback is positive. For providers committed to closing the feedback loops to their communities, some had concerns about going public with negative feedback, particularly if a provider located close to them was taking a different approach. Evidence on the wider benefits of learning is clear that participating in learning can have positive effects on health and civic engagement www.learningbenefits.net 14 15 2020 Public Services Hub at the RSA; ‘The further education and skills sector in 2020: a social productivity approach’ www.lsis.org.uk/2020PSH 30 Shaping the Community Scorecard Action Research Project These issues highlight a key emerging challenge for providers about how to discharge their responsibilities for responsiveness, to develop more dynamic engagement with their communities, and empower customers and partners to interact with them to co-design services and programmes. Many of the providers involved in the community scorecard research envisaged presenting their information in a web-based format. The potential of technology to support providers’ information and communication strategies is, we know, being explored across the sector. To give just two examples, providers are using Facebook, amongst other methods, to promote courses, communicate with students, and create online learning communities. Others are trialling ‘Trip Advisor’ style approaches to enabling students to rate and comment on courses. There is scope for information strategies to go beyond the transactional – the presentation of different types of information and traditional surveys – to harness the interactive potential of technology. In so doing, providers would open up opportunities to gather more systematic market research and intelligence and to use this to further develop deliberative engagement with their communities. If, for example, prospective learners and employers were able to register their interest in courses or programmes that were not currently available, providers would have a way of understanding latent demand and could choose to respond, subject to numbers, costs, availability of staff and resources, etc. Alternatively, the provider could assist with matching – employers with potential recruits or apprentices, or volunteer tutors with interested students – becoming a facilitator of a wider learning community and communities of interest, contributing to the Big Society as well as opening up new opportunities for business development. The responsiveness of this approach could become part of the provider’s overall narrative to both customers and strategic partners, visibly connecting it to the concerns and priorities of the communities it serves. Proposal 5: Sector providers should explore innovative approaches to developing dynamic engagement with their communities, including how to harness the potential of technology, and the possible economies of scale of working across the sector. Strategic direction and ownership There was variation in the level of senior leadership involvement across the providers involved in the action research. Where there was senior leadership, the work was more likely to be used to support the strategic planning of the institution. This is perhaps not surprising but does demonstrate the importance of senior leadership if information is genuinely going to be used to account to the community for a provider’s responsiveness, contribution and value – such approaches need to be embedded and made a priority within an institution. Senior leadership also seems to be important for practical reasons. Firstly, the research identified that it takes time to develop and agree what information should be made public. This was the case where existing information was used as well as where new research was carried out. Therefore sufficient staff time needs to be allocated. Secondly, and related, there may be implications for staff development, both for conducting primary research and for making sense of existing information. Proposal 6: Sector leaders should consider the implications of the shift to greater community responsiveness and public information for their approaches to accounting for their contribution to the economic and social life of their communities. Action Research Project Shaping the Community Scorecard 31 Community Scorecard Annex 1 Purpose 1. This paper proposes an initial outline of the purpose and scope of the community scorecard for learning and skills sector providers. 2. It aims to: • locate the community scorecard proposal in the context of The UK Commission’s overall proposals to provide more public information on the performance of the sector; • suggest how evidence about the role of providers in their communities can inform the development of the scorecard; • frame the purpose of the community scorecard, suggest the key audiences for it, and some of the benefits that could flow from it; and • outline an action research project to develop prototypes with colleges and other providers over the summer / autumn, with support from the UK Commission. Background The Policy Perspective 3. The UKCES’s paper ‘Towards Ambition 2020: skills, jobs, growth’, published in October 2009, makes a series of proposals for ‘building a more strategic, agile and labour market led employment and skills system’. At the heart of these proposals is an intention to “increase trust in, and authority to, learning providers”. Providing more public information on the performance of providers, and increasing their accountability to their communities is seen as one way to do this. The aim is for citizens, communities and employers, and members of local public service boards to be empowered to become far more directly involved in shaping outcomes and improving performance of the employment and skills system, and better informed to make choices. 4. In other public systems, such as local government, health and policing, there have, in recent years, been moves to make information on performance publicly available and to empower citizens and communities to become far more directly involved in shaping outcomes and improving performance. This direction of travel is set to continue as a central feature of public service reform. 5. The UK Commission’s proposals include: • requiring that all publicly-funded learning providers consult widely and collaboratively with employers and other stakeholders in their community, and use available labour market information to shape annually their mix of provision to meet the needs of their labour market; • requiring all publicly-funded learning programmes to provide public quality labelling on key outputs and outcomes such as learner success rates, destinations, wage gain, quality and satisfaction levels for learners and employers; 32 Shaping the Community Scorecard Action Research Project • creating a new and public institutional performance framework for learning providers – a balanced scorecard – based on aggregate outcomes / destinations, satisfaction levels and quality; and • balancing the institutional performance framework with “evidence of the economic, social and labour market characteristics of the local catchment area”, to allow “for the recognition of post-code effects and distance travelled”. 6. LSIS has an interest in and experience of supporting and encouraging colleges and providers to play a strategic role in supporting their local communities and has agreed to work with the UKCES on the development of a ‘community scorecard’ to address the fourth proposal. The locality perspective 7. A number of recent developments have raised the profile and challenged the practice of learning and skills providers as strategic partners in their local communities. These include: • the Total Place pilots – LSIS commissioned and worked closely with AoC to explore the involvement of colleges in the pilots; • the new duty on colleges to promote economic and social wellbeing in their localities – LSIS was commissioned by by the Department for Business, Innovation and Skills to develop guidance for the sector; • opening up college premises to the community to support informal learning – LSIS was commissioned by BIS to produce a good practice guide; • tackling and preventing gun, knife and gang-related problems in the sector - LSIS has produced a good practice website to support the sector; and • LSIS’s support programme for equalities, diversity and community development. 8. This work provides a body of evidence from the sector that can inform the development of the community scorecard. It illustrates the breadth of potential community development activities undertaken by providers and indicates that different providers tend to prioritise different dimensions or elements, according to their mission, the scope of their provision, and their local economic and social contexts. 9. Rather than offering the community scorecard as another framework, we will seek to develop it as a spectrum or range of potential ways for institutions to develop their contribution to communities, and consider their reputation and communications within their locality. We want it to be a stimulus for providers to consider how to secure strategic gain and to play their part in local public service coalitions. Ultimately, this means providing more information to their local community, and therefore strengthening local accountability – which is what we consider to be the key purpose of the community scorecard. Purpose of the community scorecard 10. To offer a spectrum of ways for further education providers to articulate and enhance their value to citizens and partners in their local area. This could encompass: • establishing the fit between the mission of the organisation and local strategic priorities; • raising awareness of the wider contribution that providers of further education make to securing social and economic wellbeing in their localities; Action Research Project Shaping the Community Scorecard 33 • communicating the economic and social returns to learning beyond learning outcomes; • enhancing the accountability of providers to citizens and communities; • providing a focus for developing leaders of learning and skills as strategic partners in local public service coalitions; and • providing a stimulus to develop partnerships for better integrated delivery, efficiencies, and improved outcomes for learners. Audiences 11. The key audiences for the community scorecard are likely to be: • local authorities and other public service parners; • citizens, and community and interest groups; • learners,parents and carers; • employers and employer organisations; and • the media and general public. Benefits 12. For providers, the community scorecard offers an opportunity to: • gain greater credibility and recognition as a strategic partner with significant local public sector partners; • influence policies, strategies and budgets in the locality; • raise their profile with key stakeholders and position colleges and providers in wider local education partnerships and as key players in a broad local public service coalition; • contribute solutions to complex issues that help to secure effective outcomes for learners and the wider community; • develop college provision to meet local strategic needs; • promote a positive local learning culture, stimulate ambition, and encourage demand for learning; • secure the contribution of further education to reducing inequality and promoting social mobility, as part of a community partnership; • demonstrate the wider value of learning to individuals, communities and employers; and • encourage citizens to take pride in ‘their’ local further education provider. 13. For learners, employers and the community, the community scorecard should: 34 Shaping the Community Scorecard Action Research Project • provide a vehicle for stronger public debate about the role of learning and skills in securing local prosperity and wellbeing; • offer better information about support and opportunities across learning and skills sector institutions and other public services; • ensure better alignment of provision with need, working across the public, private and third sectors; and • identify opportunities to secure improved value for money for tax payers through more effective scrutiny of services. 14. For local authorities and public service partners, the community scorecard may present opportunities to: • harness the leadership capability within the further education sector to help shape local priorities and plans and develop innovative approaches to delivering effective public services; and • place strategies for transforming lives through learning at the heart of local vision and plans for wellbeing, prosperity, and empowerment. Complexity of local contexts 15. The community scorecard must recognise the diversity in the range and mix of institutional and administrative arrangements in different localities. 16. In terms of institutions, there is a range of FE providers – general further education colleges and sixth form colleges; independent work- based learning providers; specialist colleges; and public and third sector providers, including local authorities. These institutions have a range of missions, institutional foci, learners, provision, kitemarks (e.g. Beacon status), and constitutional / governance arrangements. Within the further education market there is inevitable competition which is likely to increase as resources get tighter and the focus on marketisation increases. 17. Administratively, the range of local government arrangements is also complex. Some areas operate under unitary councils, others with two- tier authorities. Institutions can find themselves serving multiple local authority areas, or a part of a shire county authority. In addition, there are different government arrangements in London (which has a strong regional identity and an elected mayor) and Greater Manchester / Birmingham (recently designated as city regions with formal agreements between participating local authorities and enhanced powers in relation to skills). 18. In addition, spatial, demographic and socio-economic differences inform different local community priorities, a point the UK Commission seeks to capture in their proposal that all publicly-funded learning providers should “consult widely and collaboratively with employers and other stakeholders in their community, and use available labour market information to annually shape their mix of provision to meet the needs of their labour market”. 19. This diversity underlines the necessity to avoid any prescriptive design for the community scorecard. Action Research Project Shaping the Community Scorecard 35 Rather it needs to provide a spectrum of opportunities from which providers can select the most appropriate mix for their specific context. Design and piloting 20. Our approach is to engage the sector in shaping the development of the community scorecard from the beginning. 21. We propose to invite ‘willing’ providers – from FE colleges and the work-based and ACL sectors – to join a small action research / piloting group. 22. Based on an agreed outline, the focus of the action research / piloting would be to: • establish a clear narrative about the purpose of the community scorecard; • develop the key elements of the community scorecard spectrum from which providers can select; • develop a range of impact measures appropriate to providers’ contexts; • consider how the community scorecard will secure its legitimacy, including the role of trustees, and who will need to be involved in authorising it; • consider how to use the scorecard, what it will look like, how it will be communicated, and where it might be displayed; • trial aspects of the scorecard to provide examples of how it could work in practice; • identify barriers to the effective implementation of the scorecard; • identify what further support is required for colleges, providers, and other public sector partners to ensure the community scorecard adds value; and • identify key systems issues for further investigation with the UK Commission and the Department for Business, Innovation and Skills. 36 Shaping the Community Scorecard Action Research Project Proposed action research methodology and timescale May / early June Discussion with the UK Commission re: • scale of projects • agreement on selection criteria. July Invitation to become a project Selection of providers 9 September Inaugural development seminar to: 1 day seminar • discuss purpose, scope, etc • agree specific areas of focus of different projects • action plan processes / approaches. September / October / November Projects develop practice supported by LSIS associate Early December Second seminar to: 1 day seminar • share progress / approaches • agree implications for purpose and scope • capture key lessons • agree options and proposals for taking forward the scorecard Mid-end January Final reports from each project Summative report on the action research February - March Dissemination Action Research Project Shaping the Community Scorecard 37 Registered Office: Friars House, Manor House Drive Coventry CV1 2TE t 024 7662 7900 e enquiries@lsis.org.uk www.lsis.org.uk LSIS258 ©Learning and Skills Improvement Service 2011 Company number 06454450 Registered charity number 1123636 September 2011 38 Shaping the Community Scorecard Action Research Project