Quality Assurance within Higher Education Institutions Professor Phil Cardew Pro Vice Chancellor (Academic) London South Bank University Objectives of the Day • To establish the concepts of ‘standards’ and ‘quality’ and their place within a higher education institution. • To consider core processes of benchmarking, reporting and review needed to manage quality and standards. • To discuss the relationship between management processes and resourcing and quality assurance systems. • To consider the interrelationship between internal and external assurance systems. • To consider reporting mechanisms, and risk assessment and management from a management perspective. Agenda 0930: Welcome, introductions, discussion of core concepts. 1000: Quality assurance at programme level: establishing and assessing to a standard. 1100: Coffee Break 1130: Validation, monitoring and review: process and reporting. 1230: Lunch 1300: Student engagement: feedback and representation. 1400: Using reports, risk assessment and management. 1430: Plenary Discussion Quality assurance at programme level: establishing and assessing to a standard. Standards and Quality • What is ‘a standard’? – Thresholds attainment – Levels of achievement – Benchmarking and equity • What is ‘quality’? – Customer service models – Enhancement The Building Blocks of degree awards – working with institutional variation: • ‘One size fits all’ approaches. • Establishing ‘labels’ – understanding structures. • Common ‘labels’: – Programmes and courses – Frameworks and Pathways – Modules and Units • Variability of approach • Modification Structures of delivery • • • • • Full-time and part-time Single honours and combined honours Distance, blended and distributed delivery Delivery by partner institutions Multi-site delivery – Comparability – Assessment of standards • Research degrees and learning contracts • Relationship to academic regulations How do we establish a standard at programme level? • • • • • National Qualifications Frameworks Subject benchmark statements Professional body requirements Employer requirements External examiners – Academic – Professional Standards, levels, awards • • • • Award outcomes – graduate attributes Levels within an award Exit qualifications Assessment – Type – Variation – Loading • Embedded, dual and articulated awards. Employability • ‘Professional’ and ‘Academic’ qualifications • Subject knowledge, technical ability, specialist skills, core skills • Currency of knowledge • Work-based learning Conclusions • No ‘one size fits all’ or ‘standard’ model • However – especially in early stages – consistent approach pays dividends • Important to establish an ‘outcomes based’ approach • Clear understanding of overall learning outcomes • Clear understanding of level and progression • Clear assessment strategy Coffee Break Validation, monitoring and review: process and reporting. Basic Questions What are we trying to do ? Why are we doing it ? PURPOSES How are we going to do it ? Why is this the best way to do it ? How will we know it works ? How can it be improved ? METHOD REASON OPTIMISATION EFFECTIVENESS ENHANCEMENT The Building Blocks • Validation – programme approval • Annual monitoring: – Action planning – Relationship to other processes • Periodic review: – Cycle of operation – ‘End of cycle’ and ‘mid-cycle’ Working with collaborative partners • Types of relationship – ‘Flying faculty’ – Part-franchise – Franchise – Validation – Accredited Partner • Approval of delivery • Periodic review Validation • Initial approval in principle: – – – – Strategic ‘fit’ within overall academic portfolio Clarity of award title Market Desirability for professional and/or employment market • Validation event: – Programme specification – External involvement • Academic • Professional • Employer Annual Monitoring • Cyclical action planning • Responding to data: – External Examiners’ report(s) – Progression and Award Statistics – Module Evalution Questionnaire results – National Student Survey – Employment Statistics • ‘Sign off’ of minor modifications Periodic Review • Relationship between review, validation, monitoring and minor modifications: – Incremental change and re-validation – Stability of award title and learning outcomes – (Advantage of frameworks and pathways) • Gives experience of programme over a longer time-scale • Allows for ‘major’ changes • MUST include appropriate externality Action Planning and Reporting • • • • • • Identifies short and medium-term actions Includes responsibility Identifies activities to be undertaken Includes review point Establishes benefits of activity Reports on: – Conclusion of activity – Results of action Conclusions • Nested activities – not separate processes • Should establish continuum of evidenced action planning • Can work in clusters of programmes as well as individual programmes • Should lead to clear, concise reports • MUST include externality in all aspects and at all points. Lunch Student engagement: feedback and representation. • Why engage students with quality processes? – Identify strengths and weaknesses of delivery from a student perspective – Engage with aspects of delivery outside teaching: • • • • Classroom and lecture space Library IT Specialist equipment – Engage with assessment, marking, moderation and feedback to students Basic Methods • • • • • Module Evaluation Annual Surveys Course Boards Student Meetings Student involvement in Periodic Review: – In meetings – As Reviewers • Senior engagement with the Students’ Union Module Evaluation • • • • • • Standard questions Scoring Similar timescales of delivery Anonymous completion Comments as well as scores Standard reports – – – – Module Programme Department Faculty • Focus on under-performing modules Annual Surveys • • • • New entrants International Students National Student Survey Postgraduate Surveys: – Taught programmes – Research Students • ‘Pulse’ surveys Course Boards and Student Meetings • • • • • • • Elected representatives Training Timescales for meetings Standard Agendas Gathering information Feedback Relationship to other processes – External examining – Annual monitoring Students within validation and review processes • Student meetings – – – – Engagement with new proposals Feedback on existing courses Recent graduates reflecting on employability Engagement with department – responsiveness to feedback etc • Students as reviewers – Experience on QAA reviews – Training – Limits of process Conclusions • Student input adds value to processes. • Needs to happen in collaboration with Students’ Union (or a Student Society). • Representatives need training. • Need to establish clear understanding of goals of engagement. • Needs careful handling not to patronise or antagonise. • Need to reassure staff that they are in control of their programmes – but that student input is valuable! Using reports, risk assessment and management. What are the aims of quality assurance processes? • Confirmation of standards • Reassurance that processes have been completed • Reflection on performance (data monitoring) • Enhancement of future delivery (programme structure and quality of delivery/environment). What should processes focus on? • Specialist understanding of the academic discipline • Statistical data – progression and achievement • Feedback from external examiners • Feedback from students • Employment statistics • Resources What should we avoid? • • • • Long and tedious reports with nothing to say. Repetition from previous years. Narrative description with no analysis. ‘Open-ended’ action planning. What should we promote? • Focused reports. • Clear analysis of data. • Action plans which show monitoring and completion of actions. • Forward planning related to analysis. • Responsiveness to feedback. • Development, not stagnation. Becoming ‘risk aware’ • Can we focus only on key areas (programmes) of risk? • What should lead to investigation? – Threats to standards – Poor progression and/or achievement – Negative feedback (from examiners or students) – Poor satisfaction – Poor employability – Lack of action Reporting as part of a cycle • Reflection on previous year’s report (and actions). • Analysis of data – including comparison with past performance (what is ‘direction of travel’?) • What action is needed as a result? • Who will do it – by when? Conclusions • Reporting need not be a huge burden (either to the author or the reader). • Must have clear outcomes and be useful. • Must be used. • Must have place in future activity and reflection. • Poor performance must be dealt with (both in terms of activity and reporting). Basic Questions What are we trying to do ? Why are we doing it ? PURPOSES How are we going to do it ? Why is this the best way to do it ? How will we know it works ? How can it be improved ? METHOD REASON OPTIMISATION EFFECTIVENESS ENHANCEMENT Plenary Discussion