UWE presentation

advertisement
Public involvement in
research: assessing impact
through a realist evaluation
invoNET 21 February 2012
David Evans, Vito Laterza & Rosie Davies
on behalf of the UWE/Coventry team
Acknowledgements
This project was funded by the National
Institute for Health Research (NIHR)
Health Services & Delivery Research
programme (project number 10/2001/41).
The views expressed are those of the
authors and not necessarily those of the
NHS, the NIHR or the Department of
Health.
UWE/Coventry team
•
•
•
•
•
•
•
•
•
•
•
•
•
Professor David Evans (principal investigator), UWE
Professor Jane Coad, Coventry
Dr Jane Dalrymple, UWE
Ms Rosie Davies, research partner
Ms Chris Donald, research partner
Professor Sarah Hewlett, UWE
Mr Vito Laterza, UWE
Dr Amanda Longley, UWE
Professor Pam Moule, UWE
Dr Katherine Pollard, UWE
Dr Jane Powell, UWE
Ms Ruth Sayers, research partner
Ms Cathy Rice, research partner
Focus of this session
• Initial thoughts on reflexivity
• Background to our project, design and
conceptual framework
• Reflections on early learning
– Theory development and testing
– Objectivity and contamination
– Working with research partners
Reflexivity and reflective practice
• “Reflexivity” – not the same as reflective practice:
• Thinking about ourselves (i.e. reflective practice) and
learning from that (i.e. do things better with reflective
practice)
• But also linking ourselves to our research participants
(not just what you can do better, but thinking about
others through our own experience)
• Thinking about thinking (i.e. epistemology): reflecting
upon basic structures and processes that are behind the
taken-for-granted everyday reality we live in and study in
our research contexts
Background
• Team based at University of the West of
England (UWE) and University of Coventry
• Grew out of existing UWE Service User and
Carer Involvement in Research initiative
• Nine academic researcher and four
research partner (user) co-applicants/coresearchers
Design
•
•
•
•
Realist evaluation framework
18 month project
Eight case studies
Mainly qualitative methods
– Semi-structured interviews (c. 5 participants
per case study x 3 interviews over one year)
– Observation
– Documentary analysis
– Consensus workshops
• Economic analysis
Realist evaluation
• Policy driven by an underlying theory of
how an initiative is supposed to work
• Role of the evaluator to compare theory
and practice
• “What works for whom in what
circumstances and in what respects?”
• Look for regularities of context,
mechanism and outcome (CMO)
(Pawson 2006; Pawson & Tilley 1997; 2008)
Levels of public involvement in
research theory
• Policy level – what do DH, NIHR and other senior R&D
stakeholders think are effective mechanisms leading to
desired policy outcomes?
• Programme/project level – what do stakeholders (e.g. PI’s,
research teams, research partners) think involvement
contributes to their desired outcomes?
• Academic level – what are the dominant academic theories
in the literature about public involvement mechanisms in
research and whether/how they work?
• Our research team – what do we think are the contextspecific and generalisable mechanisms leading to desired
policy outcomes?
Our CMO theory – to date
Context
Mechanism
Outcome
•Leadership on
involvement by the PI or
other senior member of
the research team
•Attitudes of trust and
respect towards the public
involved
•Culture of valuing and
support for involvement
•Infrastructure that
supports involvement, e.g.
policy on payment and
expenses
•Involvement throughout a
research project
•Long-term involvement
•Training and support
•Linking involvement to
decision making
•Budget for involvement
•Defined roles
Impact on research design
and delivery:
•Project design
•Research tools
•Recruitment
•Data collection
•Analysis
•Writing up
•Dissemination
Our task – articulate and test public
involvement in research theory
• Articulate policy-level programme theory from
policy documents, actions, etc – it’s about research
quality not empowerment
• Synthesise what we know about context,
mechanisms and outcomes in practice from the
literature – recognising complexity and uncertainty
• Simplify and express theory in a testable form for
testing in case studies
• Collect case study data and analyse
• Revise theory and repeat process
Reflections on theory
development and testing
• Difficulty of categorising factors as context
or mechanism
• Multiple contextual factors and
mechanisms making causal attribution
difficult
• Time period for data collection shorter than
project timescales
• Impact may be diffuse not specific
• Outcomes may be quite limited
‘Objectivity’ and ‘contamination’
At this stage, three related reflexive effects emerged:
•
Leading members of the team are also experts who provide advice and
training on public involvement in research to researchers and PIs
(i.e. “now that you study us, can we still ask for advice?”)
•
The questions we are asking are triggering processes of reflection that are
likely to have an impact on the ongoing processes of shaping public
involvement structures and mechanisms (i.e. “Mmm that’s a good question,
I haven’t thought about that” or “I will certainly consider these issues
further”)
•
Varying responses on potential issues of overlap (or “contamination”) from
the participants themselves (i.e. some do not seem to be particularly
concerned, while others are more concerned about keeping our study
process from influencing the ongoing process of public involvement under
study)
Reflections on ‘objectivity’ and
‘contamination’
Implications for questions of ‘objectivity’ in qualitative health research:
•
Reflexive effects need to be taken into account and productively explored
as data, rather than discarded as “unwanted bias”: this is why we decided to
keep reflection going, rather than have a unilateral policy on this (i.e. we will
continue to give advice and we are also aware that participants might
change their behaviour due to the questions and reflections emerged from
data collection; these effects will be followed up, where possible)
•
We will respect research participants’ wishes on the matter: if they want us
to put specific measures in place to reduce any possible influence, we aim
to accommodate that
•
We will protect research participants’ confidentiality and anonymity in all
cases, and this will always take precedence on questions of reflexivity and
objectivity in any case.
Involving research partners
• Contributed from design stage onwards
• Each research partner works with one academic
researcher and Vito on two of the eight case studies
• Research partners meet together as a group in addition
to attending full team meetings
• Involvement in all aspects of the project, including theory
building, conducting interviews and analysis
• Reflection on the team’s process of involvement, issues
and outcomes included as data
• Support from Vito and named researcher on team
Reflections on involving
research partners
• Role development in the institution at UWE
– Extending formal arrangements because of our need
for research passports
• Difficult to keep track of impact!
• Differences: levels of experience in research,
life experiences, kinds of contributions, between
case study teams ...
• Complex roles and relationships
Contact details
• David Evans, Professor in Health Services
Research (Public Involvement)
David9.Evans@uwe.ac.uk
• Vito Laterza, Research Fellow
Vito.Laterza@uwe.ac.uk
• Rosemary Davies, Research Partner
Rosemary3.Davies@uwe.ac.uk
Download