Measuring Public Innovation: Toward a common statistical approach Public Innovation Metrics

advertisement
Measuring Public Innovation:
Toward a common statistical approach
Nordic project on the development of
Public Innovation Metrics
Carter Bloch
Brainstorming session on measuring innovation in
Education, June 11 2009, Paris
København – Århus
www.damvad.dk
Background for project – funding support
• Project initiated by: Danish Ministry of Science,
Technology and Innovation
• Other Nordic contributors:
–
–
–
–
–
Nordic Innovation Centre (NICe)
Research Council of Norway
Innovation Norway
VINNOVA
SALAR (Swedish Assoc. of municipalities and
regions)
– Finnish Ministry of Enterprise and Employment
Participants
• Denmark:
– DAMVAD (Carter Bloch, Torben Vad, Mark Riis, Lydia
Jørgensen)
– CFA (Peter S. Mortensen, Ebbe Graversen)
– Statistics Denmark (Jens Brodersen)
– Danish Agency for Science, Technology and Innovation,
Denmark (Thomas Alslev Christensen, Jesper Rasch and Hanne
Frosch)
• Norway:
– NIFU-STEP (Johan Hauknes, Stig Slipersæter)
– Statistics Norway (Frank Foyn, Lars Wilhelmsen)
• Finland: Statistics Finland (Mikael Åkerblom)
• Sweden: Statistics Sweden (Roger Björkbacka, Per Annerstedt)
• Iceland: RANNIS (Thorvald Finnbjørnsson)
• Collaboration with UK (NESTA; DIUS)
• Hope to collaborate with other countries within NESTI Task Force
• Eurostat Pilot studies (Denmark and Finland have applied)
Background for project
• National interest in Public sector innovation in
Nordic (and other) countries
– Demographic changes necessitate innovation
– Competition with private service providers
– Better quality services to citizens
• However lack of systematic data on public sector
innovation.
– Hinders efforts to better understand and to
promote public sector innovation
Main objectives
• Develop framework and questionnaire for
collecting internationally comparable data on
innovation in the public sector
– Conceptual framework
– Survey methodology
– Studies of user needs
– Respondent interviews and testing
Main objectives
• Primary focus: ’generic’ survey instrument that
can be applied across govt levels and public
sector activities
• Goal: include main public sector activities, all
three levels of govt, front-line service delivery
institutions (hospital wings, schools, etc)
• Examine option of additional sector-specific
modules
• Examine how innovation data can be used
together with output data (often from other
sources).
– Project website: www.mepin.eu
Time line for project work
• Started in November 2008
• Nov-Feb, 2009: Background work
• March-Aug, 2009:
– Meetings with user groups,
– interview respondents, cognitive testing
– conceptual framework/indicators and survey
methodology.
– Pilot questionnaire.
• Fall 2009:
– Small-scale testing of questionnaire
– Deliverables on first stage of work
– Workshop for preliminary results (November).
• 2010:
– Pilot test studies
Implications for Indicators - Overview
• Innovations – definition and types
– Oslo Manual as starting point
• Unsuccessful innovations
• Innovation outputs (qualitative only)
• Innovation input (Oslo Manual as starting point, though quant
measures likely more difficult for public sector)
• The Innovation Process
– Innovation capability (what do organisations do to structure and
promote their innovation activities – and how able and ready are they?)
– Linkages (by type of partner; more than just cooperation)
– Drivers and barriers to innovation (actors and factors)
• ’Cross-cutting themes’ to be covered: ICT, HR, Procurement
Implications for Indicators/pilot questionnaire
•
•
Innovations – definition and types
– Oslo Manual, Product-process-organisational seems like a suitable
starting point. However, a number of questions on whether these types
can be distinguished
• Marketing innovations?
– Perception by many that changes mandated by policy directives, rules,
cuts are not automatically innovations
– Ask for examples – disseminate examples
Unsuccessful innovations
– Innovation projects that have been abandoned, Implemented
innovations that have failed
– Ask for examples
– Give reasons why not implemented or why a failure
– Impacts of these unsuccessful innovation projects (learning effects vs. nomore-experiments attitude)
Innovation inputs
• Oslo Manual as a starting point (activities y/n; resources)
• Questioned: validity/reliability and interest in quantitative estimates
• How to collect? OM ”approach”?
– Estimates based primarily on budgets/accounts of innov.- projects?
– Estimates also based on loose estimates?
– Separation between R&D and non R&D (in-house; extramural)?
respondents may be seeing def of R&D for the very first time…
• Additional questions related to procurement here?
The Innovation Process
• We ask about innovation inputs and outputs;
other questions essentially are collecting data on
the innovation process:
– Innovation capability (what do organisations do to
structure and promote their innovation activities – and
how able and ready are they?)
– Linkages
– Drivers and barriers (actors and factors)
Innovation capability
some potential example ‘questions’
•
•
•
•
•
•
•
•
•
•
•
Innovation strategy
Specific goals, targets for innovation activities
Development department
Activities organised in innovation projects
Individuals charged with supporting the development and
implementation of innovative ideas
Procedures for reviewing/assessing innovative ideas for
further development and implementation
Regular evaluation of innovation strategy, innovation
processes
Systematic procedure for gathering external knowledge
Part of staff work time explicitly devoted to innovation
Innovation-related training/courses for mgmt, staff
Staff incentives for generating innovative ideas
Linkages
Interest in linkages appears to go beyond ”Cooperation: Y/N”
This suggests asking small set of interaction-related questions
by type of partner (potential examples):
• Businesses
– Innovation cooperation
– Collaboration in provision of services
– Outsourcing
– Use of external innovation specialists
• Users
– Innovation cooperation
– Analysing user needs
– Meetings/hearings with users
– Gather information on users through daily operations
• Other public institutions
– Innovation cooperation with public research inst.
– Innovation cooperation with other public inst.
Drivers and Barriers
• Can this be formulated as one question;
ie where respondent can mark whether each impact is
positive (a driver) or negative (a barrier)?
• Split the drivers/barriers in actors and factors?
• Long lists of potential drivers/barriers available
from former ad-hoc surveys.
Other areas that can potentially be treated as
cross cutting themes
• ICT
• HR
• Procurement practices
Download