General Comments

advertisement
FUNDING COUNCILS’ REVIEW OF RESEARCH ASSESSMENT
Liverpool JMU Response
General Comments
JMU welcomes the review of the RAE, in particular the opportunity it represents to allow a wider and more
relevant view of research to emerge.
The process and approach and approach of the RAE at present is problematic. A system, in which academics
undertake and quality assess their own work, is fundamentally flawed. Citations are largely by academics, grant
applications reviewed by academic peers, academic papers are refereed by academics. Not only is the current
peer panel approach narrow in terms of what is valued but it is devalued by those benefiting from the results
actually producing them.
We need to reflect on the notion of expert and move from a system when academics are assessing the
academic worth of the work submitted to one in which expert and possibly lay involvement assesses the quality
of work that is of value to the UK as a whole.
a. What is the purpose of the RAE?
The purpose of the RAE should be to inform the allocation of resource to research activity within universities
consistent with the needs of the UK in the 21st century. This role is difficult to disentangle from other resource
streams for research (such as the Research Councils) so we will focus here on the particular role of the funding
councils within the dual support system of funding background activity and allowing sufficient flexibility in
research funding for new or unfashionable areas to be supported. We would argue that this implies widening the
scope of the review to take into account broader issues than those narrowly associated with restricted views of
academic excellence. Whilst a clear requirement is a system that supports world class innovative research,
wherever it is found, it is also necessary to ensure that the research activity underpinning knowledge transfer in
all its various forms (teaching of students to spin out) is also supported. The RAE should therefore be capable
of identifying leading edge research and research which has genuine relevance to the knowledge economy.
We would argue for a strong focus on the needs of the funding council’s for research assessment and the
avoidance of overloading of the assessment process with the needs of other stakeholders. If other stakeholders
find the outcomes of RAE’s useful, that is to be welcomed but it would be a mistake to impose on an already
difficult and onerous task requirements that might distract from appropriate and accurate assessment. The
funding councils should be clear about exactly what any RAE measures but, if they do this, cannot be held
responsible for misuse by other agencies and should not compromise their needs in an attempt to make good
the shortcomings of others.
JMU is of the view that given its role in dual support the methodology for RAE support should be separated from
that of other funding agencies. Co-ordination of activity should be a matter of policy and strategy not an inbuilt
part of the data collection exercise. It is of course possible to use and exploit data from other sources but that
should not prejudice the role and function of the RAE itself.
b. How often should research be assessed?
The current time between assessments seems appropriate. We do not feel that there is any compelling reason
why research in different subjects should be assessed at the same time. A rolling programme of assessment
would be reasonable. We would argue that much larger groupings of subjects could be assessed together as
this would avoid some of the issues of fragmentation, territoriality and inconsistency of assessment between
UoA’s found in the current system.
c. What is excellence in research?
Excellence in research must be related to the reasons why we undertake research. Excellence is not an isolated
quality of research but can only be judged against a view of why the research is needed. Excellent research is
research that fits that need. Having said this it is possible to identify an important general characteristic of
excellent research – it is of use either directly in its own right or it is of importance in the development of new
knowledge, that is, it is useful to other researchers in the development of further knowledge. In terms of
excellence on this latter characteristic it would be appropriate to look judge against international norms.
It is vital that a broader understanding of the nature of research within universities and the reasons that research
is undertaken becomes part of the assessment process. At present the view of research is too heavily focussed
on a model of research support derived from big science (implying spatial concentration of research is in itself a
good thing) and on the requirement to compete against international elites supported under completely different
funding arrangements. We would note that such a model does not take into account, for example, the funding
requirements of low cost research subjects in the modern world of wide access information resources, emerging
areas of research and the emergence of distributed (often interdisciplinary) research groupings. We would note
that the exploitation off knowledge developed within the UK and elsewhere in the world requires a wide-ranging
base of research of national excellence in addition to leading edge research. As it is clear that the funding
councils should be nurturing research of national standing either the RAE should be capable of identifying where
it takes place or a mechanism for its support that does not require an RAE should be in place. In a similar
manner the cost effective role of research in academic job satisfaction (and hence recruitment, retention and
motivation of staff) and in broadly based knowledge transfer should also be recognised in any funding
mechanism for research RAE.
Excellence should be judged by fitness for purpose against both international norms and these broader aims for
research.
d. Should research assessment determine the proportion of funding directed towards each
subject?
No. Research assessments should inform such decisions but the level of support for different research areas
should be strategic need. An important caveat here is that research areas should be very broadly drawn (e.g.
bio-sciences, information technology) so that micro management is avoided and the allocation of funding is still
appropriate to the needs of dual-support. We would argue against allocations based purely on historic, external
or volume measures.
e. Should each institution be assessed in the same way?
This is an odd question. We assume that assessment is of the quality of research (we would argue for the
quality of research of individual researchers) not institutions. One possibility is to assess different research in
different ways in which case, de facto, institutions would be assessed differently dependent on the nature of the
research they submitted. All assessment should be consistent, fair and on an equal basis for comparable
research.
f.
Should each cognate area be assessed in the same way?
Yes.
g. Discretion.
Institutions should be free to opt in or out of assessment as they see fit. As responsibility for funding is currently
devolved to institutions it is difficult to envisage a system that would not devolve responsibility for submission to
institutions other than a system that allowed no discretion. For example a reasonable system might well be to
assess all individuals and allocate funding on an individual basis, aggregated to subject or institutional level.
This would avoid many game-playing, equal opportunity and ratings issues.
h. Equality of Opportunity.
This is a vital issue that the current RAE fails fundamentally to address. The current system enforces a narrow,
blind and unhelpful view of research excellence that is predicated on the research career progression of a very
particular type of individual. Any system that does not broaden both the notion of research and criteria for
excellence will inevitably systematically disadvantage a number of different groups within universities.
i.
Priorities.
Not burdensome.
Rigorous.
Transparent.
Expert Review
We have already noted our concerns over the use of peers. We welcome the adoption of the term expert review
and support the view that such a panel, properly constituted could respond to the questions of excellence in
terms of purpose and use, and consequently view research quality more rationally, usefully and accurately.
We would argue that the assessment should combine the past, based on evidence of output and achievement,
with plans against which expected performance can be evaluated.
The data used will reflect the purpose of research support that we have discussed above. Criteria should be
output rather than input based and hence concentrated on achievement, contracts delivered, knowledge
transferred and publications written.
We acknowledge that the significance of the level of assessment might vary depending on the discipline, lone
scholar to research team. We are however emphatically opposed to the assessment being at the level of
Institution.
We would support, subject to further work, assessment of individuals.
We have noted above our views on subject level assessment.
Expert review is well understood and predictable, but well known to be subject to the bias of self-similarity (“good
researchers tend to be like me”). The range of reviewers thus needs to reflect the reasons why research is
being supported and not just included those who are selected on the basis of problematic judgements of
academic excellence.
Algorithm
We would not support the sole use of algorithms in the assessment of research.
Such an approach would almost certainly result in the distortion of research activity to fit the algorithm used.
The use of metrics, where available and appropriate, in conjunction with other methods is reasonable. We
would argue strongly that if research income is to be used as a metric that it should only be used alongside a
value for public money metric.
Self Assessment
Self-assessment has positive attributes, as it would permit institutions to adopt their own approach and level. Its
major strength lies in the value of the process itself which would include strategy development, monitoring and
review. Many institutions currently undertake self-assessment as part of their normal research monitoring
activities. It would however be subject to several of the problems associated with criteria, validity and verification,
and the overall validation process, with its associated difficulties of expert and peer review, alluded to above.
Fundamentally however, we feel that it is, taken alone, an inappropriate approach in circumstances where it is
the sole determinant of a system which allocates substantial funding.
Historical Ratings
We strongly oppose the adoption of an historical approach. It is basically flawed as it takes no account of the
current quality of research, its strategic value or timely responses to changes in the distribution of research
excellence at any level.
Any assessment system should be responsive to change and there is little or no evidence to suggest that the
cost benefits of a static assessment system would be outweighed by the disbenefits of allocating funding in the
wrong places due to the inevitable time lags caused by such a system. There is no evidence that quality of
research or strategic need changes at such a slow pace that such a system could be justified even when
modified by use in conjunction with other systems.
Download