NULL: No difference between (actual RAE grade

advertisement
THES article I UoA 21 & the RAE
The RAE – Environmentally Unfriendly?
Just before Christmas, the Research Assessment Exercise (RAE) results were released. Overall, this
five-yearly nation-wide assessment of the strength of research in UK universities revealed the best
results yet. For Environmental Sciences however, in the main it was very bad news indeed. It is
resulting in the closure of schools, and appears to be in the process of shutting the subject down in
most, except the very few, large universities. There are bigger underlying issues here about how
universities are funded, and whether the model applied to Higher Education from the 1980s until the
present day is appropriate or in the national interest. However, this article will focus on this quiet
crisis which is going on in UK higher education only inasmuch as the RAE is part of it.
The Exercise
Under the RAE, 59 peer review subject panels were set up to review 69 subject areas (Units of
Assessment or UoAs), and these considered in total 2597 submissions from university schools,
departments and research groups. Each of these submissions was graded from 1 to 5* (see Table I)
where 5* is internationally excellent, and 1 is virtually no national excellence. The procedures that
would be followed were published in advance, and form the basis on which the exercise was
conducted. The immediate outcomes of the exercise are threefold:
(1) each submission has a rating, and hence (it is assumed) an indication of research strength and quality.
(2) most of the major research funders (e.g. The UK Research Councils) use these ratings (either directly or
indirectly) as an arbiter of whether and to what extent they will fund research or research proposals from a
given school/department/university. The degree to which subjects are funded also becomes an issue when an
entire area comes out as Environmental Sciences has.
(3) there is immediately an easy way to categorise departments and universities by the use of such scores, hence
many of the national newspapers produce league tables of universities.
The RAE looks at research outputs over a five-year period, so RAE2001 looked at research from
1996-2001. The 2001 exercise is the third national exercise, and followed previous exercises in
1996 and 1992. Although the UK Research Councils advised that RAE2001 should be the last RAE,
in their own words they reported1 that “..it [the RAE] was having the effect of stifling rather than
promoting research..”, the vice chancellors of the universities (“Universities UK”) lobbied strongly
for its retention. (“Universities UK” is the former Committee of Vice Chancellors & Principals
(CVCP) augmented by those of the new universities). It is thought that they did this because the ex
CVCP institutions were loathe to lose a system of research funding over which they [or their
universities] had large amounts of control. Even a casual inspection of research funding of UK
universities shows that these original institutions receive the overwhelming share of funds, and that
the RAE has strengthened that position further.
The main strengths2 (or the “jewels in the crown”) of the RAE are:
(1) it promises comparability, consistency, continuity and credibility across the spectrum of UK research, so that a
5* in anthropology or history, should be comparable to a 5* in chemistry or environmental sciences.
(2) it is supposed to be an “absolute” rather than a “relative” process, the grading that a submission receives is not
supposed to depend on that received by any other, nor on the identity of the department making the submission
There are also potential weaknesses. Unless the quality control procedures exercised by the
organisation running the RAE are rigorous, both of the above strengths can be lost, and the results
1
Reference from THES in office box
Paraphrased and summarised from RAE2001: Assessment panels’ criteria and working methods. RAE Team
publication
2
1
THES article I UoA 21 & the RAE
of the whole exercise compromised to a greater or lesser degree. As both the health of UK science,
and (literally) people’s livelihoods can depend on the outcomes of this exercise, it is incumbent on
HEFCE3. to take great care in how the RAE is set up and run. It is a situation where things must
both be done, and be seen to be done. For these reasons, reported problems in any aspect of the
running of this exercise should be the subject of rigorous investigation, review and if necessary
modification. It is perhaps important to stress that the RAE as a salient peer review system must
also be seen to work well. All UK university scientists are very dependent on peer review processes
like the RAE. We are part of these processes as both reviewers and the reviewed. Our papers are
published by peer review, our grants gained by peer review, in fact much of any individual’s future
is determined by peer review. It is important then, that there is confidence in these processes. If the
RAE is unsatisfactory it can bring the whole peer review edifice into question.
History
After each of the previous RAEs there has been serious controversy over one particular subject Environmental Sciences (UoA21). In RAE1996 it was judged by a joint panel with UoA20 Earth
Sciences. After that exercise there was a mass meeting of the subject area (under the auspices of
CHES – The Committee of Heads of Environmental Sciences) at which HEFCE effectively agreed
to change the panel status (from joint to single subject) and hence address issues around the
composition of the panel. In RAE2001 it again was judged in a joint panel with UoA20 (although
two members, including the chair, were eminent environmental scientists). This time, UoA21 came
out by almost any parameter, as apparently the weakest of all UoAs in the exercise. So weak in fact,
that the average grade for this UoA is below the threshold for government funding. With the
possible exception of Nursing, this must be the only UoA to effectively close down its own subject
area. Although not strictly correct, if the numerical codes in Table I are used for RAE grades, the
average for UoA21 comes out at 3.26 – this compares to averages for the whole exercise of 4.10, or
from UoA20 of 4.38. (On average Environmental Sciences was more than 2 RAE grades lower than
that for the whole exercise).
As Higher Education in the UK (or at least sectors of it) has been significantly under-funded over
recent years, the funding that comes with the RAE has become a major component of the operating
plans of many Schools and Universities. This is especially true for the new universities which were
floated on a single day in 1992 with no reserves, just their collateral assets (buildings etc.). Hence,
although the UoA21 results have had major impacts in many institutions with Environmental
Sciences groupings being dismantled and in some cases staff being made redundant, these effects
are enhanced in the new universities.
Quantitative Analysis of the results
As the implications of the results of UoA 21 are so great for the RAE generally, and this area of UK
science particularly, some scrutiny of the results is warranted. Moreover, it seems appropriate that
as a judgement is being made about a science area, the tools of science should be used to prepare
that judgement. Accordingly, below a statistical approach is presented which establishes that the
results of UoA21 are different to those from the rest of the RAE and UoA20, and a logic tree is then
used (see Figure 2) to investigate reasons for the differences. The sample of university
schools/departments submitted to any UoA are self selected, and universities chose how to “carve
up” their research effort to this exercise. These choices were only revealed when the submissions
were in. There has been a suggestion that the sample of institutions which comprised UoA21 in this
exercise does not well represent the actual sample which is UK Environmental Sciences. It is true
3
The exercise is the joint responsibility of the national funding bodies - in England it is the Higher Education Funding
Council of England – HEFCE, in Scotland and Wales, SHEFC and HEFCW respectively, and in Northern Ireland it is
the Department of Higher & Further Education, Training, & Employment for Northern Ireland. In this article this group
will simply be referred to as HEFCE
2
THES article I UoA 21 & the RAE
that one of the major players (Lancaster) chose to submit elsewhere, and some other institutions did
likewise, but in reality, by far the majority of Environmental Science schools were represented4.
The results of UoA21 are anomalous
In any collection of different numbers (including RAE grades) one number has to be smaller than
the rest. The fact that UoA21 is lowest in the overall ranking of all UoAs is therefore, in itself, not
necessarily a problem. What is a major problem however, is that both in terms of the distributions
and medians, (see Figure I) UoA21 is significantly different from both the results of the rest of the
exercise, and from those of UoA20 – Earth Sciences, (which was dealt with by the same panel). It
is important to realise what this implies - the chances of the UoA21 results coming from the body of
the RAE results as a chance outlier is ~ 1 in 75,000. This result is important for HEFCE and
everybody else. That this difference exists at all points to a loss of one or both of the main strengths
of the exercise.
What if the results are “correct”?
Figure 2a shows a logic tree for analysing the reasons for the difference between UoA21 and the
rest of the RAE. Logically, this highly significant difference could reflect either a real difference
between UoA21 and the rest of the RAE, or could be something which is nothing to do with the
samples themselves. For now, we will restrict ourselves to the first option. If in addition the
difference above is not due to some sort of artefact in the samples, (formally a sample limitation),
then there are only two options to explain these results:
(a) either Environmental Sciences in the UK is indeed the weakest of all subjects – but this does not address the fact
that the data is so different from that from the rest of the RAE, nor that outcomes (grades) given to most of the
UoA21 institutions cannot be reconciled to the definitions of those grades. Whilst the basic HEFCE definitions of
grades (see Table 1) are so inconsistent with the situation “on the ground”, it is extremely difficult to interpret the
outcomes as credible
(b) or, paradoxically, the UoA21 panel is the only panel which worked correctly. Figure 3 shows the grades from the
whole exercise. The distribution displays considerable skewness (distortion toward the higher grades). This is in
itself problematic, and should ring alarm bells. It has been speculated that this shows how much UK academia has
improved in terms of its research. However, it is very difficult to derive such a skewed distribution by that
mechanism. It is more likely to indicate that the (overwhelmingly single) subject panels “talked themselves up” –
i.e. came out with higher grades across the board for their subject areas. The lack of effective quality assurance
procedures is the most likely cause. Although unpalatable to celebrated UK academics, the shape of this
distribution is potentially good evidence that HEFCE has not taken sufficient care to ensure comparability across
the range of subject coverage, and hence has compromised both the main aims of the RAE.
Could the difference be due to a sampling problem?
We could alternatively assume that the difference between UoA21 and the rest of the exercise is due
to a limitation in the samples. Figure 2a lists a number of possible artefacts gleaned from
conversations with Heads of UoA20 and 21 submissions, as well as others in the area.
Chance has already been quantified as negligible above (see Figure 1). Many expressed the view
that UoA21 was either harder or easier than UoA20. However, this is logically inconsistent either
way. It seems clear that provided we do not subscribe to conspiracy theory (that some knew in
advance that UoA21 would be harder or UoA20 easier) then all would have favoured the easier
option.
One factor, (it is widely held), underlying the difference between the results of UoA21 and those of
the rest of the exercise or UoA20 is that UoA21 was primarily composed of new universities.
Hence, because these have not had the sustained high density funding that is necessary to initiate
4
There are about 45 undergraduate Environmental Sciences type courses in the UK, and possibly 90% have a
substantial research base. Of these maybe ~15% did not submit into UoA21.
3
THES article I UoA 21 & the RAE
and sustain a powerful research base, then UoA21 should be much weaker than all the others. It is a
beguiling idea, and one which at first glance is obviously correct and generally applicable across the
RAE, see Figure 4, (Table 3 contains the definition of the university age or “Type” variable).
However, were this idea the underlying reason for this particular difference, one would expect to
see all UoAs with a similar new university composition to that of UoA21 with similar outcomes.
Such a comparison (see Figure 5) shows significant difference between these six other UoAs5 and
UoA21. This effectively rules out sample composition (in terms of new universities) as the primary
reason for the difference between UoA21 and the rest of the exercise.
Size of department was the other factor that many people from the Earth and Environmental
Science communities cited as crucial. This is alarming because if size were such an important
factor, this contradicts specific advice from both from HEFCE6 “..There is no reason, in principal,
why a submission from a small group, or comprising small groups, cannot obtain a high rating…”,
and also informal advice from the panel prior to the submission. Several Earth Sciences colleagues
pointed to the Oxburgh Review of Earth Sciences which resulted in the closure of many Earth
Science departments to form a few super-departments. They posed a question like “Did you not
realise that this will happen to Environment as well?” Interestingly, most scientists now realise that
“Big is not Beautiful” when it comes to creativity in research, but still many academics and civil
servants who are involved at the core of the UK research funding scene are bedazzled by this idea.
Although that group of people is relatively small, they have a very powerful hold on the purse
strings and pop up in the RAE, Research Council peer-review committees, anywhere in fact where
there is money or influence. Even a cursory look at grants awarded from The Natural
Environmental Research Council reveals that most of the grants go to members (past or present) of
their own committee structure. This is not a conspiracy – just the recognition of a culture. Where
there is under-funding of a research area, the competition for the scarce funds accelerates and this
culture springs up. Once the main players are identified, then an equilibrium situation will result
which is difficult to break, e.g. there is an annual meeting between the Head of NERC and the heads
of what he consider the chief “client” departments. There has even been a word developed for this
type of culture – “grantsmanship”. So it should not be surprising that “Big is Beautiful” should
prevail as an attitude within this part of the UK Science community. Figure 6 shows that size of
submission is a strong indicator of grade across the whole exercise – but it is steeper in the science
area, and steepest of all in UoA21. Whilst in certain individual cases very small submissions did
obtain high grades7, overall the picture is clear.
Whilst this type of analysis may be interesting, it is simply trying to re-invent whichever wheel the
UoA21 panel themselves used. As it is unlikely we will ever know what they actually did, there is
limited use in pursuing this further with any of the other HEFCE data. The immediate conclusion
for this analysis is that despite their advice and RAE guidance to the contrary, clearly size was an
important factor, and that this is so in UoA21 more than in any other UoA.
This leaves us with only two conclusions from the assumption the significant difference between
the UoA21 and the rest of the RAE is due to a real difference in the samples. They are:
a Size of submission gives some insight into UoA21 panel operation, and was certainly more important for this
panel than for others
b..New university composition of submissions is not the reason for oddity of UoA21 results
5
UoA26 General Engineering was not different, it was also a joint panel, see Figure 5.
verbatim from RAE2001: Assessment panels’ criteria and working methods. RAE Team publication
7
The submission from Exeter University contained just one person and obtained a grade of 4.
6
4
THES article I UoA 21 & the RAE
These conclusions seem rather interesting, because it seems none of the widely thought
explanations for the anomalous UoA21 results hold up in the cold light of day. That they are
anomalous is beyond doubt, and is in itself evidence that the Quality Assurance systems put in place
by HEFCE have failed. Alongside the indicative evidence from the distribution of results for the
whole exercise, it is potentially very bad news, and it should be acknowledged as such by
academics and HEFCE alike.
A peer review process that went astray?
However, in the end, it is a peer review process that has made a judgement. To challenge such a
process is difficult unless there is some other independent evidence that supports the statistical
analysis of the HEFCE data. Happily, in this case, such data exists. Our peers (on the panels) who
reviewed UK academia in the RAE are my colleagues in this and other UK universities. Those
universities that chose to submit into UoA21 did so alongside their other submissions into other
UoAs. The required data comes from the predictions and results to all UoAs of those universities
which submitted into UoA21.
If the predicted grades across all UoAs could be obtained for those institutions who made
submissions into UoA21, the difference in predicted and actual grades could then be tested for
UoA21 vs the rest. I have obtained some of this data. There are very stringent confidentiality
conditions attached to the raw data (Public Relations considerations of the donor universities), and
hence it appears in condensed form in Table 2. RAE grades were coded from 1-7, where 1 = RAE
grade 1, and 7 = RAE grade 5* and it is the differences between predicted and actual grades that are
presented. So a difference of 0 means an accurate prediction, -1 means optimistic by 1 grade, etc.
From the condensed data it seems that this sample of universities were mildly optimistic in their
predictions of UoA grades (not including UoA21) by ~0.2 grades, whereas for UoA21, this
optimism swells to ~ 1.2 grades. The accuracy of the predictions of UoA21 and the other UoAs are
different with a confidence of 99.95%. (The standard deviation data from the two samples also
indicates a similar story with respect to accuracy of predictions). This data at last begins to shed
some light onto the conundrum which is the UoA21 results – could it be that the members of the
panel were themselves not representative of those they were peer-reviewing? It is an uncomfortable
fact that the institutions of only 2 members of this panel actually submitted into UoA21, with at
least one of the significant institutions in the sector eschewing it. This hardly makes for
“ownership” of the subject area. It is also a fact that most of the panel members were earth
scientists. It seems to me that by definition peer review must include “peers” on both sides of the
fence – the evidence here is that HEFCE have failed in this fundamental underpinning of the RAE
in this area.
Did the UoA21 panel malfunction?
There is one other area where we should probably go, and as the data is leading there it seems
reasonable. Figure 2b contains the logic tree for the assumption that there is no genuine difference
between the samples. This does not imply any sort of conspiracy or lack of integrity by any panel
member per se., Instead, it is an exploration of the role of the “culture” in the decisions that have
been taken. However, it is very bad news for the peer review process that is the RAE in this area.
There are three possibilities for difference in grades which are not to do with differences in research
quality between the samples. The chance of the UoA21 result being a genuine outlier have already
been assessed (see above) as negligible. Equally the possibility of some architectural feature of the
RAE exercise which somehow discriminates against UoA21 seems remote. The joint panel issue is
probably the only relevant point here, and this has been alluded to above in terms of subject
5
THES article I UoA 21 & the RAE
ownership. It is difficult to formulate any quantitative statistical test which might shed light on this.
Which means in short order we come back to some aspect of the function of the UoA21 panel.
Of the remaining possibilities, subject bias has been alluded to above, and may have played a part.
The process adopted by this panel may have been different from that adopted by other panels –
certainly we are aware that this panel sent out a great deal of material – maybe more than any other
in the RAE - for advice. (Understandable, given the subject range of the submissions and panel
expertise). There is opinion that this process alone would give depressed assessments because those
receiving the material will have been judging full submissions in their own areas, and hence might
unconsciously compare the UoA21 “bit” submissions as of lower quality that the “full” submissions
from their home panel. Inappropriate criteria may have been applied, or appropriate criteria applied
too harshly. I certainly have no information on what went on in the panel to express an opinion one
way or the other. Finally, personal or political bias. I personally have confidence in the integrity of
those panel members I know, but also acknowledge that bias need not be conscious, and think the
discussion on “culture” above is relevant.
The final issue here is again one of HEFCE (and the UoA21 panel) appearing to breach their own
procedures. There is of course the general Quality Assurance failure apparent by the significance
level of the difference between UoA21 and the rest, as well as what appears to be the failure in
UoA21 to provide what is normally thought of as peer review. However in an interesting twist
HEFCE seem to have indicated in a formal response to a request for information under the Data
Protection Act that in UoA21 they did not follow their published procedure in assessment of
submissions. These procedures state8 “..The work [my emphasis] of each researcher will be
individually graded…..and these grades will form the principal basis for the determination of the
overall grades for each of submissions specified research groups..”. To do this UoA21 sent out
forms to reviewers to assess each of the individuals submitted (interestingly, not only their papers,
but the researchers themselves). This then leaves a paper trail to enable panels to meet the Quality
Assurance requirement that2 “..Panels must be able to show in all cases how their quality
judgements relate to the evidence before them and to the criteria which they have stated they will
apply..”. Without this paper trail it could be said that the primary basis of the RAE gradings were
“discussions around a table”. However, in their own words on this very point, HEFCE have stated
“Neither the funding councils nor the panels retain any of the panel’s comments on the work of
individual researchers”. Yet without these written documents (which we know existed, because
some of us had to fill them out), the UoA21 panel cannot meet their own Quality Assurance
requirement (above). We also now learn that institutions were compared to assist in the
determination of final grade – this makes the RAE in UoA21 a relative rather than absolute
exercise, which in itself breaches yet another of the underpinning principals of the exercise. (One
wonders how many structural underpinnings need to be breached before the entire edifice falls.)
The aftermath and way ahead
In the aftermath of the exercise, schools struggle to reconcile the grades they have with the
definitions of those grades. The feedback which at the outset was promised to be “..summarising the
reason for the rating awarded, with reference to the published criteria..” has been delayed and is still
awaited – but we hear that it will not be particularly useful to help us find out how (in the main) our
submissions did so badly. Many institutions are in the position of having “failed” in UoA21 after
major investment, and are not able to find out exactly why they have failed, or more importantly
how to correct it for next time. Meanwhile the sector shrinks as schools close and institutions
redirect scarce resources elsewhere. There is no appeal on the RAE, unless a member of the
judiciary agrees with the thrust of this article and overturns the results of UoA21 in the courts. I
believe that the arguments (statistical and otherwise) for doing so are compelling.
8
Guidelines to submissions in UoA21, RAE Team
6
THES article I UoA 21 & the RAE
What though for next time? Although there are many suggestions that could be made, it seems to
me there are some clear lessons that stand out from the rest.

However the peer review system for the next RAE works, the peer reviewers must be representative peers of
those they are reviewing, which minimally means that their own institutions should be in the group they are
reviewing.

The goal-posts must be declared before the exercise runs, and there should be real transparency before, rather
than artificial visibility afterward. Prior declared procedures and principals should be inviolate.

HEFCE must institute effective quality control on all aspects of the RAE process from definitions of the
subjects, through to timing and nature of the feedback. There seems to be strong evidence that HEFCE quality
assurance procedures have to some degree failed, and it is alarming how there seem so many structural routes
left open by which results of panels can be tainted. Joint panels are a potential weakness.

Apart from the under-funding of this UK science sector, it seems to me that there are urgent issues for
government in the funding of universities. It cannot be in the national interest that powerful committees like the
Universities UK can overturn a recommendation by the Research Councils on the future of university research
funding which seem to be motivated by greed. It is also of concern that the lack of funding in this area is
encouraging the domination of the research funding nexus by relatively few universities, effectively preventing
the entry of other universities into the funding stream.

There is clear evidence that the results from UoA 21 are flawed (just by their difference) and that is a serious
issue for large parts of the Environmental Science sector in the UK
Acknowledgements
I would like to thank the large number of colleagues in the Earth and Environmental Science
community who have taken the time to talk to me, and pass sensitive institutional data to me.
Especial thanks go to my Colleagues in the School, especially Professor Vic Truesdale, Drs Andrew
Rendell, Peter Grebenik and Tim Shreeve. Finally I also thank the various statisticians both here at
Brookes and elsewhere who exercised patience with my questions, and Dr Nicola Madge who spent
time making me think a little more like a social scientist!!
7
THES article I UoA 21 & the RAE
LIST OF TABLES, FIGURES AND LEGENDS
Table 1 Definitions of the RAE grades
Table 2 Difference between predicted and actual grades for all UoAs from those
universities† who submitted to UoA21.
† Not all of the 34 institutions gave any or all of this data
Table 3 Codes used to define “Type” variable for university age
Figure 1 Results of Kolmogorov-Smirnov (distribution) and Mann-Witney-U
(medians) tests for difference between the results of UoA21 and the rest of the RAE
results, and the UoA20 results.
Figure 2(a) Logic Tree for assumption that there is a difference in the quality of
research between UoA21 and the rest of the RAE.
Figure 2(b) Logic Tree for assumption that there is no difference in the quality of
research between UoA21 and the rest of the RAE.
Figure 3 The distribution of grades‡ from the RAE2001.
‡ For codes used see Table 1.
Figure 4 The effect of age of university on RAE outcome§
§ For list of codes see Table 3. For this analysis non-teaching Research Institutes, Charities, non-university colleges or special
projects (e.g. University of Highlands & Islands project) are not included.
Figure 5 (top) a plot of the average “Type variable” for each UoA. (bottom) Results
of Kolmogorov-Smirnov tests for difference between each of those areas with a Type
variable of 1.78 or less with UoA21 (Type variable = 1.54)
Figure 6 The affect of the size of the submission on RAE outcome#
# The definition of “science” has been taken to include: UoAs 1,6-9, 13-21, 25-32.
8
THES article I UoA 21 & the RAE
Table 1 Definitions of the RAE grades
RAE grade
5*
5
4
3a
3b
2
1
Number
5.5
5
4
3.5
3
2
1
Definition (paraphrased)
International Excellence in > 50% of the work
International Excellence in up to 50% of the work
National Excellence in ~ 100% of the work
National Excellence in > 66% of the work
National Excellence in > 50% of the work
National Excellence in up to 50% of the work
National Excellence in virtually none of the work
Table 2 Difference between predicted and actual grades for all UoAs from those
universities† who submitted to UoA21.
† Not all of the 34 institutions gave any or all of this data
NULL: No difference between (actual RAE grade-predicted RAE grade) for UoA
21 and (others-UoA21)
mean UoA21
-1.219
Data is in the form of (actual RAE gradepredicted RAE grade) for all UoAs that each of
mean (others-21)
-0.222
the UoA21 submitting institutions submitted
stdev UoA21
0.826
to. The distribution of these differences should
stdev (others-21)
0.071
be normal around a mean of zero. The test is
21 count
21
whether the differences for UoA21 are
(others-21) count
11
significantly different than those for the other
t coeff
3.963
UoAs. Result is that they are different at
t crit (99.95%)
3.646
99.95% confidence level.
Table 3 Codes used to define “Type” variable for university age
Age of Institution
(in more or less current form)
1990- (the “New” universities)
1960-1989 (the “Robbins” universities)
1900-1959 (the “Redbrick” universities)
-1899 (the “Old” universities)
Type variable
1
2
3
4
9
THES article I UoA 21 & the RAE
Figure 1 Results of Kolmogorov-Smirnov (distribution) and Mann-Witney-U
(medians) tests for difference between the results of UoA21 and the rest of the RAE
results, and the UoA20 results.
Results for UoA21 seem different from
those of other areas and RAE generally
Medians of UoA 21
different (99.999%
CL.) from rest of
RAE (~1 in 74,000).
Also different
(99.97% CL.) from
UoA20 (~1 in 3000).
Distributions of
UoA 21 different
(99.98% CL
.)
from UoA20, rest of
RAE (~1 in 5000)
Kolmogorov-Smirnov
Mann-Witney-U
YES
10
THES article I UoA 21 & the RAE
Figure 2(a) Figure 2(a) Logic Tree for assumption that there is a difference in the
quality of research between UoA21 and the rest of the RAE.
Results for UoA21 are different from
those of other areas in RAE

NO
YES
Difference does not
reflect genuine
difference in samples
Difference reflects
genuine difference in
samples
Sampling
limitation
Genuine
Difference
Chance
21 Easy
Newer
1 in 74,000
logically fails
21 still different
Other
size
11
In UK,
Research in
Environment
al Sciences of
lowest
quality.
Environment
al Sciences
only panel to
function
correctly RAE flawed
THES article I UoA 21 & the RAE
Figure 2(b) Logic Tree for assumption that there is no difference in the quality of
research between UoA21 and the rest of the RAE.
Results for UoA21 are different from
those of other areas in RAE

NO
YES
Difference does not
reflect a genuine
difference in samples
Difference reflects
genuine difference in
samples
UoA 21 Panel
Chance
0.001% likely
RAE process
overall
Criteria
Criteria
Process
People
Subjects
Appropriate, but
harsh
Inappropriate
Inappropriate
Personal or
political bias
Subject Bias
12
Joint
Panels
THES article I UoA 21 & the RAE
Figure 3 The distribution of grades‡ from the RAE2001.
‡ For codes used see Table 1.
Results of RAE 2001
Fr
eq
ue
nc
y
800
600
400
200
0
1
2
3
3.5
4
RAE grade
13
5
5.5
More
THES article I UoA 21 & the RAE
Figure 4 The effect of age of university on RAE outcome§
§ For list of codes see Table 3. For this analysis non-teaching Research Institutes, Charities, non-university colleges or special
projects (e.g. University of Highlands & Islands project) are not included.
3000
Av
ran 2000
k
1000
4.00
Av 3.00
typ
e 2.00
1.00
0
0.00
0
1
2
3
4
Average type
0
2
4
RAE grade
14
6
THES article I UoA 21 & the RAE
Figure 5 (top) a plot of the average “Type variable” for each UoA. (bottom) Results
of Kolmogorov-Smirnov tests for difference between each of those areas with a Type
variable of 1.78 or less with UoA21 (Type variable = 1.54)
6.00
av.
typ 4.00
e
2.00
0.00
0
20
40
60
80
UoA
UoA
Name
11
Other Studies & Professions
allied to medicine
Built Environment
Library & Information Management
Art & Design
Comm, Culture & Media Studies
33
61
64
65
15
Results distribution
different from that
of UoA21
CL 80%
CL 95%
CL 95%
CL 99%
CL 80%
THES article I UoA 21 & the RAE
Figure 6 The affect of the size of the submission on RAE outcome#
# The definition of “science” has been taken to include: UoAs 1,6-9, 13-21, 25-32.
Av 50.00
era 40.00
ge
# 30.00
sta 20.00
ff
10.00
Whole Exercise
Science
UoA21
0.00
0
4
2
RAE grade
16
6
Download