Mandated online RAE CVs Linked to University Eprint Archives:

Increasing the predictive power of the UK Research Assessment Exercise while making it cheaper and easier

Stevan Harnad, Les Carr, Tim Brody (Southampton University) & Charles Oppenheim (Loughborough University)

ABSTRACT: Being the only country with a national research assessment exercise http://www.hero.ac.uk/rae/submissions/, the UK is today in a unique position to make a very small procedural change that will generate some very large benefits. The Funding Councils should mandate that in order to be eligible for Research Assessment and funding, all UK research-active university staff must maintain (I) a standardised online RAE-CV, including all designated RAE performance indicators, chief among them being (II) the full-text of every refereed research paper, self-archived in the university’s online Eprint Archive and linked to the CV for online harvesting, digitometric analysis and assessment. This will (i) give the UK Research Assessment Exercise (RAE) far richer, more sensitive and more predictive measures of research productivity and impact, for far less cost and effort (both to the RAE and to the universities preparing their RAE submissions), (ii) increase the uptake and impact of UK research output, by increasing its visibility, accessibility and usage, and (iii) set an example for the rest of the world that will almost certainly be emulated, in both respects: research assessment and research access.

Figure 1

Figure 1: Predicting RAE Ratings from Citation Impact (Smith & Eysenck 2002)

The UK already has a Research Assessment Exercise (RAE), every 4 years. The RAE costs a great deal of time and energy to prepare and assess, for both universities and assessors (time and energy that could be better used to actually do research, rather than preparing and assessing RAE returns).

For many areas of research, an important and predictive measure of research impact is the "Journal Impact Factor" (JIF) of the journal in which the article appears: the average number of citations its articles receive annually. (For core journals in all subject areas the JIF can be obtained from the Institute of Scientific Information's Journal Citation Reports service, for which the UK has a national site license: http://wos.mimas.ac.uk/jcrweb/.) The number of times a paper has been cited (hence used) is a measure of the importance and uptake of that research.

AppleMark

Figure 2: Predicting Citation Impact From Usage Impact (Physics ArXiv)

The JIF figures only indirectly in the RAE: Researchers currently have to submit 4 publications for the 4-year interval. It is no secret that departments (informally) weight candidate papers by their JIFs in deciding on what and whom to submit. Although it is always stressed by the RAE  panels that they will not judge papers by the journals in which they appeared (but by the quality of their content), it would nevertheless be the strange RAE reviewer who was indifferent to the track-record, refereeing standards, and rejection-rate of the journal whose quality-standards a paper has met. (For books or other kinds of publications, see below; in general, peer-reviewed journal- or conference-papers are the coin of the research realm, especially in scientific disciplines.)

Statistical correlational analyses on the numerical outcome of the RAE using average citation frequencies predicts departmental outcome ratings remarkably closely. Smith & Eysenck (2002), for example, found a correlation of as high as .91 in Psychology (Figure 1).  Oppenheim and collaborators (1995, 1998; Holmes & Oppenheim 2001) found correlations of .80 and higher in other disciplines.

The power of the indirect journal-based JIF has not yet been tested for predicting RAE rankings, but it is no doubt correlated with the well-demonstrated RAE predictive power of direct author-based citation counts (average or total). Journal-impact is the blunter instrument, author- or paper-impact the sharper one (Seglen 1992). But a natural conclusion is that the reliability and validity of RAE rankings can and should be maximized by adding and testing as many candidate predictors as possible within a weighted multiple regression equation.

AppleMark

Figure 3: New Online Performance Indicators

Nor is there any reason why the RAE should be done, at great effort and expense, every 4 years! Since the main determining factor in the RAE outcome ratings is research impact, there is no reason why research impact should not be continuously assessed, using not only author- and paper-citation counts and the JIF, but the many other measures derivable from such a rich research-performance-indicator database. There is now not only a method to assess UK research impact (i) continuously, (ii) far more cheaply and effortlessly for all involved, and (iii) far more sensitively and accurately (Figures 2-4), but doing the RAE this new way will also dramatically enhance UK research impact itself, (iv) increasing research visibility, usage, citation and productivity, simply by maximizing its accessibility (Figure 5-8).

AppleMark

Figure 4: Time-Course of Citations and Usage (Physics ArXiv)

The method in question is to implement the RAE henceforth online-only, with only two critical components: (a) a continuously updated and continuously accessible RAE-standardized online CV (containing all potential performance indicators: publications, grants, doctoral students, presentations, etc.) http://paracite.eprints.org/cgi-bin/rae_front.cgi for every researcher plus (b) a link from each CV to the full digital text of every published paper -- books discussed separately below – self-archived in that researcher's university Eprint Archive (an online archive of that institution's peer-reviewed research output). http://www.eprints.org/self-faq/#institution-facilitate-filling (See the free, open-source software developed at Southampton to allow universities to create their own institutional Eprint Archives: http://software.eprints.org/ )

AppleMark

Figure 5: Open Online Full-Text Access Enhances Citations by Dramatically  (Computer Science)

Currently, university peer-reviewed research output -- funded by government research grants, the researcher's time paid for by the researcher's institution -- is given, free, by all researchers, to the peer-reviewed journals in which it appears. The peer-reviewed journals in turn perform the peer-review, which assesses,improves and then certifies the quality of the research (this is one of the indirect reasons that the RAE depends on peer-reviewed journal publications) (Figure 7). There is a hierarchy of peer-reviewed journals, from those with the highest quality standards (and hence usually the highest rejection rates and impact factors) at the top, grading all the way down to the lowest-quality journals at the bottom http://wos.mimas.ac.uk/jcrweb/. The peers review for free; they are just the researchers again, wearing other hats. But it costs the journals something to implement the peer reviewing: http://preprints.cern.ch/archive/electronic/other/agenda/a01193/a01193s5t11/transparencies/Doyle-peer-review.pdf.)

 

AppleMark

Figure 6: The Vast and Varied Influence of Research Impact

 Partly because of the cost of peer review, but mostly because of the much larger cost of print-on-paper and its dissemination, plus online enhancements, journals charge tolls (subscriptions. licenses, pay-per-view) for access to researchers' papers (even though the researchers gave them the papers for free http://www.ecs.soton.ac.uk/~harnad/Tp/resolution.htm). The result is a great loss of potential research impact, because most institutions cannot afford to pay the access-tolls for most peer-reviewed journals (there are 20,000 in all, across disciplines), but only to a small and shrinking proportion of them http://www.arl.org/stats/index.html.

Hence the second dramatic effect of revising the RAE to turn it into online continuous assessment based on the institutional self-archiving of all UK peer-reviewed research output is that it will make all that UK research accessible to all would-be users worldwide whose access is currently blocked by access-toll-barriers (Figure 8). If RAE mandates self-archiving, university departments will mandate it too. Here, for example, is the draft Southampton self-archiving policy: http://www.ecs.soton.ac.uk/~lac/archpol.html

The UK full-text peer-reviewed research archives will not only be continuously accessible to all potential users, but the access will be continuously assessable, in the form not only of continuously updated impact estimates based on the classical measure of impact, which is citations, but usage will also be measured at earlier stages than citation, namely downloads ("hits," Figure 2)[1] of both peer-reviewed “postprints” and pre-refereeing “preprints”. Many powerful new online measures of research productivity and impact will also develop around this rich UK research performance database (Figure 3,4), increasing the sensitivity and predictiveness of the RAE analyses more and more. (See the online impact-measuring scientometric search engines we have developed at Southampton: http://citebase.eprints.org/cgi-bin/search and http://opcit.eprints.org )

And all that is needed for this is for RAE to move to online submissions, mandating online CVs linked to the full-text draft of each peer-reviewed publication in the researcher's institutional Eprint Archive. http://www.eprints.org/self-faq/#research-funders-do Reference-link-based impact-assessment engines like citebase and Web of Science http://wos.mimas.ac.uk/ can then be used by RAE to derive ever richer and more accurate measures of research productivity and impact (Figure 3), available to the RAE continuously. Universities could continuously monitor and improve their own research productivity and impact, using those same measures. And the rest of the world could see and emulate the system, and its measurable effects on research visibility, uptake and impact.

Just a few loose ends: Books are usually not give-aways, as peer-reviewed research is, so full-text self-archiving is probably not viable for book output (apart from esoteric monographs that produce virtually no royalty revenue). But even if the book's full-text itself cannot be made accessible online, its metadata and references can be! Then the citation of books by the online peer-reviewed publications becomes a measurable and usable estimate of their impact too! For disciplines whose research and productivity does not consist of text but of other forms of digital output, both online usage counts and citations by text publications can still be used to estimate impact; and there are always the further kinds of performance indicators in the standardized RAE-CV that can be used to design discipline-specific metrics.

The UK is uniquely placed to move ahead with this and lead the world, because the RAE is already in place. The Netherlands has no formal RAE yet, but it is about to implement a national system of open research archiving for all of its universities called DARE: http://www.ecs.soton.ac.uk/~harnad/Hypermail/Amsci/2356.html It is just a matter of time before they too realize that a marriage between a national network of DARE-style institutional Eprint Archives and CVs plus a national RAE-style research assessment exercise make a natural, perhaps even an optimal combination.

But although the naturalness and optimality -- indeed the inevitability -- of all this is quite transparent, it is a fact that research culture is slow to change of its own accord, even in what is in its own best interests. That, however, is precisely why we have funding councils and research assessment: To make sure that researchers do what is best for themselves, and best for research, and hence also best for the supporters (and beneficiaries) of research, namely, tax-paying society: The institutional self-archiving of research output, for the sake of maximizing research access and impact, has been much too slow in coming, even though it has already been within reach for several years. The UK and the RAE are now in a position to lead the world research community to the optimal and the inevitable. http://www.eprints.org/self-faq/#research-funders-do

We at Southampton and Loughborough, meanwhile, keep trying to do our bit to hasten the optimal/inevitable for research and researchers. At Loughborough we are clearing the way for universal self-archiving of university research output by sorting out the copyright issues (and non-issues http://www.lboro.ac.uk/departments/ls/disresearch/romeo/index.html). At Southampton we are planning to harvest all the metadata form the submissions to RAE 2001 http://www.hero.ac.uk/rae/submissions/ into RAEprints, a "meta-archive" that is intended to demonstrate what RAE returns would look like if this RAE upgrade proposal were adopted. Of course (i) RAEprints will contain only four papers per researcher, rather than their full peer-reviewed research output, (ii) it will only contain the metadata for those papers (author, title, journal-name), not the full-text and the all-important references cited. But we will also try to enhance the demo by adding as much of this missing data as we can – both from Journal Citation Reports http://wos.mimas.ac.uk/jcrweb/ and from the Web itself, to at least give a taste of the possibilities: Using paracite http://paracite.eprints.org/ an on-line citation-seeker that goes out and tries to find peer-reviewed full-text papers on the web, we will "stock" RAEprints with as much as we can find -- and then we will invite all the RAE 2001 researcher/authors to add their full-texts to RAEprints too!

We hope that the UK Funding Councils will put their full weight behind our recommended approach (Figure 9) when they publish their long-awaited review of the RAE process http://www.ra-review.ac.uk/

AppleMark

Figure 7: The Limited Impact Provided by Toll-Based Access Alone

AppleMark

Figure 8: Maximizing Research Impact Through Self-Archiving of University Research Output

AppleMark

Figure 9: What Needs to be Done to Fill the Eprint Archives

 

References

Berners-Lee, Tim. & Hendler, Jim. (2001). Scientific publishing on the "semantic web." Nature, 410, 1023-1024

Garfield, Eugene (1979). Citation indexing: Its theory and applications in science, technology and the humanities. New York. Wiley lnterscience

Harnad, S. (2001) The Self-Archiving Initiative. Nature 410: 1024-1025 http://www.nature.com/nature/debates/e-access/Articles/harnad.html

Harnad, Stevan, & Carr, Les (2000). Integrating, navigating, and analysing open eprint archives through open citation linking (the OpCit project). Current Science. 79(5). 629-638

Harnad, Stevan. (2001) Research access, impact and assessment. Times Higher Education Supplement 1487: p. 16. http://www.ecs.soton.ac.uk/~harnad/Tp/thes1.html

Harnad, S. (2003) Electronic Preprints and Postprints. Encyclopedia of Library and Information Science Marcel Dekker, Inc.
http://www.ecs.soton.ac.uk/~harnad/Temp/eprints.htm

Harnad, S. (2003) Online Archives for Peer-Reviewed Journal Publications. International Encyclopedia of Library and Information Science. John Feather & Paul Sturges (eds). Routledge. http://www.ecs.soton.ac.uk/~harnad/Temp/archives.htm

Hodges, S., Hodges, B., Meadows, A.J., Beaulieu, M. and Law, D. (1996) The use of an algorithmic approach for the assessment of research quality, Scientometrics, 35, 3-13.

Holmes, Alison & Oppenheim, Charles (2001) Use of citation analysis to predict the outcome of the 2001 Research Assessment Exercise for Unit of Assessment (UoA) 61: Library and Information Management. http://www.shef.ac.uk/~is/publications/infres/paper103.html

Jaffe, Sam (2002) Citing UK Science Quality: The next Research Assessment Exercise will probably include citation analysis. The Scientist 16 (22), 54, Nov. 11, 2002. http://www.the-scientist.com/yr2002/nov/prof1_021111.html

Ingwersen, P., Larsen, B. and Wormell, I. (2000) Applying diachronic citation analysis to ongoing research program evaluations. In: The Web of Knowledge : a Festschrift in Honor of Eugene Garfield / Cronin, B. & Atkins, H. B. (eds.). Medford, NJ: Information Today Inc. & The American Society for Information Science.

Oppenheim,  Charles (1995) The correlation between citation counts and the 1992 Research Assessment Exercises ratings for British library and information science departments, Journal of Documentation, 51:18-27.

Oppenheim,  Charles (1996) Do citations count? Citation indexing and the research assessment exercise, Serials, 9:155-61, 1996.

Oppenheim, Charles (1998) The correlation between citation counts and the 1992 research assessment exercise ratings for British research in genetics, anatomy and archaeology, Journal of Documentation, 53:477-87. http://dois.mimas.ac.uk/DoIS/data/Articles/julkokltny:1998:v:54:i:5:p:477-487.html

Seglen, P. O. (1992). The skewness of science. Journal of the American Society for Information Science, 43, 628-638 http://www3.interscience.wiley.com/cgi-bin/abstract/10049716/START

Smith,  Andrew, & Eysenck,  Michael (2002) “The correlation between RAE ratings and citation counts in psychology,” June 2002 http://psyserver.pc.rhbnc.ac.uk/citations.pdf.

Warner, Julian (2000) Research Assessment and Citation Analysis. The Scientist 14(21), 39, Oct. 30, 2000. http://www.the-scientist.com/yr2000/oct/opin_001030.html

Thelwall,  Mike (2001) Extracting macroscopic information from Web links. Journal of the American Society for Information Science 52(13)  (November 2001) http://portal.acm.org/citation.cfm?id=506355&coll=portal&dl=ACM

Zhu, J., Meadows, A.J. & Mason, G. (1991) Citations and departmental research ratings. Scientometrics, 21, 171-179

 

 

 

 

 

 

 

 

 



[1] Both usage statistics and citation statistics are open to potential abuse. See http://citebase.eprints.org/help/ - impactwarning and http://www.ecs.soton.ac.uk/~harnad/Hypermail/Amsci/2642.html The data nevertheless still have considerable signal value, its effect-size can be estimated, and ways of detecting and correcting for abuses will evolve as these new measures become more important.