435
Views
12
CrossRef citations to date
0
Altmetric
Research Article

Is external research assessment associated with convergence or divergence of research quality across universities and disciplines? Evidence from the PBRF process in New Zealand

, &
Pages 3919-3932 | Published online: 14 Feb 2020
 

ABSTRACT

Performance-based research evaluations have been adopted in several countries both to measure research quality in higher education institutions and as a basis for the allocation of funding across institutions. Much attention has been given to evaluating whether such schemes have increased the quality and quantity of research. This paper examines whether the introduction of the New Zealand Performance-Based Research Fund process produced convergence or divergence in measured research quality across universities and disciplines between the 2003 and 2012 assessments. Two convergence measures are obtained. One, referred to as β-convergence, relates to the relationship between changes in average quality and the initial quality level. The second concept, referred to as σ-convergence, relates to the changes in the dispersion in average research quality over time. Average quality scores by discipline and university were obtained from individual researcher data, revealing substantial β- and σ-convergence in research quality over the period. The hypothesis of uniform rates of convergence across almost all universities and disciplines is supported. The results provide insights into the incentives created by performance-based funding schemes.

JEL CLASSIFICATION:

Acknowledgments

We are grateful to the New Zealand Tertiary Education Commission for providing the data used here. An earlier version of this paper was presented at the New Zealand Association of Economists Conference in July 2019. We have benefited from the suggestions of an anonymous referee.

Disclosure statement

No potential conflict of interest was reported by the authors.

Supplementary material

Supplemental data for this article can be accessed here.

Notes

1 Hicks (Citation2012) defines the main characteristics of these systems as: (i) research is assessed, (ii) research evaluation is ex post, (iii) research output and/or impact is evaluated, (iv) part of government allocation of university research funding depends on the outcome of the evaluation, and (v) the scheme is a national or regional system. They vary by coverage and assessment methods, which may be based on bibliometric data or peer review.

2 On the aims of the NZ PBRF see NZ Tertiary Education Commission (Citation2002), Ministry of Education (Citation2012), and Smart and Engler (Citation2013).

3 The dollar amounts received were, however, weighted according to whether an academic discipline fell within one of three categories, as discussed in Section II.

4 As in other country studies, caution must be exercised in attributing causal impacts of such research assessment schemes. In the NZ case, there are no comparable data on pre-PBRF research quality. Further, since the scheme was introduced across all universities simultaneously, it is not possible to pursue a ‘treatment versus control’ approach.

5 The assessment and scoring method used in the New Zealand PBRF system from 2003 to 2012 are described in more detail and critically evaluated in Buckle and Creedy (Citation2019b).

6 The recognition that new researchers may take time to establish their research, publications and academic reputations led to the introduction in 2006 of the new categories, C(NE) and R(NE), which applied to new and emerging researchers who did not have a full six-year portfolio. The following analysis does not distinguish the NE categories, since neither numerical scores nor funding was affected.

7 Since the grade for R-rated staff is zero, their number affects only the denominator in (2).:

8 This dataset is not publicly available. The TEC publishes summary results of each assessment round; see, for example, New Zealand Tertiary Education Commission (Citation2013).

9 The number of staff who did not submit a portfolio (NP) do not enter into the calculation of AQSs derived in this paper because NPs cannot be identified by discipline. Including NP-staff in the Buckle and Creedy (Citation2018, Citation2019a) confirm conclusions regarding the substantial increase in AQSs across all universities.

10 Only QCs, not component scores, are available. Also, the dataset does not include information on part-time status; hence eh = 1 was adopted in calculating AQSs. Values do not differ substantially from those reported by TEC.

11 The amounts received also depended on the discipline to which a researcher belongs, with three separate categories given financial weightings of 1, 2 and 2.5; see Appendix A. Roa et al. (Citation2009) estimate that, following the 2006 PBRF, universities received $34,166 per year per A-researcher (discipline weighting = 1) and $6,832 per year for each C-researcher in the same discipline. These weightings remained the same throughout the period.

12 Indirect financial effects on universities could result from research assessment schemes such as the PBRF. For example, Biancardi and Bratti (Citation2019) find that universities that performed better in the Italian Research Evaluation Exercise benefited from more students with higher entry qualifications, with effects stronger for top performing universities.

13 Writing z=q/μ, if only random proportional changes in the ratio, z, occur, then (dz/z)dt=u which in discrete time converts to logztlogzt1=ut. To allow for relative size to have a systematic contribution, add a term, βlogzt1 to the right-hand side. Hence, when qt1/μt1<1, so that q is initially below the geometric mean, the logarithm of the relative value, z, is negative. Hence if β<1, the proportional change is positive and larger than when q is initially above the geometric mean.

14 For example, if β = – 0.7 the stable variance, σq2, is about 10% higher than σu2, but if β = – 0.5 the difference rises to 33%, and when β = – 0.2 the difference becomes 178% higher (and 426% higher when β = – 0.1).

15 As shown in , Lincoln University has no observations for Education and Law, so that the total number of universities and disciplines shown in is 70 rather than 72.

16 Buckle and Creedy (Citation2019a) show that AQS improvements involved all universities reducing the proportion of the lowest QC researchers (R-rated) and increasing A- and B-rated staff. Only VUW reduced, rather than increased, the share of C-rated researchers.

17 Random shocks in this case might include, for example, the 2010–11 Canterbury earthquakes which may have affected the ability of Canterbury and Lincoln to recruit staff after 2011. In NZ’s small university system this may have indirectly affected recruitment practices, including other universities recruiting higher-quality staff from CU and LU.

18 Regressions in use 87 observations: 70 for the 8 universities and 9 disciplines within universities, less two missing observations for Law and Education at Lincoln, plus 17 AQSij values, averaged across all universities and all disciplines. This enables parameters for each university’s growth and convergence to be compared directly with the average across all universities or disciplines rather than adopting one university and discipline as the omitted variable. However, adjusted-R2s must be interpreted cautiously since they are somewhat inflated by the inclusion of individual observations and their cross-university or cross-discipline averages. For example, running the regression specification in on the 70 AQSij observations yields a slightly lower adjusted-R2 = 0.940.

19 As mentioned earlier, both CU and LU experienced severe infrastructure damage from the 2010–11 Canterbury earthquakes. However, there are no significant shift dummy effects identified for CU, suggesting, it is hard to argue that an earthquake effect fully explains the LU shift.

20 These data are reported in Appendix A, Tables A1 and A2.

21 Results obtained for non-log versions of these included variables, and when mean, rather than median, age was used were uniformly inferior to those reported in , with no variables even close to statistical significance at the 10 per cent level.

22 Testing whether the shift and slope dummy variables included in become redundant when new variables are added to regressions confirms that this is not the case. A nested model always supports the inclusion of all dummy variables over the age or scale variables.

23 From , the total convergence effect is given by: d(AQS growth)/d(logAQS2003) = –0.796 + 0.016*logN2003 = –0.796 + 0.074 = –0.722, using mean logN2003 = 4.638.

Log in via your institution

Log in to Taylor & Francis Online

PDF download + Online access

  • 48 hours access to article PDF & online version
  • Article PDF can be downloaded
  • Article PDF can be printed
USD 53.00 Add to cart

Issue Purchase

  • 30 days online access to complete issue
  • Article PDFs can be downloaded
  • Article PDFs can be printed
USD 387.00 Add to cart

* Local tax will be added as applicable

Related Research

People also read lists articles that other readers of this article have read.

Recommended articles lists articles that we recommend and is powered by our AI driven recommendation engine.

Cited by lists all citing articles based on Crossref citations.
Articles with the Crossref icon will open in a new tab.