2,404
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Knowledge capital theory: a critical analysis using Lakatos’ idea of research programmes

ORCID Icon

ABSTRACT

This paper introduces and uses Lakatos’ idea of research programmes to summarise and critically evaluate academic discourse towards knowledge capital theory. The analysis uses rational reconstruction to formulate the components of the hardcore and protective belt of knowledge capital theory. By critically surveying the literature, it challenges the validity of the claims made by Erik Hanushek, Ludger Woessmann and the OECD that there is a causal link between cognitive development and economic growth. It concludes by stating that knowledge capital theory in its current form is degenerate and should be considered a high-risk research programme and that more sophisticated theories are required to be developed to explain current phenomena.

Introduction

The OECD and ILSAs

Since the early 2000s there has been a meteoric rise to prominence in academic discourse, international education policy discourse and traditional media of the Organisation for Economic Cooperation and Development’s (OECD) international large scale assessments (ILSA), primarily the Programme for International Student Assessment (PISA). Increasingly their other testing products are also gaining more widespread attention such as: the International Early Learning and Child Well-Being Study (IELS), Assessment of Learning Outcomes in Higher Education (AHELO) and the Programme for International Assessment of Adult Competencies (PIAAC) (Moss et al. Citation2016; Shahjahan and Morgan Citation2016). In 2010, the World Bank’s survey Skills Towards Employment and Productivity (STEP) was launched with a goal of being closely ‘linked to PIAAC’ (Educational Testing Services (ETS) Citation2020), which helps feed into the World Bank’s new Human Capital Index. While not being an ILSA, the Human Capital Index uses the data generated by the OECD’s ILSAs (World Bank Citation2018). These new measures have undoubtedly superseded other ILSAs which focused mainly on ascertaining the effective delivery of intended curricula, such as The Progress in International Reading Literacy Study (PIRLS), The Trends in International Mathematics and Science Study (TIMMS) and Literacy Assessment and Monitoring Programme (LAMP) for global recognition and policy influence.

The introduction of PISA for Schools, which allows for school-level measurements and gives teachers and leaders access to data to use OECD ‘best practice’ to inform their pedagogy and curriculum is problematic (Lewis Citation2017). Lewis (Citation2017) argues that the OECD’s influence extends much deeper now than at the national policy level. This change is due to PISA for Schools, and there is a risk teachers will apply the same narrowly normative changes as their national policymakers have done (Carvalho and Costa Citation2015; Breakspear Citation2012).

The question of cognitive development and GDP Growth

The success of this new global education regime, led by the OECD and supported by the World Bank, derives from claims made by Erik Hanushek and Ludger Woessmann (hereafter H&W). H&W claim that gains in cognitive development of a population are closely correlated to future GDP per capita growth. H&W also claim that the strength of the evidence is so strong that the relationship between the two can now be considered causal (OECD Citation2015). The relationship has since been utilised as a means for predicting GDP gains and centring policy recommendations to target students’ cognitive development as measured by ILSA test scores.

H&W outlined their claims in three important publications (Hanushek and Woessmann Citation2008, Citation2015; OECD Citation2015). The first work in 2007 published by the World Bank presented their research that claimed a correlation between tests scores from 1964–2003 and GDP per capita growth for the period 1960–2000 across fifty major economies. Their second paper built upon and expanded upon the first. The third, which was published by the OECD, did not contribute any further data, but notably considered the supposed causation as if it were now merely an empirical fact.

The attractiveness of their model’s predictions are obvious to national and economic policymakers. At heart, those at the OECD are econometricians (Aloisi and Tymms Citation2017). The causal relationships advocated by H&W, therefore fall nicely into this narrative.

Improving scores would not only justify the ‘aggressive’ education reform, but that in the medium to longer-term the reforms will pay for themselves and much else besides (Komatsu and Rappleye Citation2017). These claims were seized upon and incorporated by the OECD to justify international assessment benchmarking and education reform. The assertions on causality between test scores and per capita GDP growth formed the basis of the third paper published by the OECD and authored by H&W. There are many positive political and economic dividends aggressive reforms would provide. It is not difficult to understand why governments, often democratically elected, have used these ideas as much of their basis to provide a narrative of ongoing school improvement leading to wider increases in competitiveness and strength of their economies.

The critical difficulty before H&W was that human capital theory is much more complicated than knowledge capital theory and considers many other factors as playing a significant role in supporting economic gains. Human capital theorists continued to be unsure if increasing education quality or expanding access contributed most to economic betterment (Komatsu and Rappleye Citation2017). The arguments over quality vs access were intractable for most of the period from the 1970s-1990s. Previously policymakers had to operate with a complex array of priorities. Policymakers needed to consider countries’ stage of development, changing contexts, means of measurement, feasibility etc. The vital distinction advocated by knowledge capital theorist such as H&W is: the quality of education is what matters most (OECD Citation2015). Once the position that what mattered the most was citizens’ cognitive development, then, arguments about the developmental stage and other priorities begin to recede (Komatsu and Rappleye Citation2017). PISA and other ILSAs provide the longitudinal and cross-national data needed to demonstrate many of the claims made by knowledge capital theory. The empirical work and the longitudinal sample of over 50 countries have become the foundation of a new logic for forming education policy (Rutkowski and Rutkowski Citation2016). What works for high-income countries is plausibly considered to work for low-income countries, given the right implementation. The discussions over priorities of what should be changed has given way to technical discussions regarding how to measure test scores and how to improve them.

Previously the concept of ‘universal basic skills’ had multiple interpretations. The OECD’s definition has begun to be considered the de facto definition. The OECD defines universal basic skills as those ‘basic skills’ measured by Level 1 PISA tasks (OECD Citation2015). Since a straightforward testing instrument had become available, it was coupled with H&W’s strong statistical relationship. Consequently, some quite extraordinary extrapolations have been made. For instance, by 2030, Ghana would increase current GDP by 944% (Schleicher and Qian Citation2015, 10). They go on to extrapolate that all lower-middle-income countries could expect a gain of 627% (Schleicher and Qian Citation2015, 55–56). The change required, render these numbers somewhat meaningless, but it serves as a foretaste of what could be achieved. Given the potentially astronomical GDP gains by improving Level 1 PISA tasks, the OECD’s necessary policy reform becomes a magic bullet for national policymakers.

There are concerns related to the legitimacy of: the mandate of the OECD (d’Agnese, Vasco Citation2015), methodological robustness of the testing instruments used in PISA (Zhao Citation2020; Goldstein Citation2018) and the rationale of the purpose of education underpinning the entire enterprise (Komatsu and Rappleye Citation2019).

In the next section, Lakatos’ (Citation1978) notion of research programmes will frame current critical academic discourse. It will recommend that the knowledge capital research programme on which the OECD’s and World Bank’s ISLA’s are based appears to be in a degenerate state.

Lakatos’ notion of research programmes

The knowledge capital theory research programme

The assertion by H&W that increases in per capita GDP is causally related to improved cognitive development is challenged on a fundamental level by the Duhem-Quine thesis (DQ). DQ asserts that, ‘it is not possible to falsify single hypotheses because it is invariably conjunctions of hypotheses which are being tested.’ (Cross Citation1982, 320). Therefore, if a particular hypothesis is challenged with empirical evidence, the most we can say is that ‘… the conjunction of the particular hypotheses with a set of auxiliary hypotheses is false’ (p320). We cannot be sure that the anomalous empirical evidence is not explained by one or more of the auxiliary hypotheses. The foremost issue is that the participants in the debates about knowledge capital theory have primarily ignored the jointness of hypothesis testing. The central focus of knowledge capital theory has been mostly in single target hypotheses, such as, ‘better qualified teachers improve student outcomes’, which are considered in isolation from supporting auxiliary hypotheses. The assumption in this line of reasoning is that it is possible to falsify a single target hypothesis, such as falsification theory (Popper Citation1963), which DQ contends is wrong. We must therefore use supportive auxiliary hypotheses in conjunction to appraising any target hypothesis. The only truth that falsification establishes is the denial of the hypothesis. A denial does not necessarily imply any single ‘opposite’ hypothesis, but instead, entails multiple alternative conjunctions of hypotheses; some may, in fact, contain the target hypothesis.

The starting point for using Lakatos (Citation1978) is to view the networks of hypotheses being tested as whole structures. The most important subgroup of hypotheses is called the hardcore. The collection of hardcore hypotheses defines a research programme. By agreement of researchers the hardcore is protected from being challenged empirically or from descriptive forms of arguments by a negative heuristic. The negative heuristic is how problems are not solved. The remainder of the hypotheses constituting a research programme are called the protective belt, which, as the name suggests, protects the hardcore from challenge. If hardcore assumptions are changed, it no longer describes the same research programme.

Underlying the research programme is a dynamic force named the positive heuristic, or, how problems are solved. The positive heuristic determines how the protective belt is adapted to protect the hardcore from empirical or logical challenges. Therefore, the research programme evolves rationally over time in response to the challenges made to the protective belt. A key tenant of Lakatos is that progressive research programmes can predict novel and more precise data or facts. Conversely, a degenerate programme shows a lack of growth; the belt of auxiliary hypothesis, supported by the positive heuristics do not lead to valid or novel predictions.

In contrast to Kuhn and Hawkins (Citation1963) advocated that science followed long periods of ‘normal science’ to ‘paradigm change’. Earlier, Kuhn rejected the idea the paradigm change occurs without scientific reasons. Later, however, Kuhn argued for a more relativist interpretation of scientific change, and that scientist did not use rules to arrive at their decisions to choose the most successful paradigm. This view faced criticism, since, it seemed to suggest science was, in reality, irrational. Lakatos developed his notion of a progressive research programme to reconcile the rational positivist process of falsification of Popper and the more relativist Kuhn (Blaug Citation1975; Maxwell Citation2005; O’Donohue, Ferguson, and Naugle Citation2003). Importantly, when ‘paradigm shifts’ do occur, Lakatos argued it is as a result from a research programme transitioning from a degenerate state to a more progressive state. Consequently, the paradigm shift in Lakatosian research programmes are mostly rational, not irrational, as Kuhn suggests. It is for these reasons that the Lakatosian research programme is used to evaluate knowledge capital theory.

Lakatos used the term rational reconstruction. It refers to what the researchers in a research programme have been committed to historically and logically. A rational reconstruction of knowledge capital theory follows:

The hardcore (HC)

  1. Increasing knowledge capital (as measured by student test scores) leads to increased GDP per capita growth.

  2. Knowledge capital can be measured using testing instruments.

  3. Purpose of education is for economic gain on the level of the agent (individual person) in an economic system.

  4. Educational reform policies successfully increase knowledge capital.

Positive heuristics

  1. Go forth and construct theories in how to increase knowledge capital.

  2. Go forth and develop predictions based on the hardcore assumptions.

  3. Develop comparative large scale testing instruments to measure knowledge capital growth.

  4. Devise recommendations for national education policymakers.

Negative heuristics

  1. Do not build theories in which the primary purpose of education is not for economic gain.

  2. Do not construct theories in which schools do not affect student outcomes.

  3. Do not construct theories in which the measuring of knowledge capital is not possible or is invalid.

  4. Do not test hardcore propositions.

A critical summary of academic discourse. The hardcore (HC) of knowledge capital theory

Following a rational reconstruction of the knowledge capital research programme, the next section uses a ‘Point: Counterpoint’ style to survey the current academic debate concerning each of the hardcore’s four parts.

HC1: increasing knowledge capital (measured by student test scores) leads to increased GDP per capita growth

Temporal mismatch

Ramirez et al. (Citation2006) and Kamens (Citation2015) showed that the original analysis by H&W compared test scores and economic growth from the same period. They argued forcefully that the temporal comparisons by H&W are not appropriate. The period for the study of economic growth should follow that of the testing since it takes time for these students to enter the workforce. Both also showed by using similar data; the strong statistical link was no longer evident for test scores to impact future economic growth.

In response, H&W re-analysed their data and compared the testing period of 1964–1984 and the economic growth between 1985–2007 (OECD Citation2015). In contrast to Ramirez et al. (Citation2006) and Kamens (Citation2015), they were able to re-confirm and strengthen their argument about their original claims’ validity. The reasons why the three sets of researchers came up with differing results is a little unclear. It appears to revolve around different types of data and analysis used (Komatsu and Rappleye Citation2017). Kamens (Citation2015) used test score data for a specific point in time, while H&W (OECD Citation2015) used long-term average test scores. H&W (OECD Citation2015) processed their data by interpolating and extrapolating the missing data, Ramirez et al. (Citation2006) did not use a similar technique. Lastly, Sahlgren (Citation2014) simply puts it down to Ramirez et al. (Citation2006) not having access to the most up to date data and that they analysed too few countries for the trend to be apparent.

Extrapolating historical trends and making future projections

Klees (Citation2016) and Stromquist (Citation2016) have many criticisms of H&W’s work, but we will concentrate on those raising doubt regarding the internal validity of H&W’s claims for our purposes. Both authors take issue that H&W used the link between historical test scores and economic growth and subsequently extrapolated it for the future. Both Klees (Citation2016) and Stromquist (Citation2016) argued that it is not appropriate, as many factors affect economic growth, and H&W did not control for them. Klees (Citation2016) cautions that policymakers should not take the projections given by H&W seriously. While a formal retort has still not been forthcoming from H&W, Komatsu and Rappleye (Citation2017) suggest that H&W have in a way already pre-empted the criticism. They note that H&W argued that additional factors appeared to be inconsequential (OECD Citation2015, 91) and that they had already considered the relationship being a predictive one and concluded that it was (OECD Citation2015, 90). A formal reply by H&W has yet to appear in print, but this will likely be their response.

Validity of statistical claims made by H&W

Much deeper methodological concerns have been raised regarding the inappropriate criticism of the parameters used by Hanushek and Woessmann (Citation2015), (Klees Citation2016; Komatsu and Rappleye Citation2017, Citation2019; Stromquist Citation2016). A methodological point has long been made of being cautious of the explanatory power given by the statistical significance (p) and the slope of the regression line for very large samples of data (Nuzzo Citation2014). What matters most is the explanatory power of test scores (Komatsu and Rappleye Citation2017). It is how much the variation seen in economic growth is explained by test scores which will produce a more robust analysis. The more robust analysis is given not by statistical significance (p), but by the determination coefficient (R2). Komatsu and Rappleye (Citation2017) reviewed the same historical data and expanded on it to include the most recent PISA and PIAAC data (Rappleye and Komatsu Citation2019). They showed that when considering R2, the apparent relationship between test scores and economic growth made by H&W becomes invalid. The work of Klees (Citation2016), Ramirez et al. (Citation2006) and Stromquist (Citation2016) barely mention R2, if at all, and so Komatsu and Rappleye (Citation2017) contend the criticism is mostly misdirected. The focus on statistical significance allows H&W (OECD Citation2015) to quickly counter criticisms, since, given the large sample sizes, H&W are easily able to provide statistically significant results. A rigorous response by H&W regarding R2 analysis by Komatsu and Rappleye (Citation2017) would be beneficial to move the argument forward.

HC2: knowledge capital can be measured using testing instruments

Potential assessment bias

PISA assessments have faced criticism for potentially being biased and that this bias has impacted the performance of different groups of students (Zhao Citation2020). There are multiple areas which have come under scrutiny for potentially being biased; question formats, language, culture, constructs and text types (Hopfenbeck et al. Citation2018; Solheim and Kjersti Citation2018; Zhao Citation2020). PISA tests are 18% longer in German than in English; however, the assessment length stays the same (Shiel and Eivers Citation2009). There are linguistic divergences between Greek textbooks and PISA items, and as a result, this could help explain poor Greek performance in science (Hatzinikita, Dimopoulos, and Christidou Citation2008). Reading assessments are biased against boys in Nordic countries because the test items are ‘girl-friendly’ due to the type of writing required to document their reading comprehension (Solheim and Kjersti Citation2018).

Additionally, PISA is less comparable across Asian or Middle Eastern countries due to cultural and linguistic reasons (Grisay, Gonzalez, and Monseur Citation2009). In response, there have been some adjustments of late of the methodologies used to counter some of the above criticism. It is to this response by PISA we address next.

Item response model

There are concerns regarding the item response model used by PISA. The first was the lack of item-by-country interactions; without these, important and interesting differences could be ‘smoothed out’ which could jeopardise cross-national comparability (Goldstein Citation2018). Simply put, some testing items are treated as harder, for example, in Scotland than in England, which on a conceptual level seems difficult to justify, but not so given the biases discussed above. In 2015 PISA introduced item-by-country interactions. In a recent analysis of the 2015 PISA data set; there was no evidence to suggest that the inclusion of the item-by-country interactions provided a particular advantage or disadvantage for any of those countries in the sample (Jerrim et al. Citation2018).

Until 2015 the method for treating items not reached was to consider them as being incorrect. A longstanding concern was that those students from disadvantaged backgrounds were the most likely not to finish within the time limit; disproportionately affecting low achieving students (Bridgeman et al. Citation2004). Post 2015 PISA items not reached no longer contribute to proficiency scores. Jerrim et al. (Citation2018) found once again that this change had almost no significant impact on the results.

Rasch model

A recurring criticism of PISA is that it uses the Rasch model, a well-known psychometric model named after the mathematician and statistician George Rasch (1901–1980). Kreiner and Christensen (Citation2014) stated that PISA did not meet the Rasch model’s necessary criteria to be used appropriately. They argued that as a result of using an inappropriate model for the data, the resulting country rankings were not robust. Other researchers analysing the same data assert that the PISA data is more appropriately interpreted using a multidimensional model with two factors (Hopfenbeck et al. Citation2018). Jerrim et al. (Citation2018) re-analysed the data and found no significant difference when using the Rasch model or supposedly more robust multilevel model with two factors. Jerrim et al., state that the vocal criticism of the methods used in analysing PISA data appear to be ‘overblown’ and the models used previously were ‘good enough’ (Citation2018, 37). This contribution is a significant blow to those critics. Surprisingly, it has yet to feature prominently within recent reviews of the literature. For instance, the absence of any reference to Jerrim et al. (Citation2018) in the recent work by Zhao (Citation2020) who is a vocal critic, is puzzling. However, Jerrim et al. (Citation2018) do caution that the robustness of the analysis may not be replicated when a more sophisticated three factor multi-level analysis is applied.

Sampling problems

Appropriately sampling populations so that they are genuinely representative is an area that has received close attention (Zhao Citation2020). Students with specific characteristics are excluded from the assessments as long as the total exclusion rate is below 5% (Hopfenbeck et al. Citation2018). However, during the 2012 cycle, for example, eight educations systems had an actual exclusion rate of above 5%, making robust comparisons more difficult (Rutkowski and Rutkowski Citation2016). The types of pupils who are excluded matters too.

Students with disabilities are omitted. Disabled pupils are, in reality, a vital indicator of the equity and quality an education system exhibits. Since students with disabilities and special educational needs can be excluded; there is little incentive for their inclusion in improvement initiatives, since they won’t contribute to rising PISA scores. This practice further marginalises those pupils with disabilities and special needs (Schuelka Citation2013).

Those students already marginalised for economic and societal reasons, such as the children of migrant workers in Shanghai who are excluded from the public school system (Chen and Feng Citation2013). These migrant students are not included in the sample, leading to considerable criticism of Shanghai’s ‘real’ performance (Loveless Citation2014). Those 15-year-old students not enrolled in schools are also not included because PISA is administered in schools. PISA aims to capture a measurement of the human capital development of those 15-year-olds; this means that those 15-year-olds who will also contribute to the workforce but not enrolled in school are not included.

Lastly, to students included, PISA does not differentiate between the starting ages of students in their respective school systems. Therefore, in some countries, students have had more exposure to their curriculum (Zhao Citation2020). Further, it is a test of 15-year-olds and so includes students at different cohort grade levels. There are significant concerns that these shifts of the proportions of 15-year-olds in different grade levels are a crucial factor in explaining between-cycle-variations (Aloisi and Tymms Citation2017). The OECD notes that in response to such concerns regarding exposure to the curriculum that unlike TIMMS or PEARLS, PISA does not assess curriculum focused skills and knowledge, however, exposure time to the curriculum would surely matter in PISA results (Hopfenbeck et al. Citation2018).

HC3: purpose of education is for economic gain on the level of the agent (individual person) in an economic system

Before considering knowledge capital, it is instructive to consider the role of education in the labour market from the perspectives of the different theories available, the summary of what follows uses the recent work of Shields and Sandoval Hernandez (Shields and Hernandez Citation2020) as a jumping-off point.

Human capital theory

Human capital theory has more or less become the de facto model for conceptualising the labour market and education (Shields and Hernandez Citation2020). The interplay between the labour market and knowledge was developed as a framework for understanding the variation in education and earnings of individuals (Mincer Citation1958). In its original form, the wages of a worker were seen as a return on investment (Becker Citation1994). Modern formulations of human capital theory consider the role of highly skilled labour and analyse the skills premium it can demand (Shields and Hernandez Citation2020). Knowledge capital theory at heart considers the cognitive and skills development of the worker as one of the factors in determining wage levels.

Other theories have emerged in addition to the classical theories of human capital (Shields and Hernandez Citation2020). Signalling refers to the idea of educational qualifications being used by employees to demonstrate their abilities using a trusted third party verification to potential employers (Spence Citation1973). The apparent status of the third party, such as an admission to a highly ranked university alone, can work to signal comparative advantage (Hoekstra Citation2009). The flipside to signalling is screening (Stiglitz Citation1975), whereby an employer sets ‘wage contracts’ for consideration for a particular salary, a specific educational qualification is required (Stiglitz and Weiss Citation1990; Stiglitz Citation1975). Both theories suggest that unlike human capital theory, part of the role of qualifications is to reduce uncertainty in the job market (Shields and Hernandez Citation2020). Those with the qualifications from a reputable institution are also likely to have higher cognition and skills levels. In all three theories above, education holds a unique role in labour markets; it is seen as both an absolute and a positional good (Matthews and Hirsch Citation1977). Absolute goods can be thought of as the size of a cake, with increasing the size of the pie for all, positional goods can be thought of as increasing the individual’s slice of the pie (Shields and Hernandez Citation2020).

Knowledge capital is an extremely narrow version of human capital theory; only the cognitive ability measured using test scores that matters (Rappleye and Komatsu Citation2019). Shields and Sandoval Hernandez (Shields and Hernandez Citation2020) used both qualifications, income and cognitive levels as measured by PIAAC to produce an international comparison of the roles cognitive skills and qualifications have on income. Interestingly, they found that broadly the level of cognitive skills is comparable with qualifications in their explanatory power for income; their results held for both the individual level and those as an outcome of the school system. However, they also found that there was a differing ratio of signalling to cognitive skills in different countries. The United States, in particular, appears to be a unique case where signalling of qualifications is rewarded disproportionately by the labour market over the actual cognitive skills workers possess.

The relationship between skills, signalling and so forth in national economies is now even more complicated due to globalisation. Previously, workers may have been able to accrue a ‘skills premium’ in the labour market. The increased mobility of workers from different countries suggests that this premium is now in question (Lauder, Brown, and Cheung Citation2018).

The forecasted impacts of automation further call into question specialist knowledge in certain high-risk industries (Hawksworth, Berriman, and Goel Citation2018; Kattan, Macdonald, and Harry Citation2018). Supply and demand of skills and qualifications is just one part of the picture. The composition of national economies, labour market institutions, the regulatory environment, historical and geographical contexts all play a role and as such the picture is much more complicated (Shields and Hernandez Citation2020). This is an interesting counterpoint to H&W’s assertion that cognitive development is the primary determinant of future GDP growth.

While there are indications that cognitive development does have an impact on individual earned income, the relationship between these skills and productive GDP development remains unclear. This further calls into question the use of knowledge capital rather than more nuanced theories to explain the role of education and the labour market.

The social compact promoted by the OECD from the viewpoint of the individual citizen of ‘… more learning leads to more earning’ is no longer tenable and has been fiercely opposed and criticised (Brown, Lauder, and Cheung Citation2020; Lauder Citation2015; Brown, Lauder, and Ashton Citation2010). In Bourdieu’s (Citation2013) terms, the OECD uses a performative semiotic construction of globalisation concepts; it implies that only a neo-liberal kind of globalisation is valid and denies other accounts (Sellar and Lingard Citation2014). In an increasingly interdependent globalised economy, the pressure on systems is to be both economically prosperous and providing high-quality education. It is difficult to imagine that a developed economy would turn its back on this education model despite the inconclusive nature of the evidence that is purported to support it.

HC4: educational reform policies borrowed from more successful countries successfully increase knowledge capital

Policy borrowing and the importance of context

Different countries have often adopted policies on the OECD’s advice without a clear understanding or appreciation for local cultural, historical or socioeconomic considerations being taken into account (Salokangas and Kauko Citation2015). Finland is an illustrative example of a country which has leant many policies. Salokangas and Kauko (Citation2015) explain that many of the Finnish comprehensive school model features were as a result of a longstanding political consensus which was agreed by the main political parties in the 1970s (Simola Citation2015). Therefore, education has not featured as a core political issue that has contributed to stability (Kauko et al. Citation2015). One of the central tenants of the Finnish model, is of a decentralised school system. The decentralisation, however, resulted from an economic depression in the 1990s and the need to make it easier for national-level decision-makers to delegate the detail of the budget cuts down to the local level. It was initially not a deliberate initiative to improve education quality (Simola Citation2015).

The political and cultural consensus in Finland has begun to show cracks. School choice initiatives have been trialled, without formal testing scores, comparative league tables, or school inspection reports. As a result, parents are exercising school choice based on families’ socioeconomic background that attend particular schools (Salokangas and Kauko Citation2015). In particular, in urban areas, this has led parents to conceptualise a ‘bad’ school from superficial criteria (Kosunen Citation2014). Many parents cite socially and ethnically mixed student populations as reasons to change school. However, the policy borrowing and lending from countries like Finland continues apace, even when prominent scholars from Finland advise against the de-contextualisation of policy and its subsequent importation into wildly different contexts (Sahlberg Citation2011).

The OECD professes not to encourage such policy borrowing. It states that it is up for the individual national governments to carefully consider which policies are applicable in their context (Schleicher Citation2018). In the same book, Schleicher speaks with pride on the phenomena of ‘PISA Shock’ and its effect on policymakers, see the case of Germany as an illustrative example (Gruber Citation2006; Ringarp Citation2016). In Sweden, PISA data was used selectively for party political purposes to legitimise their reform agenda and policy borrowing arguments. As a result, reforms are narrowed down to practice or teaching-oriented problems (Lundahl and Serder Citation2020). Who should countries choose to borrow from? In their overview of policy discourse Davis, Wilson, and Dalton (Davis, Wilson, and Dalton Citation2020) found that PISA performance was consistently sighted as a reason for policy borrowing from countries such as Finland and more recently East Asian cultures.

One example of a policy reform initiated by analysing top-performing countries was that of teacher recruitment. The policy recommended that schools recruit teachers from high school graduates’ top echelons because that is what high performing systems do (Barber and Mourshed Citation2007). These recommendations, however, resulted from erroneous observations (Zhao Citation2018). Research has shown that the top graduates tend to benefit the top-performing students in a system but at the expense of low performing students (Grönqvist and Vlachos Citation2016). Therefore, the policy initiative may raise PISA scores, but at what cost? (Zhao Citation2017).

The rise in policy borrowing from East Asian cultures has increased recently but is framed in a manner which continues the legacy of Orientalism (You Citation2019). Countries, like the UK, are attempting to replicate some of the success of countries in mathematics on PISA. One system they have looked at closely is that of as Shanghai’s (Chen and Feng Citation2013; d’Agnese, Vasco Citation2015). Policymakers have studied how mathematics is taught in England and China to try and replicate China’s consistently high scores (Miao and Reynolds Citation2017). This cross-national attraction eventually leads to policy borrowing (Davis, Wilson, and Dalton Citation2020). It is ironic that from the Chinese perspective, the strong performance on the PISA rankings has developed into a barrier for reform. The reform in China intends to lower the pre-eminence of the high stakes test-taking in university admissions (Ross and Wang Citation2010). This examination based culture has been a feature of China’s education system for hundreds of years, and reform has centred on increasing creativity and improving student wellbeing (Zhao Citation2014). But are policy reforms worth it? Do they lead to improvement in PISA scores?

Does policy change account for changes in ILSA scores?

An examination of some specific case studies where reforms have been cited for the reason for changes in PISA test scores will be addressed below to try to highlight some of the criticism against the idea that policy reform will lead to improved PISA scores.

In 2001, Qatar based their reforms on OECD recommendations (Guarino and Tanner Citation2012). The reforms included a significant intervention regarding the formation of publically funded independent schools (Zellman et al. Citation2009). After the reforms, Qatar increased its PISA score by up to 75 points, and in TIMMS 2007–2011 the change was over 100 points. Several points should be considered before apportioning the change to the scores to the reforms alone.

The first point is that the achievement gap remained unchanged between public schools’ performance, which included the new independently funded school and private schools. Private schools did better, even though they were not affected by the policy reforms (Cheema Citation2015). Secondly, the adjusted PISA scores showed that the gains between 2009–2012 are because of the student populations’ socioeconomic and demographic changes (OECD Citation2014). Lastly, the main driver of the improvement was the performance of immigrant students in Qatar. Immigrants are a group who routinely outperform local students in Qatar (Cheema Citation2014).

A case where reforms of the curriculum were more plausible was that of Ireland. In PISA 2009, they experienced a significant drop in reading and mathematics scores, and this was ascribed to, like Qatar, as a result of changes to the demography being sampled (Cosgrove and Cartwright Citation2014). Science scores however stayed level. As a result, the Educational Research Centre in Dublin suggested that earlier reforms to the primary and secondary curriculum could have mitigated the demography changes and kept the scores stable (Perkins et al. Citation2011). It is an interesting assertion, given that the policies implemented at the time were focused on reading and mathematics and not science. Even the evidence of implementing those curriculum reforms was best considered moderately successful (Varley, Murphy, and Veale Citation2008). Conversely, Ireland’s mathematics scores did not change when the curriculum was first revised in 2000, but scores did begin to change when examinations were modified to be more PISA-like (Perkins et al. Citation2013). A trend of teaching to the test that incidentally is not relegated to Ireland alone (Breakspear Citation2012).

A case where scores drastically decreased following educational reform is that of Argentina. Between 1994 and 1998, new curricula were introduced, following these reforms, a new framework to help ensure the system was more equitable was introduced in 2004 (UNESCO-IBE Citation2006). From 2000 to 2006, Argentina experienced a 45 point drop in reading scores (Aloisi and Tymms Citation2017). The narrow viewpoint of considering educational reform and student reform as an isolated system inevitably leads to oversimplification. During the same period under question, Argentina underwent a severe economic crisis leading to significant job losses, a sovereign debt default and social unrest. Is it unreasonable to consider that those students who spent a large proportion of their education during this tumultuous period would do worse in 2006 than those students in 2000? It is important to consider the social impact in part as an explanation rather than attributing it to catastrophic reading and literacy reform alone. Realistically, it is difficult for policymakers to justify keeping reforms which have appeared to fail so abjectly in improving performance, despite evidence that the reasons for failure a varied and complex.

The causes of between cycle’s fluctuations in PISA scores are essential to understand. While PISA should be commended for stability in the results, year on year errors might account for 20 points or more, being just a fifth of a standard deviation is an impressive feat for such a large scale assessment (Aloisi and Tymms Citation2017). Unfortunately at the same time, countries go through the PISA shocks and recriminations from falling down the tables. A particular example following a new scaling approach in 2015 may have affected several countries’ trend comparisons (OECD Citation2016). Korea was reported to have ‘lost’; 13 points between 2009 and 2015, wherein fact these are within the margin of error. These potentially non-significant changes are not portrayed as such in the media, and as a result, the pressure for policy change to find solutions increases (Harris and Jones Citation2017).

The effect of policy changes on PISA scores is critical in assessing whether policy borrowing based on strong school system performance on PISA tests is valid. When considering socioeconomic and demographic predictors, there was a correlation of p = 0.99 of a countries performance between two cycles of PISA (Aloisi and Tymms Citation2017). In other words, almost all of the variations seen in the scores could be attributed to changing proportions of grade levels accounting for the 15-year-olds and their socioeconomic status. In the same study, Aloisi and Tymms (Citation2017, 206) showed that the impact of reforms on test scores amounted to annual effect size of around just p = 0.02, suggesting very little correlation. A cautionary note must be raised here, earlier in this paper attention was drawn to the proposed use of R2 and not p to give a more robust analysis, no R2 analysis was included to accompany p values in this instance.

In defence of PISA, they do publish adjusted scores based on demographic and other factors. However, these scores are not published prominently in the league tables for public consumption, but are relegated to an appendix. Further, while Aloisi and Tymms’s (Citation2017) conclusions bring into question much of the evidence used by policymakers to justify policy borrowing and reform, importantly, it is a model that they have recently developed, and it suggests correlation; causation would need further research.

Conclusion

This paper considers the question: is the knowledge capital research programme in a degenerate state? The criticisms of the hardcore of the knowledge capital research programme would seem to suggest that it is. The hardcore is being consistently and effectively criticised, and it has yet to give novel predictions. The case of (Komatsu and Rappleye Citation2017; Rappleye and Komatsu Citation2019) in particular should be of considerable concern to proponents of knowledge capital theory, a criticism that has yet to be met with a response by H&W. Further, the work of Aloisi and Tymms (Citation2017) on the impact of policy reform on PISA performance requires further attention.

The influence of H&W within the OECD and the World Bank is considerable. The role the two organisations play as trans-national policy actors, means the likelihood of a new research programme supplanting it appears remote. Lakatos’ research programme structures this analysis. Yet, the degenerate knowledge capital research programme appears to be following more along the lines of Kuhn; the question is not if a paradigm shift will happen, but a question of when and at whose insistence given the influential proponents of the current paradigm dominating academic discourse. What may replace it? The work of Brown, Lauder, and Cheung (Citation2020) and their New Human Capital Theory deserves serious debate as a potential new paradigm.

A limitation of this paper is that it uses the knowledge capital research programme concerning ILSAs. The knowledge capital perspective nevertheless represents an extreme and limited interpretation of human capital. Using Lakatos’ research programme to analyse whether the more representative human capital research programme concerning ILSAs is in a degenerate state is the next logical progression of this paper. Implicit in both analysis’ is a suggestion that underlying both research programmes is a lack of theory in both the selection, operationalisation and explanation of the relationships between the variables it uses. An attempt to develop a theory of scientific method that supports alternative research programmes to take ILSAs forward would be constructive to the ongoing academic discourse.

Disclosure statement

No potential conflict of interest was reported by the author.

References

  • Aloisi, C., and P. Tymms. 2017. “PISA Trends, Social Changes, and Education Reforms.” Educational Research and Evaluation 23 (5–6): 180–220. doi:10.1080/13803611.2017.1455290.
  • Barber, M., and M. Mourshed. 2007. “How the World’s Best-Performing School Come Out on Top.” In September 2007 Reports-Social Sector, 56. New York.
  • Becker, G. S. 1994. “Human Capital Revisited.” In Human Capital: A Theoretical and Empirical Analysis with Special Reference to Education. 3rd ed, 15–28. Chicago: University of Chicago Press.
  • Blaug, M. 1975. “Kuhn versus Lakatos, or Paradigms versus Research Programmes in the History of Economics.” History of Political Economy 7 (4): 399–433. doi:10.1215/00182702-7-4-399.
  • Bourdieu, P. 2013. “Distinction: A social critique of the judgement of taste.” In Food and Culture: A Reader, edited by C. Counihan and P. Van Esterik, 30–39. New York: Routledge. https://doi.org/10.4324/9780203079751
  • Breakspear, S. 2012. “The Policy Impact of PISA: An Exploration of the Normative Effects of International Benchmarking in School System Performance.” OECD Journals 1–32. doi:10.1787/19939019.
  • Bridgeman, B., A. Mcbride, and W. Monaghan. 2004. “Testing and Time Limits (Report No. RDC-01).” In R&D Connections, 1–6. Princeton, NJ: ETS Research & Development.
  • Brown, P., H. Lauder, and D. Ashton. 2010. The Global Auction: The Broken Promises of Education, Jobs, and Incomes. Oxford: Oxford University Press.
  • Brown, P., H. Lauder, and S. Y. Cheung. 2020. The Death of Human Capital?: Its Failed Promise and How to Renew It in an Age of Disruption. Oxford: Oxford University Press.
  • Carvalho, L. M., and E. Costa. 2015. “Seeing Education with One’s Own Eyes and through PISA Lenses: Considerations of the Reception of PISA in European Countries.” Discourse: Studies in the Cultural Politics of Education 36: 638–646. doi:10.1080/01596306.2013.871449.
  • Cheema, J. R. 2014. “The Migrant Effect: An Evaluation of Native Academic Performance in Qatar.” Research in Education 91 (1): 65–77. doi:10.7227/RIE.91.1.6.
  • Cheema, J. R. 2015. “The Private–public Literacy Divide amid Educational Reform in Qatar: What Does PISA Tell Us?.” International Review of Education 61 (2): 173–189. doi:10.1007/s11159-015-9479-8.
  • Chen, Y., and S. Feng. 2013. “Access to Public Schools and the Education of Migrant Children in China.” China Economic Review 26: 75–88. doi:10.1016/j.chieco.2013.04.007.
  • Cosgrove, J., and F. Cartwright. 2014. “Changes in Achievement on PISA: The Case of Ireland and Implications for International Assessment Practice.” Large-scale Assessments in Education 2 (1): 17. doi:10.1186/2196-0739-2-2.
  • Cross, R. 1982. “The Duhem-Quine Thesis, Lakatos and the Appraisal of Theories in Macroeconomics.” The Economic Journal 92 (366): 320–340. doi:10.2307/2232443.
  • d’Agnese, Vasco. 2015. “PISA’s Colonialism: Success, Money, and the Eclipse of Education.” Power and Education 7 (1): 56–72. doi:10.1177/1757743814567387.
  • Davis, E. R., R. Wilson, and B. Dalton. 2020. “Another Slice of PISA: An Interrogation of Educational Cross-national Attraction in Australia, Finland, Japan and South Korea.” Compare: A Journal of Comparative and International Education 50 (3): 309–331. doi:10.1080/03057925.2018.1510305.
  • Educational Testing Services (ETS). 2020. “Human Capital and International Large Scale-Assessment.” Educational Testing Services (ETS), Accessed 26 June 2020. https://www.ets.org/research/topics/ilsa
  • Goldstein, H. 2018. “Measurement and Evaluation Issues with PISA.” In The PISA Effect on Global Educational Governance, edited by L. Volante, 258. New York and London: Routledge.
  • Grisay, A., E. Gonzalez, and C. Monseur. 2009. “Equivalence of Item Difficulties across National Versions of the PIRLs and PISA Reading Assessments.” IERI Monograph Series: Issues and Methodologies in Large-Scale Assessments 2: 63–174.
  • Grönqvist, E., and J. Vlachos. 2016. “One Size Fits All? The Effects of Teachers‘ Cognitive and Social Abilities on Student Achievement.” Labour Economics 42: 138–150. doi:10.1016/j.labeco.2016.08.005.
  • Gruber, K. H. 2006. “The German ‘PISA-shock’: Some Aspects of the Extraordinary Impact of the OECD’s PISA Study on the German Education System.” In Cross-national Attraction in Education: Accounts from England and Germany, 195–208. Vol. 16. Oxford: Symposium Books.
  • Guarino, C. M., and J. C. Tanner. 2012. “Adequacy, Accountability, Autonomy and Equity in a Middle Eastern School Reform: The Case of Qatar.” International Review of Education 58 (2): 221–245. doi:10.1007/s11159-012-9286-4.
  • Hanushek, E. A., and L. Woessmann. 2008. “The Role of Cognitive Skills in Economic Development.” Journal of Economic Literature 46 (3): 607–668. doi:10.1257/jel.46.3.607.
  • Hanushek, E. A., and L. Woessmann. 2015. The Knowledge Capital of Nations: Education and the Economics of Growth. 1 edn., Vol. 1, 55–70. CESifo book series. Cambridge, MA: MIT Press.
  • Harris, A., and M. Jones. 2017. “Leading Educational Change and Improvement at Scale: Some Inconvenient Truths about System Performance.” International Journal of Leadership in Education 20 (5): 632–641. doi:10.1080/13603124.2016.1274786.
  • Hatzinikita, V., K. Dimopoulos, and V. Christidou. 2008. “PISA Test Items and School Textbooks Related to Science: A Textual Comparison.” Science Education 92 (4): 664–687. doi:10.1002/sce.20256.
  • Hawksworth, J., R. Berriman, and S. Goel. 2018. “Will Robots Really Steal Our Jobs?: An International Analysis of the Potential Long Term Impact of Automation.” Pricewaterhouse Coopers 1–47. doi:10.1080/15374416.2013.822308.
  • Hoekstra, M. 2009. “The Effect of Attending the Flagship State University on Earnings: A Discontinuity-Based Approach.” Review of Economics and Statistics 91 (4): 717–724. doi:10.1162/rest.91.4.717.
  • Hopfenbeck, T. N., J. Lenkeit, Y. E. Masri, K. Cantrell, J. Ryan, and J.-A. Baird. 2018. “Lessons Learned from PISA: A Systematic Review of Peer-Reviewed Articles on the Programme for International Student Assessment.” Scandinavian Journal of Educational Research 62 (3): 333–353. doi:10.1080/00313831.2016.1258726.
  • Jerrim, J., P. Parker, A. Choi, A. K. Chmielewski, C. Sälzer, and N. Shure. 2018. “How Robust are Cross-Country Comparisons of PISA Scores to the Scaling Model Used?” Educational Measurement: Issues and Practice 37 (4): 28–39. doi:10.1111/emip.12211.
  • Kamens, D. H. 2015. “A Maturing Global Testing Regime Meets the World Economy: Test Scores and Economic Growth, 1960–2012.” Comparative Education Review 59 (3): 420–446. doi:10.1086/681989.
  • Kattan, R. B., K. Macdonald, and A. P. Harry. 2018. “Automation and Labor Market Outcomes: The Pivotal Role of High-Quality Education.” Policy Research Working Paper - World Bank. doi:10.1596/1813-9450-8474.
  • Kauko, J., J. Corvalán, H. Simola, and A. Carrasco. 2015. “The Historical Dynamics in Chilean and Finnish Basic Education Politics.” In In Contrasting Dynamics in Education Politics of Extremes, edited by A. Pitman and M. A. Pereyra, 29–52. Boston: Brill | Sense.
  • Klees, S. J. 2016. “Human Capital and Rates of Return: Brilliant Ideas or Ideological Dead Ends?” Comparative Education Review 60 (4): 644–672. doi:10.1086/688063.
  • Komatsu, H., and J. Rappleye. 2017. “A New Global Policy Regime Founded on Invalid Statistics? Hanushek, Woessmann, PISA, and Economic Growth.” Comparative Education 53 (2): 166–191. doi:10.1080/03050068.2017.1300008.
  • Komatsu, H., and J. Rappleye. 2019. “Refuting the OECD-World Bank Development Narrative: Was East Asia’s ‘Economic Miracle’ Primarily Driven by Education Quality and Cognitive Skills?” Globalisation, Societies and Education 17 (2): 101–116. doi:10.1080/14767724.2019.1577718.
  • Kosunen, S. 2014. “Reputation and Parental Logics of Action in Local School Choice Space in Finland.” Journal of Education Policy 29 (4): 443–466. doi:10.1080/02680939.2013.844859.
  • Kreiner, S., and K. B. Christensen. 2014. “Analyses of Model Fit and Robustness. A New Look at the PISA Scaling Model Underlying Ranking of Countries according to Reading Literacy.” Psychometrika 79 (2): 210–231. doi:10.1007/s11336-013-9347-z.
  • Kuhn, T. S., and D. Hawkins. 1963. “The Structure of Scientific Revolutions.” American Journal of Physics 31 (7): 554–555. doi:10.1119/1.1969660.
  • Lakatos, I. 1978. “History of Science and Its Rational Reconstructions.” In The Methodology of Scientific Research Programmes: Philosophical Papers, edited by G. Currie, I. Lakatos, and J. Worrall, 102–138. Cambridge: Cambridge University Press.
  • Lauder, H. 2015. “Human Capital Theory, the Power of Transnational Companies and a Political Response in Relation to Education and Economic Development.” Compare: A Journal of Comparative and International Education 45 (3): 490–493. doi:10.1080/03057925.2015.1027515.
  • Lauder, H., P. Brown, and S.-Y. Cheung. 2018. “Fractures in the Education–economy Relationship: The End of the Skill Bias Technological Change Research Programme?” Oxford Review of Economic Policy 34 (3): 495–515. doi:10.1093/oxrep/gry008.
  • Lewis, S. 2017. “Governing Schooling through ‘What Works’: The OECD’s PISA for Schools.” Journal of Education Policy 32 (3): 281–302. doi:10.1080/02680939.2016.1252855.
  • Loveless, T. 2014. “PISA’s China Problem Continues: A Response to Schleicher, Zhang, and Tucker.” Accessed 21 June 2020. https://www.brookings.edu/research/pisas-china-problem-continues-a-response-to-schleicher-zhang-and-tucker/
  • Lundahl, C., and M. Serder. 2020. “Is PISA More Important to School Reforms than Educational Research? The Selective Use of Authoritative References in Media and in Parliamentary debates.” Nordic Journal of Studies in Educational Policy 6 (3): 54–83. doi:10.1080/20020317.2020.1831306.
  • Matthews, R. C. O., and F. Hirsch. 1977. “Social Limits to Growth.” The Economic Journal 87 (347): 574. doi:10.2307/2231564.
  • Maxwell, N. 2005. ““Popper, Kuhn, Lakatos and Aim-oriented Empiricism.”.” Philosophia 32 (1–4): 181–239. doi:10.1007/BF02641621.
  • Miao, Z., and D. Reynolds. 2017. “Improving Maths Teaching Cross-Nationally.” In The Effectiveness of Mathematics Teaching in Primary Schools: Lessons from England and China, 152–160. London: Taylor & Francis Group.
  • Mincer, J. 1958. “Investment in Human Capital and Personal Income Distribution.” Journal of Political Economy 66 (4): 281–302. doi:10.1086/258055.
  • Moss, P., G. Dahlberg, S. Grieshaber, S. Mantovani, H. May, A. Pence, S. Rayna, B. B. Swadener, and M. Vandenbroeck. 2016. “The Organisation for Economic Co-operation and Development’s International Early Learning Study: Opening for Debate and Contestation.” Contemporary Issues in Early Childhood 17 (3): 343–351. doi:10.1177/1463949116661126.
  • Nuzzo, R. 2014. “Scientific Method: Statistical Errors.” Nature 506 (7487): 150–152. doi:10.1038/506150a.
  • O’Donohue, W., K. E. Ferguson, and A. E. Naugle. 2003. “The Structure of the Cognitive Revolution: An Examination from the Philosophy of Science.” The Behavior Analyst 26 (1): 85–110. doi:10.1007/BF03392069.
  • OECD. 2014. PISA 2012 Results: What Students Know and Can Do. Vol. I, Revised edition, February 2014. Paris: OECD Publishing.
  • OECD. 2015. “Universal Basic Skills.” Universal Basic Skills: What Countries Stand to Gain. Paris: OECD Publishing. doi:10.1787/9789264234833-en.
  • OECD. 2016. PISA 2015 Results. Vol. I, pp. 492. Paris: OECD Publishing. doi:10.1787/9789264266490-en.
  • Perkins, R., G. Moran, J. Cosgrove, and G. Shiel. 2011. “PISA 2009: The Performance and Progress of 15-year-olds in Ireland.” Paper presented at the Submission to the Joint Oireachtas Committee on Education and Skills, Dublin.
  • Perkins, R., G. Shiel, B. Merriman, J. Cosgrove, and G. Moran. 2013. “Learning for Life: The Achievements of 15-year-olds in Ireland on Mathematics.” Reading Literacy and Science in PISA 2012, Dublin.
  • Popper, K, 1963. ”Truth, Rationality, and the Growth of Scientific Knowledge.” In Conjectures and Refutations: The Growth of Scientific Knowledge, 291–336. London: Routledge & Kegan Paul.
  • Ramirez, F. O., X. Luo, E. Schofer, and J. W. Meyer. 2006. “Student Achievement and National Economic Growth.” American Journal of Education 113 (1): 43–56. doi:10.1086/506492.
  • Rappleye, J., and H. Komatsu. 2019. “Is Knowledge Capital Theory Degenerate? PIAAC, PISA, and Economic Growth.” Compare: A Journal of Comparative and International Education 1–19. doi:10.1080/03057925.2019.1612233.
  • Ringarp, J. 2016. “PISA Lends Legitimacy: A Study of Education Policy Changes in Germany and Sweden after 2000.” European Educational Research Journal 15 (4): 447–461. doi:10.1177/1474904116630754.
  • Ross, H., and Y. Wang. 2010. “The College Entrance Examination in China: An Overview of Its Social-Cultural Foundations, Existing Problems, and Consequences.” Chinese Education & Society 43 (4): 3–10. doi:10.2753/ced1061-1932430400.
  • Rutkowski, L., and D. Rutkowski. 2016. “A Call for A More Measured Approach to Reporting and Interpreting PISA Results.” Educational Researcher 45 (4): 252–257. doi:10.3102/0013189x16649961.
  • Sahlberg, P. 2011. “PISA in Finland: An Education Miracle or an Obstacle to Change?” Center for Educational Policy Studies Journal 1 (3): 119–140.
  • Sahlgren, G. H. 2014. “Incentive to Invest: How Education Affects Economic Growth,”1–43. London: Adam Smith Institute. https://www.adamsmith.org/research/incentive-to-invest-how-education-affects-economic-growth#
  • Salokangas, M., and J. Kauko. 2015. “Borrowing Finnish PISA Success? Critical Reflections from the Perspective of the Lender.” Educação E Pesquisa 41 (spe): 1353–1366. doi:10.1590/S1517-9702201508144214.
  • Schleicher, A., and T. Qian. 2015. “Education Post-2015: Knowledge and Skills Transform Lives and Societies.” In Universal Basic Skills: What Countries Stand to Gain, edited by E. A. Hanushek and L. Woessmann, 9–14. Paris: OECD.
  • Schleicher, A. 2018. “Word Class: How to Build a 21st-century School System.” OECD 296. doi:10.1787/4789264300002.
  • Schuelka, M. J. 2013. “Excluding Students with Disabilities from the Culture of Achievement: The Case of the TIMSS, PIRLS, and PISA.” Journal of Education Policy 28 (2): 216–230. doi:10.1080/02680939.2012.708789.
  • Sellar, S., and B. Lingard. 2014. “The OECD and the Expansion of PISA: New Global Modes of Governance in Education.” British Educational Research Journal 40 (6): 917–936. doi:10.1002/berj.3120.
  • Shahjahan, R. A., and C. Morgan. 2016. “Global Competition, Coloniality, and the Geopolitics of Knowledge in Higher Education.” British Journal of Sociology of Education 37 (1): 92–109. doi:10.1080/01425692.2015.1095635.
  • Shiel, G., and E. Eivers. 2009. “International Comparisons of Reading Literacy: What Can They Tell Us?” Cambridge Journal of Education 39 (3): 345–360. doi:10.1080/03057640903103736.
  • Shields, R., and A. S. Hernandez. 2020. “Mixed Signals: Cognitive Skills, Qualifications and Earnings in an International Comparative Perspective.” Oxford Review of Education 46 (1): 111–128. doi:10.1080/03054985.2019.1687436.
  • Simola, H. 2015. “The Finnish Education Mystery.” In The Finnish Education Mystery: Historical and Sociological Essays on Schooling in Finland, edited by H. Simola, 207–244. 1sted. New York: Routledge.
  • Solheim, O. J., and L. Kjersti. 2018. “Can Test Construction Account for Varying Gender Differences in International Reading Achievement Tests of Children, Adolescents and Young adults?–A Study Based on Nordic Results in PIRLS, PISA and PIAAC.” Assessment in Education: Principles, Policy and Practice 25 (1): 107–126. doi:10.1080/0969594X.2016.1239612.
  • Spence, M. 1973. “Job Market Signaling.” The Quarterly Journal of Economics 87 (3): 355. doi:10.2307/1882010.
  • Stiglitz, J., and A. Weiss. 1990. “Sorting Out the Differences between Signaling and Screening Models.” Paper presented at the National Bureau of Economic Research, Cambridge, MA.
  • Stiglitz, J. E. 1975. “The Theory of Screening, Education, and the Distribution of Income.” American Economic Review 65: 283–312. doi:10.2307/1804834.
  • Stromquist, N. P. 2016. “Using Regression Analysis to Predict Countries’ Economic Growth: Illusion and Fact in Education Policy.” Real-World Economics Review 76: 65–74.
  • UNESCO-IBE. 2006. World Data on Education: Sixth Edition 2006/07-Argentina, 48. London: Routledge.
  • Varley, B. J., C. Murphy, and Ó. Veale. 2008. “Science in Primary Schools.” Phase 1 Final Report 1–215. doi:10.4324/9780203421178.
  • World Bank. 2018. World Development Report 2018: Learning to Realize Education’s Promise 18-20. Washington, DC: World Bank. doi:10.1596/978-1-4648-1096-1
  • You, Y. 2019. “The ‘New Orientalism’: Education Policy Borrowing and Representations of East Asia.” Compare 50 (5): 1–22. doi:10.1080/03057925.2018.1559038.
  • Zellman, G. L., W. Ryan Gery, R. Karam, L. Constant, H. Salem, G. Gonzalez, N. Orr, et al. 2009. “Implementation of the K-12 Education Reform in Qatar’s Schools.” Monograph. Paper presented at the RAND Corporation, Santa Monica.
  • Zhao, Y. 2014. Who’s Afraid of the Big Bad Dragon? Why China Has the Best (And Worst) Education System in the World. San Fransisco, CA: Jossey-Bass.
  • Zhao, Y. 2017. “What Works May Hurt: Side Effects in Education.” Journal of Educational Change 18 (1): 1–19. doi:10.1007/s10833-016-9294-4.
  • Zhao, Y. 2018. “Shifting the Education Paradigm: Why International Borrowing Is No Longer Sufficient for Improving Education in China.” ECNU Review of Education 1 (1): 76–106. doi:10.30926/ecnuroe2018010105.
  • Zhao, Y. 2020. “Two Decades of Havoc: A Synthesis of Criticism against PISA.” Journal of Educational Change 21 (2): 245–266. doi:10.1007/s10833-019-09367-x.