1,767
Views
9
CrossRef citations to date
0
Altmetric
Original Article

Performance-based accountability: exploring Ghanaian teachers perception of the influence of large-scale testing on teaching and learning

& ORCID Icon
Pages 333-350 | Received 04 Nov 2021, Accepted 03 Aug 2022, Published online: 15 Aug 2022

ABSTRACT

Undergirded by “economic theory of the principal–agent problem”, the study investigated secondary school teachers’ perception of the influence of large-scale testing accountability on teaching and learning. Cross-sectional survey design was used. Simple random sampling was also employed to select 200 teachers for this study. The exploratory factor analysis results showed that although large-scale testing accountability serves as a motivational factor for teachers to be more productive, it puts pressure on teachers and students. Teachers reported teaching to the test (i.e. washback) and spending a lot of instructional time preparing students for the test. Teachers suggested that large-scale testing results should not be seen as the sole basis for teacher effectiveness. Rather, they indicated that large-scale testing results should be used to foster equitable distribution of educational resources and support teachers’ professional development to enhance teaching and learning. Implications for policy and practice are discussed.

Introduction

Accountability could be understood as either being held accountable or giving an account (Lingard, Sellar, & Lewis, Citation2017). Depending on the phenomenon, accountability could be seen as either answerability or responsibility (Gregory, Citation2003). Answerability accentuates that people or some group of individuals should be held accountable for either or not achieving a required and agreed upon goals. Responsibility highlights the conceptions that people or group of individuals must be held accountable for upholding norms or rules that are mostly unclear and very complicated to specify. Researchers have argued that tensions between answerability and responsibility have shaped accountability policies in most countries (Olssen & Peters, Citation2005; Suspitsyna, Citation2010). In education, accountability has garnered global attention largely because of how it has historically been perceived as a one-way responsibility of educators to demonstrate to governments that outcomes have been realised or that policies have been either or not effectively implemented (Edmonton, Citation2005). Contemporary, educational scholars perceive accountability within education as not static, but rather a complex and frequently changing social process (Lingard et al., Citation2017). In lieu of the constantly evolving notion regarding accountability, there are varying conceptualisation and approaches to accountability, which are concurrently at play and in competition for recognition within school systems. Among these accountability conceptualisations are but not limited to bureaucratic or performance-based accountability (Darling-Hammond, Citation1989; Firestone & Shipps, Citation2007), intelligent accountability (O’Neill, Citation2013; Sahlberg, Citation2010), genuine accountability (Darling-Hammond, Wilhoit, & Pittenger, Citation2014) and professional accountability (Darling-Hammond, Citation1989; Firestone & Shipps, Citation2007).

Despite different operationalisation and conceptualisation of accountability in education (Gilbert, Citation2011), it basically narrows down to key actors and agents within the school system (i.e. schools, headteachers, students and teachers) who are being held accountable for student performance, especially in a large-scale testing (Brill, Grayson, Kuhn, & O’Donnell, Citation2018). For this study, we focused on bureaucratic accountability that over a decade has been conceptualised as performance-based accountability (Firestone & Shipps, Citation2007). Bureaucratic accountability has historically been understood as hierarchically structured in a bureaucratic manner. In this approach to accountability “priorities determined at the superior level are followed, and supervisory control is exercised intensively with a clear understanding for the need to follow orders” (Stone, Citation1989, p. 6). Contemporary, bureaucratic or performance-based accountability approach largely refers to the use of data from large-scale student testing as a mechanism or tool to hold school systems accountable (Popham, Citation1999). Although extant literature has demonstrated that one major aim of large-scale testing is to support student success by checking those in charge of educating students (Decker & Bolt, Citation2008; Kellaghan, Greaney, & Murray, Citation2009; Klinger, DeLuca, & Miller, Citation2008), unfortunately large-scale testing has come to be a tool for accountability across the globe, and testing has shifted from an instrument for making significant decisions about learners to a tool for holding schools and educational systems accountable (Brill et al., Citation2018; Chung & Chea, Citation2016; Hargreaves, Citation2020; Kirkpatrick & Zang, Citation2011; Liu, Luo, & Tang, Citation2021; Rahman, Seraj, Hasan, Namaziandost, & Tilwani, Citation2021).

In sub-Saharan Africa, large-scale testing has been found to be imperative to all educational stakeholders and policy makers (Browne, Citation2016). Besides evaluating the quality of schools and educational systems, large-scale testing also determines the future career of students (Browne, Citation2016). Large-scale testing has been used in Sub-Saharan Africa as a gate keeping tool because of the limited and scarce resources. Accountability issues, in response to political, social, and economic pressures, have in over the years gained recognition and importance in government administrations in many countries. This accountability system varies, unfolding with different speeds and impact (Hopmann Citation2007; Kellaghan et al., Citation2009). Increasingly political accountability demands have led to proliferation of testing requirements, which put teachers and students under constant monitoring to make sure learning is effectively and successfully happening (Barksdale-Ladd & Thomas, Citation2000; Hargreaves, Citation2020; Hoffman, Assaf, & Paris, Citation2001; Kellaghan et al., Citation2009; Liu et al., Citation2021; Popham, Citation2003; Rahman et al., Citation2021).

Large scale testing: advocates and critics

Due to the diverse usage of large-scale testing in schools and educational systems, it has become one of the highly contentious topics in educational jurisdictions around the globe. Activists of large-scale testing are of the view that these tests put the need pressure on educators to work effectively and efficiently. Advocates of large-scale testinguphold that it offers students with knowledge of their own learning and also teachers with deeper understanding of students’ progress (Carnoy & Loeb, Citation2002; Hamilton & Gonzales, 2003; Roderick, Jacob & Bryk, Citation2002). Researchers also argued that high-stake standardised exams are strongly linked to improvements in an average performance in international standardised assessments (Bergbauer, Hanushek, & Woessmann, Citation2018; Yang, Luo, Vadillo, Yu, & Shanks, Citation2021). Similarly, other researchers assert that when teachers and schools are held responsible for student performance, performing schools and teachers are identified and rewarded and underperforming schools can be supported with extra resources to ensure equitable distribution of resources (Elbousty, Citation2009; Jones & Egley, Citation2007; Marchant & Paulson, Citation2005).

Advocates further argue that performance-based accountability programmes are to reduce the achievement disparities between students from high socio-economic status and those from low socio-economic status. Accountability systems are based on the premise that if schools are held accountable for scores on large-scale tests, student attainment will increase, mainly in schools that serve a large number of students from poor homes. Moreover, awards tied to student achievement are supposed to encourage and stimulate administrators, teachers, and students to improve educational practices (i.e. instructional practices and learning strategies) in order to achieve positive results. The testing results are aimed at identifying learners and schools in need of remedy, thus permitting a more effective use of resources (Hamilton & Koretz, Citation2002).

In contrast, critics of large-scale tests argue that large-scale testing accountability augment students’ probability of low achievement and hold teachers with unequal and inequitable resources to the same standards, narrow and garble the curriculum, and coagulate social class and ethnic discrepancies (Berliner, Citation2011; Bracey, Citation2000; Burger & Krueger, Citation2003; Mika, Citation2005; Nichols & Berliner, Citation2008; Rahman et al., Citation2021). Additionally, they contend that high stakes testing accountability transmits the information that the central aim of learning is to get good results (Berliner, Citation2011; Klenowski, Citation2011; Klinger & Rogers, Citation2011; Nichols & Berliner, Citation2008). Hargreaves (Citation2020) maintained that high stake testing has led to “teaching to the test, cultural bias, avoidance of innovation, dilemmas of whether to include highly vulnerable students in the testing process or not, and emotional ill-being among students and teachers” (P. 414). Similarly, extant literature shows that high-stake testing reduces significant amount of time spent in deep-rooted learning and vital creative skills including self-monitoring skills, organisation of knowledge, problem-solving etc (Ananda & Ashadi, Citation2021; Berliner, Citation2011; Klenowski, Citation2011; Klinger & Rogers, Citation2011; Zhao, Citation2015). Irrespective of others perception, the importance and efficiency of high-stake testing, it is unquestionable that these testing programmes have had substantial adverse impacts on students, teachers, and school administrators. For example, countries develop large-scale assessments that hold principals and students accountable. In these systems, students do not graduate unless they pass an exit exam and principals do not get raises unless student achievement improves to a predetermined level. Principals feel the pressure to improve student achievement and fear losing their contracts if they fail to make academic gains (Berliner, Citation2011; Kellaghan et al., Citation2009; Liu et al., Citation2021; Prytula, Noonan, & Hellsten, Citation2013).

Moreover, critics of high-stakes accountability do not believe that large-scale testing when used for making high decisions raise the achievement disparities between privileged and underprivileged students (von Zastrow & Janc, Citation2004). Rather, they contend that the pressure for schools to meet adequate yearly progress and prevent punishments urges teachers to concentrate only on the content and skills assessed on the tests. This political accountability leads to narrowing of the curriculum and thus restricts the experiences of students, notably students who are in schools that serve larger populations of underprivileged students (Thompson, Citation2001). Additionally, the pressure to increase large-scale test results might adversely affect student placement decisions (Baidoo-Anu, Citation2022; Baidoo-Anu, Gyamerah, & Chanimbe, Citation2022), such as class retention and special education detection. Critics also claim that test-based accountability policies make it complicated for schools to hire and retain qualified teachers (Rotberg, Bernstein, & Ritter, Citation2001).

The Ghanaian context

Various countries, including Ghana have made large-scale testing the centrepiece of their school reform (Anane, Citation2015). In lieu of this, large-scale testing is prevalent in the educational system in Ghana. Students at all levels of education from kindergarten through to junior high school are required to take these tests. Irrespective of the grade level, teacher and national-mandated tests are very competitive and results are used for high-stake decisions: they determine students’ progress to the next grade level (Baidoo-Anu, Citation2022; Baidoo-Anu et al., Citation2022). At present, national-mandated tests such as the Basic Education Certificate Examination (BECE) (i.e. school-leaving examination all Grade 9 students must write and pass before progressing to secondary school) and West African Senior School Certificate Examination (WASSCE) (i.e. qualifying exams that all Grade 12 students write and pass before progressing to have post-secondary education) have become high-stakes (Amoako, Citation2019). Outcomes of these national mandated tests are also used for measuring the quality and effectiveness of school, and teachers. Although several studies have investigated the influence of large-scale testing accountability on curriculum and learning at the basic level (Amoako, Citation2019), there are limited studies on the influence of large-scale testing accountability on senior high-school teachers teaching and students learning. As a preliminary investigation, this study unpacks secondary school teachers’ perception of the influence of large-scale testing accountability on teaching and learning and suggests ways that data from large-scale testing could be used to move learning forward. Based on the purpose of the study, two research questions were formulated to guide the study.

1. What are senior high-school teachers’ perceptions of the influences of large-scale testing accountability on teachers teaching and learning?

2. What do teachers suggest are effective ways to use data from large-scale testing?

Theoretical framework

Based on the overarching objective of the study, the economic theory of the principal–agent problem was employed to provide a theoretical lens for the study. Economic theory of the principal–agent problem was proposed by Grossman and Hart in 1983. This theory has been used extensively to provide a justification for the proliferation of the usage of data-driven accountability in schools and educational systems (Figlio & Loeb, Citation2011; Rothstein, Jacobsen, & Wilder, Citation2008). Principal–agent problem happens when one person or group who is the agent/s (such as schools, teachers) acts on behalf of another person or group who is the principal/s (such as parents, governments). According to the theory, suboptimal outcomes occur when there is misalignment in the aims and incentives of the principals (i.e. parents and governments) and the agents (i.e. schools and teachers). Theory of the principal–agent problem was deemed appropriate for this study since in bureaucratic or performance-based approaches to accountability “priorities determined at the superior level are followed, and supervisory control is exercised intensively with a clear understanding for the need to follow orders” (Stone, Citation1989, p. 6). The theory provides understanding of how principal (i.e. governments) uses data-driven accountability to force agents (i.e. schools, teachers and headteachers) to align their goals and objectives to the goals and objectives of the government and educational policy makers. Thus, through data-driven accountability, principals monitor agents' performance and ensure alignments in their aims and incentives to make sure they work to meet their (i.e. principals) set goals.

In this study, we regard the government, Ghana education service, parents and educational policy actors as the core “principals” in the education system, while secondary school teachers are the key agents. That is, in this study secondary-school teachers are the key agents or groups that act (i.e. teach secondary school students) on behalf of Ghana education service, parents, educational policy actors and the government. Because most Ghanaians measure the quality of education by number of students who pass national mandated high-stakes testing, for political reasons it appears successive governments are more concerned about students passing these national mandated large-scale testing at all costs (Baidoo-Anu et al., Citationforthcoming). The desire for students to pass the exams at all costs may bring a misalignment between political aims of the government (i.e. students passing national mandated exams) and teachers who want to achieve the intended learning goals or objectives by promoting formative assessment practices. This misalignment introduces “principal–agent problem”. Thus, misalignment in the aims and incentives (i.e, the need to pass national large-scale exams at all costs) between teachers and governments or educational policy makers results in “principal-agents problem” between teachers and educational policy makers (Jerrim & Sims, Citation2021). To solve this “principal–agent problem” in education, Figlio and Loeb (Citation2011) stated that government and educational policy makers (i.e. principals) uses data-driven accountability to force teachers (i.e. agents) to align their goals and objectives to the goals and objectives of the government and educational policy makers. Figlio and Loeb (Citation2011) emphatically stated;

The information content in school accountability systems can provide a powerful mechanism for overcoming the principal-agent problem. Assessing schools against the common metric of standardized student test scores provides policy makers and members of the general public with independent information regarding how well schools and school districts (and potentially teachers) are doing in comparison to their peers and outside performance standards. Measuring and reporting school performance and attaching positive and negative consequences to meeting or failing to meet performance objectives provides incentives that encourage educators to concentrate on the subjects and materials that are being measured and to potentially alter the methods through which they educate students. The measurement and reporting of a school’s progress allow policy makers to assess how successful a school has been in meeting the state’s achievement goals (p. 386).

Gleaning from the theory, school-based accountability is premised on the idea that public pressure from openly reported data will lead to an enhanced student success and achievement (Figlio & Loeb, Citation2011; Rothstein et al., Citation2008; Supovitz, Citation2009). However, extant literature explicates that data-driven accountability has unintended negative consequences on the agents (i.e. teachers) (Grupe & Nitschke, Citation2013; Smith & Amick, Citation1989). For example, Smith and Amick (Citation1989) found that accountability puts too much stress on teachers. Similarly, Grupe and Nitschke (Citation2013) maintained that close monitoring and punishments attached to accountability leads to stress and anxiety. Further, data-driven accountability has also been found to cause loss in job autonomy (Jerrim & Sims, Citation2021). Thus, teachers do not have the liberty to work due to data-driven accountability. Finally, data-driven accountability forces teachers to do more work such as preparing students for tests (Perryman & Calvert, Citation2019). This adds extra work to teachers and puts them under pressure, consequently affecting classroom teaching and learning.

Methods

Research design, sampling and participants

Cross-sectional survey design was employed in this study. Desiring to give participants an equal chance to be part of the study, simple random sampling, specifically the lottery method was used to select the teachers from the three secondary schools in the Birim Central Municipal in Ghana. The secondary schools were also randomly selected using lottery method. In each school, one teacher was trained and engaged as a research assistant to assist with the administration and data collection. During the administration of the instruments, respondents were informed on the purpose of the study and the need to provide honest responses to the items since the study is for academic purposes and also be useful to the school, teachers and students. Teachers were assured that their responses would be kept confidential, and their identity kept anonymous. They were also made to sign the consent form and were apprised that they can withdraw from the study anytime without consequences. It took 7 weeks to collect the data from the respondents. Out of the 227 participants contacted, 200 secondary school teachers completed and returned the questionnaire. This shows 88.1% response rate for this study. Teachers were prompted to respond to survey items based on their individual experience. Out of the total participants, 50 (25%) were females and 150 (75%) were males. Majority 154 (77%) of the teachers had a bachelor’s degree. Moreover, regarding teaching experience, preponderance of the teachers indicated that they have taught for within 6–10 years. This is evident that teachers in this study are experienced teachers who are in the position to respond to the question and provide credible information concerning large-scale accountability influence. The analysis of teachers demographics is presented in .

Table 1. Demographics of the participants.

Instruments and data analysis

Teachers’ perception of large-scale testing accountability inventory was developed and used in this study. The inventory was made up of 20 items that sought to understand teachers’ perception of the influence of large-scale testing accountability on teaching and learning. To provide validity evidence, the underpinning conceptual dimensions for the survey were identified based on (a) prior research study, focusing on influence of high-stake testing curriculum implementation in Ghana (Amoako, Citation2019), and (b) previous surveys and literature on how high-stake testing impacts teachers’ classroom practices (Hamilton & Koretz, Citation2002; Pedulla et al., Citation2003; Ritt, Citation2016; Stecher, Citation2002) as well as (c) influence of data-driven accountability on teachers’ classroom practices (Figlio & Loeb, Citation2011; Jerrim & Sims, Citation2021). Further, three expert researchers were recruited to provide open-ended feedback as well as edits on the wording of items of the questionnaire. The expert further streamlined the items in alignments with research questions and enhanced the clarity of the item wordings. The final survey was pilot tested to check understanding and vagueness and fix any errors that would be due to the framing and development of the items. After the pilot-testing, five items were removed.

In order to understand the structural patterns emanating from teachers’ responses to the influence of large-scale testing accountability on teaching and learning, exploratory factor analysis was conducted. Before starting the factor analysis, data were inspected to ensure suitability for factor analysis. Principal axis factor analysis was employed because it is legitimate for short-scale response options and small sample sizes (Bandalos & Finney, Citation2018). Oblique rotation (oblimin) was used to examine the component correlation matrix. Exploratory factor analysis was conducted on the 20 items that examine teachers’ perception of large-scale accountability on teaching and learning. The Kaiser–Meyer–Olkin see ) measure validated the sampling adequacy for the analysis, KMO = .634 (“middling” according to Hutcheson & Sofroniou, Citation1999). This means the sample was adequate for carrying out factor analysis. Bartlett’s test was significant (.000) indicating that the assumption of sphericity was met.

Table 2. Kaiser–Meyer–Olkin measure of sampling adequacy.

Moreover, a preliminary analysis was done to obtain eigenvalues for each factor in the data. Seven factors had eigenvalues over Kaiser’s criterion of 1 explaining 64.04% of the variance. In order to determine the best factor structure pattern, Monte Carlo PCA for Parallel Analysis (Watkins, Citation2000) and scree plot (see ) were used. The results of this analysis supported three factors, explaining 38.54% of the variance and with eigenvalues more than the corresponding criterion values in a random data set, generated with a matrix of the same size (20 variables × 200 participants).

Figure 1. Scree plot

Figure 1. Scree plot

Factor loadings below 0.4 were suppressed because they did not represent significant values (Field, Citation2013). Items (12, 13, 7, 5 and 3) did not load onto any factor and were thus deleted. The factors were named based on the items that clustered around the three factors. All the three factors had an internal consistency (i.e. Cronbach’s alpha) greater than 0.70, indicating good reliability (Field, Citation2013). The three-factor structure was used to understand teachers’ perceived influence on large-scale testing accountability on teaching and learning.

Regarding research question two, participants responses to open-ended questionnaires were thematically analysed. First, the authors organised and (re)read the data. Responses were coded suggestions-by-suggestions, and a code list was generated and refined through an iterative process. Codes were grouped into subthemes in line with our research questions and then into larger themes to describe teachers suggestions. A second ratter was used to review the code list and code 10% of the data. Disparities in the data were co-discussed by the authors until we reached a consensus.

Results

Research question one

presents the factor structure and loadings, reliability coefficients (i.e. Cronbach’s alpha), and variance explained. A three-factor structure was identified to understand teachers’ perceived influence of large-scale testing accountability on teaching and learning. Factor 1 which was termed as motivation to teachers centred on six items that imply that teachers perceive large-scale testing accountability as a way of motivating teachers and students. Teachers perceive large-scale accountability as a way of boosting their morale in school. Also, large-scale testing motivates previously unmotivated students to learn. Factor two was termed pressure on teachers. This factor was made up of five items that implies that large-scale accountability puts pressure on teachers. Teachers perceive that large-scale testing accountability forces them to teach to the test. Teachers further indicated that they feel pressure from headteachers to raise students’ scores. Moreover, school administrators and parents also put pressure on them to increase students’ scores. Finally, factor three was termed pressure on students. The third factor was made up of three items. Teachers indicated that large-scale accountability put pressure on students, thereby increasing student test anxiety. Teachers also spend extra time preparing teachers for the test, which consequently increases the number of instructional hours. The reliability estimates for the four factors ranged from 0.702 to 0.74: (a) motivate teachers and students: α = .743, (b) pressure on teachers: α = .702, (c) pressures on students: α = .714. Overall, these three factors accounted for 38.54% of variance in teachers’ perceptions of large-scale testing accountability.

Table 3. Exploratory factor analysis of perceived influences of large-scale testing accountability (high-stake) on teaching and learning.

Research question two

An open-ended question was added to allow teachers to suggest ways that large-scale assessment can be used. The question was “Please, use the space provided below to suggest ways that large scale testing can be used effectively”. Teachers’ responses to this question have been presented in . Teachers suggested that large-scale testing should not be seen as the sole basis for teacher effectiveness. Rather, they indicated that large-scale testing results should be used to foster equitable distribution of educational resources and foster teachers’ professional development to enhance teaching and learning. Moreover, teachers indicated that large-scale testing results should be used to investigate school and educational policies related to curriculum and instructional practices. Teachers further suggested that large-scale testing results should be reported in a manner that promotes professional collegiality and foster unity among schools, teachers and students.

Table 4. Teachers’ suggestions for effective use of large scale-testing.

Discussions

Undergirded by theory of the principal–agent problem, the study investigated Brim municipality secondary school teachers’ perceived influence of large-scale testing accountability on teaching and learning. Specifically, the study sought to (a) understand the perceived influences of large-scale testing accountability on senior high-school teachers teaching and learning in Birim central Municipal and (b) teachers suggestions on effective ways to use data from large-scale testing.

Regarding perception of teachers on large-scale testing accountability, the study found that teachers perceive large-scale testing accountability as way to increase their morale in schools and motivate previously unmotivated students. However, they indicated that they feel pressured by the use of data from large-scale testing as a measure of teachers and schools quality and effectiveness. According to the theory of “principal–agent problem” explanation to data-driven accountability, large-scale testing accountability is intended to encourage teachers’ productivity and enhance student success and achievement (Figlio & Loeb, Citation2011; Rothstein et al., Citation2008; Supovitz, Citation2009). Moreover, large-scale testing also offers students with the knowledge of their own progress in their learning and teachers with a deeper understanding of students’ progress and effectiveness of their own instructional practices (Carnoy & Loeb, Citation2002; Hamilton & Gonzales, 2003; Roderick, Jacob & Bryk, Citation2002). Notwithstanding large-scale accountability has unintended negative consequences on the agents (i.e. teachers) (Grupe & Nitschke, Citation2013; Smith & Amick, Citation1989). Similarly, Grupe and Nitschke (Citation2013) maintained that close monitoring and punishments attached to accountability leads to stress and anxiety. Further, data-driven accountability has also been found to augment feeling of more responsibility, less job autonomy and authority to decide quality practices and teacher effectiveness (Jerrim & Sims, Citation2021; Prytula et al., Citation2013). Thus, teachers do not have the liberty to decide which teaching practice work best for their students due to data-driven accountability.

The findings further confirm the work of McMillan, (Citation2005) who found that when large-scale testing results are used for making high stake decisions, it forces teachers to teach to the test, with the primary purpose of increasing students’ scores on external tests. The results further support the work of Kellaghan et al. (Citation2009). According to Kellaghan et al. (Citation2009) large-scale testing puts pressure on teachers to improve student achievement and fear of losing their contracts if they fail to make academic gains. This result further resonates with previous research (Browne, Citation2016; Berliner, Citation2011; Kellaghan et al., Citation2009; Prytula et al., Citation2013), who found that teachers and headteachers are also under constant pressure to show parents that they are preparing their children to be successful in a test that has the ability to open or otherwise the door to the next level of education.

Teachers in our study also indicated that large-scale testing accountability does not only affect teachers but adversely affect students. The findings support the works of researchers (Ho, Citation2012) who found that high-stakes test burden and put psychological stress on students, often leading to feeling of failure in life (Baidoo-Anu & Adomaa Acquah, Citation2021). For example, in China more than 80% of Chinese students recounted worries about high stake testing, leading to headaches and abdominal pain among students (Hesketh et al., Citation2010). Similarly, high stake testing has been found to increase students’ probability of educational failure, hold students and teachers with unequal and inequitable resources to the same standards, and garbles with the curriculum, and congeal socioeconomic class and ethnic inequalities (Nichols & Berliner, Citation2008; McMillan, Citation2005; Burger & Krueger, Citation2003; Bracey, Citation2000). Reviewing how large-scale testing shapes students’ motivation, Harlen and Deakin Crick (Citation2003) found that large-scale testing has “particularly strong and devastating impact” (p. 196) on learners who are underperforming. Researchers contend that large-scale testing accountability communicates information that denotes that the main objective of learning is to have high scores on tests (Berliner, Citation2011; Klenowski, Citation2011; Klinger & Rogers, Citation2011; Burger & Krueger, Citation2003; Bracey, Citation2000; McMillan, Citation2005; Nichols & Berliner, Citation2008). No matter the perceptions others have about the value and utility of large-scale testing, it is indisputable that these testing programmes have had substantial effects on students, teachers, and school administrators. For example, some states developed large-scale assessments that hold head teachers and students accountable. In these systems, students do not graduate unless they pass an exit exam and headteachers do not get raises unless student achievement improves to a predetermined level. Headteachers feel the pressure to improve student achievement and fear losing their contracts if they fail to make academic gains (Berliner, Citation2011; Kellaghan et al., Citation2009; Klenowski, Citation2011; Prytula et al., Citation2013). Similarly, other researchers point out that when schools are held answerable for student attainment, performing schools can be given recognition and award and failing schools can be identified for additional assistance and resources (Elbousty, Citation2009; Perkins & Wellman, Citation2008; Jones & Egley, Citation2007; Marchant & Paulson, Citation2005; Stecher, Citation2002). Suggesting ways of improving the use of large-scale testing, teachers indicated that large-scale testing should not be seen as the sole basis for teacher effectiveness and school quality. Rather, large-scale testing results should be used to enhance equitable distribution of educational resources and foster teachers’ professional development to enhance classroom practices. Teachers are concerned about how the results from large-scale testing are used to evaluate or determine schools and teachers’ effectiveness. Teachers stated that using large-scale testing results to make high-stake decisions create pressure that forces them to teach to test. Investigating the effect of high stakes testing in Ghana basic education, Amoako (Citation2019) reported that teachers engaged students in a narrowed curriculum because of the accountability pressure attached to Basic Education Certificate Examination (BECE). Moreover, Oduro (Citation2015) argued that the majority of teachers conceive assessment as mainly for accountability purposes that tend to obscure the improvement (formative) function. This view of assessment has restricted teachers in exploring ways of involving their pupils in the assessment process and making them take ownership of their own learning.

Implications for policy and practice

This study has accentuated the need for government and educational policy makers in Ghana to rethink the use of data from large-scale testing to make high stake decision regarding teachers and students at the secondary school level. When these large-scale testing are used for making high-stake decisions about teachers and students, it does not only put pressure on teachers but also on students because data from national large-scale testing are also used to make high-stake decisions about students. Volante (Citation2006) maintained that the advantage or disadvantage of large-scale testing lies in its use. Therefore, to leverage large-scale testing, results should be used to examine school policies related to curriculum and instruction and promote better allocation of state educational resources. The fundamental value of large-scale testing programmes is contingent on their capability to promote, and not hinder student and teacher learning. To do this, large-scale testing must be seen as significant, but not as measures for high-stakes decisions about teachers, students and schools (Volante, Citation2006). Therefore, Ghana Education Service should consider making large-scale testing low stakes. Data from large-scale testing could be used to foster better allocation of educational resources and encourage teachers’ professional development to improve instruction. Again, evaluation of school and teachers’ performance on the basis of students’ performance on large-scale testing creates tension and puts pressure on teachers to teach to the test that eventually narrows the curriculum (Amoako, Citation2019). Hence, to avoid unhealthy competition among schools and teachers, LSA results should be reported in a manner that encourages professional collegiality. Educational policy makers and politicians should not assume that large-scale testing alone can on its own increase teachers’ productivity and students’ achievement (Black & Wiliam, Citation1998; Kohn, Citation2000). Rather, encouraging and supporting teachers to engage in quality classroom assessment has the potential of moving learning forward and increasing students’ success. Merely providing teaching and learning materials (Adane, Citation2013; Baidoo-Anu, Citation2018a, Citation2018b), educational infrastructures (Baidoo-Anu & Mensah, Citation2018), would have little impact on students’ overall academic performance if teachers do not engage in assessment practices that move learning forward (Black & Williams, Citation2010). Irrespective of the assessment literacy of teachers, if large-scale testing results are not used appropriately, it will negatively affect teachers’ assessment practices (Volante, Citation2006). To avoid wash back (i.e. teaching to the test) (Carless, Citation2011), and other negative effect of large-scale testing, educational stakeholders in Ghana should eschew using large-scale testing results for making high stake decisions like solely determining teachers’ and schools’ effectiveness, rewarding or punishing teachers and schools, streaming or tracking students, determining students educational progress etc.

Limitation and suggestion for future research

The sampling of a small size in one municipality in Ghana, however, limits the generalisability of the study’s findings, which is a limitation of the study. Again, the study relied largely on questionnaires as the main data collection instruments can introduce biases. This can happen for a variety of reasons, including social desirability bias and attempting to protect privacy. However, assuring participants of anonymity and confidentiality of their responses helps reduce the probability of this problem occurring. Notwithstanding, this study has provided some unique compelling evidence. Future research could consider increasing the sample size to allow for broader generalisation of the results.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

We have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Notes on contributors

David Baidoo-Anu

David Baidoo-Anu is currently a PhD in Education Candidate at the Faculty of Education, Queen’s University in Kingston, Ontario Canada. He earned his Master of Philosophy in educational measurement and evaluation from the University of Cape Coast-Ghana. Previously a part-time lecturer at Presbyterian University College- Ghana, his enormous interest in research, particularly classroom assessment coupled with his experience directed him to the Queen’s Assessment and Evaluation Group (AEG). He believes that if classroom assessment is appropriately used, it has the power to support equitable access to education and propel learning forward for diverse students. David’s current research focuses on culturally responsive classroom assessment. Particularly, he draws on culturally situated understandings of assessment and empirical evidence showing the positive impacts of quality classroom assessment practices on raising students’ achievement to advocate for systemic assessment reforms. The aim of his doctoral research is to support quality and consistent assessment practices in schools across cultural context and ultimately enhance the assessment experiences of all students regardless of their background. Isaac Ennu BaidooIsaac Ennu Baidoo is a mathematics teacher at Atweaman Senior High School in Akyem Manso-Ghana. He is also a senior Tutor at the University of Cape Coast Distance Education Program. He holds Master of Education in Teacher from University of Cape Coast- Ghana. Isaac is currently completing Master of Philosophy in educational measurement and evaluation from the University of Cape Coast-Ghana. Isaac’s research focus on the influence high-stake testing on teaching and learning with specific interest in mathematics.

Isaac Ennu Baidoo

Isaac Ennu Baidoo is a mathematics teacher at Atweaman Senior High School in Akyem Manso-Ghana. He is also a senior Tutor at the University of Cape Coast Distance Education Program. He holds Master of Education in Teacher from University of Cape Coast- Ghana. Isaac is currently completing Master of Philosophy in educational measurement and evaluation from the University of Cape Coast-Ghana. Isaac’s research focus on the influence high-stake testing on teaching and learning with specific interest in mathematics.

References

  • Adane, L. O. (2013). Factors affecting low academic achievement of pupils in Kemp Methodist Junior High School in Aburi, Eastern region (Doctoral dissertation, University of Ghana).
  • Amoako, I. (2019). What’s at stake in high-stakes testing in Ghana: Implication for curriculum implementation in basic schools. International Journal of Social Sciences & Educational Studies, 5(3), 72.
  • Ananda, E., & Ashadi, A. (2021). High-stakes testing and English teachers’ role as materials developers: Insights from vocational high schools. Studies in English Language and Education, 8(1), 115–130.
  • Anane, E. (2015). Influence of accountability pressure on science, mathematics and English language teachers’ classroom practice in senior high school in Ghana. Ghana Journal of Education, Issues and Practices, 1, 44–63.
  • Baidoo-Anu, D. (2018a). Perceived school environmental, home conditions and academic performance of junior high school students in Asikuma-Odoben-Brakwa district, Ghana. Journal of Education, Society and Behavioural Science, 24(2), 1–7.
  • Baidoo-Anu, D. (2018b). Students and teachers attitudes responsible for poor academic performance of junior high school students. Asian Journal of Education and Social Studies, 1–8.
  • Baidoo-Anu, D. (2022). Between-school streaming: Unpacking the experiences of secondary school teachers and students in category C schools in Ghana. International Journal of Educational Research Open, 3, 100–188.
  • Baidoo-Anu, D., & Adomaa Acquah, P. (2021). Social challenges of adolescent secondary school students in Ghana: Evidence from Cape Coast schools. International Journal of Adolescence and Youth, 26(1), 529–540.
  • Baidoo-Anu, D., Asamoah, D., Gyamerah, K., Asare, E., & Quainoo, A. E. (Forthcoming). Tracking students increase or decrease achievement gap? Comparing academic performance of high and low streamed secondary schools in Ghana.
  • Baidoo-Anu, D., Gyamerah, K., & Chanimbe, T. (2022). Secondary school categorization in Ghana: Silent plights of students and implications for equitable learning. Journal of Human Behavior in the Social Environment, 1–18. doi:10.1080/10911359.2022.2061665
  • Baidoo-Anu, D., & Mensah, G. E. (2018). The perceptions of junior high school students and teachers towards teaching and learning of integrated science at Komenda-Edina-Eguafo-Abrim district. Asian Journal of Education and Social Studies, 1–8.
  • Barksdale-Ladd, M. A., & Thomas, K. F. (2000). What’s at stake in high-stakes testing: Teachers and parents speak out. Journal of Teacher Education, 51(5), 384–397.
  • Bergbauer, B. A., Hanushek, E. A., & Woessmann, L. (2018). Testing. NBER Working Paper No. 24836. National Bureau of Economic Research.
  • Berliner, D. C. (2011). Rational responses to high stakes testing: The case of curriculum narrowing and the harm that follows. Cambridge Journal of Education, 41(3), 287–302.
  • Black, P., & Wiliam, D. (1998). Assessment and classroom learning. Assessment in Education: Principles, Policy & Practice, 5(1), 7–74.
  • Black, P., & Wiliam, D. 2010 Inside the black box: Raising standards through classroom assessment: Formative assessment is an essential component of classroom work and can raise student achievement. Phi Delta Kappan Sept. 2010, p. 81+. Gale OneFile: CPI.Q (Canadian Periodicals).https://link.gale.com/apps/doc/A237749007/CPI?u=queensulaw&sid=CPI&xid=25782744. Accessed 18 Oct. 2019
  • Bracey, G. W. (2000). The TIMSS “final year” study and report: A critique. Educational Researcher, 29(4), 4–10.
  • Brill, F., Grayson, H., Kuhn, L., & O’Donnell, S. (2018). What impact does accountability have on curriculum,standards and engagement in education? A literature review. NFER. Accessed 09/01/2020 from https://www.nfer.ac.uk/media/3032/nfer_accountability_literature_review_2018.pdf
  • Browne, E. (2016). Evidence on Formative Classroom Assessment for Learning. K4D Helpdesk Report. Brighton, UK: Institute of Development Studies.
  • Burger, J. M., & Krueger, M. (2003). A balanced approach to high-stakes achievement testing: An analysis of the literature with policy implications. : International Electronic Journal for Leadership in Learning 7(4), 1-20.
  • Carless, D. (2011). From testing to productive student learning: Implementing formative assessment in Confucian-heritage settings. Abingdon UK: Routledge/Taylor and Francis.
  • Carnoy, M., & Loeb, S. (2002). Does external accountability affect student outcomes? A cross-state analysis. Educational Evaluation and Policy Analysis, 24(4), 305–331.
  • Chung, P. J., & Chea, H. (2016). South Korea’s accountability policy system and national achievement test. In W. C. Smith (Ed.), The global testing culture: Shaping education policy, perceptions, and practice (pp. 249–260). Oxford, England: Symposium Books.
  • Darling-Hammond, L. (1989). Accountability for professional practice. Teachers College Record, 91(1), 59–80.
  • Darling-Hammond, L., Wilhoit, G., & Pittenger, L. (2014). Accountability for college and career readiness: Developing a new paradigm. Education Policy Analysis Archives, 22(86), 1.
  • Decker, D. M., & Bolt, S. E. (2008). Challenges and opportunities for promoting student achievement through large-scale assessment results: Research, reflections, and future directions. Assessment for Effective Intervention, 34(1), 43–51.
  • Edmonton (2005). Accountability in Education. Revtrieved on 23 July from https://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=&ved=2ahUKEwj0xqyC0K_5AhVTKkQIHUOgBwYQFnoECAMQAQ&url=https%3A%2F%2Fwww.teachers.ab.ca%2FSiteCollectionDocuments%2FATA%2FIssues%2520In%2520Education%2FEmerging%2520Issues%2FAccountability%2520Discussion%2520Paper.pdf&usg=AOvVaw0VVUuZfrSH1LsiSJMx4pLN
  • Elbousty, Y. (2009). High stakes testing literature review and critique.NERA Conference Proceedings 2009. 16.https://opencommons.uconn.edu/nera_2009/16
  • Field, A. (2013). Discovering statistics using IBM SPSS statistics. Londo: sage.
  • Figlio, D., & Loeb, S. (2011). School Accountability. In E. A. Hanushek, S. Machin, & L. Woessmann (Eds.), Handbooks in Economics (Vol. 3, pp. 383–421). Netherlands: North-Holland.
  • Firestone, W., & Shipps, D. (2007). Orchestration, coherence and the problem of conflicting accountabilities. Managing Change in the Public Services, 212.
  • Gilbert, R. (2011). Pass or Fail: Assessing contemporary educational reform. Social Alternatives, 30(4), 3.
  • Gregory, R. (2003). Accountability in modern government. Handbook of Public Administration, 557–568.
  • Grupe, D., & Nitschke, J. (2013). Uncertainty and anticipation in anxiety. An integrated neurobiological and psychological perspective. Nature Reviews. Neuroscience, 14(7), 488–501.
  • Hamilton, L., & Koretz, D. (2002). Tests and their use in test-based accountability systems. In L. Hamilton, B. Stecher, & S. Klein (Eds.), Making sense of test-based accountability in education. Santa Monica: RAND.
  • Hancock, G.R., Stapleton, L.M., & Mueller, R.O. (Eds.). (2018). The Reviewer’s Guide to Quantitative Methods in the Social Sciences (2nd ed.). New York:Routledge. https://doi.org/10.4324/9781315755649
  • Hargreaves, A. (2020). Large-scale assessments and their effects: The case of mid-stakes tests in Ontario. Journal of Educational Change, 21(3), 393–420.
  • Harlen, W., & Deakin Crick, R. (2003). Testing and motivation for learning. 10(2), 169–207.
  • Hesketh, T., Zhen, Y., Lu, L., Dong, Z., Jun, Y., & Xing, Z. (2010). Stress and psychosomatic symptoms in Chinese school children: Cross-sectional survey. Archives of Disease in Childhood, 95(2), 136–140.
  • Ho, S. E. (2012). Asia-Pacific education system review series no. 5: Student learning assessment. Hong Kong, China: UNESCO.
  • Hoffman, J. V., Assaf, L. C., & Paris, S. G. (2001). High-stakes testing in reading: Today in Texas, tomorrow? The Reading Teacher, 54(5), 482–492.
  • Hopmann, S. T. (2007). Hopmann, S. T. (2007). No Child, No School, No State Left Behind: Comparative Research in the Age of Accountability. In Hopmann, S. T. (). Hopmann, S. T. (S. T. Hopmann, G. Brinek, and M. RetzlHopmann, S. T. (). Hopmann, S. T. (S. T. Hopmann, G. Brinek, and M. RetzlEds.), PISA According to PISA (pp. 363–415). Vienna: Lit-Verlag.
  • Hutcheson, G. D., & Sofroniou, N. (1999). The multivariate social scientist: Introductory statistics using generalized linear models. London:Sage.
  • Jerrim, J., & Sims, S. (2021). School accountability and teacher stress: International evidence from the OECD TALIS study. Education Assessment Evaluation Acc. doi:10.1007/s11092-021-09360-0
  • Jones, B. D., & Egley, R. J. (2007). Learning to take tests or learning for understanding? Teachers’ beliefs about test-based accountability. New Yor: The Educational Forum.
  • Kellaghan, T., Greaney, V., & Murray, S. (2009). Using the results of a national assessment of educational achievement (Vol. 5). WashWashington DC: World Bank Publications.
  • Kirkpatrick, R., & Zang, Y. (2011). The negative influences of exam-oriented education on Chinese high school students: Backwash from classroom to child. Language Testing in Asia, 1(36). doi:10.1186/2229-0443-1-3-36
  • Klenowski, V. (2011). Assessment for learning in the accountability era: Queensland, Australia. Studies in Educational Evaluation, 37(1), 78–83.
  • Klinger, D. A., DeLuca, C., & Miller, T. (2008). The evolving culture of large-scale assessments in Canadian education. Canadian Journal of Educational Administration and Policy, 76, 1–34.
  • Klinger, D. A., & Rogers, W. T. (2011). Teachers’ perceptions of large-scale assessment programs within low-stakes accountability frameworks. International Journal of Testing, 11(2), 122–143.
  • Kohn, A. (2000). The case against standardized testing: Raising the scores. In Ruining the schools (pp. 1–25). Portsmouth, NH: Heinemann.
  • Lingard, B., Sellar, S., & Lewis, S. (2017). Accountabilities in schools and school systems. Oxford Research Encyclopedia of Education. doi:10.1093/acrefore/9780190264093.013.74
  • Liu, H., Luo, L., & Tang, W. (2021). Kindergarten teachers’ experiences of stress under a high-stake inspection regime: An exploration in the Chinese context. International Journal of Educational Research, 109, 101850.
  • Maddolyn, R. (2016). The impact of high-stakes testing on the learning environment. Retrieved from Sophia, the St. Catherine University repository website: https://sophia.stkate.edu/msw_papers/658
  • Marchant, G. J., & Paulson, S. E. (2005). The relationship of high school graduation exams to graduation rates and SAT scores. Education Policy Analysis Archives, 13, 6.
  • McMillan, J. H. (2005). The Impact of High-Stakes Test Results on Teachers' Instructional and Classroom Assessment Practices. Online Submission. Retrieved on 9 May from https://eric.ed.gov/?id=ED490648
  • Mika, S. (2005). Testing testing: Do schools do too much. Retrieved from Gallup, Inc. website http://www.gallup.com/poll/19552/testing-testing-schools-too-much.aspx
  • Nichols, S. L., & Berliner, D. C. (2008). Why has high-stakes testing so easily slipped into contemporary American life? Phi Delta Kappan, 89(9), 672–676.
  • O’Neill, O. (2013). Acting on principle: An essay on Kantian ethics. UK: Cambridge University Press.
  • Oduro, O. E. (2015). Assessment in mathematics classrooms in Ghana: A study of teachers’ practices. Sussex: University of Sussex.
  • Olssen, M., & Peters, M. A. (2005). Neoliberalism, higher education and the knowledge economy: From the free market to knowledge capitalism. Journal of Education Policy, 20(3), 313–345.
  • Pedulla, J. J., Abrams, L. M., Madaus, G. F., Russell, M. K., Ramos, M. A., & Miao, J. (2003). Perceived effects of state-mandated testing programs on teaching and learning: Findings from a national survey of teachers. https://files.eric.ed.gov/fulltext/ED481836.pdf.
  • Perkins, G., & Wellman, N. (2008). The impact of high-stakes testing: Recommendations for professional school counselors and other support personnel. Academic Leadership: The Online Journal, 6(4), 15.
  • Perryman, J., & Calvert, G. (2019). What motivates people to teach, and why do they leave? Accountability, performativity and teacher retention. British Journal of Educational Studies. https://doi.org/10.1080/00071005.2019.1589417
  • Popham, W. J. (1999). Why standardized tests don’t measure educational quality. Educational Leadership, 56, 8–16.
  • Popham, W. J. (2003). Test better, teach better: The instructional role of assessment. USA: Ascd.
  • Prytula, M., Noonan, B., & Hellsten, L. (2013). Toward instructional leadership: Principals’ perceptions of large-scale assessment in schools. Canadian Journal of Educational Administration and Policy, 140, 1–30.
  • Rahman, K., Seraj, P. M. I., Hasan, M., Namaziandost, E., & Tilwani, S. A. (2021). Washback of assessment on English teaching-learning practice at secondary schools. Language Testing in Asia, 11(1), 1–23.
  • Ritt, M. (2016). The impact of high-stakes testing on the learning environment. Social Work Master’s Clinical Research Papers. 669. https://ir.stthomas.edu/ssw_mstrp/669
  • Roderick, M., Jacob, B. A., & Bryk, A. S. (2002). The impact of high-stakes testing in Chicago on student achievement in promotional gate grades. Educational Evaluation and Policy Analysis, 24(4), 333–357.
  • Rotberg, I. C., Bernstein, K. J., & Ritter, S. B. (2001). No child left behind: Views about the potential impact of the Bush Administration’s education proposals. USA: Center for Curriculum, Standards, and Technology, Institute for Education.
  • Rothstein, R., Jacobsen, R., & Wilder, T. (2008). Grading education: Getting accountability right. NY: Teachers College Press and the Economic Policy Institute.
  • Sahlberg, P. (2010). The secret to Finland’s success: Educating teachers. Stanford Center for Opportunity Policy in Education, 2, 1–8.
  • Smith, M., & Amick, B. (1989). Electronic monitoring at the workplace: Implications for employee control and job stress. In S. L. Sauter, J. J. Hurrell Jr, & C. L. Cooper (Eds.), Job Control and Worker Health (pp. 275–289). UK: Wiley.
  • Stecher, B. M. (2002). Consequences of large-scale, high-stakes testing on school and classroom practice. Making Sense of test-based Accountability in Education, 79–100.
  • Stone, D. A. (1989). Causal stories and the formation of policy agendas. Political Science Quarterly, 104(2), 281–300.
  • Supovitz, J. (2009). Can high-stakes testing leverage educational improvement? Prospects from the last decade of testing and accountability reform. Journal of Educational Change, 10(2–3), 211–227.
  • Suspitsyna, T. (2010). Accountability in American education as a rhetoric and a technology of governmentality. Journal of Education Policy, 25(5), 567–586.
  • Thompson, S. (2001). The Authentic Standards Movement and its Evil Twin. Phi Delta Kappan, 82(5), 358–362. doi:10.1177/003172170108200504
  • Volante, L. (2006). An alternative vision for large-scale assessment in Canada. Journal of Teaching and Learning, 4(1). doi:10.22329/jtl.v4i1.89
  • von Zastrow, C. E., & Janc, H. (2004). Academic atrophy: The condition of the liberal arts in America’s public schools.Washington DC: Council for Basic Education.
  • Watkins, M. W. (2000). Monte carlo PCA for parallel analysis [Programa informático]. State College, PA: Ed and Psych Associates.
  • Yang, C., Luo, L., Vadillo, M. A., Yu, R., & Shanks, D. R. (2021). Testing (quizzing) boosts classroom learning: A systematic and meta-analytic review. Psychological Bulletin, 147(4), 399.
  • Zhao, Y. (2015). Lessons that matter: What we should learn from Asian school systems. Retrieved on August 23, 2021 from http://www.mitchellinstitute.org.au/reports/lessons-that-matter-what-should-we-learn-from-asias-schoolsystems/