169
Views
0
CrossRef citations to date
0
Altmetric
Research Article

Are Online and Paper Tests Comparable? Evidence from Statewide K-12 Tests

Pages 1-13 | Published online: 05 Feb 2024
 

ABSTRACT

We investigate two research questions using a recent statewide transition from paper to computer-based testing: first, the extent to which test mode effects found in prior studies can be eliminated; and second, the degree to which online and paper assessments offer different information about underlying student ability. We first find very small mode effects for a more recent transition in Massachusetts. Second, we investigate the predictive evidence of validity for paper and online tests for predictions of future test scores and grades. We generally find minimal differences for the extent to which scores on paper tests can differentially predict future online versus paper test scores. Finally, online and paper test scores are similarly predictive of future grade point average. We conclude that the online test penalty can vary substantially by test and that extreme care should be taken when administering online tests to some students and paper tests to others.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Notes

1 Previous work spanning several states, including Massachusetts, suggests that test scores across different standards or assessment regimes provide similar information about students’ underlying ability (Backes et al., Citation2018).

2 Boston, Worcester, and Springfield had the option of assigning individual schools to the online or paper format. Otherwise, districts selected a single test administration for the entire district. In November 2015, the Massachusetts State Board of Education voted to discontinue the PARCC assessment and implement a redeveloped version of the MCAS in all schools beginning in 2017.

3 originally displayed in Backes and Cowan (Citation2019). These sample tests may be found at https://dc.mypearsonsupport.com/practice-tests/.

4 As noted below, because some specifications use twice-lagged test scores, the actual underlying data used are 2010 through 2018.

5 This conversion was performed by the Massachusetts Department of Elementary and Secondary Education in order to ensure that PARCC scores could be used for accountability purposes along with MCAS scores. Briefly, this process was conducted in two steps. First, comparable representative samples of MCAS and PARCC students were selected by ensuring that the prior performance and demographic variables for each sample looked similar to the overall mean from the final year that each student took the MCAS; i.e., 2014. Second, an equipercentile method was used to identify comparable test scores across the two different tests using student achievement percentiles for each test. For the full technical description, please see Massachusetts Department of Elementary and Secondary Education (Citation2016).

6 For instance, test results under PARCC (2015 and 2016) are compared to test results using the MCAS 1.0 (2013 and 2014); similarly, test results under MCAS 2.0 (2017 and 2018) are compared to results using the PARCC (2015 and 2016).

7 All estimates were generated using StataCorp (Citation2021).

8 Schools could obtain a waiver to administer the MCAS tests on paper, but waivers were granted only for cases in which online testing was not possible at a school (e.g., due to lack of technology required) or if every student in the school would take a paper-based test due to testing accommodations. For more information, see http://www.doe.mass.edu/news/news.aspx?id=25169.

Additional information

Funding

This research was funded by IES Research Grant R305A170119 and supported by the National Center for the Analysis of Longitudinal Data in Education Research (CALDER), which is funded by a consortium of foundations. For more information about CALDER funders, see www.caldercenter.org/about-calder. The data were provided by the state of Massachusetts and the state had the right to review the paper prior to public release in order to ensure that the paper does not disclose any personally identifiable information provided by the state. We thank Elana McDermott, Bob Lee, Dan Goldhaber, and three anonymous referees for their feedback. All errors are our own.

Log in via your institution

Log in to Taylor & Francis Online

PDF download + Online access

  • 48 hours access to article PDF & online version
  • Article PDF can be downloaded
  • Article PDF can be printed
USD 53.00 Add to cart

Issue Purchase

  • 30 days online access to complete issue
  • Article PDFs can be downloaded
  • Article PDFs can be printed
USD 400.00 Add to cart

* Local tax will be added as applicable

Related Research

People also read lists articles that other readers of this article have read.

Recommended articles lists articles that we recommend and is powered by our AI driven recommendation engine.

Cited by lists all citing articles based on Crossref citations.
Articles with the Crossref icon will open in a new tab.