769
Views
0
CrossRef citations to date
0
Altmetric
Teaching Reproducibility

Teaching Reproducible Methods in Economics at Liberal Arts Colleges: A Survey

, &
Pages 296-302 | Published online: 15 Sep 2023

Abstract

Economics has become increasingly empirical and, alongside this shift, has come more demand for improved transparency and reproducibility in empirical economic research. In this article, we distribute a survey to almost 1500 economics faculty from the top 161 liberal arts colleges with an economics major (according to U.S. News & World Report) in the United States to determine the prevalence of teaching reproducible methods in undergraduate economics, summarize the most-common methods of instruction, and determine the intended student learning objectives. We find that of the economics faculty at liberal arts colleges who teach these reproducible methods, most do so in advanced upper-level (42%) and basic econometrics (31%) courses. Those faculty report teaching reproducibility using the following methods: transparent coding (85%), organizational skills (78%), and producing replication documentation (47%) through individual research projects (82%), homework assignments (55%), and/or workshops (33%). We conclude with some qualitative text analysis to shed light on the intended learning objectives and find that research skills (59%) and the importance of reproducibility (37%) are the most common reasons cited for teaching these methods.

1 Introduction

The lack of successful replicability and reproducibility in empirical research across the sciences and social sciences has received significant attention (Dewald, Thursby, and Anderson Citation1986; King Citation1995; Hamermesh Citation2007; Vilhuber Citation2021; Vilhuber et al. Citation2022).Footnote1 Approximately 90% of respondents to a recent survey conducted by Nature believe that there is a “slight” or “significant” replicability crisis (Baker Citation2016). In economics, researchers have found that less than half to two-thirds of papers studied are replicable (Camerer et al. Citation2016; Chang and Li Citation2017). The American Statistical Association (ASA) suggests that authors of research involving statistical analysis have an ethical responsibility to “promote sharing of data and methods” and “make documentation suitable for replicate analyses” available (ASA Citation2016). Thus, there have been increasing calls for more transparency by researchers (Bell and Miller Citation2013; Stodden et al. Citation2016; Coffman, Niederle, and Wilson Citation2017; Höffler Citation2017) and professional organizations including the National Academies of Sciences, Engineering, and Medicine and the National Science Foundation (Bollen et al. Citation2015).

Failure of research to be reproducible and/or replicable comes from a variety of sources including falsified data, data mining, search for statistical significance (“p-hacking”), validity and repeatability of experiments, incomplete documentation, and unintentional errors in data entry, cleaning, coding, or analysis. Reproducibility refers to the ability of a researcher to duplicate the results of a prior study using the same materials and procedures used by the original investigator. Reproducibility requires documentation that allows an independent researcher to reproduce every step of the data management and analysis process and replicate the results presented in the study (Ball and Medeiros Citation2012). Replicability refers to the ability of a researcher to duplicate the results of a prior study if the same procedures are followed but new data are collected (Bollen et al. Citation2015).Footnote2

In part, replicability in economics has become increasingly important as the discipline trends toward empirical research (Hamermesh Citation2013; Angrist et al. Citation2017). This pattern is also reflected in the undergraduate economics curriculum, with more courses in econometrics being required (6% in 1980 and 54% in 2020) and more degrees classified as “Econometrics and Quantitative Economics” (1% of economics degrees conferred in 2012 and 22% in 2019) (Siegfried and Wilkinson Citation1982; Marshall, Underwood, and Hyde Forthcoming; Marshall and Underwood Citation2020, Citation2022). The calls for increased transparency and trends in the discipline have led economic journals and organizations to promote policies requiring data and/or corresponding documentation for published work. (American Economic Association Citation2019, Citation2020; Vilhuber et al. Citation2022).

Despite the growing emphasis on replication documentation in empirical economic research, far less work has given primacy to the prevalence and assessment of reproducibility in teaching economics (Vilhuber Citation2020). There is little research on the degree to which transparency in research is taught in the economics classroom. In statistics and data science, where reproducibility is arguably more intrinsic, more is known, and more teaching resources are available (Çetinkaya-Rundel and Ellison Citation2021; Dogucu and Çetinkaya-Rundel Citation2022; Horton et al. Citation2022; Ostblom and Timbers Citation2022). In the social sciences, including economics, a small, but growing list of resources are available for teaching reproducibility and establishing pedagogical best practices (Orozco et al. Citation2020; Ball et al. Citation2022; Rethlefsen et al. Citation2022). Assessing how reproducible teaching methods are incorporated into undergraduate economics programs has been challenging, as has determining whether teaching reproducibility is beneficial to students’ understanding of data analysis and econometrics (Vilhuber Citation2020). Therefore, we examine the prevalence of reproducible methods in undergraduate liberal arts economics programs, as well as what reproducible methods are being taught, how instructors are teaching reproducible methods, and why reproducible methods are used.

2 Data and Methods

We survey nearly 1500 economics faculty from the top 161 liberal arts colleges (LACs) in the United States that offer an economics degree, as ranked and classified by the U.S. News and World Reports (USNWR) in June 2021.Footnote3 Our goal is to understand the prevalence of teaching reproducibility in undergraduate economics. We survey only LACs to manage the scope of the project and because economics professors at LACs may be more able to undertake new teaching methods with smaller class sizes, relative to professors at larger institutions. This is consistent with research showing that larger class sizes were found to be highly associated with specific pedagogical lecture styles and grading schema (Allgood, Walstad, and Siegfried Citation2015). However, large Ph.D.-granting universities are responsible for 75% of undergraduate economics degrees conferred every year (Marshall, Underwood, and Hyde Forthcoming), so this analysis is intended to be descriptive and should be viewed as a baseline for the prevalence of, and motivations for, teaching reproducible methods in economics. Participants were sent a link in the summer and fall of 2021 to an online survey asking whether they teach reproducible research methods in their courses, and if so, which methods of instruction are used and what are the intended learning objectives.Footnote4 The survey was distributed to 1474 potential respondents. 253 respondents agreed to participate and completed some or all the survey for a response rate of 17%. A copy of the survey instrument, copies of the distribution emails, and the data and code needed to reproduce our results are included in the supplementary material.

3 Results

Liberal arts colleges are often discussed in the context of their rankings and the rank of 50 is a common ranking threshold which, as others have suggested, may be particularly salient because it represents the last ranking on the first page in the printed version of the USNWR (Meyer, Hanson, and Hickman Citation2017). To that end, we organize many of the results according to this ranking threshold. We received responses from faculty at 98 of the top 161 LACs. Faculty from the most highly ranked LACs were more likely to respond to the survey. We received at least one response from 47 of the top 50 institutions and 51 responses from the remaining 111 institutions outside the top 50. Thus, 72% of respondents are from the top 50 and 44% are from the top 25 LACs. Thus, highly-ranked LACs may be over-represented in the sample.

This suggests the first of two likely limitations of the sample. First, more highly ranked institutions have larger economics faculty, so faculty from those institutions may be over-represented in the sample, as noted above. Based on our survey distribution list derived from faculty listings on institutional websites, institutions in the top 50 have around 17 economics faculty on average, while those outside the top 50 have about 9 economics faculty. Thus, as expected, we have 3.8 responses per institution in the top 50 and only 1.5 responses per institution outside the top 50. However, faculty response rates among those institutions with at least one response are identical inside and outside the top 50: 25% of faculty responded, on average. Second, faculty teaching these reproducible methods may be more likely to respond to a survey asking about it. This may lead to upwardly biased estimates of prevalence. Furthermore, if faculty at the top 50 LACs are more likely to teach these methods due to either higher incentives for quality teaching or lower teaching loads, then these limitations may be compounding and further bias our sample toward higher prevalence of teaching these methods.

We find that 51% of respondents (129 of 253 respondents) have taught some form of reproducible methods, as defined in Box 1, in at least one course over their past three years of teaching. This varies only slightly by institution rank, with 53% of respondents from the top 50 LACs teaching these methods and 45% outside the top 50. However, given the limitations of the sample noted above and a response rate of 17%, this is likely an overestimation of the prevalence of teaching reproducibility in economics at LACs. Only around 9% (129 of 1474 possible survey participants) of those that were sent the survey noted that they teach reproducible methods, so actual prevalence at LACs is likely between 9% and 51%.

Of those faculty teaching these methods, we asked in which course(s) they teach these methods, which methods they teach, and which teaching tools they use to integrate these methods into their course(s). Faculty teaching these methods, reported doing so in 167 total courses, for an average of 1.3 courses per faculty respondent; however, most (74%) teach reproducible methods in only one course.

Box 1. What are reproducible methods?

Reproducibility refers to the ability of a researcher to duplicate the results of a prior study using the same materials and procedures as were used by the original investigator. Replicability refers to the ability of a researcher to duplicate the results of a prior study if the same procedures are followed but new data are collected (Vilhuber Citation2020).

Despite teaching these methods in 167 total courses, respondents only included complete course information, including course type and title, for 137 courses. shows the most common courses and how this varies by institution rank. In the full sample, these methods are mostly taught in courses other than econometrics and statistics. Courses in the “other” category primarily include advanced electives (many with econometrics prerequisites), research methods courses (including honors theses), and senior seminar/capstone courses.Footnote5 Faculty at the top 50 LACs are most likely to teach these methods in these other courses, while faculty outside the top 50 are most likely to teach these methods in econometrics.

shows which methods are most prevalent, how often faculty teach more than one method, and how this varies by institution rank. Most respondents teach coding for reproducibility (85%) and organizational skills (78%) while teaching how to produce replication documentation is also quite common (47%). This does not vary much by ranking, except for organizational skills which are more commonly taught at LACs outside the top 50 (but are very common at all institutions) and “other” methods which are more common among the top 50 (pre-analysis plans and pre-registration were the mostly commonly cited method among those few responses).Footnote6 The vast majority of respondents (82%) teach more than two of the listed methods in their courses and 14% report teaching four or more.

The methods used are consistent across course types but there are some notable differences, as shown in . While coding for reproducibility remains the most common method taught, this varies from 79% in advanced econometrics up to 94% in introductory statistics. More notably, teaching organizational skills varies from only two-thirds of basic econometrics courses to 94% of introductory statistics courses, likely reflecting the vertical nature of these skills. Students are taught these skills at lower levels of the curriculum and likely are assumed as prior knowledge at upper-levels.

Table 3 Methods by course type.

Similarly, more advanced research skills, such as using dynamic documents and facilitating accessibility and sharing (GitHub, Open Science Framework, Dataverse, etc.) are more prevalent in more advanced courses. The clustering of the two most common methods, coding for reproducibility and organizational skills, is also consistent with the vertical nature of these skills. Across all courses, these two methods are taught together 70% of the time, likely due to their complementarity. However, these are taught together more often in introductory statistics (89%) compared to advanced econometrics (60%) courses.

Faculty primarily teach these methods through student projects or research papers (82%) but homework exercises (55%), workshops or labs (53%), and group work (45%) are also common methods of instruction. shows how these instructional methods vary by course type. Student projects or research papers are most common in both basic econometrics (79%) and other (88%) courses, while workshops or labs are most common in advanced econometrics (78%) and introductory statistics (72%) courses. These differences may reflect differences in the expected student learning outcomes.

Table 4 Teaching tools by course type.

To gauge motivation for teaching these methods, faculty respondents were asked, “Why do you teach these methods in this course? In other words, what are the student learning objectives?” and “What skills do you want students to learn through these methods? Why might they be useful in the future, regardless of vocational choice?” and were able to provide an open-ended response. Of the 253 survey responses, 104 faculty answered at least one of these questions. We combined the qualitative responses to these two questions and then tagged them with up to seven possible categories via inductive coding, as shown in .Footnote7 The most common reason cited for teaching these methods is for students to learn research skills and methods (59%) followed by the general importance of reproducibility (37%). There are some slight differences in these motivations and desired learning outcomes based on institution rank. Faculty outside the top 50 are slightly more likely to list data, organization, and career skills (44% compared to 20%) in discussing the intended learning outcomes through these methods.

Table 5 Learning objectives.

The resulting learning objectives closely reflect the findings of the National Academies of Sciences, Engineering, and Medicine (Citation2018) on instilling data acumen as the most critical task in educating future data scientists. The key concepts involved in developing data acumen include “mathematical foundations, computational foundations, statistical foundations, data management and curation, data description and visualization, data modeling and assessment, workflow and reproducibility, communication and teamwork, domain-specific consideration, ethical problem solving” (National Academies of Sciences, Engineering, and Medicine Citation2018). These similarities suggest that economics faculty may look to data science educators and curricula for ways to incorporate reproducible methods more widely in the economics curriculum.

4 Conclusion

Transparency and reproducibility in empirical economic research has gained more visibility and traction over the last decade. Educating current and future generations of economists is crucial for continued improvement in the reproducibility of economic research (Vilhuber Citation2021). To achieve this progress, educators have a responsibility to set high standards of reproducibility through the norms and incentives they provide to students. This has the potential to create a “trickle-up” effect on the discipline more broadly (Ball and Medeiros Citation2012; Ball et al. Citation2022). More and more undergraduate students are developing data and empirical skills across a variety of disciplines and thus the importance of reproducibility will continue to grow (Schwab-McCoy, Baker, and Gasper Citation2021; Vilhuber et al. Citation2022).

While some explicit training in reproducible methods in economics has emerged over the past decade, no systematic survey of economic educators had been conducted on this topic. We survey nearly 1500 economics faculty from the top 161 liberal arts colleges with an economics major in the United States and find that 51% of survey respondents teach these reproducible methods, mostly in advanced upper-level electives (42%) and basic econometrics (31%) courses.

These results suggest that while some educators value and prioritize teaching reproducible methods, there is room for improvement and continued outreach to facilitate and encourage instructors to incorporate these methods in their teaching. The similarities in the desired learning objectives noted here and those proposed as necessary to develop data acumen in future data scientists suggests that economic educators have many opportunities to collaborate and develop pedagogical best practices alongside faculty in data science and other adjacent fields. To what extent these methods are taught at other types of institutions, in graduate programs, and in other empirical disciplines needs further exploration. Finally, the question of whether teaching reproducible methods at the undergraduate level trickles up to the field more broadly remains open.

Supplementary Materials

A copy of the survey instrument, copies of the distribution emails, and the data and code needed to reproduce our findings are available here: https://figshare.com/s/ae6ec51b730d6e7564f5.

Acknowledgments

We would like to thank Richard Ball for helpful input and feedback on the survey instrument and thorough feedback on an early draft of this manuscript. We also thank our discussant and participants at the 2022 Conference on Teaching and Research in Economic Education (CTREE) and the 2022 Southern Economics Association Annual Meeting for their valuable feedback. Finally, we thank the editor, associate editor, and several reviewers whose feedback substantially improved the clarity of this manuscript.

Data Availability Statement

The authors confirm that a copy of the survey instrument, copies of the distribution emails, and the data and code needed to reproduce our findings are available here: https://figshare.com/s/ae6ec51b730d6e7564f5.

Disclosure Statement

No potential conflict of interest was reported by the authors.

Notes

1 There is also a smaller literature that defends the credibility of scientific research claiming that the prevalence of problematic research is less than currently perceived (Fanelli Citation2018). However, they do not deny the replicability and reproducibility issues altogether.

2 Replicability can be statistical or scientific. Statistical replication examines the same population with the same model but a different sample; while scientific replication, studying a different population and sample, is more relevant for economists (Hamermesh Citation2007).

3 All faculty (including visiting and adjunct faculty but excluding emeritus faculty) listed on the institutional economics department websites were included.

4 The initial email was sent on June 24, 2021, using Qualtrics. Weekly reminders were sent until the survey closed on July 17, 2021. We then resent the survey on October 26, 2021 and sent one additional reminder. The protocol 973. Teaching Reproducibility in Undergraduate Economics was verified by the Dickinson College Institutional Review Board as Exempt according to 45CFR46.101(b)(1): (1) Educational Practices on June 29, 2021.

5 In the survey, we distinguish between these course types by listing the course options in this way: “(a) an introductory statistics course students take before econometrics; (b) a basic econometrics course; (c) an advanced statistics or econometrics course, taken after basic econometrics; or (d) other. Please specify.” For those respondents who selected “other” courses, the course titles (when provided) are listed in Appendix B, Table B1. Additionally, we cannot distinguish between economic statistics courses (typically taught by economics faculty) and probability and statistics courses (typically taught by mathematics/statistics faculty). Given that our respondents are economics faculty, it is likely that these are primarily economic statistics courses, but we do not have that information in our sample.

6 These “other” methods include: Reproducing the results of well-known articles in econometrics; unit root and cointegration analysis and forecasting; preregistration of experiments and data analysis plans; Stata do-files; discussing the idea of reproducibility as a research standard; pre-analysis plans, pre-registration; discussion of replication studies and replication of a research study; version control; research plans and self-memos.

7 See Appendix A for some common phrases included in each category.

References

  • Allgood, S., Walstad, W. B., and Siegfried, J. J. (2015), “Research on Teaching Economics to Undergraduates,” Journal of Economic Literature, 53, 285–325. DOI: 10.1257/jel.53.2.285.
  • American Economic Association. (2019), “Data and Code Availability Policy.” Retrieved January 11,2022, available at https://www.aeaweb.org/journals/policies/data-code.
  • American Economic Association. (2020), “Data and Code Availability Policy,” AEA Papers and Proceedings, 110, 776–778. https://www.jstor.org/stable/27042656
  • American Statistical Association (ASA). (2016), Ethical Guidelines for Statistical Practice. Alexandria, VA: ASA, Committee on Professional Ethics. Available at http://www.amstat.org/ASA/Your-Career/Ethical-Guidelines-for-Statistical-Practice.aspx.
  • Angrist, J., Azoulay, P., Ellison, G., Hill, R., and Lu, S. F. (2017), “Economic Research Evolves: Fields and Styles,” American Economic Review, 107, 293–297. DOI: 10.1257/aer.p20171117.
  • Baker, M. (2016), “1,500 Scientists Lift the Lid on Reproducibility,” Nature, 533, 452–454. DOI: 10.1038/533452a.
  • Ball, R., and Medeiros, N. (2012), “Teaching Integrity in Empirical Research: A Protocol for Documenting Data Management and Analysis,” The Journal of Economic Education, 43, 182–189. DOI: 10.1080/00220485.2012.659647.
  • Ball, R., Medeiros, N., Bussberg, N. W., and Piekut, A. (2022), “An Invitation to Teaching Reproducible Research: Lessons from a Symposium,” Journal of Statistics and Data Science Education, 30, 209–218. DOI: 10.1080/26939169.2022.2099489.
  • Bell, M., and Miller, N. (2013), “How to Persuade Journals to Accept Your Replication Paper,” available at https://politicalsciencereplication.wordpress.com/2013/09/11/guest-blog-how-to-persuade-journals-to-accept-yourreplication-paper/
  • Bollen, K., Cacioppo, J. T., Kaplan, R. M., Krosnick, J. A., and Olds, J. L. (2015), Social, Behavioral, and Economic Sciences Perspectives on Robust and Reliable Science [Report of the Subcommittee on Replicability in Science Advisory Committee to the National Science Foundation Directorate for Social, Behavioral, and Economic Sciences]. National Science Foundation. Available at https://www.nsf.gov/sbe/AC_Materials/SBE_Robust_and_Reliable_Research_Report.pdf.
  • Camerer, C. F., Dreber, A., Forsell, E., Ho, T.-H., Huber, J., Johannesson, M., Kirchler, M., Almenberg, J., Altmejd, A., Chan, T., Heikensten, E., Holzmeister, F., Imai, T., Isaksson, S., Nave, G., Pfeiffer, T., Razen, M., and Wu, H. (2016), “Evaluating Replicability of Laboratory Experiments in Economics,” Science, 351, 1433–1436. DOI: 10.1126/science.aaf0918.
  • Çetinkaya-Rundel, M., and Ellison, V. (2021), “A Fresh Look at Introductory Data Science,” Journal of Statistics and Data Science Education, 29, S16–S26. DOI: 10.1080/10691898.2020.1804497.
  • Chang, A. C., and Li, P. (2017), “A Preanalysis Plan to Replicate Sixty Economics Research Papers That Worked Half of the Time,” American Economic Review, 107, 60–64. DOI: 10.1257/aer.p20171034.
  • Coffman, L. C., Niederle, M., and Wilson, A. J., (2017), “A Proposal to Organize and Promote Replications,” American Economic Review, 107, 41–45. DOI: 10.1257/aer.p20171122.
  • Dewald, W., Thursby, J., and Anderson, R. (1986), “Replication in Empirical Economics: The Journal of Money, Credit and Banking Project,” American Economic Review, 76, 587–603. https://www.jstor.org/stable/1806061.
  • Dogucu, M., and Çetinkaya-Rundel, M. (2022), “Tools and Recommendations for Reproducible Teaching,” Journal of Statistics and Data Science Education, 30, 251–260. DOI: 10.1080/26939169.2022.2138645.
  • Fanelli, D. (2018), “Opinion: Is Science Really Facing a Reproducibility Crisis, and Do We Need It to?,” Proceedings of the National Academy of Sciences of the United States of America, 115, 2628–2631. DOI: 10.1073/pnas.1708272114.
  • Hamermesh, D. S. (2007), “Viewpoint: Replication in Economics,” Canadian Journal of Economics/Revue Canadienne D’économique, 40, 715–733. DOI: 10.1111/j.1365-2966.2007.00428.
  • Hamermesh, D. S. (2013), “Six Decades of Top Economics Publishing: Who and How?” Journal of Economic Literature, 51, 162–172. DOI: 10.1257/jel.51.1.162.
  • Höffler, J. H. (2017), “Replication and Economics Journal Policies,” American Economic Review, 107, 52–55. DOI: 10.1257/aer.p20171032.
  • Horton, N. J., Alexander, R., Parker, M., Piekut, A., and Rundel, C. (2022), “The Growing Importance of Reproducibility and Responsible Workflow in the Data Science and Statistics Curriculum,” Journal of Statistics and Data Science Education, 30, 207–208. DOI: 10.1080/26939169.2022.2141001.
  • King, G. (1995), “Replication, Replication,” PS: Political Science & Politics, 28, 444–452. DOI: 10.2307/420301.
  • Marshall, E. C., and Underwood, A. (2020), “Is Economics STEM? Trends in the Discipline from 1997 to 2018,” The Journal of Economic Education, 51, 167–174. DOI: 10.1080/00220485.2020.1731387.
  • Marshall, E. C., and Underwood, A. (2022), “Is Economics STEM? Process of (Re)Classification, Requirements, and Quantitative Rigor,” The Journal of Economic Education, 53, 250–258. DOI: 10.1080/00220485.2022.2075508.
  • Marshall, E. C., Underwood, A., and Hyde, A. (Forthcoming), “Requirements of the Undergraduate Economics Major: An Update and Comparison,” Journal of Economic Education.
  • Meyer, A. G., Hanson, A. R., and Hickman, D. C. (2017), “Perceptions of Institutional Quality: Evidence of Limited Attention to Higher Education Rankings,” Journal of Economic Behavior & Organization, 142, 241–258. DOI: 10.1016/j.jebo.2017.07.036.
  • National Academies of Sciences, Engineering, and Medicine. (2018), Data Science for Undergraduates: Opportunities and Options, Washington, D.C.: The National Academies Press. DOI: 10.17226/25104.
  • National Academies of Sciences, Engineering, and Medicine. (2019), Reproducibility and Replicability in Science, Washington, D.C.: National Academies Press. DOI: 10.17226/25303.
  • Orozco, V., Bontemps, C., Maigné, E., Piguet, V., Hofstetter, A., Lacroix, A., Levert, F., and Rousselle, J. M. (2020), “How to Make a Pie: Reproducible Research for Empirical Economics and Econometrics,” Journal of Economic Surveys, 34, 1134–1169. DOI: 10.1111/joes.12389.
  • Ostblom, J., and Timbers, T. (2022), “Opinionated Practices for Teaching Reproducibility: Motivation, Guided Instruction and Practice,” Journal of Statistics and Data Science Education, 30, 241–250. DOI: 10.1080/26939169.2022.2074922.
  • Rethlefsen, M., Norton, H. F., Meyer, S. L., MacWilkinson, K. A., Smith, P., and Ye, H. (2022), “Interdisciplinary Approaches and Strategies from Research Reproducibility 2020: Educating for Reproducibility,” Journal of Statistics and Data Science Education, 30, 219–227. DOI: 10.1080/26939169.2022.2104767.
  • Schwab-McCoy, A., Baker, C. M., and Gasper, R. E. (2021), “Data Science in 2020: Computing, Curricula, and Challenges for the Next 10 Years,” Journal of Statistics and Data Science Education, 29, S40–S50. DOI: 10.1080/10691898.2020.1851159.
  • Siegfried, J. J., and Wilkinson, J. T. (1982), “The Economics Curriculum in the United States: 1980,” American Economic Review, 72, 125–138. https://www.jstor.org/stable/1802317.
  • Stodden, V., McNutt, M., Bailey, D. H., Deelman, E., Gil, Y., Hanson, B., Heroux, M. A., Ioannidis, J. P. A., and Taufer, M. (2016), “Enhancing Reproducibility for Computational Methods,” Science, 354, 1240–1241. DOI: 10.1126/science.aah6168.
  • Vilhuber, L. (2020), “Reproducibility and Replicability in Economics,” Harvard Data Science Review, 2. DOI: 10.1162/99608f92.4f6b9e67.
  • Vilhuber, L. (2021), “Report by the AEA Data Editor,” AEA Papers and Proceedings, 111, 808–817. DOI: 10.1257/pandp.111.
  • Vilhuber, L., Son, H. H., Welch, M., Wasser, D. N., and Darisse, M. (2022), “Teaching for Large-Scale Reproducibility Verification,” Journal of Statistics and Data Science Education, 30, 274–281. DOI: 10.1080/26939169.2022.2074582.
  • Vilhuber, L., Connolly, M., Koren, M., Llull, J., and Morrow, P. (2020), A Template README for Social Science Replication Packages [Zenodo Version Number: v1.0.0]. DOI: 10.5281/ZENODO.4319999.

Appendix A:

Learning Objectives

The categories listed below are the result of qualitative inductive coding of the (n = 104) open-ended responses received for questions 10 and 11 of the survey from those faculty that noted teaching reproducible methods in at least one course over the past three years. The phrases listed under each category are those common and/or repeated phrases from responses coded in each category. Responses were categorized in all applicable categories, up to seven possible, however, this varies from only 1 to a maximum of 5 in the sample.

10: Why do you teach these methods in this course? In other words, what are the student learning objectives?

11: What skills do you want students to learn through these methods? Why might they be useful in the future, regardless of vocational choice?

Data management and analysis skills

Data analysis

Big Data

Data management

Cleaning data

Data wrangling

Organizational Skills

File management

Documentation methods

Organizational skills

Working directories

Relative directory/file paths

Codebook development

Coding/Software/Syntax

Coding skills

Programming skills

Basic coding

Careful coding

Commenting code

Programming skills

The Importance of Reproducibility

Best practices

Reproducible methods

Reproducible results

Replicability

Accountability

Research Skills/Methods

Research methods

Formal research paper

Empirical investigation

Scientific research

Original research

Quantitative reasoning

Empirical research

Empirical methods

Research integrity

Research skills

Literature review

Econometrics/Statistical Skills

Interpretation skills

Econometric methods

Forming testable hypothesis

Interpreting multivariate regressions

Testing for significance

Addressing endogeneity issues

Generating summary statistics

Running regressions

Creating graphs

Statistical analysis

Descriptive statistics

Data visualization

Career/Graduate School Prep

Meet current expectations

Needed skills

Appendix B:

Other Courses

lists course titles, when provided by respondents, for courses where reproducible methods are taught, which are included in the “Other” category in = 57). Some respondents provided course titles, some provided course numbers only, and some provided neither. There were also a few duplicate course titles which we have only included here once. As a result, we list 39 course titles below.

Table B1 Other courses.