In Short

  • Institutions can begin to answer if they have a diverse enrollment and accessible entry pathways by examining (1) the diversity of STEM enrollees and (2) the success rate of STEM transferees, disaggregated by relevant demographic factors.

  • They can assess the (3) adoption of evidence-based educational practices and (4) equitable participation in high-impact STEM practices to know if they have equitable learning environments that foster learning for all students.

  • Institutions have evidence that they are fostering inclusion and belonging for all STEM students if (5) STEM students report feeling included and supported and (6) students persist in STEM.

  • Finally, institutions can know that they are ensuring equitable outcomes if (7) there are no equity gaps in the percentage of students who attain STEM credentials or (8) longer time-to-degree for diverse STEM students (see ).

In an era where STEM expertise is desperately needed to help solve critical world challenges, ensuring the success of students in STEM, particularly those who have been historically marginalized, underserved, and/or excluded, has never been more pressing. In response to this challenge, the authors, informed by the work of the National Academies for Sciences, Engineering, and Medicine (NASEM) Roundtable on Systemic Change in Undergraduate STEM Education (NASEM, Citationn.d.), offer key guidance for educators from 2-year and 4-year colleges and universities who want to better assess equitable student success in STEM. While we have focused on STEM in this article, this guidance can be applied to all disciplines, academic programs, and institution-wide efforts.

In this article, we provide four reflective questions and eight measurable indicators to help guide the development and refinement of how equity is being measured. The questions serve as a compass for institutions seeking to better support their STEM students. We provide indicators and evidence for each question drawn from NASEM’s Indicators for Monitoring Undergraduate STEM Education (NASEM, Citation2018). Our goal is to offer a concise, targeted resource that fuels data-informed conversations within your academic program, department, college, or university regarding STEM student support.

Figure 1. Eight Equity Indicators for Institutions to Ensure Equitable Stem Student Success, Adapted from the National Academies of Sciences, Engineering, and Medicine (2018)

Figure 1. Eight Equity Indicators for Institutions to Ensure Equitable Stem Student Success, Adapted from the National Academies of Sciences, Engineering, and Medicine (2018)

For this article, we define equitable student success as fostering diversity, nurturing an equitable and inclusive learning environment, cultivating a sense of belonging, and actively eliminating bias in student outcomes. This definition and the associated indicators build on the data required by the U.S. Department of Education (e.g., student admissions, enrollment, retention rate, and graduation rate; Integrated Postsecondary Education Data System, Citationn.d.) and emphasize the importance of additional data and evidence for a more holistic assessment of student success. We invite institutional leaders at all levels to embrace our broader definition of equitable student success and to cultivate a culture of data collection and use to ensure equitable outcomes for students pursuing STEM fields.

A key principle underlying this guidance is that institutions should analyze student experiences and student outcomes by examining the data and evidence that are available at their institution. Where feasible, this evidence can be disaggregated by demographic categories standardized across institutional units (e.g., admissions, departments, and financial aid offices). Further, we urge institutions to explore these data at the intersections of race, ethnicity, gender, and other demographic characteristics, where possible. If institution-wide data are not available for an indicator, an institution can consider a smaller-scale data collection activity and/or conduct a focus group with faculty and students. While there is no one-size-fits-all approach to data disaggregation, institutions may consider demographic categories including, but not limited to, race, ethnicity, gender, gender expression, sexual orientation, religion, language, geographic region, ability status, parenting status, socioeconomic status, and veteran status, as well as whether students are full time or part time, first generation, or have transferred for their data collection. Intersections of these categories are important to examine where possible while protecting student privacy. Institutions can weigh which of these categories are available for analysis, most salient for their student populations, and may be most relevant for examining equity gaps (i.e., in enrollment, STEM graduation rates, access to research experiences, or time-to-degree disparities).

Additionally, a cross-campus approach with an integrated series of supports is important. Leaders, including presidents, provosts, vice provosts of undergraduate education, executive vice presidents and vice presidents of instruction, vice presidents of student affairs, and deans, provide powerful and meaningful leadership when they champion the collection and use of student data for program improvement and appropriately resource these efforts. Through collaboration with other stakeholders, they can establish institutional targets for program outcomes to increase accountability, celebrate success, and identify areas for future growth.

As research has amply demonstrated, a top-down approach alone is not sufficient (Holcombe et al., Citation2021). An effective collaborative approach requires the engagement of institutional data stewards or institutional research officers, directors of student success who have data on student engagement beyond the classroom, department leaders entrusted with overseeing undergraduate courses, Centers for Teaching and Learning, and STEM Education Center staff (Carlisle & Weaver, Citation2018, Citation2020; Horii et al., Citation2015; Wright et al., Citation2018) who can help facilitate reflective dialog (Kruse & Louis, Citation1993) and implementation of evidence-based and inclusive pedagogical practices. Department chairs and faculty, including Visiting faculty, Instructors, Teaching Assistants, teaching professors, adjunct faculty, and Lecturers or VITAL faculty (Lee et al., Citation2023; Levy, Citation2019), are central to this effort. STEM faculty interact most closely with their STEM students as classroom instructors, research and career mentors, and often as academic advisors. Finally, campuses would benefit from including students in these reflective dialogs to ensure that the desired outcomes of the STEM students are being met (Ellis-Nelson et al., Citation2023). These dialogs can also be an educational opportunity for students to build their data expertise.

Reflective Questions, Indicators, and Evidence

These reflective questions and eight indicators provide a structured approach for institutions to identify areas for improvement and to support student success effectively. The indicators have been adapted from NASEM’s Indicators for Monitoring Undergraduate STEM Education (Citation2018); we indicate the alignment of each indicator to this report in brackets. Undoubtedly, different institutions will focus on different indicators and types of evidence that are best aligned with their institutional mission, available resources, student populations served, and local context.

Reflective Question 1: Does your institution, college, and/or STEM department/academic program encourage a diverse and equitable enrollment, and establish accessible and supportive entry pathways for all potential students, particularly into STEM programs?

Indicator 1: Diversity of enrollees in STEM programs compared to the diversity of enrollees across academic fields. [Aligned with Indicators 2.2.1 and 2.2.2]

Evidence:

  • If available, review historical data showing trends in the diversity of students enrolled in STEM versus non-STEM degree or certificate programs over a specified period.

  • Examine disaggregated demographic data showing the percentages of students of differing race, ethnicity, gender, transfer status, first-generation status, socioeconomic status, and other relevant backgrounds and identities who enroll in STEM certificate and degree programs versus those enrolled in non-STEM programs. These data should allow for an intersectional analysis.

Indicator 2: Outcomes of STEM transfer students compared to all transfer students (STEM and non-STEM programs at the institution). [Aligned with Indicator 3.2.2]

Evidence:

  • Students are increasingly transferring between institutions. Regardless of the transfer path, determine the ratio of students who transfer into the institution and declare a STEM-related major to the total number of students who transferred to the institution.

  • Evaluate the acceptance rate of transfer credits, especially in STEM fields, for incoming transfer students. Disaggregate these data for deeper analysis.

  • Compare the performance of transfer students in courses where they completed the prerequisite at a different institution with students who completed the prerequisite at their current institution.

  • Compare the use of support services, such as advising, orientation programs, and learning communities, by STEM transfer students to all transfer students and to all STEM students.

Reflective Question 2: Does your institution, college, and/or STEM department/academic program develop equitable learning environments that foster learning for all students in STEM programs?

Indicator 3: Extent of the implementation of evidence-based educational practices in STEM courses. [Aligned with Indicator 1.1.1]

Evidence:

  • Gather survey data on student perceptions of the learning environment, including their engagement and satisfaction in STEM courses.

  • Evaluate whether resources, rewards, and accountability systems are structured to encourage and sustain instructors’ use of evidence-based practices (TEval, Citationn.d.). Look for indications of institutional recognition programs for teaching, funds allocated for faculty professional development in teaching, and usage of validated measures (such as observation protocols and teaching portfolios) that go beyond just student evaluations to assess teaching performance. Evidence may include the following:

  • Percentage of instructors (tenure track and nontenure track) who report annual engagement in education-related professional development.

  • Average funding available for instructional development and support per instructor.

  • Percentage of instructors who report using evidence-based instructional practices to support student learning.

  • Percentage of instructors who are observed to use evidence-based instructional practices as documented by observation protocols such as Classroom Observation Protocol for Undergraduate STEM (Smith et al., Citation2013), Reformed Teaching Observation Protocol (RTOP; Sawada et al., Citation2002), Teaching Dimensions Observation Protocol (Hora et al., Citation2013), and Practical Observation Rubric to Assess Active Learning (PORTAAL; Eddy et al., Citation2015).

  • Assess the extent to which departments consider the use of evidence-based teaching in decisions for hiring, merit, retention, and promotion.

Indicator 4: Extent to which there is equitable student access to and participation in evidence-based educational programs and experiences. [Aligned with Indicator 2.1.3]

Evidence:

  • Determine if the institution or department has clearly defined minimum criteria for educational experiences and methods for monitoring access to these experiences (Estrada, Citation2014). Key data points could include the following:

  • Percentage of students who have engaged in an experiential learning experience such as many of the Association of American Colleges & Universities’s High Impact Practices (Citation2023; Kuh, Citation2008) or other authentic experiential learning experiences as defined by the institution. Examples include course-based research experiences, National Science Foundation’s Community College Undergraduate Research Initiative (Citationn.d.), and first-year research initiatives aimed at ensuring every STEM student has access to an authentic research experience.

  • Percentage of students who have an assigned mentor with whom they meet regularly.

  • Percentage of departmental and/or institutional funds allocated for this work and/or for the students to engage in these opportunities.

Reflective Question 3: Does your institution, college, and/or STEM department/academic program foster a sense of inclusion and belonging for all students in your program?

Indicator 5: Extent to which students feel included and have a sense of belonging in their STEM academic programs. [Aligned with Indicator 2.4.1]

Evidence:

  • Assess the extent to which the curriculum in STEM programs is inclusive and culturally responsive by reviewing the curriculum and syllabi to identify the inclusion of diverse perspectives, theories, or applications in STEM (Estrada et al., Citation2018; Student Experience Project, Citation2022).

  • Evaluate the accessibility of learning materials and resources by reviewing course materials and learning resources for accessibility (considering disabilities, language proficiency, etc.).

  • Disaggregate data on the recipients of financial support to enable intersectional analysis.

  • Evaluate the extent to which classrooms are inclusive and support diversity through surveys or focus groups to measure students’ perceptions of classroom inclusivity.

  • Survey students on their satisfaction, sense of belonging, and perspectives, which could include regularly conducting climate surveys on students’ belonging, support systems, and access to resources within their program and/or institution.

  • Investigate faculty and staff professional development efforts focused on fostering an inclusive and supportive classroom environment.

Indicator 6: Extent to which students pursuing STEM credentials and degrees persist in STEM programs, from course to course, and year to year. [Aligned with Indicators 2.1.2 and 3.2.1]

Evidence:

  • Track incoming students’ degree aspirations and the declaration of STEM degrees through surveys or other measures capturing the specific STEM fields incoming students intend to pursue. Analyze trends over time in degree aspirations among different demographic groups.

  • Calculate the number of students, disaggregated by demographic characteristics, who switch their declaration away from a STEM major, and at what point in their academic career this occurs.

  • Investigate barriers or challenges faced by students who switch their declaration away from a STEM major, such as course availability, academic advising, or particular courses that have high Drop/Fail/Withdraw rates.

  • Monitor if students stay in their declared STEM major, their course load per semester, Grade Point Average, and progression/grades in subsequent courses.

  • Track the success rates (grades, class completions) of STEM students over time, disaggregated by demographic categories and other relevant factors.

  • Review student support services usage data (tutoring, academic advising, mental health services) to see if there is a correlation between service usage and retention in STEM programs.

Reflective Question 4: Does your institution, college, and department/academic program ensure equitable outcomes for all students, particularly in STEM programs?

Indicator 7: Percentage of students across demographic categories who complete a credential or degree in STEM compared to the percentage of students who complete a credential or degree across academic fields. [Aligned with Indicator 3.3.1]

Evidence:

  • Compare the graduation rates of STEM students with the institution’s overall graduation rates to identify potential gaps or disparities. Examine the data disaggregated by demographic categories and attend to relevant and available intersectional analyses.

  • Monitor the number and percentage of students who complete a STEM degree or certificate program.

  • For students who begin in a 2-year program and transition into a 4-year program, track their degree completion rates compared to 4-year-only students.

  • Examine postgraduation metrics, such as job placements in STEM fields, graduate-school acceptances, and other indicators of success in the field.

Indicator 8: Time-to-degree for students in STEM programs compared to time-to-degree for students across academic programs, disaggregated by demographic categories and other relevant factors. [Aligned with Indicator 2.2.3]

Evidence:

  • Calculate the time required for students to earn their STEM degrees, measured by time to degree at 100 percent (2 years for 2-year institutions or 4 years for 4-year institutions), 150 percent, and 200 percent. Disaggregate the data by race, ethnicity, gender, transfer status, first-generation status, socioeconomic status, and other relevant characteristics. Assess whether there is a correlation between longer time-to-degree and other factors, such as part-time enrollment, work commitments, or family responsibilities.

  • Calculate the average academic terms (semesters or quarters) needed to obtain a STEM degree or credential, disaggregated by relevant factors.

An Example of These Reflective Questions in Action

To effectively engage with the reflective questions, one valuable approach is to use the Partnership for Undergraduate Life Sciences (PULSE) tools (Brancaccio-Taras et al., Citation2016; PULSE Rubrics, Citation2023). During the 2022–2023 academic year, the University of Iowa College of Engineering embarked on a journey, prompted by then-dean Nembhard, to discuss and assess the current state of affairs within each department (Vigmostad, Citation2023). Specifically, they were tasked with assessing their department’s current consensus score for the 13 criteria on the PULSE Diversity, Equity, and Inclusion (DEI) Rubric. This rubric encompasses four distinct categories:

  • Curriculum (aligned with Reflective Question 1).

  • Assessment (aligned with Reflective Question 2).

  • Faculty Practice and Support (aligned with Reflective Questions 2 and 3).

  • Climate for Change (aligned with Reflective Question 4).

To facilitate this assessment, faculty discussion of the PULSE DEI Rubric occurred at departmental faculty meetings led by the department chairs or the chair of the department’s DEI committee. These discussions were useful in guiding the faculty to the critical juncture where they sought the empirical data required to address the reflective questions effectively, resulting in a data-driven approach to enhancing DEI within the College of Engineering.

Additional Resources and Tools

  • The National Academies of Sciences, Engineering, and Medicine have three reports that will be useful for readers: Barriers and Opportunities for 2-year and 4-year STEM Degrees (Citation2016); Indicators for Monitoring Undergraduate STEM Education (Citation2018); and Advancing Antiracism, Diversity, Equity, and Inclusion in STEMM Organizations (Citation2023).

  • The Sloan Equity and Inclusion in STEM Introductory Courses (SEISMIC) Equity Learning Communities Report (Citationn.d.) outlines how demographic student data might be analyzed for insights into course outcomes, grade distribution patterns, and grade anomalies. They provide source code via GitHub so that institutions can run their analysis.

  • The Association of Public and Land-Grant Universities’s Data Maturity Index has questions and criteria that can help institutional leaders evaluate how advanced their institutional data culture is for data governance, quality, analysis, dissemination, and application (Powered by Publics, Citation2023).

  • The National Institutes of Health and the Howard Hughes Medical Institute convened the Joint Working Group on Improving Underrepresented Minorities (URMs) Persistence in STEM, which made five recommendations for institutions to increase URM persistence in STEM (Estrada et al., Citation2016). The first recommendation was to track success and failure and collect data to increase institutional accountability.

  • There is a growing body of research on how to support departmental change that includes departmental change principles (Quan et al., Citation2019; Reinholz et al., Citation2017), frameworks (Corbo et al., Citation2016; Reinholz & Apkarian, Citation2018), and approaches (Andrews et al., Citation2021; Ngai et al., Citation2020; Reinholz et al., Citation2017, Citation2019).

  • In addition to the department-level resources from PULSE (PULSE Rubrics, Citation2023) in the life sciences, the physics community has also developed a comprehensive set of resources for physics departments in their Effective Practices for Physics Programs (EP3) Guide and Toolkits for engaging in systemic improvement (EP3 Guide, Citation2023).

Final Remarks

We invite institutional leaders at all levels to champion a culture of smart data use in pursuit of equitable student success. The reflective questions, indicators, and evidence outlined in this document serve as a guide for your institution and/or department to assess how well you are serving STEM students. As a starting place for this transformative journey, NASEM recommended benchmarking the demographic representation of STEM undergraduate degree/certificate earners compared to the demographic representation of all undergraduate graduates during the same period within the institution. NASEM recommended this comparison group because these students have successfully navigated higher education. However, this is merely a benchmark; your institution, college, department, or academic program may want to benchmark against peer institutions or your institution may want to choose a more ambitious target that defines successfully and equitably serving all students.

Acknowledgments

We thank Archie Holmes, Susan Singer, Mark Mitsui, and the members and staff of the National Academies for Sciences, Engineering, and Medicine Roundtable on Systemic Change in Undergraduate STEM Education for their contributions in helping select the reflective questions, final eight indicators, and evidence. We thank Susan Elrod and Marco Molinaro for their suggested revisions. The authors are members of the NASEM Roundtable on Systemic Change in Undergraduate STEM Education.

Disclosure Statement

No potential conflict of interest was reported by the author(s).

Additional information

Funding

This work was supported by the National Institutes of Health (Common Fund and Office of Scientific Workforce Diversity under award U01GM132174-01) and the National Science Foundation (grant number 1834518). The article represents the authors’ views and does not necessarily reflect the views of the National Academies of Sciences, Engineering, and Medicine, the Association of Public and Land-Grant Universities, the National Institutes of Health, or the National Science Foundation. 

Notes on contributors

Kacy Redd

Kacy Redd is the Associate Vice President for research and STEM education at the Association of Public and Land-Grant Universities (APLU), a member association of nearly 250 public research universities. At APLU, Dr. Redd leads the Network of STEM Education Centers (NSEC), the backbone for the National Science Foundation Inclusion across the Nation of Communities of Learners of Underrepresented Discoverers in Engineering and Science (INCLUDES) Aspire Alliance, and works with the vice presidents of research and change agents in STEM education across the country.

Mica Estrada

Mica Estrada is the Associate Dean of Diversity, Inclusion and Outreach and Professor at the University of California at San Francisco’s School of Nursing, in the Department of Social and Behavioral Sciences and the Institute for Health and Aging. She leads, as Principal Investigator, several national longitudinal studies, which implement and assess interventions aimed to increase student persistence of diverse learners in science, technology, engineering, mathematics, and medicine careers.

Harriet B. Nembhard

Harriet B. Nembhard, the sixth President of Harvey Mudd College, is a distinguished industrial engineer and collaborative academic leader who has a passion for strengthening the role of education and innovation in meeting the needs of society. With a background in health care systems, she brings perspectives centered on wellbeing to her role. President Nembhard is dedicated to inclusive excellence and broadening participation in STEM education. She is committed to preparing students to become leaders, emphasizing the impact of their work and equipping them with skills in civic engagement and building a joyful life.

Courtney Ngai

Courtney Ngai is the Associate Director of the Office for Undergraduate Research and Artistry at Colorado State University. In addition to supporting student engagement in research and artistry experiences, she is a Principal Investigator on several grants investigating organizational change in higher education. She recently launched her work for a National Science Foundation CAREER award, which involves characterizing the culture of chemistry from the perspectives of those who hold identities that are historically excluded from STEM.

References

  • Andrews, T. C., Brickman, P., Dolan, E. L., & Lemons, P. P. (2021). Every tool in the toolbox: Pursuing multilevel institutional change in the DeLTA project. Change: The Magazine of Higher Learning, 53, 25–32.
  • Association of American Colleges & Universities. (2023). High impact practices. https://www.aacu.org/trending-topics/high-impact
  • Brancaccio-Taras, L., Pape-Lindstrom, P., Peteroy-Kelly, M., Aguirre, K., Awong-Taylor, J., Balser, T., Cahill, M. J., Frey, R. F., Jack, T., Kelrick, M., Marley, K., Miller, K. G., Osgood, M., Romano, S., Uzman, J. A., & Zhao, J. (2016). The PULSE Vision & Change Rubrics, Version 1.0: A valid and equitable tool to measure transformation of life sciences departments at all institution types. CBE—Life Sciences Education, 15(4), ar60. https://doi.org/10.1187/cbe.15-12-0260
  • Carlisle, D. L., & Weaver, G. C. (2018). STEM education centers: Catalyzing the improvement of undergraduate STEM education. International Journal of STEM Education, 5(1), 47. https://doi.org/10.1186/s40594-018-0143-2
  • Carlisle, D. L., & Weaver, G. C. (2020). The role of centers: Disrupting the status quo while stabilizing initiatives in undergraduate STEM. Change: The Magazine of Higher Learning, 52(1), 60–70. https://doi.org/10.1080/00091383.2020.1693852
  • Community College Undergraduate Research Initiative. (n.d.). NSF’s Community College Undergraduate Research Initiative. https://www.ccuri.us/
  • Corbo, J. C., Reinholz, D. L., Dancy, M. H., Deetz, S., & Finkelstein, N. (2016). Framework for transforming departmental culture to support educational innovation. Physical Review Physics Education Research, 12(1), 010113. https://doi.org/10.1103/PhysRevPhysEducRes.12.010113
  • Eddy, S. L., Converse, M., & Wenderoth, M. P. (2015). PORTAAL: A classroom observation tool assessing evidence-based teaching practices for active learning in large science, technology, engineering, and mathematics classes. CBE—Life Sciences Education, 14(2), ar23. https://doi.org/10.1187/cbe.14-06-0095
  • Effective Practices for Physics Programs Guide. (2023). Effective practices for physics programs. https://ep3guide.org/
  • Ellis-Nelson, L., Ikegwuonu, E., Frederick, A., Michaels, J. (2023). Equity roundtables: A guide for institutions. Association of Public and Land-Grant Universities. https://www.aplu.org/wp-content/uploads/Equity-Roundtable-Guide_FINAL.pdf
  • Estrada, M. (2014). Ingredients for improving the culture of STEM degree attainment with co-curricular supports for underrepresented minority students. National Academies of Sciences White Paper, 28. http://sites.nationalacademies.org/cs/groups/dbassesite/documents/webpage/dbasse_088832.pdf
  • Estrada, M., Burnett, M., Campbell, A. G., Campbell, P. B., Denetclaw, W. F., Gutierrez, C. G., Hurtado, S., John, G. H., Matsui, J., McGee, R., Okpodu, C. M., Robinson, T. J., Summers, M. F., Werner-Washrune, M., & Zavala, M. (2016). Improving underrepresented minority student persistence in STEM. Cell Biology Education, 15(3), es5, 1–10.
  • Estrada, M., Eroy-Reveles, A., & Matsui, J. (2018). The influence of affirming kindness and community on broadening participation in STEM career pathways. Social Issues and Policy Review, 12(1), 258–297. https://doi.org/10.1111/sipr.12046
  • Holcombe, E., Kezar, A., Elrod, S. L., & Ramaley, J. (Eds.). (2021). Shared leadership in higher education: Framework and models for responding to a changing world. Stylus.
  • Hora, M. T., Oleson, A., & Ferrare, J. J. (2013). Teaching Dimensions Observation Protocol (TDOP) user’s manual.Wisconsin Center for Education Research.
  • Horii, C., Redd, K., Ouellett, M., Finkelstein, N., Beach, A., Carlisle, D., & Weaver, G. (2015, November). Collaborating at the centers: Report from a STEM education transformation workshop involving leaders of centers for teaching and learning and STEM education centers. In Collaborating at the Centers Workshop. https://doi.org/10.31219/osf.io/pknbf
  • Integrated Postsecondary Education Data System. (n.d). U.S. Department of Education, National Center for Education Statistics. https://nces.ed.gov/ipeds/
  • Kruse, S. D., & Louis, K. S. (1993). An emerging framework for analyzing school-based professional community. Paper presented at the annual meeting of the American Educational Research Association, Atlanta, GA.
  • Kuh, G. D. (2008). High-impact educational practices: What they are, who has access to them, and why they matter. Association of American Colleges & Universities.
  • Lee, M., Volpe Horii, C., E. Austin, A., Avery, L., DeSanctis, M., Finkelstein, N., Miller, E., & Schaal, B. (2023). An instructional-workforce framework for coordinated change in undergraduate education. Change: The Magazine of Higher Learning, 55(1), 54–63. https://doi.org/10.1080/00091383.2023.2151809
  • Levy, R. (2019). VITAL faculty: A growing workforce in colleges and universities. Mathematical Association of America.
  • National Academies for Sciences, Engineering, and Medicine. (2016). Barriers and opportunities for 2-year and 4-year STEM degrees: Systemic change to support students’ diverse pathways. National Academies Press.
  • National Academies for Sciences, Engineering, and Medicine. (2018). Indicators for monitoring undergraduate STEM education. National Academies Press.
  • National Academies for Sciences, Engineering, and Medicine. (2023). Advancing antiracism, diversity, equity, and inclusion in STEMM organizations: Beyond broadening participation. National Academies Press.
  • National Academies for Sciences, Engineering, and Medicine. (n.d). Roundtable on systemic change in undergraduate STEM education. National Academies Press. https://www.nationalacademies.org/our-work/roundtable-on-systemic-change-in-undergraduate-stem-education
  • Ngai, C., Pilgrim, M. E., Reinholz, D. L., Corbo, J. C., & Quan, G. M. (2020). Developing the DELTA: Capturing cultural changes in undergraduate departments. CBE Life Sciences Education, 19(2), ar15. https://doi.org/10.1187/cbe.19-09-0180
  • Partnership for Undergraduate Life Sciences Rubrics. (2023). PULSE. https://pulse-community.org/rubrics
  • Powered by Publics. (2023). The Data Maturity Index. Association of Public and Land-Grant Universities. https://www.aplu.org/wp-content/uploads/Data-Maturity-Index-March-2023-APLU-PxP.pdf
  • Quan, G. M., Corbo, J. C., Finkelstein, N. D., Pawlak, A., Falkenberg, K., Geanious, C., Ngai, C., Smith, C., Wise, S., Pilgrim, M. E., & Reinholz, D. L. (2019). Designing for institutional transformation: Six principles for department-level interventions. Physical Review Physics Education Research, 15(1), 010141. https://doi.org/10.1103/PhysRevPhysEducRes.15.010141
  • Reinholz, D. L., & Apkarian, N. (2018). Four frames for systemic change in STEM departments. International Journal of STEM Education, 5(1), 1–10. https://doi.org/10.1186/s40594-018-0103-x.
  • Reinholz, D. L., Corbo, J. C., Dancy, M. H., & Finkelstein, N. (2017). Departmental action teams: Supporting faculty learning through departmental change. Learning Communities Journal, 9(1), 5–32.
  • Reinholz, D. L., Pilgrim, M. E., Corbo, J. C., & Finkelstein, N. (2019). Transforming undergraduate education from the middle out with departmental action teams. Change: The Magazine of Higher Learning, 51, 64–70.
  • Sawada, D., Piburn, M. D., Judson, E., Turley, J., Falconer, K., Benford, R., & Bloom, I. (2002). Measuring reform practices in science and mathematics classrooms: The reformed teaching observation protocol. School Science and Mathematics, 102(6), 245–253. https://doi.org/10.1111/j.1949-8594.2002.tb17883.x.
  • Sloan Equity and Inclusion in STEM Introductory Courses (SEISMIC). (n.d). SEISMIC Equity Learning Communities Report DEMO. GitHub. https://htmlpreview.github.io/?https://github.com/vsfarrar/SEISMIC-equity-measures/blob/main/archive/questions_version3_demo.html
  • Smith, M. K., Jones, F. H., Gilbert, S. L., & Wieman, C. E. (2013). The Classroom Observation Protocol for Undergraduate STEM (COPUS): A new instrument to characterize university STEM classroom practices. CBE Life Sciences Education, 12(4), 618–627. https://doi.org/10.1187/cbe.13-08-0154
  • Student Experience Project. (2022). Student Experience Project resource hub. https://studentexperienceproject.org/resources/
  • Teval. (n.d). Teaching evaluation resources. https://teval.net/resources.html
  • Vigmostad, S. (2023). Unpublished memo to NASEM Roundtable on Systemic Change in Undergraduate STEM Education.
  • Wright, M. C., Lohe, D. R., & Little, D. (2018). The role of a center for teaching and learning in a de-centered educational world. Change: The Magazine of Higher Learning, 50(6), 38–44. https://doi.org/10.1080/00091383.2018.1540826