6,455
Views
3
CrossRef citations to date
0
Altmetric
Health Sciences

Comparing Student Performance in a Graduate-Level Introductory Biostatistics Course Using an Online versus a Traditional in-Person Learning Environment

&
Pages 105-114 | Published online: 11 Dec 2020

Abstract

Our study compared the performance of students enrolled in a graduate-level introductory biostatistics course in an online versus a traditional in-person learning environment at a school of public health in the United States. We extracted data for students enrolled in the course online and in person from 2013 to 2018. We compared average quiz and final exam scores between students in the two learning environments adjusting for demographic characteristics and prior academic performance using linear mixed models. Data were available for 1461 (83.1%) students learning online and 298 (16.9%) students learning in person. After adjusting for sex, race/ethnicity, age, quantitative GRE score, undergraduate GPA, and math refresher score, we found quiz scores for students learning online were about 2.5% lower than those for students learning in person, on average, with a 95% confidence interval ranging from 4.9% lower to 0.02% higher. Differential performance was even closer to equality for the final exam where scores for students learning online were about 0.9% higher with a 95% confidence interval ranging from a 3.9% reduction to 5.8% improvement. These estimates suggest comparable student performance can be achieved in a graduate-level introductory biostatistics course among students learning online and in person. Supplementary materials for this article are available online.

1 Introduction

Imagine a world where the traditional classroom vanished and the only option available for teaching and learning was online. That imaginary world became a reality in 2020 as a result of the Coronavirus disease (COVID-19) pandemic, introducing a time when learning via online platforms became a critical necessity for students at all stages of their education and for the financial viability of universities across the nation. Every teacher, regardless of their field of study or expertise, was teaching online, and every student, regardless of their age, sex, or race/ethnicity, was learning online. Given this seemingly overnight increase in prominence, quality assessments of the online learning environment in comparison with the traditional in-person learning environment are of paramount importance (Kim and Bonk Citation2006; Mathieson Citation2010).

Even before the pandemic, online education was on the rise. According to the Education Department’s National Center for Education Statistics report from 2018, 38.0% of graduate (and 33% of all college) students took at least one online course, where 9.1% (and 17.6%) took both online and in-person courses, and 28.9% (and 15.4%) took only online courses, which were all higher than earlier reports (Ginder et al. Citation2018). Online education has expanded to accommodate learning styles of specific cohorts such as the Net Generation (i.e., people born between 1982 and 1991), who were the “first to grow up with digital and cyber technologies” (Barnes et al. Citation2007, p. 1), were “constantly exposed to computer-based technology” (Sandars and Morrison Citation2007, p. 85), and were inclined to be “independent” and “autonomous” learners (Barnes et al. Citation2007). For such students, many have contemplated whether an online education would be comparable to traditional in-person learning (Larreamendy-Joerns and Leinhardt Citation2006; Mathieson Citation2010; Means et al. Citation2010; Xu and Jaggars Citation2014; Wladis et al. Citation2015).

As the availability and sophistication of online courses has grown, faculty and students alike have had to decide whether they wanted to branch out into the digital world (Atchley et al. Citation2013). Faculty (and students) have questioned whether they would have the technical skills needed to use online tools, the ability to successfully teach (or learn) course content asynchronously, the aptitude to effectively interact with one another online and participate in synchronous discussions, and a suitable personality that would shine on screen. Addressing these questions across science, technology, engineering, and mathematics (STEM) fields is especially critical, where many continue to advocate for the in-person learning environment (Noble Citation2004; Smith et al. Citation2008). Even advocates for online STEM courses acknowledge a need for additional support far beyond what is required in the traditional classroom, especially for women and younger students (Wladis Citation2015).

The literature addresses these questions by examining undergraduate-level courses in non-statistical disciplines (McFarlin Citation2008; Winquist and Carlson Citation2014), undergraduate-level statistics courses (McLaren Citation2004; Ward Citation2004; Dutton and Dutton Citation2005; Summers et al. Citation2005; Wilson Citation2013; Gundlach et al. Citation2015; Milic et al. Citation2016), and graduate-level courses in non-statistical disciplines (Caywood and Duckett Citation2003; Rovai and Jordan Citation2004). The few studies that have investigated graduate-level introductory biostatistics courses have been limited to small sample sizes over a limited number of terms (Evans et al. Citation2007; McGready and Brookmeyer Citation2013; Loux et al. Citation2016). We aim to contribute to the ongoing national discussion about the relative comparability of online and in-person learning environments by providing a detailed analysis of data from our experiences teaching graduate-level introductory biostatistics.

Our graduate-level introductory biostatistics course “Biostatistical Applications for Public Health” (BAPH) has been a Master of Public Health (MPH) core course required of all incoming MPH students at a school of public health within a private university in Washington, DC. This course was taught exclusively in person until 2013, when a new online degree program was created and included BAPH as a core course. Wanting all students to have comparable educational experiences in this fundamental core course, the primary in-person course director designed the online version of BAPH. This course director developed and taught all sections of BAPH that were examined in our study to remove instructor variability. Both the online and in-person class sizes varied from approximately 50–150 students with a wide range of mathematical and statistical backgrounds. Students self-selected the version of the course in which they wanted to enroll based on the mode of delivery, so there was no random assignment of students to the online and in-person courses.

The goal of our study was to provide an objective comparison of learning outcomes between students enrolled in online and in-person versions of the same graduate-level biostatistics course. While our study was not randomized, several important sources of bias were controlled by the unique nature of our study. First, the online and in-person course materials were designed by the same course director, which eliminated differences due to variation in the course content. Second, the same course director delivered content in both learning modes, which controlled for variation in teaching styles. Third, the courses both had the same basic structure: Students learning in person attended a weekly lecture, while students learning online were expected to watch the lectures asynchronously at their convenience (where the online recordings mirrored in-person lectures); both groups then attended equivalent synchronous breakout lab sessions led by different lab instructors. Fourth, this was a multi-year study where data were compiled over a lengthy period from the 2013 to 2018 academic years, which allowed us to take year-to-year variation of the student body into account. Student demographic characteristics and academic performance upon entry into the MPH program were also available for adjusted analyses. Therefore, our study provides a unique look at the comparison of online and in-person learning environments among students with certain common characteristics.

2 Methods

2.1 Course Description

BAPH is a three-credit graduate-level introductory course in the application of biostatistical principles to the critical analysis of studies in the health sciences literature. In this course, students are taught how to select an appropriate statistical method based on a research question, perform calculations of basic statistical procedures for estimation and inference, interpret the results of statistical analysis, and understand numbers in published literature and health news.

2.2 Online versus in-Person Learning Environments

As mentioned previously, the online and in-person versions of this course were created and taught by the same course director. All the online asynchronous material was purposefully designed and recorded to match the style of the highly regarded in-person lectures. Both online and in-person sections covered the exact same content. For all students, the methods of instruction included lectures, required and recommended readings, and class and small lab group discussions. All students were expected to engage in some independent learning each week, which included reviewing assigned material, preparing for class discussions, working on assignments, studying for exams, and completing group work and independent assignments. Nongraded practice problems were assigned from Bernard Rosner’s Fundamentals of Biostatistics textbook every week to prepare them for quizzes and exams.

Students learning in person attended lectures—delivered by the same course director—as an entire class once a week for two hours. Following lectures, they broke out into smaller groups of approximately 20 students for a one-hour lab session led by lab instructors. These lab instructors were Teaching Assistants (current students) or Instructional Assistants (graduates from the program) who had successfully completed the course. During these interactive lab sessions, students worked in small groups to complete and present solutions to nongraded lab exercises that reinforced the material presented during lecture, while the lab instructor moderated discussions.

Each week, students learning online independently navigated through two hours of asynchronous material stored on a cloud-based software-as-a-service platform provided by 2 U, an educational technology company. They received a handout to accompany the recorded video lectures, which closely followed the PowerPoint lecture slides of the students learning in person. As students listened to the recordings, they were cued to pause the video, look at their handout, and attempt to solve a problem on their own before watching the solution. Knowledge checks and flip books were dispersed throughout the asynchronous material to diversify the flow. Students learning online also met via Adobe Connect or Zoom in smaller groups of approximately 15 students once a week for a two-hour synchronous live session led by an instructor holding a master’s or PhD degree. These live sessions resembled the breakout lab sessions for the students learning in person, with additional time devoted to more thorough question-and-answer discussions of the asynchronous material. As these live session instructors assumed the same role as the in-person TAs and IAs, we will also refer to them as lab instructors hereafter. The course director, holding a PhD, monitored the progress of all students and lab instructors.

While BAPH was a three-credit course that required the same workload for all students regardless of delivery mode, the length of the terms differed for students learning online versus those learning in person. The in-person course ran for 15 weeks on a semester schedule, which included 13 weeks of new lecture material followed by a review session during week 14 and a final exam during week 15. The online course spanned 11 weeks on a quarterly schedule, which included 10 weeks of new lecture material followed by a review session and a final exam during week 11.

Both online and in person, student performance was evaluated using lab participation (based on attendance and completion of in-lab exercises), a math refresher, in-class and take-home quizzes, and a final exam. The math refresher was a free-response assignment used to assess students’ baseline mathematical preparedness. All students in overlapping terms were asked the same questions on the in-lab exercises, math refresher, quizzes, and the final exam. The materials were organized differently to accommodate the difference in term lengths, but the nature of the problem sets was generally the same otherwise. Quiz and exam questions were problem solving in the form of multiple choice, and students were instructed to show all work to receive full credit. Partial credit was awarded on quizzes for incorrect answers with partially correct approaches, but all-or-nothing grading was used for the final exam. For consistency, the lab instructors strictly followed the grading criteria set up by the course director.

Students learning in person had to complete three in-class quizzes and three take-home quizzes. In-class quizzes, which were designed to take between 30 and 45 minutes, were given during the first hour of the lecture period. Solutions to the quizzes were discussed immediately after all students had submitted their quizzes at the one-hour mark. Students learning online had to complete five quizzes (one every other week). The quizzes were designed to take between 45 and 60 min, but students were given two hours in case they experienced any technical difficulties with the online data entry system or scanning in their work (in an attempt to relieve technophobia-related anxiety). Online quizzes were slightly longer (with respect to number of questions and length of time) than the in-person quizzes because there were five instead of six in total. Quizzes focused on course content covered in the previous two weeks, but the final exam was cumulative, covering all topics.

All course materials for the in-person version were housed in Blackboard Learn, the university’s official learning management system (LMS). All materials for the online version (both asynchronous and synchronous) were housed in 2U’s LMS. Students were encouraged to post general questions directly to the Discussions section on Blackboard or the course wall on 2U’s platform, where the course director and lab instructors would post answers in a timely manner.

2.3 Study Design

This was a retrospective study. Demographic and prior academic performance data for all students who were enrolled in BAPH under the direction of the same course director between the Fall semester of 2013 and Fall semester of 2018 were extracted by the university’s Office of Institutional Research and Planning and the Office of Recruitment and Admissions for students learning in person and by 2 U analysts—a third-party vendor who administered and managed the online admissions process and was contractually required to archive admissions applications—for students learning online.

For in-person applicants, an Institutional Research Analyst and an Assistant Director of Admissions extracted the data from the Schools of Public Health Application Service (SOPHAS), which is a centralized application service that archives application data. If the SOPHAS application did not include the requested data, they searched in the university’s electronic filing cabinet (or database) for archiving student information. They de-identified all data to protect the students’ privacy and confidentiality. They entered all de-identified data in encrypted (password protected) Excel files. There was no link between study code numbers and direct identifiers. The files were emailed securely using ZixCorp and stored on the university’s enterprise file sharing service for online cloud storage and collaboration with two-step authentication to secure access. The university’s Office of Human Research approved our request for a waiver of consent and an exemption from Institutional Review Board (IRB) review.

The course director provided the Offices with student scores on the math refresher, quizzes, final exam, and lab participation to be de-identified and merged with the demographic and prior academic performance data, including sex (female, male), race/ethnicity (White, Black, Asian, Hispanic, not reported), age (years), student hours (full-time, part-time), quantitative Graduate Record Examination (GRE) scores, and undergraduate grade point average (GPA). We included a lab instructor variable indicating which (a) TA/IA instructed the students learning in person during the small breakout lab sessions (determined by the first letter of the student’s last name) or (b) instructor taught the students learning online during the live sessions.

2.4 Statistical Methods

We used two primary outcomes in our study to assess differences in student performance by learning environment (i.e., online versus in-person). The first outcome was quiz average, which was calculated as the average of five quizzes for those learning online and the average of six quizzes for students learning in person. The discrepancy in number of quizzes was only due to the difference in term length (i.e., 10 weeks online versus 14 weeks in person). The second outcome of interest was the final exam score, which was calculated on a scale from 0 to 100 points. In the descriptive analysis, we summarized these outcomes on their original scale. In the statistical models used to compare learning environments, however, Likelihood Ratio tests revealed an improved fit for both outcomes after a natural log-transformation.

The primary independent variable was learning environment. Other predictors of interest included sex (female, male), race/ethnicity (White, Black, Asian, Hispanic, not reported), age (years), student hours (full-time, part-time), academic year (categorical from 2013 to 2018), quantitative GRE score (above 50th percentile, below 50th percentile, not reported), undergraduate GPA (above 3.0, below 3.0, not reported), and math refresher score. We used the 50th percentile and 3.0 as respective cutoffs for GRE and GPA to reflect how admissions decisionmakers at the university typically use this information. Relevant summary statistics for these predictors, both aggregated and stratified by learning environment, can be found in Section 3. We conducted inference for simple comparisons using Wilcoxon Rank-Sum tests for continuous variables and Pearson’s Chi-square test of independence for categorical variables.

We modeled the natural log-transformed quiz average and final exam score using separate linear mixed effects models. The covariance structure was implied through specification of two random effects. The first random effect was for lab instructor to control for variability that may be attributed to a given instructor within the course. The second random effect was for academic year to account for year-to-year random variability in the student body. We compared models with and without these random effects, as well as models with only one of the two, and chose to include both random effects since this model had the lowest Akaike information criterion (AIC).

In the mixed model analyses, we included several additional fixed effects to adjust for sex, race/ethnicity, age, quantitative GRE score, undergraduate GPA, and math refresher score. In the model for quiz scores, we also adjusted for math refresher score as a linear spline after scatterplot investigation and a Partial F-test revealed an improved fit with different slopes before and after the breakpoint of 80. The mathematical form of each model has been provided in Appendix ∼ A.

All statistical tests were two-sided. Significance levels were set to 0.05. Interval estimates were calculated to have a 95% confidence level. We performed all analyses using SAS/STAT® software, Version 9.4 of the SAS System for Windows.Footnote1

3 Results

Descriptive analyses in revealed our sample to have more female and more White students with an average age of about 27 years. Students learning online and in person showed little overall difference in sex distribution. The race distributions were similar, but students learning online had a slightly lower percentage of White and a slightly higher percentage of Black students compared to those learning in person. Students learning online tended to be older. also summarizes the students’ academic backgrounds. In both groups, the majority were full-time, with an average undergraduate GPA of 3.28 and an average quantitative GRE score in about the 46th percentile. Comparing students learning online and in person, we found that the proportion of full-time students was lower online, the mean GPA was lower online, and the quantitative GRE tended to be lower online. Lastly, summarizes course assessments based on the math refresher, average quiz score, and final exam score. The average math refresher, average quiz scores, and the final exam scores, were all, on average, lower for students learning online.

Table 1 Descriptive statistics for the students’ demographic characteristics, academic background, and performance on course assessments.

Results from our primary analyses are shown in , which summarizes the estimates from the linear mixed models comparing students learning online to those learning in person for the two outcomes of interest (i.e., log-transformed average quiz score and final exam score). Interactions between key demographic variables and learning environment were investigated, but we did not find evidence to support their inclusion. After adjusting for the aforementioned predictors, we found that quiz scores for students learning online were about 2.5% lower than quiz scores for students learning in person, on average, with a 95% confidence interval ranging from about 4.9% lower to about 0.02% higher, which just barely includes equal performance within the interval. Differential performance was near equality for the final exam where scores for students learning online were about 0.9% higher with a 95% confidence interval ranging from about a 3.9% reduction to about a 5.8% improvement suggesting no evidence of differential performance for this comparison. Given that the scores were out of 100 points, these estimates are not indicative of a large degree of differential performance. Little or any material difference in performance was also observed between sexes with an estimated comparison between females and males that was near 1 with interval estimates that did not show evidence substantially favoring either a positive or negative effect.

Table 2 Exponentiated estimates with corresponding confidence intervals from two adjusted mixed models with log-transformed average quiz score and final exam score as the outcomes.

In addition to the finding regarding our primary comparison of online versus in-person learning environment, the model revealed some other interesting results. After adjustment, those who scored below the 50th percentile on the quantitative GRE had about 4.3% lower average performance on the quizzes compared to those who scored above the 50th percentile on the quizzes. The 95% confidence interval for this comparison ranged from as much as a 5.7% to as little as a 2.6% reduction. For the final exam, this comparison was estimated to be about an 8.0% reduction with a 95% confidence interval ranging from as much as an 11.7% to as little as a 4.1% reduction. The estimated results for the quantitative GRE effect indicate the possibility of a substantial differential performance associated with better scores on this exam. For example, the 95% confidence interval indicates that being below the 50th percentile may be associated with a reduction of roughly 3–10 points from a score of 90 on a scale of 100.

For GPA, those with an undergraduate GPA below 3.0 had about a 4.1% lower average quiz performance compared to those with a GPA above 3.0. The 95% confidence interval for this comparison ranged from as much as 5.6% to as little as a 2.6% reduction. The final exam showed a similar reduction, however its 95% confidence interval allowed for up to a 0.4% improvement associated with this comparison. The GPA estimate is not quite as strong as the GRE estimate, though the 95% confidence interval still indicates a reduction associated with having a GPA below 3.0 of roughly 2 to 5 points from a score of 90.

With regard to race/ethnicity, Black students appeared to fare worse on both assessments compared to White students. We found a reduction of 4.7% on average for the quizzes with a 95% confidence interval ranging from as much as a 6.3% to as little as a 3.0% reduction. For the final exam, the reduction was 7.4% with a 95% confidence interval ranging from as much as 11.6% to as little as 3.0%. A 7.4% reduction in performance for these students would lead to approximately a 7-point reduction in score from a grade of 90, which could have a material impact on the student’s grade.

For age, a negative, though not meaningful, association was found for both quiz and final exam performances. In both cases, we estimated about a 0.15% decline in performance per year increase in age. While the 95% confidence interval did not include 1 for quiz scores, this estimate is not large and does not imply a meaningful association with quiz performances. The 95% confidence interval did include 1 for final exam performance.

In the model for quiz scores, a scatterplot investigation suggested inclusion of linear spline for math refresher score with a knot at 80, which we included in the model based on evidence suggesting an improved fit from a Likelihood Ratio Test (F = 45.16, 1 numerator df, p < 0.0001). Similar investigation in the model for final exam did not reveal the necessity for such an effect. The math refresher score results used different slopes below versus above a score of 80. The positive slopes indicated that those with better mathematical preparation tended to perform better. The positive magnitude of the slopes was greater for math refresher scores above 80 than below. The slope for math refresher scores above 80 indicated an improvement of about 0.7% per one-point increase of the math refresher score with a 95% confidence interval ranging from as little as 0.6% to as much as 0.8%. As an example, this estimate would imply about a 5-point improvement in quiz score over a score of 85 associated with a 10-point improvement in math refresher score from 80 to 90. For the final exam, the slope coefficient indicated an improvement of about 0.6% per one-point increase of the math refresher score with a 95% confidence interval ranging from as little as 0.5% to as much as 0.7%. The results from the math refresher and quantitative GRE both showed a substantial association between prior mathematical experience and improved performance, which is consistent with our anecdotal experiences.

4 Discussion

In analyzing our data from students enrolled in a graduate-level introductory biostatistics course from 2013 to 2018, our data analysis showed little difference in performance when comparing students in an online environment to those learning the same content in a traditional in-person environment. Our results were adjusted for important predictors that allowed for a comparison among students with similar demographic and academic characteristics. In addition, our random effects control for variability among lab instructor and academic class year. An interesting advantage to our study is that since the lead professor for all classes was the same, we were able to control for content delivery directly through the study design. A disadvantage of our study is the fact that there was no randomization of students to online or in-person learning environments. Students self-selected into the online or in-person environments. Hence, causality cannot be inferred from our results due to the possibility of imbalance in other unmeasurable influential variables which may be associated with both learning environment and academic performance in an introductory biostatistics class.

Several studies reported similar findings to ours. At the undergraduate level, comparable performance levels between students learning online and in person have been reported for courses in statistics (Yablon and Katz Citation2001; Summers et al. Citation2005), business statistics (McLaren Citation2004), earth science (Werhner Citation2010), programming (Zacharis Citation2010), personal finance (Ary and Brune Citation2011), business application software (Wagner et al. Citation2011), psychology (Arviso Citation2019), history (Arviso Citation2019), algebra (Arviso Citation2019), and environmental science (Paul and Jefferson 2019). At the graduate level, comparable performance levels between students learning online and in person have been reported for courses in applied statistics for industry (Stephenson Citation2001), behavior management (Caywood and Duckett Citation2003), biostatistics (Evans et al. Citation2007; McGready and Brookmeyer Citation2013), and research methods (Holmes and Reid Citation2017). Further supporting our findings, Ary and Brune (Citation2011) found pre-course GPAs and ACT scores, as opposed to the learning environment, predicted student performance.

4.1 Where Students Learning in Person Outperformed Students Learning Online

Contrary to our results, Paden (Citation2006) found undergraduate students learning in a traditional in-person section of an introductory math course performed significantly better than those learning online; however, they only adjusted for sex in their analysis. Sue (Citation2005) observed this difference in outcomes among introductory business statistics students—although the discrepancy was attributed to online students taking the exam in person since they were accustomed to being online. Tanyel and Griffin (Citation2014) reported similar differences across multiple disciplines offered in their College of Arts and Sciences, College of Business and Economics, and the School of Education over 10 years.

4.2 Where Students Learning Online Outperformed Students Learning in Person

On the other hand, Atchley et al. (Citation2013) found students enrolled in online courses spanning different disciplines (e.g., psychology, English, computer information systems, and accounting) performed significantly better academically than their counterparts enrolled in the traditional in-person versions of the courses. Their study, however, did not include any statistics courses, and they did not control for any other variables in their analysis. Graham and Lazari (Citation2018) found that students enrolled in an online college algebra course performed significantly better on the departmental final exam than those students enrolled in the traditional in-person class, but they also failed to control for other factors. Similarly, Dutton and Dutton (Citation2005) found undergraduate students enrolled in an online business statistics class achieved significantly higher final exam scores and final course grades than students enrolled in the traditional in-person version of the class, noting that students enrolled online were significantly older with higher GPAs. They also identified GPA and student effort (i.e., homework average) as other “positive predictors of success” (Dutton and Dutton Citation2005).

4.3 What’s in a Name: The “Online” Course

Some of these differences in student performance could be attributed to the actual design of the “online” course, which has been loosely defined and widely diverse (Sun and Chen Citation2016). In some studies, the descriptions of their “hybrid,” “blended,” and “flipped” statistics courses more closely resembled the design of our “online” course than other “online” courses (Utts et al. Citation2003; Ward Citation2004; Wilson Citation2013; Winquist and Carlson Citation2014; Gundlach et al. Citation2015). As we observed in our study, they found student performance comparable in such “hybrid” and traditional courses (Utts et al. Citation2003, Ward Citation2004).

4.4 Our Strengths

Few studies used data from multiple semesters in their analysis (Stephenson Citation2001; McLaren Citation2004; Wagner et al. Citation2011; Tanyel and Griffin Citation2014). Our study compared the performance of biostatistics students in the two learning environments using five years’ worth of data from 25 different offerings (5 in-person, 20 online) in overlapping terms. The same course director who taught the course in person designed and recorded the asynchronous material for the online version of the course, thus removing instructor variability and ensuring that the students would cover the same content and receive the same assessments. This is a strength over Tanyel and Griffin’s (Citation2014) 10-year study, which (as they acknowledged) was limited by the small number of instructor-matched online and in-person courses available during the same semester.

4.5 Our Limitations

Just as Mathieson (Citation2010) documented “methodological limitations” in her systematic review comparing the performance of undergraduate and graduate students enrolled in online and in-person statistics courses (e.g., “lack of randomization, lack of generalizability, differences in baseline characteristics between groups, and failure to adjust for confounders”), our study also has some limitations (p. 3). Fewer students enrolled in the course in person than online annually, so the sample size is smaller for the in-person subset. Actual class sizes for the lab sessions, however, were about the same. Due to incomplete student records from admissions, data were missing on undergraduate GPA (4.2% online, 1.7% in person) and quantitative GRE scores (36.4% online, 12.1% in person). Employment status, occupation, and undergraduate major were unavailable for most of the students, so these factors could not be used in our analysis. Students self-selected which version of the course they wanted to enroll in based on the mode of delivery, so there was no random assignment of students to the online and in-person courses. Just as we were not able to randomize students, all the other studies referenced reported the same limitation. Stack (Citation2015), however, was able to achieve a “quasi-randomization of students into sections” of a criminology course “due to an administrative error in the course schedule,” and he still did not find a significant difference in performance between students enrolled in the traditional in-person and online sections of the course. We reference Stack (Citation2015) to show that (1) the only study we found that did achieve randomization did so serendipitously and (2) they still reached the same conclusion we did regarding learning environments.

4.6 Future Considerations

When polling our BAPH students at the beginning of the semester to determine how they feel about taking the course, most students typically replied that they are “scared” or “terrified.” Thus, we continuously search for ways to reduce this fear factor and improve our course. Just as Devaney (Citation2010) found graduate students enrolled in statistics courses online experienced less anxiety and developed better attitudes than those learning in person, we could explore ways of delivering content online and in person to create a more comfortable learning environment. Being in a comfortable learning environment can help a student build their confidence. Among first-year public health graduate students who completed an introductory biostatistics course, Loux et al. (Citation2016) found that those students from the “flipped” class (resembling the structure of our online course) were just as confident in their ability to apply biostatistics as those from the traditional in-person class. Moreover, among undergraduate and graduate students who completed at least one online course of any discipline at a university in the Midwest, Eom (Citation2006) discovered that satisfied students performed better in the course. In the future, we would like to examine end-of-semester evaluations to explore student satisfaction and how it relates to student performance in our course.

As suggested in the literature, such improvements in student satisfaction and performance may involve blending the best features of online and traditional in-person courses (Chen and Jones Citation2007; Gundlach et al. Citation2015; Chingos et al. Citation2017). Given the improved performance of students enrolled in “hybrid” or “blended” statistics courses reported in the literature, we could consider offering the traditional in-person version as a “hybrid” course (Wilson Citation2013; Winquist and Carlson Citation2014; Gundlach et al. Citation2015; Milic et al. Citation2016). This design would closely model the online version with the exception that the live synchronous sessions would be in person as opposed to online.

To do so, we would need to expand our analysis to identify critical success factors. In designing our BAPH course online and in person, care was taken to balance lectures, interactive small group labs, readings, practice exercises, formulas, figures, and interpretations in both learning environments. While we cannot identify which of these features primarily led to student success in our study, Eom and Ashill (Citation2016) found that a student’s performance primarily depends on the course design, instructor, and dialogue. Their findings agree with Volery and Lord (Citation2000), who identified technology and instructor as the critical success factors for an online management course. In their qualitative assessment of online courses presented in the literature, Sun and Chen (Citation2016) also concluded that strong instructors who implement a high-quality course design while building a sense of community lead to a successful online learning environment. This is further supported by Barnes et al. (Citation2007) who recognize that “to be human is to learn, and we learn from good teachers” (p. 6). Identifying these critical success factors in our BAPH course designed and taught by the same course director would allow us to further improve the course and create the optimal learning environment for our students.

5 Conclusion

What was learned from this study could facilitate improvement of educational programming at our university and perhaps elsewhere, which is of paramount importance as online education continues to gain prominence. Our study has shown some evidence that learning outcomes may not be affected materially when presented in an online environment as compared with the traditional in-person standard. Discovering that comparable student performance can be achieved in an online setting is especially important at a time when an unforeseeable event, such as a public health crisis, abruptly forces nearly all university courses to be delivered exclusively online regardless of teacher and student preference or experience. Further research is needed, however, before universities can expand their online offerings more extensively as other considerations of the student experience may also contribute to their willingness to enter an online curriculum. There is still much to be learned, and we have only begun to see the full picture of how educational instruction will evolve in the coming decade.

Software

SAS and all other SAS Institute Inc. product or service names are registered trademarks or trademarks of SAS Institute Inc. in the USA and other countries. ® indicates USA registration.

Financial Disclosure

The authors declare that they have no financial relationship relevant to this article to disclose.

Supplemental material

Supplemental Material

Download PDF (83.5 KB)

Acknowledgments

We would like to thank all the students enrolled in the graduate-level introduction to biostatistics courses for contributing data for this study, as well as the lab instructors for their hard work in this course and dedication to these students. We would also like to thank the analysts from the Office of Institutional Research and Planning, the Office of Recruitment and Admissions, and 2U for their assistance with the data extraction. Finally, we would like to thank the Research Assistant who conducted a preliminary exploratory data analysis, as well as the university for providing funding through the Flexible Research Development Award to support this Research Assistant.

Supplementary Materials

The supplementary material includes the description of the mathematical form of each model used in our analysis.

Conflict of Interest

The authors declare that they have no potential conflicts of interest relevant to this article to disclose.

Data Sharing

In the IRB application, we declared that all data would be aggregated/summarized in all publications, and no individual-level results would be communicated to protect the students’ privacy and confidentiality. Since there was no contact with students during the data extraction, we did not provide students with consent documents to obtain necessary permissions to share the data.

Additional information

Funding

This study received partial funding through the university’s Flexible Research Development Award to support a Research Assistant to conduct a preliminary exploratory data analysis.

Notes

1 Copyright [copyright] 2016 SAS Institute Inc. SAS and all other SAS Institute Inc. product or service names are registered trademarks or trademarks of SAS Institute Inc., Cary, NC, USA.

References

  • Arviso, C. C. (2019), “Comparison of Adaptive Online Education with Adaptive Traditional Face-to-Face Education at Navajo Technical University: A Quantitative Study,” Ph.D. diss., Northcentral University, ProQuest Dissertations Publishing, UMI No. 22624713. Available at http://proxygw.wrlc.org/login?url=https://search-proquest-com.proxygw.wrlc.org/docview/2312801710?accountid=11243.
  • Ary, E. J., and Brune, C. W. (2011), “A Comparison of Student Learning Outcomes in Traditional and Online Personal Finance Courses,” MERLOT Journal of Online Learning and Teaching, 7, 465–474.
  • Atchley, W., Wingenbach, G., and Akers, C. (2013), “Comparison of Course Completion and Student Performance through Online and Traditional Courses,” International Review of Research in Open and Distributed Learning, 14, 104–116. DOI: 10.19173/irrodl.v14i4.1461.
  • Barnes, K., Marateo, R. C., and Ferris, S. P. (2007), “Teaching and Learning with the Net Generation,” Innovate: Journal of Online Education, 3, 1. Available at https://nsuworks.nova.edu/innovate/vol3/iss4/1.
  • Caywood, K., and Duckett, J. (2003), “Online Vs. On-Campus Learning in Teacher Education,” Teacher Education and Special Education: The Journal of the Teacher Education Division of the Council for Exceptional Children, 26, 98–105. DOI: 10.1177/088840640302600203.
  • Chen, C., and Jones, K. (2007), “Blended Learning versus Traditional Classroom Settings: Assessing Effectiveness and Student Perceptions in an MBA Accounting Course,” Journal of Educators Online, 4, 1–15, DOI: 10.9743/JEO.2007.1.3.
  • Chingos, M. M., Griffiths, R. J., Mulhern, C., and Spies, R. R. (2017), “Interactive Online Learning on Campus: Comparing Students’ Outcomes in Hybrid and Traditional Courses in the University System of Maryland,” The Journal of Higher Education, 88, 210–233. DOI: 10.1080/00221546.2016.1244409.
  • Devaney, T. A. (2010), “Anxiety and Attitude of Graduate Students in On-Campus Vs. Online Statistics Courses,” Journal of Statistics Education, 18, 15. DOI: 10.1080/10691898.2010.11889472.
  • Dutton, J., and Dutton, M. (2005), “Characteristics and Performance of Students in an Online Section of Business Statistics,” Journal of Statistics Education, 13, 24. DOI: 10.1080/10691898.2005.11910564.
  • Eom, S., and Ashill, N. (2016), “The Determinants of Students’ Perceived Learning Outcomes and Satisfaction in University Online Education: An Update*,” Decision Sciences Journal of Innovative Education, 14, 185–215. DOI: 10.1111/dsji.12097.
  • Eom, S. B., Wen, H. J., and Ashill, N. (2006), “The Determinants of Students’ Perceived Learning Outcomes and Satisfaction in University Online Education: An Empirical Investigation,” Decision Sciences Journal of Innovative Education, 4, 215–235. DOI: 10.1111/j.1540-4609.2006.00114.x.
  • Evans, S. R., Wang, R., Yeh, T.-M., Anderson, J., Haija, R. M., McBratney-Owen, P. M., Peeples, L., Sinha, S. K., Xanthakis, V., Rajicic, N., and Zhang, J. (2007), “Evaluation of Distance Learning in an ‘Introduction to Biostatistics’ Class: A Case Study,” Statistics Education Research Journal, 6, 59–77.
  • Ginder, S. A., Kelly-Reid, J. E., and Mann, F. B. (2018), “Enrollment and Employees in Postsecondary Institutions, Fall 2017; and Financial Statistics and Academic Libraries, Fiscal Year 2017: First Look (Provisional Data) (NCES 2019-021rev),” Report, U. S. Department of Education, National Center for Education Statistics.
  • Graham, V., and Lazari, A. (2018), “College Algebra - Online Section Versus Traditional Section,” Georgia Journal of Science, 76, 1–6.
  • Gundlach, E., Richards, K. A. R., Nelson, D., and Levesque-Bristol, C. (2015), “A Comparison of Student Attitudes, Statistical Reasoning, Performance, and Perceptions for Web-Augmented Traditional, Fully Online, and Flipped Sections of a Statistical Literacy Class,” Journal of Statistics Education, 23, 1–33, DOI: 10.1080/10691898.2015.11889723.
  • Holmes, C. M., and Reid, C. (2017), “A Comparison Study of On-campus and Online Learning Outcomes for a Research Methods Course,” The Journal of Counselor Preparation and Supervision, 9, DOI: 10.7729/92.1182.
  • Kim, K.-J., Bonk, C. J. (2006), “The Future of Online Teaching and Learning in Higher Education: The Survey Says …,” Educause Quarterly, 29, 22–30. Available at http://www.educause.edu/ero/archive/eq.
  • Larreamendy-Joerns, J., and Leinhardt, G. (2006), “Going the Distance With Online Education,” Review of Educational Research, 76, 567–605. DOI: 10.3102/00346543076004567.
  • Loux, T. M., Varner, S. E., and VanNatta, M. (2016), “Flipping an Introductory Biostatistics Course: A Case Study of Student Attitudes and Confidence,” Journal of Statistics Education, 24, 1–7. DOI: 10.1080/10691898.2016.1158017.
  • Mathieson, K. (2010), “Comparing Outcomes between Online and Face-to-Face Statistics Courses: A Systematic Review,” In Proceedings of the Eighth International Conference on Teaching Statistics. Available at http://iase-web.org/Conference_Proceedings.php?p=ICOTS_8_2010.
  • McFarlin, B. (2008), “Hybrid Lecture-Online Format Increases Student Grades in an Undergraduate Exercise Physiology Course at a Large Urban University,” Advances in Physiology Education, 32, 86–91. DOI: 10.1152/advan.00066.2007.
  • McGready, J., and Brookmeyer, R. (2013), “Evaluation of Student Outcomes in Online Vs. Campus Biostatistics Education in a Graduate School of Public Health,” Preventive Medicine, 56, 142–144. DOI: 10.1016/j.ypmed.2012.11.020.
  • McLaren, C. H. (2004), “A Comparison of Student Persistence and Performance in Online and Classroom Business Statistics Experiences,” Decision Sciences Journal of Innovative Education, 2, 1–10. DOI: 10.1111/j.0011-7315.2004.00015.x.
  • Means, B., Toyama, Y., Murphy, R., Bakia, M., and Jones, K. (2010), “Evaluation of Evidence-Based Practices in Online Learning: A Meta-Analysis and Review of Online Learning Studies,” Report, U. S. Department of Education, Office of Planning, Evaluation, and Policy Development, Policy and Program Studies Service. Available at https://eric.ed.gov/?id=ED505824.
  • Milic, N. M., Trajkovic, G. Z., Bukumiric, Z. M., Cirkovic, A., Nikolic, I. M., Milin, J. S., Milic, N. V., Savic, M. D., Corac, A. M., Marinkovic, J. M., and Stanisavljevic, D. M. (2016), “Improving Education in Medical Statistics: Implementing a Blended Learning Model in the Existing Curriculum,” PLoS One., 11, e0148882. DOI: 10.1371/journal.pone.0148882.
  • Noble, J. V. (2004), “For Science and Math, Skip the Online Course,” Communications of the Association for Computing Machinery, 47, 12–13.
  • Paden, R. R. (2006), “A Comparison of Student Achievement and Retention in an Introductory Math Course Delivered in Online, Face-to-Face, and Blended Modalities,” Ph.D. diss., Capella University, ProQuest Dissertations Publishing, UMI No. 3237076. Available at http://proxygw.wrlc.org/login?url=https://search-proquest-com.proxygw.wrlc.org/docview/304916870?accountid=11243.
  • Paul, J., and Jefferson, F. (2019), “A Comparative Analysis of Student Performance in an Online vs. Face-to-Face Environmental Science Course From 2009 to 2016,” Frontiers of Computer Science, 1, 7. DOI: 10.3389/fcomp.2019.00007.
  • Rovai, A. P., and Jordan, H. (2004), “Blended Learning and Sense of Community: A Comparative Analysis with Traditional and Fully Online Graduate Courses,” The International Review of Research in Open and Distributed Learning, 5, 1–13, DOI: 10.19173/irrodl.v5i2.192.
  • Sandars, J., and Morrison, C. (2007), “What is the Net Generation? The Challenge for Future Medical Education,” Medical Teacher, 29, 85–88. DOI: 10.1080/01421590601176380.
  • Smith, G. G., Heindel, A. J., and Torres-Ayala, A. T. (2008), “E-learning Commodity or Community: Disciplinary Differences between Online Courses,” Internet and Higher Education, 11, 152–159. DOI: 10.1016/j.iheduc.2008.06.008.
  • Stack, S. (2015), “Learning Outcomes in an Online vs Traditional Course,” International Journal for the Scholarship of Teaching and Learning, 9, 5. DOI: 10.20429/ijsotl.2015.090105.
  • Stephenson, W. R. (2001), “Statistics at a Distance,” Journal of Statistics Education, 9 [online]. Available at ww2.amstat.org/publications/jse/v9n3/stephenson.html.
  • Sue, V. (2005), “Comparing Online and Traditional Courses,” Academic Exchange Quarterly, 9, 30–34.
  • Summers, J. J., Waigandt, A., and Whittaker, T. A. (2005), “A Comparison of Student Achievement and Satisfaction in an Online Versus a Traditional Face-to-Face Statistics Class,” Innovative Higher Education, 29, 233–250. DOI:/10.1007/s10755-005-1938-x. DOI: 10.1007/s10755-005-1938-x.
  • Sun, A., and Chen, X. (2016), “Online Education and its Effective Practice: A Research Review,” Journal of Information Technology Education: Research, 15, 157–190. DOI: 10.28945/3502.
  • Tanyel, F., and Griffin, J. (2014), “A Ten-Year Comparison of Outcomes and Persistence Rates in Online versus Face-to-Face Courses,” available at https://www.westga.edu/∼bquest/2014/onlinecourses2014.pdf.
  • Utts, J., Sommer, B., Acredolo, C., Maher, M. W., and Matthews, H. R. (2003), “A Study Comparing Traditional and Hybrid Internet-Based Instruction in Introductory Statistics Classes,” Journal of Statistics Education, 11, available at ww2.amstat.org/publications/jse/v11n3/utts.html. DOI: 10.1080/10691898.2003.11910722.
  • Volery, T., and Lord, D. (2000), “Critical Success Factors in Online Education,” International Journal of Educational Management, 14, 216.
  • Wagner, S. C., Garippo, S. J., and Lovaas, P. (2011), “A Longitudinal Comparison of Online Versus Traditional Instruction,” MERLOT Journal of Online Learning and Teaching, 7, 68–73.
  • Ward, B. (2004), “The Best of Both Worlds: A Hybrid Statistics Course,” Journal of Statistics Education, 12, DOI: 10.1080/10691898.2004.11910629.
  • Werhner, M. J. (2010), “A Comparison of the Performance of Online versus Traditional On-Campus Earth Science Students on Identical Exams,” Journal of Geoscience Education, 58, 310–312. DOI: 10.5408/1.3559697.
  • Wilson, S. G. (2013), “The Flipped Class: A Method to Address the Challenges of an Undergraduate Statistics Course,” Teaching of Psychology, 40, 193–199. DOI: 10.1177/0098628313487461.
  • Winquist, J. R., and Carlson, K. A. (2014), “Flipped Statistics Class Results: Better Performance Than Lecture Over One Year Later,” Journal of Statistics Education, 22, DOI: 10.1080/10691898.2014.11889717.
  • Wladis, C., Conway, K. M., and Hachey, A. C. (2015), “The Online STEM Classroom-Who Succeeds? An Exploration of the Impact of Ethnicity, Gender, and Non-traditional Student Characteristics in the Community College Context,” Community College Review, 43, 142–164. DOI: 10.1177/0091552115571729.
  • Xu, D., and Jaggars, S. S. (2014), “Performance Gaps between Online and Face-to-Face Courses: Differences Across Types of Students and Academic Subject Areas,” Journal of Higher Education, 85, 633–659. 2016), DOI: 10.1353/jhe.2014.0028.
  • Yablon, Y. B., and Katz, Y. J. (2001), “Statistics Through the Medium of the Internet: What Students Think and Achieve,” Academic Exchange Quarterly, 5, 17–22.
  • Zacharis, N. Z. (2010), “The Impact of Learning Styles on Student Achievement in a Web-Based Versus an Equivalent Face-to-Face Course,” College Student Journal, 44, 591–597.