ArticlePDF Available

Abstract and Figures

Significance Low persistence in educational programs is a major obstacle to social mobility. Scientists have proposed many scalable interventions to support students learning online. In one of the largest international field experiments in education, we iteratively tested established behavioral science interventions and found small benefits depending on individual and contextual characteristics. Forecasting intervention efficacy using state-of-the-art methods yields limited improvements. Online education provides unprecedented access to learning opportunities, as evidenced by its role during the 2020 coronavirus pandemic, but adequately supporting diverse students will require more than a light-touch intervention. Our findings encourage funding agencies and researchers conducting large-scale field trials to consider dynamic investigations to uncover and design for contextual heterogeneity to complement static investigations of overall effects.
Content may be subject to copyright.
Scaling up behavioral science interventions in
online education
René F. Kizilcec
a,1,2
, Justin Reich
b,1,2
, Michael Yeomans
c,1,2
, Christoph Dann
d
, Emma Brunskill
e
, Glenn Lopez
f
,
Selen Turkay
g
, Joseph Jay Williams
h
, and Dustin Tingley
f,i
a
Department of Information Science, Cornell University, Ithaca, NY 14850;
b
Comparative Media Studies/Writing, Massachusetts Institute of Technology,
Cambridge, MA 02139;
c
Harvard Business School, Harvard University, Cambridge, MA 02138;
d
Machine Learning Department, Carnegie Mellon University,
New York, NY 10004;
e
Computer Science Department, Stanford University, Stanford, CA 94305;
f
Office of the Vice Provost for Advances in Learning, Harvard
University, Cambridge, MA 02138;
g
School of Computer Science, Queensland University of Technology, Brisbane City, QLD 4000, Australia;
h
Department of
Computer Science, University of Toronto, Toronto, M5S 1A1 ON, Canada and
i
Department of Government, Harvard University, Cambridge, MA 02138
Edited by Susan T. Fiske, Princeton University, Princeton, NJ, and approved May 12, 2020 (received for review December 5, 2019)
Online education is rapidly expanding in response to rising demand
for higher and continuing education, but many online students
struggle to achieve their educational goals. Several behavioral science
interventions have shown promise in raising student persistence and
completion rates in a handful of courses, but evidence of their
effectiveness across diverse educational contexts is limited. In this
study, we test a set of established interventions over 2.5 y, with one-
quarter million students, from nearly every country, across 247 online
courses offered by Harvard, the Massachusetts Institute of Technol-
ogy, and Stanford. We hypothesized that the interventions would
produce medium-to-large effects as in prior studies, but this is not
supported by our results. Instead, using an iterative scientific pro-
cess of cyclically preregistering new hypotheses in between waves
of data collection, we identified individual, contextual, and tem-
poral conditions under which the interventions benefit students.
Self-regulation interventions raised student engagement in the
first few weeks but not final completion rates. Value-relevance
interventions raised completion rates in developing countries to
close the global achievement gap, but only in courses with a global
gap. We found minimal evidence that state-of-the-art machine
learning methods can forecast the occurrence of a global gap or
learn effective individualized intervention policies. Scaling behav-
ioral science interventions across various online learning contexts
can reduce their average effectiveness by an order-of-magnitude.
However, iterative scientific investigations can uncover what
works where for whom.
behavioral interventions
|
scale
|
online learning
Behavioral scientists have argued that it is possible to in-
tervene and modify personal habits, decisions, and thought
patterns that contribute to social problems (1). Behavioral sci-
ence interventions have been developed to promote a variety of
prosocial behaviors, such as healthy eating habits, physical ac-
tivity, getting medical check-ups, voting, and achievement in
schools and colleges. While these interventions are usually low-
costto participants and policy-makersthey are still thought
to be effective because they target the psychological mechanisms
underlying peoples behavior (2). The ubiquity of networked
devices has made it even easier to implement these interventions
at large scale and to run field experiments that reveal their
broader impact.
In this study, we conducted one of the largest global field
experiments in higher education, with one-quarter million stu-
dents across nearly every country, to examine the scalability of
several behavioral science interventions that improved outcomes
for thousands of students in our own prior research. Online
education is rapidly expanding to address problems of educa-
tional access and meet the rising economic demands for pro-
fessional development and retraining. For all this growth, many
online students struggle to achieve their goals. Course comple-
tion rates are often low: Around 20% in Harvard University, the
Massachusetts Institute of Technology (MIT), and Stanford
University massive open online courses (MOOCs) among stu-
dents who intend to complete (3, 4).
Online learning environments are well-suited to test the scal-
ability of behavioral interventions. They have a well-defined
outcome (course completion), requiring sustained effort. Stu-
dent progress is continuously tracked through a common soft-
ware platform. Improving outcomes in online learning through
targeted support holds great promise for human capital devel-
opment around the world. National education platforms have
started using online courses to supplement college STEM (sci-
ence, technology, engineering, and math) instruction (5) and
students who complete MOOCs report benefits ranging from
earning credit toward a degree to enhanced skills in a current job
or finding a new job (6, 7). Moreover, there is evidence that
students can transfer skills learned from MOOCs into real-world
settings: They deploy new programming skills into open-source
software projects, participate in scholarly activity following a
research methods course, and develop new school initiatives
after an education leadership course (810).
Following the joint Common Guidelines for Educational Re-
search from the National Science Foundation (NSF) and Institute
Significance
Low persistence in educational programs is a major obstacle to
social mobility. Scientists have proposed many scalable inter-
ventions to support students learning online. In one of the
largest international field experiments in education, we itera-
tively tested established behavioral science interventions and
found small benefits depending on individual and contextual
characteristics. Forecasting intervention efficacy using state-of-the-
art methods yields limited improvements. Online education pro-
vides unprecedented access to learning opportunities, as evidenced
by its role during the 2020 coronavirus pandemic, but adequately
supporting diverse students will require more than a light-touch
intervention. Our findings encourage funding agencies and re-
searchers conducting large-scale field trials to consider dynamic in-
vestigations to uncover and design for contextual heterogeneity to
complement static investigations of overall effects.
Author contributions: R.F.K., J.R., M.Y., C.D., E.B., S.T., J.J.W., and D.T. designed research;
R.F.K., J.R., M.Y., C.D., and G.L. performed research; R.F.K., J.R., M.Y., C.D., and G.L.
analyzed data; and R.F.K., J.R., M.Y., and C.D. wrote the paper.
The authors declare no competing interest.
This article is a PNAS Direct Submission.
This open access article is distributed under Creative Commons Attribution-NonCommercial-
NoDeriv atives L icense 4. 0 (CC BY-N C-ND).
1
R.F.K., J.R., and M.Y. contributed equally to this work.
2
To whom correspondence may be addressed. Email: kizilcec@cornell.edu, jreich@
mit.edu, or myeomans@hbs.edu.
This article contains supporting information online at https://www.pnas.org/lookup/suppl/
doi:10.1073/pnas.1921417117/-/DCSupplemental.
www.pnas.org/cgi/doi/10.1073/pnas.1921417117 PNAS Latest Articles
|
1of6
PSYCHOLOGICAL AND
COGNITIVE SCIENCES
Downloaded by guest on June 16, 2020
for Education Sciences (IES) (11), we designed this study as a
scale-up research impact study to examine effectiveness in a
wide range of populations, contexts, and circumstances, without
substantial developer involvement in implementation or evaluation.
We took existing interventions and deployed them with minimal
ongoing adjustment across a range of courses. The courses included
in the study sample spanned a remarkable range of subjectsfrom
poetry to data scienceand a diversity of students from nearly every
country in the world. We leveraged the unprecedented variation in
this platform to advance our understanding of how behavioral science
intervention effects vary across students in MOOCs and different
institutional and course contexts. The scalable interventions that we
tested could be feasibly implemented by administrators or instructors
hosting online courses in diverse settings.
Interventions and Prior Results
For 2.5 y, from September 2016 to May 2019, we added a ran-
domly assigned intervention module at the start of nearly all
MOOCs offered by Harvard, MIT, and Stanford (n=269,169
students across 247 courses). We had previously published re-
sults from large field experiments for three of the five inter-
ventions in this study (1214). These interventions had each been
shown to substantially improve completion rates for targeted
groups of students and replicated across courses. We scaled the
interventions consistently by embedding a survey early in the
materials of every course. After answering typical survey ques-
tions about themselves and their goals, students were randomly
assigned to receive one of the intervention activities described
below, or no activity in the control condition.
The plan-makinginterventions prompted students to con-
cretely describe when and how they will complete required
coursework for the entire course. Plan-making interventions
target peoples reluctance to forecast the procedural details of
goal pursuit (15, 16). Previous work showed effects of plan-
making on discrete behaviors like voting or doctors appoint-
ments. We had tested plan-making interventions in three online
courses on Business, Chemistry, and Political Science (n=2,053)
and found a 29% increase in course completion (from 14 to
18%) among committed English-fluent students (12). We had
preregistered the analysis (https://osf.io/wq8m5/) and predicted
the effect for this specific subpopulation. We use two versions of
a plan-making activity in the present study: A replication of the
previous paper (12), focused on long-term plans, and a short-
term variant that asks students to plan for the first week only.
The value-relevanceintervention is a motivational activity
that asks students to indicate important values and write about
how taking the course reflects and reinforces what is most im-
portant to them. The intervention builds on self-affirmation and
utility-value intervention research that has been shown to reduce
ethno-racial achievement gaps by lifting performance among
disadvantaged students (1719). Previous work tested, preregis-
tered (https://osf.io/g8bu4/), and replicated this intervention in
two online courses on Computing and Public Policy (n=3,451)
and found that it closed the global achievement gap between
students in more-developed and less-developed countries by
raising the completion rate of students in developing countries
from 17 to 41% (13). Students in the developed world remained
unaffected in one course but experienced a decline in completion
(from 32 to 23%) in the second course.
The mental contrasting with implementation intentions
(MCII) intervention prompts students to reflect on the benefits
and barriers to achieving their goal (20) and then plan ahead for
how to overcome obstacles (15). We tested and replicated this
intervention in two online courses, on Computing and on Soci-
ology (n=17,963), and found that it increased the completion
rate by 15% (from 26 to 30%) and 32% (from 5.5 to 7.3%) for
students in individualist countries (such as the United States and
Germany) (14).
The culture-specific effect for the MCII intervention led us to
hypothesize that students in less-individualist countries could
benefit from a social accountabilityintervention that prompts
them to make a plan to ask people to regularly check in about
their course progress. This strategy can foster a sense of ac-
countability that strengthens goal motivation (21). In political
and education contexts, it has been shown to increase voting and
school attendance (22), but unlike the other interventions, we
had not previously tested it in online courses.
Results
Our primary hypothesis was that the main results from previous
studiesimproving course completion rates for targeted sub-
groups of studentswould replicate in a larger sample. Overall,
we did not find new evidence for our original large- to medium-
sized effects. Specifically, the long-term planning prompts did
Year 1 Year 2
123123
4
8
12
16
20
Week After Student Be
g
ins Course
Weekly Course Activity
Control Long
Plans
Short
Plans
Value−
Relevance MCII Social
Accountability
Fig. 1. Average student activity (count of course platform events) in the first 3 wk after exposure to each intervention. Points show covariate-adjusted means
on a logarithmic scale (to match the log-transformed outcome in the regression model) with cluster-robust SE bars.
2of6
|
www.pnas.org/cgi/doi/10.1073/pnas.1921417117 Kizilcec et al.
Downloaded by guest on June 16, 2020
not improve course completion for committed English-fluent
students (year 1: β=0.19 percentage points [pp], 95% CI =
[0.69, 1.07], t=0.43, P=0.670; year 2: β=0.23 pp, 95% CI =
[1.61, 1.14], t=0.33, P=0.741). The MCII intervention did
not improve course completion for students in individualistic
countries (year 2: β=0.25 pp, 95% CI =[0.87, 1.37], t=0.44,
P=0.662), and the social accountability intervention did not
improve completion rates for students in nonindividualistic
countries (year 2: β=0.89 pp, 95% CI =[0.22, 1.99], t=1.58,
P=0.115). Finally, the value-relevance intervention did not
improve completion rates across all of the courses for students in
less-developed countries (year 1: β=0.14 pp, 95% CI =[0.753,
1.03], t=0.30, P=0.764; year 2: β=0.04 pp, 95% CI =[1.37,
1.30], t=0.05, P=0.957).
A number of recent scale-up studies have failed to find effects
of similar magnitude as the original studies, raising questions
about whether the original effect estimates generalized beyond
the original populations and contexts (2329). Although our
initial results were disappointing, our study was designed for
multiple opportunities to preregister, test, explore, and then
register and test updated hypotheses. In the process, we found
consistent evidence for two benefits from the interventions that
were more limited in scope than the original findings.
Planning Supports Short-Term Persistence. We found that the plan-
making interventions slightly increased studentsengagement
during the first few weeks in the course. To measure persistence
consistently across diverse course structures, we developed a
statistical surrogate for intermediate course progress in our it-
erative research process (see SI Appendix for details). Specifi-
cally, we compiled student clickstream logs to measure their
follow-up activity in the first weeks after their first day in the
course. Fig. 1 shows studentsdaily activity for 3 wk following
their exposure to the intervention, grouped by condition for the
first and second year of the study.
Overall, the plan-making interventions increased students
week 1 activity levels in year 1 (short plans: β=0.0437, 95% CI =
[0.0066, 0.0808], t=2.31, P=0.021; long plans: β=0.0336, 95%
CI =[0.0033, 0.0706], t=1.79, P=0.074) and in year 2 (long
plans: β=0.1057, 95% CI =[0.0496, 0.1619], t=3.69, P<0.001;
MCII: β=0.0603, 95% CI =[0.0039, 0.1167], t=2.10, P=0.036;
social accountability: β=0.0788, 95% CI =[0.0225, 0.1351], t=
2.74, P=0.006). However, as Fig. 1 suggests, these treatment
effects were short-lived. The effect on week 2 activity was at-
tenuated in year 1 (short plans: β=0.0257, 95% CI =[0.0110,
0.0624], t=1.37, P=0.169; long plans: β=0.0493, 95% CI =
[0.0127, 0.0859], t=2.64, P=0.008) and year 2 (long plans: β=
0.0685, 95% CI =[0.0134, 0.1236], t=2.44, P=0.015; MCII: β=
0.0099, 95% CI =[0.0451, 0.0649], t=0.35, P=0.724), except
for a sustained effect of the social accountability intervention (β=
0.0939, 95% CI =[0.0387, 0.1491], t=3.34, P<0.001).
These results suggest that while planning prompts raised en-
gagement in educational goal pursuit, their benefits dissipated
over the span of a full course. Short-term effects may be con-
sequential for tasks that require one-time behavior, such as
voting or going to the doctors (30, 31), but other work suggests
that planning prompts may be unreliable for more complex goals
(32, 33). Online education requires sustained effort toward
complex, long-term goals, and the effects of our plan-making
interventions attenuated after 1 to 2 wk, and were not detect-
able in the final course completion rates.
Value-Relevance Intervention Closes the Global Achievement Gap in
Courses that Have One. We found that the value-relevance in-
tervention predictably reduces the global achievement gap,
insofar as there is a global achievement gap in the course. The
gap is defined by the difference in completion rates between
students in more-developed versus less-developed countries, as
demarcated by 0.7 on the United Nations Human Development
Index (HDI) (13). The gap was large on average (as in our
previous studies), but it was not uniform across courses and
occasionally even reversed. We therefore refined our hypothesis
for the value-relevance intervention in the second year to specify
that it would only be effective in courses with a significant global
gap, defined as a 0.2 SD lower completion rate for students in
less-developed than more-developed countries in the control
condition.
In courses with a significant global gap, the value-relevance
intervention increased the average completion rate among stu-
dents in less-developed countries by 2.79 pp in the first year
(95% CI =[1.30, 4.27], t=3.68, P<0.001) and by 2.74 pp in the
second year (95% CI =[0.32, 5.17], t=2.22, P=0.026). The
effect of the intervention is significant but an order-of-magnitude
smaller than in our prior study (Table 1) (13). In courses without
a global gap (or where it was reversed), post hoc analyses in-
dicate that the intervention lowered the average completion rate
among students in less-developed countries (year 1: β=1.62
pp, 95% CI =[2.73, 0.27], t=2.86, P=0.004; year 2:
β=1.71 pp, 95% CI =[3.27, 0.16], t=2.16, P=0.031).
While our prior study found that the intervention negatively
affected students in more developed countries (13), we found no
new evidence of this back-firing effect, neither in courses with a
global gap (year 1: β=0.45 pp, 95% CI =[0.52, 1.43], t=0.91,
P=0.363; year 2: β=0.62 pp, 95% CI =[2.46, 1.22],
t=0.66, P=0.509) nor in courses without a global gap (year 1:
β=0.08 pp, 95% CI =[0.83, 0.67], t=0.21, P=0.835; year
2: β=0.94 pp, 95% CI =[0.01, 1.89], t=1.95, P=0.051). The
findings are visualized in Fig. 2.
Consistent with its theoretical underpinnings and prior results,
the value-relevance intervention specifically benefits marginal-
ized students in environments where they are at risk for en-
countering psychological barriers (1719). The original intervention
effect replicates in contexts that most resemble the original courses
that featured a global achievement gap. In contrast, in courses
without a global gap, it is counterproductive to provide a value-
relevance intervention. This result highlights the need to account
for contextual variation when scaling an intervention from a few
select research sites to a broader set of contexts (34). Table 1
presents a comparison of results from our prior studies and the
scaled-up versions in the present research.
Subsequent exploratory analyses revealed that several other
interventions significantly improved completion rates for the
same population: Students in less-developed countries in courses
with a global gap (short plans year 1: β=2.44 pp, 95% CI =
[0.96, 3.92], t=3.23, P=0.001; long plans year 1: β=2.74 pp,
95% CI =[1.27, 4.22], t=3.64, P<0.001; MCII year 2: β=2.76
pp, 95% CI =[0.35, 5.18], t=2.24, P=0.025). However, the
effect did not replicate in the second year for the plan-making
intervention (β=1.06 pp, 95% CI =[1.33, 3.44], t=0.87, P=
0.386) or the social accountability intervention (β=1.23 pp, 95%
CI =[1.16, 3.62], t=1.01, P=0.314). The identification of a
responsive subpopulation may be evidence of a common un-
derlying mechanism for several interventions (i.e., self-reflective
writing) or evidence that these students are receptive to a range
of supports. The post hoc determination of which courses pre-
sent a global gap may also contribute to this unexpected pattern
of results.
Forecasting Where and for Whom an Intervention Will Work Is
Challenging. Our findings suggest that policymakers and admin-
istrators who deploy behavioral science interventions should
consider targeting specific students and contexts, such as stu-
dents in less-developed countries in courses with a global
achievement gap. However, the global gap is a characteristic that
can only be determined after the course has run, as it depends on
differences in completion rates in the control condition. We did
Kizilcec et al. PNAS Latest Articles
|
3of6
PSYCHOLOGICAL AND
COGNITIVE SCIENCES
Downloaded by guest on June 16, 2020
not find patterns to forecast when the global gap will occur; it
appears uncorrelated with institution (Harvard, MIT, Stanford),
subject domain (STEM versus Humanities, or at the level of
department or program), or any other features we examined. In
fact, among the 79 course offerings that were repeats of courses
that had been offered previously, the presence of a global gap in
repeat offerings matched the original offering only 60.8% of the
time. A predictive model (see SI Appendix for details) with
21 course-level features in year 1 could not forecast the occur-
rence of a global gap in year 2 significantly better than random
chance (54.3% accuracy, 95% CI=[44.1, 64.4], compared to
50.0%, 95% CI =[39.8, 60.2]).
One possibility is that behavioral science interventions need to
be targeted at a fine-grained individual level, but our analysis
suggests that this would have at most a modest impact on course
completion. We performed an exploratory analysis using machine-
learning algorithms to optimize an individualized policy (SI
Appendix) using data collected in year 1. We then estimated
that the average completion rate of this personalized policy
in year 2 is 13.38% (95% CI =[12.79, 13.98]). This is slightly
but not significantly higher than the estimated average com-
pletion rate of no intervention (12.81%, 95% CI =[12.23,
13.39]) or a randomly assigned intervention (13.08%, 95% CI =
[12.74, 13.38]). To realize the potential benefit of personalized
policies for students, the field will require more effective in-
terventions and more comprehensive collection of individual-
and course-level features to identify which students will benefit
from particular supports.
Discussion
Our preregistered analyses demonstrate that a value-relevance
intervention improves course completion for students in less-
developed countries in courses with a global achievement gap.
Post hoc analysis suggests that our other interventions may
similarly improve outcomes for these specific students in these
specific contexts. Our finding that plan-making interventions
have limited benefits provides further evidence that behavioral
insights have more promise in encouraging one-time, short-lived
actions than more continuous behaviors that require sustained
effort and habit change. These conclusions are consistent with a
number of recent scale-up studies in other domains that have
found diminished scope and magnitude for behavioral inter-
ventions (2329, 32).
In our original studies, we recommended that policymakers
and online instructors consider employing the tested behavioral
interventions in their own MOOCs (1214); we now conclude
that further research is necessary to predict in advance when
these interventions will help populations of students in need. The
population of students who complete surveys in MOOCs is ex-
ceptionally diverse along certain dimensions, but may not cap-
ture aspects of other populations of interest: For example, less
engaged students or those in other educational settings. As such,
we believe caution is warranted in applying our findings beyond
motivated and self-directed students in open-enrollment courses.
The kind of large-scale research that is needed to advance this
work is not well-represented in the dominant paradigm of ex-
perimental educational research. The NSF/IES Common
Guidelines for Education Research define a trajectory for ex-
perimental research that proceeds from pilot studies in labora-
tories, to initial implementations in field sites, to scale-up studies
designed to generate reliable estimates of the ability of a fully-
developed intervention or strategy to achieve its intended out-
comesacross multiple, diverse, real-world contexts (11). Many
large grants available to researchers require that they hold their
intervention constant across contexts.
Our present study confirms a principle that is central to social
psychology and the learning sciences: Context matters. Along-
side large-scale studies that test a single, fully developed in-
tervention across multiple contexts, scale-upfunding should be
available for approaches that assume interventions will need to
Table 1. Comparison of intervention results from prior research and this research for comparable interventions and subgroups
of students
Intervention Subpopulation Prior result Present result
Plan-making (long-term) Committed English-fluent
students
β=3.9 pp, χ
2(1)
=5.2, P=0.023,
n=2,053 (3 courses)
Year 1: β=0.19 pp, t=0.43,
P=0.670, n=26,586
Year 2: β=0.23 pp, t=0.33, P=0.741,
n=10,372
Value-relevance Students in less-developed countries
in courses with a global gap
Study 1: β=3.4 course activities,
z=2.82, P=0.005, n=227
Year 1: β=2.79 pp, t=3.68,
P<0.001, n=5,974
Study 2: β=24 pp, z=2.26,
P=0.024, n=64
Year 2: β=2.74 pp, t=2.22,
P=0.026, n=2,712
Mental contrasting
with implementation
intentions
Students in individualistic
countries
Study 1: β=1.8 pp, z=2.35,
P=0.019, n=4,628
Year 2: β=0.25 pp, t=0.44,
P=0.662, n=12,879
Study 2: β=3.9 pp, z=2.41,
P=0.016, n=3,248
Note that there are several differences between the prior and present research in terms of the implementation of intervention instructions and sample
exclusion criteria. Effects denote percentage point (pp) increases in course completion except where noted.
Courses with
Global Achievement Gap
Courses without
Global Achievement Gap
Control Value−relevance
intervention Control Value−relevance
intervention
0%
5%
10%
15%
20%
Course Completion
More−developed countries Less−developed countries
Fig. 2. Average course completion rate in all waves in the value-relevance
intervention and control condition by student context (more vs. less de-
veloped country) and course context (with vs. without global achievement
gap). Bars show covariate-adjusted means with cluster-robust SE bars.
4of6
|
www.pnas.org/cgi/doi/10.1073/pnas.1921417117 Kizilcec et al.
Downloaded by guest on June 16, 2020
be constantly refined and modified to support specific groups of
people across diverse contexts. These studies would be designed
to respond to concerns of temporal validity, the notion that the
effectiveness of interventions can vary as contexts and pop-
ulations change over time (35). Rather than treating large-scale
studies as the conclusion of a research trajectory, scale-up studies
should support new research into context-level variation that
cannot be explored in small field trials. We encourage greater
focus on the characteristics of different contexts that induce
variation in the effects of interventions to advance the devel-
opment of a science of context in education. In a new paradigm,
the question of what works?would be replaced with what
works, for whom, right here?
Materials and Methods
Adopting best practices from open science, we conducted our study in four
preregistered wavesof implementation. In each wave, we preregistered
hypotheses and analysis code, collected data, conducted post hoc analyses of
heterogeneous treatment effects, and refined the preregistration for the
subsequent wave. All preregistrations, analysis code, and data are available
online at https://osf.io/9bacu/. Our study plans were reviewed and approved
by the Institutional Review Boards at Harvard University, Stanford Univer-
sity, and MIT. Participants consented to research participation in agreeing to
the terms of service as part of the site registration process for EdX (Harvard,
MIT) and Open EdX (Stanford). EdX participants consented to research
participation again before taking the survey; Open EdX prominently dis-
played a message on every course enrollment page notifying participants
that they participate in research by using the platform.
In the first year (waves 1 to 2; September 2016 to December 2017), we
tested the value-relevance and short- and long-term plan-making inter-
ventions, individually and in combination, across 153 courses. In the
second year (waves 3 to 4; January 2018 to May 2019), based on early
findings, we shortened the overall survey, eliminated the short-term plan-
making intervention, simplified the value-relevance intervention, and added
MCII and social accountability interventions. We collected data in 94 courses.
In total, the 247 focal courses include all courses offered by the three in-
stitutions during this time period, unless a course did not implement the
survey, had fewer than 100 students assigned to a condition, or less than 1%
of students assigned to a condition completed the course. The focal courses
span a wide range of subjects (22% humanities, 40% social science, 29%
STEM, 9% computer science), sizes (between 102 and 16,645 students
assigned to a condition), and completion rates (from 1 to 65% among those
assigned to condition). We define the focal sample of students for our main
analyses to be everyone assigned to a condition for the first time (SI Ap-
pendix): 199,517 students in year 1 and 69,652 students in year 2.
Our primary outcome measure, borrowed from our previous research, is
course completion operationalized as earning a passing grade in the course.
We also developed a surrogate outcome (36) to measure proximate in-
tervention effects on early course engagement in terms of the log-scaled
number of actions students performed in the first week (days 2 to 8) and the
second week (days 9 to 16). We initially created a continuous measure of
progress in the course based on the percentage of videos viewed or as-
signments completed, but found that the wide heterogeneity of course
models prevented reasonable comparisons. Treatment effects were esti-
mated using preregistered regression models with individual- and course-
level covariates and course fixed effects. Hypotheses that specify a subgroup
effect were tested by fitting the model only for those students (see SI Ap-
pendix for additional methodological details, including outcomes tested,
exclusion criteria, and model specifications).
Data Availability. We provide de-identified datasets to run many of the
analyses reported in the paper, hosted on the Harvard Dataverse, which is
linked from within our Open Science Framework (OSF) repository (https://
osf.io/9bacu/). Some variables have been anonymized (e.g., course names
have been hashed), and other variables have been binned to preserve pri-
vacy (e.g., number of courses finished or country HDI). Where possible, all of
the analyses reported in the paper are conducted on this anonymized
dataset, and we confirm that the results are substantively identical when the
same analyses are conducted on the raw dataset. However, some analyses
reported in the paper cannot be conducted on anonymous data (e.g., tables
of demographic descriptives and the adaptive learning policy), so we only
include the analysis code and results.
The full raw datasets analyzed in this study are not publicly available to
protect the privacy of personally identifiable student information, and re-
strictions apply to the availability of these data, which were used under
license from Harvard, MIT, and Stanford. The datasets are maintained sep-
arately by each institution, and contacts and/or guidelines for data requests
are available at ir.mit.edu/mitx-data-request-checklist,https://dataverse.harvard.
edu/dataset.xhtml?persistentId=doi:10.7910/DVN/RTVIEM,andhttps://online.
stanford.edu/about-us.
Code Availability. All analysis code, output, and study materials are available
at https://osf.io/9bacu/.
ACKNOWLEDGMENTS. This work would not have been possible without Jon
Daries, Kathy Mirzaei, Andreas Paepcke, and Yigal Rosen. We thank Yifei
Zheng and Tiffany Zhong for assistance coding courses. This material is based
upon work supported by National Science Foundation Grant 1646978, a
Stanford Interdisciplinary Graduate Fellowship, and a Microsoft Faculty
Fellowship.
1. C. R. Sunstein, R. H. Thaler, Nudge: Improving Deisions About Health, Wealth, and
Happiness, (Yale Univeresity Press, New Haven, 2008).
2. G. M. Walton, T. D. Wilson, Wise interventions: Psychological remedies for social and
personal problems. Psychol. Rev. 125, 617655 (2018).
3. J. Reich, MOOC Completion and Retention in the Context of Student Intent (EDU-
CAUSE Review Online, 2014) https://er.educause.edu/articles/2014/12/mooc-comple-
tion-and-retention-in-the-context-of-student-intent.
4. R. F. Kizilcec, S. Halawa, Attrition and achievement gaps in online learningin
Proceedings of the Second ACM Conference on Learning @ Scale, (Association for
Computing Machinery, 2015), pp. 5766.
5. I. Chirikov, T. Semenova, N. Maloshonok, E. Bettinger, R. F. Kizilcec, Online education
platforms scale college STEM instruction with equivalent learning outcomes at lower
cost. Sci. Adv. 6, eaay5324 (2020).
6. C. Zhenghao et al., Whos benefiting from MOOCs, and why. Harv. Bus. Rev. 25,28
(2015).
7. J. Littenberg-Tobias, J. Reich, Evaluating access, quality, and equity in online learning:
A case study of a MOOC-based blended professional degree program. SocArXiv,
10.31235/osf.io/8nbsz (7 December 2018).
8. G. Chen, D. Davis, C. Hauff, G. J. Houben, Learning transfer: Does it take place in
MOOCs? An investigation into the uptake of functional programming in practicein
Proceedings of the Third ACM Conference on Learning@ Scale, (Association for
Computing Machinery, 2016), pp. 409418.
9. Y. Wang, L. Paquette, R. Baker, A longitudinal study on learner career advancement
in MOOCs. J. Learn. Anal. 1, 203206 (2014).
10. A. Napier, E. Huttner-Loan, J. Reich, Evaluating transfer of learning from MOOCs to
workplaces: A case study from teacher education and Launching Innovation in
Schools. Revista Iberoamericana de Educación a Distancia 23 (2), 4564 (2020).
11. Joint Committee of the Institute of Education Sciences, Department of Education, and
the National Science Foundation, Common Guidelines for Education Research and
Development (IES, DOE, and NSF, Washington, DC, 2013) https://www.nsf.gov/pubs/
2013/nsf13126/nsf13126.pdf.
12. M. Yeomans, J. Reich, Planning to learn: Planning prompts encourage and forecast
goal pursuit in online educationin Proceedings of the Seventh International Con-
ference on Learning Analytics & Knowledge, (Association for Computing Machinery,
2017), pp. 464473.
13. R. F. Kizilcec, A. J. Saltarelli, J. Reich, G. L. Cohen, Closing global achievement gaps in
MOOCs. Science 355, 251252 (2017).
14. R. F. Kizilcec, G. L. Cohen, Eight-minute self-regulation intervention raises educational
attainment at scale in individualist but not collectivist cultures. Proc. Natl. Acad. Sci.
U.S.A. 114, 43484353 (2017).
15. P. M. Gollwitzer, P. Sheeran, Implementation intentions and goal achievement: A
metaanalysis of effects and processes. Adv. Exp. Soc. Psychol. 38,69119 (2006).
16. T. Rogers, K. L. Milkman, L. K. John, M. I. Norton, Beyond good intentions: Prompting
people to make plans improves follow-through on important tasks. Behav. Sci. Policy
1,3341 (2015).
17. G. L. Cohen, J. Garcia, N. Apfel, A. Master, Reducing the racial achievement gap: A
social-psychological intervention. Science 313, 13071310 (2006).
18. C. S. Hulleman, J. M. Harackiewicz, Promoting interest and performance in high
school science classes. Science 326, 14101412 (2009).
19. R. F. Kizilcec, G. M. Davis, G. L. Cohen, Towards equal opportunities in MOOCs: Af-
firmation reduces gender & social-class achievement gaps in Chinain Proceedings of
the Fourth ACM Conference on Learning @ Scale, (Association for Computing Ma-
chinery, 2017), pp. 121130.
20. G. Oettingen, Future thought and behaviour change. Eur. Rev. Soc. Psychol. 23,163
(2012).
21. A. T. Hall, D. D. Frink, M. R. Buckley, An accountability account: A review and synthesis
of the theoretical and empirical research on felt accountability. J. Organ. Behav. 38,
204224 (2017).
Kizilcec et al. PNAS Latest Articles
|
5of6
PSYCHOLOGICAL AND
COGNITIVE SCIENCES
Downloaded by guest on June 16, 2020
22. T. Rogers, N. J. Goldstein, C. R. Fox, Social mobilization. Annu. Rev. Psychol. 69,
357381 (2018).
23. P. Hanselman, C. S. Rozek, J. Grigg, G. D. Borman, New evidence on self-affirmation
effects and theorized sources of heterogeneity from large-scale replications. J. Educ.
Psychol. 109, 405424 (2017).
24. D. S. Yeager et al., A national experiment reveals where a growth mindset improves
achievement. Nature 573, 364369 (2019).
25. C. R. Dobronyi, P. Oreopoulos, U. Petronijevic, Goal setting, academic reminders, and
college success: A large-scale field experiment. J. Res. Educ. Eff. 12,3866 (2019).
26. P. Oreopoulos, U. Petronijevic, The remarkable unresponsiveness of college students
to nudging and what we can learn from it (Tech. Rep. No. w26059)(National Bureau
of Economic Research, Cambridge, MA, 2019).
27. H. Lortie-Forgues, M. Inglis, Rigorous large-scale educational RCTs are often un-
informative: Should we be concerned? Educ. Res. 48, 158166 (2019).
28. K. A. Bird et al., Nudging at scale: Experimental evidence from FAFSA completion
campaigns. NBER working paper No. 26158(National Bureau of Economic Research,
Cambridge, MA, 2019).
29. A. S. Kristal, A. V. Whillans, What we can learn from five naturalistic field experiments
that failed to shift commuter behaviour. Nat. Hum. Behav. 4,18 (2019).
30. D. W. Nickerson, T. Rogers, Do you have a voting plan? Implementation intentions,
voter turnout, and organic plan making. Psychol. Sci. 21, 194199 (2010).
31. K. L. Milkman, J. Beshears, J. J. Choi, D. Laibson, B. C. Madrian, Planning prompts as a
means of increasing preventive screening rates. Prev. Med. 56,9293 (2013).
32. C. Townsend, W. Liu, Is planning good for you? The differential impact of planning on
self-regulation. J. Consum. Res. 39, 688703 (2012).
33. J. Beshears, H. N. Lee, K. L. Milkman, R. Mislavsky, J. Wisdom, Creating Exercise Habits
Using Incentives: The Tradeoff between Flexibility and Routinization. Manag. Sci.,in
press.
34. H. Allcott, Site selection bias in program evaluation. Q. J. Econ. 130, 11171165 (2015).
35. K. Munger, The limited value of non-replicable field experiments in contexts with low
temporal validity. Soc. Media Soc. 5, 2056305119859294 (2019).
36. R. L. Prentice, Surrogate endpoints in clinical trials: Definition and operational crite-
ria. Stat. Med. 8,431440 (1989).
6of6
|
www.pnas.org/cgi/doi/10.1073/pnas.1921417117 Kizilcec et al.
Downloaded by guest on June 16, 2020
... The proliferation of mobile technology has also provided students with unprecedented access to online resources and social learning opportunities (Ferri et al., 2020). Studies in Africa have demonstrated that mLearning has the potential to supplement traditional schooling, particularly in regions where educational resources are limited or where formal education is disrupted (Kizilcec et al., 2020). Volume 18, Issue 2, 2025 Analyzing teachers' perceptions of mobile learning in senior secondary schools: A frame model perspective in sub-Saharan Africa ...
Article
The paper focuses on a systematic review of existing literature regarding the teachers’ perceptions toward mobile learning (mLearning) in public secondary schools across sub-Saharan Africa. It emphasizes that this research synthesizes previously published work rather than presenting original findings. The review identifies significant challenges and opportunities associated with mLearning implementation, particularly in the context of the COVID-19 pandemic, which has accelerated the adoption of mobile technologies in education. Despite the growing body of literature on mLearning trends in sub-Saharan Africa, there remains a notable lack of systematic reviews specifically addressing stakeholder perceptions. This study aims to fill that gap by providing evidence-based insights into the role of mLearning and its potential to shape emerging pedagogical practices. In summary, this research is pivotal for understanding how teachers perceive mLearning, thus offering a foundation for future studies and policy formulations aimed at enhancing educational practices through technology in sub-Saharan Africa.
... Consequently, it has emerged as a significant avenue for education and training. Nevertheless, it is worth noting that the completion rate of online and blended learning courses continues to be rather low, as indicated by research conducted by authors (Reich & Ruipérez-Valiente 2019;Kizilcec et al. 2020). According to Bolliger et al. (2010), the observed low rates of course completion in online educational settings could perhaps be attributed to the absence of in-person social connection, resulting in feelings of isolation and, thus, an increased likelihood of not successfully finishing the course. ...
Article
Background: Blended learning, which integrates traditional face-to-face education with digital and online elements, has gained prominence in recent years as an efficient method to enhance student learning experiences. But it is worth noting that the completion rate of online and blended learning courses continues to be rather low, as indicated by research. Purpose: The purpose of the current study is to understand the intellectual discussion around dynamic learner engagement in the era of blended learning so that the areas of improvement can be earmarked and suitable strategies can be made to bridge the gap between the expectations of the learners and the actual delivery of the course. Method: A systematic review was conducted using a replicable search strategy. The Preferred Reporting Items for Systematic Reviews and Meta-Analyses 2020 (PRISMA 2020) statement guided this study. Many keyword combinations were searched for in the title, keyword, and abstract fields, according to the search criteria to arrive at the final corpus of papers. Result: The study found that in order to cultivate learners’ engagement in blended learning, it is imperative to investigate the various aspects that impact learning engagement. Through a comprehensive analysis of existing scholarly literature, the present study has successfully discovered a range of both internal and external elements that exert a significant influence on the level of engagement exhibited by learners. Active participation by students and innovative pedagogical tactics emerged as significant aspects in a blended learning format, allowing designers and educators to provide targeted support for students. Conclusion: The study provides clear insights for instructors into the various methods to strategize around in the blended learning environment format so that they can design better pedagogical tactics and increase learner engagement through active student participation.
... The results showed that there was no significance in both interventions and that SRL interventions had short-lived effects and that continuous support is needed. They also suggested integrating context-specific information into the SRL interventions so as to induce variation in the effects of the interventions (Kizilcec, et al., 2020). ...
Article
Full-text available
This research investigates the development and implementation of meta-learning systems for computer science skill acquisition through optimized learning pathways utilizing reinforcement learning models. The researcher addresses a critical gap in computer science education, where traditional instructional approaches frequently fail to accommodate diverse learning styles, prior knowledge bases, and cognitive development patterns. The study employs a mixed-methods research design combining quantitative performance metrics with qualitative assessments of learner experiences across multiple cohorts (n=142) of undergraduate and graduate computer science students. Through the implementation of a novel adaptive learning architecture, the researcher demonstrates how reinforcement learning algorithms can effectively model the skill acquisition process, dynamically adjusting content sequencing and difficulty to optimize learning trajectories. Results indicate that students engaging with the meta-learning system exhibited significantly improved performance metrics (27.8% increase in concept mastery, p<0.001) compared to control groups following traditional curriculum structures. Furthermore, the system demonstrated remarkable capability in identifying optimal learning pathways that diverged from expert-designed sequences, particularly benefiting learners with non-traditional backgrounds. Analysis of learning behavior patterns revealed that the system's adaptive mechanisms successfully mitigated common bottlenecks in programming concept acquisition, particularly in abstract data structures and algorithmic complexity domains. The research contributes to both theoretical understanding of meta-learning principles in educational contexts and practical applications for computer science curriculum design, offering implications for intelligent tutoring systems, curriculum development, and lifelong learning frameworks in rapidly evolving technical disciplines
Preprint
Full-text available
AI-driven education, particularly Large Language Models (LLMs), has the potential to address learning disparities in rural K-12 schools. However, research on AI adoption in rural India remains limited, with existing studies focusing primarily on urban settings. This study examines the perceptions of volunteer teachers on AI integration in rural education, identifying key challenges and opportunities. Through semi-structured interviews with 23 volunteer educators in Rajasthan and Delhi, we conducted a thematic analysis to explore infrastructure constraints, teacher preparedness, and digital literacy gaps. Findings indicate that while LLMs could enhance personalized learning and reduce teacher workload, barriers such as poor connectivity, lack of AI training, and parental skepticism hinder adoption. Despite concerns over over-reliance and ethical risks, volunteers emphasize that AI should be seen as a complementary tool rather than a replacement for traditional teaching. Given the potential benefits, LLM-based tutors merit further exploration in rural classrooms, with structured implementation and localized adaptations to ensure accessibility and equity.
Article
This research aimed to explore the trends of research on online learning technology for mathematics and science. The data were obtained from Scopus database using the Publish or Perish application as a reference management tool. Subsequently, the computational mapping analysis was conducted using VOSviewer with a bibliometric approach. The keywords used as a guide for the search process were “technology” AND “online learning” AND “math” OR “science”. Over a span of ten years (2014 to 2023), a total of 458 relevant articles indexed by Scopus were identified. The results showed that research on online learning technology for mathematics and science were categorized into four terms, namely technology, online learning, mathematics, and science. The term "technology" fell within cluster 1, consisting of 149 links, a total link strength of 4936, and 324 occurrences. Online learning was categorized within cluster 2, comprising 149 links, a total link strength of 2921, and 181 occurrences. Mathematics was classified within cluster 4, with 138 links, a total link strength of 939, and 54 occurrences. Lastly, science was also categorized in cluster 4, with 129 links, a total link strength of 2420, and 167 occurrences. The data analysis revealed a fluctuation in research output on online learning technology for mathematics and science. The number of articles increased from 22 in 2014 to 25 in 2015, followed by a decrease to 17 in 2016, increased to 20, 21, 32, 40, 99, and 125 in 2017-2022, and finally decreased to 57 in 2023. These results were illustrated by VOS viewer analysis. The discussion presented can serve as a foundational point for future research on related topics.
Chapter
Full-text available
The integration of AI into Learning Management Systems (LMS) communication tools represents a major advancement in online learning. This chapter explores how AI is transforming learner interaction and engagement in distance education, starting with an overview of existing tools like discussion forums and video conferencing. It then highlights AI-driven innovations such as chatbots and virtual assistants for personalized support, sentiment analysis for better moderation, and real-time translation to enhance accessibility. AI-generated personalized recommendations and collaboration tools further enrich the learning experience and reduce educators' workload. Despite these benefits, ethical challenges like privacy and algorithmic biases arise. As AI continues to evolve, it promises to reshape not only communication but also the entire structure of education in the digital.
Article
Full-text available
Meeting global demand for growing the science, technology, engineering, and mathematics (STEM) workforce requires solutions for the shortage of qualified instructors. We propose and evaluate a model for scaling up affordable access to effective STEM education through national online education platforms. These platforms allow resource-constrained higher education institutions to adopt online courses produced by the country’s top universities and departments. A multisite randomized controlled trial tested this model with fully online and blended instruction modalities in Russia’s online education platform. We find that online and blended instruction produce similar student learning outcomes as traditional in-person instruction at substantially lower costs. Adopting this model at scale reduces faculty compensation costs that can fund increases in STEM enrollment.
Article
Full-text available
Across five field experiments with employees of a large organization (n = 68,915), we examined whether standard behavioural interventions (‘nudges’) successfully reduced single-occupancy vehicle commutes. In Studies 1 and 2, we sent letters and emails with nudges designed to increase carpooling. These interventions failed to increase carpool sign-up or usage. In Studies 3a and 4, we examined the efficacy of other well-established behavioural interventions: non-cash incentives and personalized travel plans. Again, we found no positive effect of these interventions. Across studies, effect sizes ranged from Cohen’s d = −0.01 to d = 0.05. Equivalence testing, using study-specific smallest effect sizes of interest, revealed that the treatment effects observed in four out of five of our experiments were statistically equivalent to zero (P < 0.04). The failure of these well-powered experiments designed to nudge commuting behaviour highlights both the difficulty of changing commuter behaviour and the importance of publishing null results to build cumulative knowledge about how to encourage sustainable travel. Kristal and Whillans conducted five field experiments (n = 68,915) designed to increase sustainable commuting using standard behavioural science tools. The interventions’ failures highlight the difficulty of changing commuter behaviour using this approach.
Article
Full-text available
A global priority for the behavioural sciences is to develop cost-effective, scalable interventions that could improve the academic outcomes of adolescents at a population level, but no such interventions have so far been evaluated in a population-generalizable sample. Here we show that a short (less than one hour), online growth mindset intervention—which teaches that intellectual abilities can be developed—improved grades among lower-achieving students and increased overall enrolment to advanced mathematics courses in a nationally representative sample of students in secondary education in the United States. Notably, the study identified school contexts that sustained the effects of the growth mindset intervention: the intervention changed grades when peer norms aligned with the messages of the intervention. Confidence in the conclusions of this study comes from independent data collection and processing, pre-registration of analyses, and corroboration of results by a blinded Bayesian analysis.
Article
Full-text available
My first publication as a grad student was a field experiment using Twitter “bots” to socially sanction users engaged in racist harassment. The ascendant paradigm in quantitative social science emphasizes the need for research to be “internally valid,” with a preference for randomized control trials like the one I conducted. That research project was well received, both by the political science discipline and the public, but I no longer believe that one-off field experiments are as valuable a tool for studying online behavior as I once did. The problem is that the knowledge they generate decays too rapidly (alternatively, that the realms in which it can be applied are too few) because the object of study changes too rapidly. I have been developing the concept of “temporal validity” as form of “external validity” that is particularly relevant to the study of social media. I suggest two avenues for producing more temporally valid research: (1) faster, more transparent publication (adapting the CS model of conference proceedings); (2) a “hollowing out” of empirical research, replacing medium-scale experimentation like my own work with either purely descriptive work (ethnographic or large-scale) or with massive, collaborative, replicable experimentation.
Article
Full-text available
Long-standing social problems such as poor achievement, personal and intergroup conflict, bad health, and unhappiness can seem like permanent features of the social landscape. We describe an approach to such problems rooted in basic theory and research in social psychology. This approach emphasizes subjective meaning-making—working hypotheses people draw about themselves, other people, and social situations; how deleterious meanings can arise from social and cultural contexts; how interventions to change meanings can help people flourish; and how initial change can become embedded to alter the course of people’s lives. We further describe how this approach relates to and complements other prominent approaches to social reform, which emphasize not subjective meaning-making but objective change in situations or in the habits and skills of individuals. In so doing, we provide a comprehensive theoretical review and organization of a psychologically informed approach to social problems, one that encompasses a wide-range of interventions and applies to diverse problem areas.
Article
Habits involve regular, cue-triggered routines. In a field experiment, we tested whether incentivizing exercise routines—paying participants each time they visit the gym within a planned, daily two-hour window—leads to more persistent exercise than offering flexible incentives—paying participants each day they visit the gym, regardless of timing. Routine incentives generated fewer gym visits than flexible incentives, both during our intervention and after incentives were removed. Even among subgroups that were experimentally induced to exercise at similar rates during our intervention, recipients of routine incentives exhibited a larger decrease in exercise after the intervention than recipients of flexible incentives. This paper was accepted by Yan Chen, decision analysis.
Article
As massive open online courses (MOOCs) shift toward professional degree and certificate programs, can they become a global on-ramp for increasing access to emerging fields for underrepresented groups? This mixed-methods study addresses this question by examining one of the first MOOC-based blended professional degree programs, which admitted students to an accelerated residential master's program on the basis of performance in MOOCs and a proctored exam. We found that male students and students with master's degrees were more likely to complete the online program and the blended program had more male students and more students with master's degrees than students in the existing residential program. Students who enrolled in the blended graduate program earned higher average grades than students in the residential program earned in their in-person courses (3.86 vs 3.75, p < .01). The findings of this study provide an example of how new online learning models can serve particular niches, but may not address broader equity challenges.
Article
This article presents an independent large-scale experimental evaluation of two online goal-setting interventions. Both interventions are based on promising findings from the field of social psychology. Approximately 1,400 first-year undergraduate students at a large Canadian university were randomly assigned to complete one of two online goal-setting treatments or a control task. In addition, half of treated participants were offered the opportunity to receive follow-up goal-oriented reminders through e-mail or text messages to test a cost-effective method for increasing the saliency of treatment. Across all treatment groups, we observed no evidence of an effect on grade point average, course credits, or second-year persistence. Our estimates are precise enough to discern a 7% standardized performance effect at a 5% significance level. Our results hold by subsample, for various outcome variables, and across a number of specifications.
Article
There are a growing number of large-scale educational Randomized Controlled Trials (RCTs). Considering their expense, it is important to reflect on the effectiveness of this approach. We assessed the magnitude and precision of effects found in those large-scale RCTs commissioned by the EEF (UK) and the NCEE (US) which evaluated interventions aimed at improving academic achievement in K-12 (141 RCTs; 1,222,024 students). The mean effect size was 0.06 standard deviations (SDs). These sat within relatively large confidence intervals (mean width 0.30 SDs) which meant that the results were typically uninformative (the median Bayes factor was 0.68). We argue that our field needs, as a priority, to understand why educational RCTs typically find small and uninformative effects.
Preprint
Many higher education institutions have begun offering opportunities to earn credit for in-person courses through massive open online courses (MOOCs). This mixed-methods study examines the experiences of students participating in one of the first iterations of this trend: a blended professional master's degree program that admitted students based on performance in MOOC-based online courses. We found that the blended master's program attracted a cohort of highly educated mid-career professionals from developed countries who were looking for more flexible alternatives to traditional graduate programs. Success in the online courses was correlated with higher levels of prior formal education and effective use of learning strategies. Students who enrolled in the blended graduate program reported being academically prepared for their coursework and had higher GPAs (3.86, p<0.01) than students in the residential program (3.75). The findings of this study suggest that the technological affordances of MOOC-based online and blended degrees will neither transform higher education nor solve its most stubborn equity challenges, but there may be particular niches where they provide a valuable service to learners in particular programs and contexts.