ArticlePDF Available

Theory of Change Analysis: Building Robust Theories of Change

Authors:

Abstract

Models for theories of change vary widely as do how they are used. What constitutes a good or robust theory of change has not been discussed much. This article sets out and discusses criteria for robust theories of change. As well, it discusses how these criteria can be used to undertake a vigorous assessment of a theory of change. A solid analysis of a theory of change can be extremely useful, both for designing or assessing the designs of an intervention as well as for the design of monitoring regimes and evaluations. The article concludes with a discussion about carrying out a theory of change analysis and an example.L’utilisation qui est faite de modèles de théories du changement varie grandement. Par ailleurs, il y a peu de discussion sur ce qui constitue une bonne ou solide théorie du changement. Le présent article décrit et analyse les critères de détermination de la robustesse d’une telle théorie. De plus, il discute de la façon dont ces critères peuvent servir à l’évaluation rigoureuse d’une théorie du change-ment. Une analyse approfondie d’une théorie du changement peut être extrêmement utile, autant pour concevoir ou évaluer la conception d’une intervention, que pour concevoir des évaluations et systèmes de monitorage. L’article se termine avec une discussion sur l’analyse d’une théorie du changement et un exemple.
Theory of Change Analysis:
Building Robust Theories of Change
John Mayne
Ottawa, Ontario
Abstract: Models for theories of change vary widely as do how they are used.
What constitutes a good or robust theory of change has not been discussed much.
is article sets out and discusses criteria for robust theories of change. As well, it
discusses how these criteria can be used to undertake a vigorous assessment of a
theory of change. A solid analysis of a theory of change can be extremely useful, both
for designing or assessing the designs of an intervention as well as for the design of
monitoring regimes and evaluations. e article concludes with a discussion about
carrying out a theory of change analysis and an example.
Keywords: analysis of theories of change, criteria for good theories of change, impact
pathways, theory of change
Résumé: Lutilisation qui est faite de modèles de théories du changement varie
grandement. Par ailleurs, il y a peu de discussion sur ce qui constitue une bonne
ou solide théorie du changement. Le présent article décrit et analyse les critères de
détermination de la robustesse d’une telle théorie. De plus, il discute de la façon
dont ces critères peuvent servir à l’évaluation rigoureuse d’une théorie du change-
ment. Une analyse approfondie d’une théorie du changement peut être extrêmement
utile, autant pour concevoir ou évaluer la conception d’une intervention, que pour
concevoir des évaluations et systèmes de monitorage. Larticle se termine avec une
discussion sur lanalyse d’une théorie du changement et un exemple.
Mots clés: analyse de théories du changement, caractéristiques d’une bonne théorie
du changement, chaine des résultats, théorie du changement
INTRODUCTION
eories of change (ToCs) are now widely used in evaluations. ey are the basis
of theory-based evaluations (Coryn, Noakes, Westine, & Schroter 2011; Donald-
son, 2007; Funnell & Rogers 2011; Rogers, 2007). As many have noted, the specic
models used vary greatly (James, 2011; Valters, 2014; Vogel, 2012) and there is
no overall agreement on just what comprises a ToC. Funnell and Rogers (2011,
pp.15–34) discuss the range of terms used and their histories, as does Patton (2008,
pp. 336–340). Further, what constitutes a good or solid ToC is not at all clear; the
characteristics or criteria of a robust ToC have not been widely discussed.
Corresponding author: John Mayne, john.mayne@rogers.com
© 2017 Canadian Journal of Program Evaluation / La Revue canadienne d’évaluation de programme
32.2 (Fall / automne ), 155–173 doi: 10.3138/cjpe.31122
156 Mayne
© 2017 CJPE 32.2, 155–173 doi: 10.3138/cjpe.31122
is article discusses criteria for a robust ToC and a tool for carrying out anal-
ysis of ToCs, namely eory of Change Analysis (ToCA) to assess and strengthen
ToCs.
When discussing specic aspects of ToCs and presenting examples, I will be
using the behaviour change-based ToC model shown in Figure 1. Behaviour change-
based ToCs are discussed in Mayne (2015) and the COM-B model in Mayne (2016a).1
However, the steps and principles discussed apply to theories of change generally.
SOME TERMS
Given the diversity of how terms around ToCs and results are used, let me rst a
review the terms being used here:
Results is used to include outputs, outcomes, and impacts, where impacts
are the nal outcomes aecting well-being. A result statement is the exact
text used to describe the result. e term intervention is used here to de-
scribe specic activities undertaken to make a positive dierence in out-
comes and impacts of interest. It covers policies, programs, and projects.
1 e COM-B model postulates that behaviour (B) occurs as the result of interaction be-
tween three necessary conditions, capabilities (C), opportunities (O), and motivation (M).
Behaviour
Change
Capacity Change
Reach &
Reaction
Goods & Services
/Activities
Direct
Benefits
Improved
Wellbeing
Reach
Assumptions
Capacity
Change
Assumptions
Behaviour
Change
Assumptions
Direct Benefits
Assumptions
Wellbeing
Assumptions
External
Influences
Capability Opportunity
Motivation
Timeline
Supporting
Activities
to help bring
about the
assumptions,
including the
enabling
environment
Figure 1. The COM-B Based Theory of Change
Building Robust Theories of Change 157
CJPE 32.2, 155–173 © 2017doi: 10.3138/cjpe.31122
Impact pathways describe causal pathways showing the linkages between
a sequence of steps in getting from activities to impact. An intervention
may have several pathways to impact.
A theory of change (ToC) adds to an impact pathway by describing the
causal assumptions behind the links in the pathway—what has to hap-
pen for the causal linkages to be realized. eories of change are models
of how change is expected to happen (ex ante case) or how change has
happened (ex post case).
Rationale assumptions identify the underlying hypotheses or premise(s)
on which the intervention is founded.
Causal link assumptions are the salient events or conditions necessary (or
likely necessary) for a particular causal link in a ToC to be realized; if the
assumption doesnt hold, then the expected eect from that link will not
occur. is can be a very demanding requirement, if interpreted literally.
We can rather think in probabilistic terms, whereby causal link assump-
tions can be thought of as likely necessary assumptions, events, and condi-
tions that almost always have to occur for the causal link to work.
Further discussion of these terms and other alternative terms used such as
logic models and program theory can be found in Mayne (2015).
Because they are necessary or likely necessary, causal link assumptions also
represent risks to the causal link occurring—the risk being that the assumption does
not occur, that is, is not realized. For example, if an assumption is that local govern-
ment takes some action, the risk is that it does not take the action. Consequently,
rather than listing assumptions and risks, one can just identify assumptions.
Typically, some assumptions are less likely to be realized than others. For
example, if an assumption is that some party, perhaps a local government, will
take some action that has not been taken before, and nothing is being done to en-
courage the government to do so, then that assumption is quite likely at risk—and
indeed may not be plausible. If an assumption is that a market will emerge for a
new product and nothing is being done to encourage such a market, then that as-
sumption is at risk. In addition, an assumption may be at risk because of counter
pressures trying to ensure the assumption is not realized. An assumption that
monitoring will be done by a third party may be at risk if there are other powerful
parties who do not want the monitoring to be eective.
For theory of change analysis, I will call these at-risk assumptions. In an ex ante
situation, at-risk assumptions represent potential gaps in the design of the interven-
tion and likely serious threats to the intervention working. As a result, one may want
to identify possible conrming actions that could be taken early on to give assurances
that the assumption is likely to be realized, or corrective actions that might be taken
to mitigate the at-risk assumptions. In an ex post situation, these are areas that need
special attention in evaluations to see if in fact anything was done to address the risk.
In ex ante situations, it is important to keep the timeline in mind. At-risk as-
sumptions for causal links well in the future may be less of a problem—realizing
158 Mayne
© 2017 CJPE 32.2, 155–173 doi: 10.3138/cjpe.31122
the risk, actions could be taken later to address the issue. Many assumptions would
not be expected to be at risk, such as when the assumption can be expected to be
realized based on past experience and/or research, if the intervention design is
solid, or even if it is agreed by stakeholders that it is likely to occur.
In a theory of change model, at-risk assumptions could be identied by bold-
ing the assumption and discussing it in the accompanying text.
ISSUES IN ANALYZING A THEORY OF CHANGE
In addition to their use in evaluations, ToCs have also been found useful in
designing interventions or assessing the designs of interventions (Leeuw, 2012;
Rey, Brousselle, & Dedobbeleer, 2012; Tremblay, Brousselle, Richard, & Beaudet
2013). Mayne (2015) and Mayne and Johnson (2015) discuss a variety of uses of
theories of change. Mayne (2015) and Johnson, Mayne, Grace, and Wyatt (2015)
discuss some forms of analysis of ToCs. However, no structured approach for such
analysis has yet been proposed.
ose developing theories of change use forms of analysis both during de-
velopment and aer. However, given the numerous elements of a ToC and the
various possible purposes, it is useful to undertake a structured analysis with
specic aims in sight. e theory of change analysis (ToCA) discussed here aims
at addressing two questions:
1. Does the intervention ToC appear robust? at is, is the ToC structurally
sound and plausible?
2. What are the implications for monitoring and evaluating the intervention?
ToCA is done on a proposed ToC, one that has been developed to reect how
an intervention is working or was expected to work; hence the “appear” term in the
question. Reality might suggest that the intervention and its ToC were not in fact
that robust! But a priori, before undertaking extensive data collection, we would
want to identify any evident shortcomings in the ToC and hence the intervention
design. And ex post, if we nd that a ToC that has been used to model an interven-
tion is not very robust, we might nd that helpful in explaining a less than successful
intervention and/or identifying issues that an evaluation should explore.
Criteria for Robust Theories of Change
When a ToC is being developed, the expectation is that it is not just a bunch
of ideas, but that it is well articulated, credible, plausible, and logical—that it
is robust. A robust ToC is defensible, would support a well-designed plausible
intervention design, and would provide a solid basis both for monitoring and for
theory-based evaluations.
A related idea is that of a ToC being evaluable, for which Rick Davies (2012) has
set out a list of criteria. Daviess criteria are quite broad in their coverage, meant to
Building Robust Theories of Change 159
CJPE 32.2, 155–173 © 2017doi: 10.3138/cjpe.31122
include anything that is called a ToC. And indeed, as noted above, a wide range of
models and representations have been used to depict theories of change, and several
of Davies’s criteria challenge what has been set out as to whether it is a ToC at all.
My starting point is a little dierent. In dening above what a ToC is, it is
assumed that what is being examined sets out the pathways of change as a causal
sequence of results, and assumptions behind the pathways.
us, several of Davies’s criteria are “assumed,” namely testable, explained,
complete, and inclusive. How well those criteria are addressed in a ToC is part of
the robust criteria discussed below. Most of Davies’ other criteria are covered in the
robust criteria as well. In addition, several other criteria are needed to assess the
robustness of a ToC.
A robust ToC is one that is structurally sound and plausible. A robust ToC
supports a solid and plausible intervention design: with this design, it is reason-
able to expect that the intervention, if implemented as designed, will be able to
contribute to the intended results. Criteria for a robust theory of change for an
intervention would address the following questions:
For a structurally sound ToC:
1. Is the ToC understandable? Are there pathways of results, and are causal
link assumptions set out? Is there a reasonable number of results?
2. Are the ToC results and assumptions well dened?
3. Is the timing sequence of results and assumptions plausible?
4. Is the ToC logically coherent? Do the results follow a logical sequence?
Are the causal link assumptions pre-events and conditions for the sub-
sequent eect? Is the sequence plausible or at least possible?
5. Are the causal link assumptions necessary or likely necessary?
6. Are the assumptions independent of each other (recognizing that some
assumptions may apply for more than one causal link)?
For a structurally sound ToC that is plausible:
7. Is the ToC generally agreed?
8. Are the results and assumptions, or at least the key results and as-
sumptions, measurable? What is the likely strength or status of evi-
dence?
9. Are the causal link assumptions likely to be realized? Are at-risk assump-
tions mitigated through conrming or corrective actions?
10. Are the sets of assumptions for each causal link along with the prior
causal factor plausibly sucient to bring about the eect?
11. Is the level of eort (activities and outputs) commensurate with the ex-
pected results?
12. To what extent are the assumptions sustainable?
160 Mayne
© 2017 CJPE 32.2, 155–173 doi: 10.3138/cjpe.31122
A ToC that is reasonably robust would provide a solid basis for using the
ToC (Mayne & Johnson, 2015) in (a) designing and planning an intervention,
(b) managing an intervention, (c) assessing and evaluating an intervention,
and (d) scaling up an intervention. Robustness, as imagined here, is not a 0-1
variable. Meeting all the criteria could be quite demanding. Rather, in most
cases, one would be improving a ToC over time, moving toward a more robust
version.
ere is evident need for an intervention to be plausible. At the outset, clear
gaps or aws in the design will most probably lead to a less successful inter-
vention. Evaluability assessments are now seen as exploring the plausibility of
intervention design with a view to improving the design and/or to identifying
if it makes sense to undertake an evaluation (Davies, 2013; Peersman, Guijt, &
Pasanen, 2015; Trevisan & Walser, 2014). e criteria here for a robust ToC in-
clude those used in evaluability assessments. e criteria also include those set
out for SMARTly describing outcomes (Smart, Measurable, Achieved, Relevant,
Timely) in Outcome Harvesting (Wilson-Grau & Britt, 2013).
M&E Implications
One purpose of a ToC is to provide a framework for setting out monitoring
and evaluation plans. In carrying out ToC Analysis, it becomes clear just what
needs to be monitored and paid attention to in evaluations. Questions here
would be:
1. What data on results and assumptions should be monitored?
2. What issues need attention in an evaluation?
3. What is the likely strength or current status of evidence for the various
results and assumptions, and in particular for each causal link?
Table 1 pulls these criteria together for ToCA. Each criterion is then dis-
cussed. ese criteria build on ones I suggested earlier (Mayne, 2011).
Ex Ante and Ex Post Perspectives
In carrying out ToCA, it is important to keep in mind the perspective being used,
namely if the situation is ex ante or ex post. e analysis is similar in both cases,
but the implication of the ndings will dier.
e context for the ex ante perspective is where a ToC is being developed
for an intervention that has yet to be implemented or is in the early stages
of implementation. e intent would be to develop a robust ToC to match a
plausible intervention design, so that at the outset it seems reasonable that the
intervention would bring about the expected results. In this setting, ToCA can
be used to
facilitate agreement on a ToC
identify possible gaps in the intervention design and what can be done
Building Robust Theories of Change 161
CJPE 32.2, 155–173 © 2017doi: 10.3138/cjpe.31122
identify results and assumptions that need monitoring
identify issues that a future evaluation needs to address.
In an ex post setting, the intervention has been in operation for some time
and an evaluation is to be undertaken to see the extent to which the intervention
has actually worked. Some monitoring data may have been gathered and some
changes in the intervention may have been made over time. ere is a need to
either build (reconstruct) a ToC or revise an earlier ToC to reect how the inter-
vention is now seen as working. Ex post ToCA can be used to
facilitate agreement on a robust ToC, oen a reconstructed ToC
identify current intervention design weaknesses that may explain limited
expected results being achieved
identify results and assumptions data that an evaluation needs to collect
or get from monitoring data
identify evaluation questions that need addressing in an evaluation.
Note that in this ex post scenario, ToCA itself would not be assessing if the
ToC was in fact realized. at would be done as part of the evaluation, using
something like contribution analysis.
In either case then, ToC Analysis would seek to
strengthen the ToC: identify and correct any structural weaknesses in
proposed theories of change.
strengthen the intervention design: identify weakness in intervention
design and what could be, or should have been, done to strengthen the
design
identify data needs: identify monitoring and evaluation data that need to
be collected for assessing performance of the intervention.
THE THEORY OF CHANGE ANALYSIS CRITERIA
e criteria in Table 1 can be used to assess the robustness of the ToC and the
underlying intervention. However, as noted earlier, robustness is not a 0–1 rating.
at is, because there can be dierent models for the ToC of an intervention with
dierent levels of detail, the criteria need to be applied in a sensible manner. ey
might best be thought of as guidelines for assessing the strength of a ToC and the
intervention it represents.
Overall Criteria
Understandable: e ToC and especially its pathways should be clearly evi-
dent so that readers understand the intervention in the same way. I have argued
elsewhere that a complex ToC needs to be unpacked into several nested ToC mod-
els (Mayne, 2015). Further, in any one ToC, there should be a reasonable number
of results statements, so that the ToC model is “readable” to others beyond those
162 Mayne
© 2017 CJPE 32.2, 155–173 doi: 10.3138/cjpe.31122
Table 1. Criteria for Theory of Change Analysis
Overall Criteria
Understandable Is the logic and structure of the ToC clear?
Agreed To what extent is the ToC agreed or contestable?
Level of eort Are the activities and outputs of the intervention commensurate
with the expected results?
Criteria for Each Result
Well-dened Is the results statement unambiguous?
Plausible timing Is the time frame for the result reasonable?
Logical coherence Does the result follow logically from the previous result? Is the
sequence plausible or at least possible?
Measurable Is there a need to measure the result? How can the results be
measured? What is the likely strength or status of evidence for
the result being realized?
M&E implications What are the implications for monitoring and evaluation?
Criteria for Each Assumption
Well-dened Is the assumption unambiguous?
Logical coherence Is the assumption a precondition or event for the eect sought?
Justied What is the justication for the assumption as being necessary
or likely necessary?
Realized Is it plausible that the assumption will be realized? Are there
at-risk assumptions that should be addressed?
Sustainable Is the assumption sustainable?
Measurable Is there a need to measure the assumption? How can the
assumption be measured? What is the likely strength or status of
evidence for the assumption being realized?
M&E implications What are the implications for monitoring and evaluation?
Criteria for Each Causal Link
Independence Are the assumptions for the link independent from each other?
A sucient set Are the set of causal link assumptions along with the prior
causal factor sucient to bring about the eect? Is the link
plausible?
Strength/Status of
evidence
What is the strength or current status of evidence for the causal
link being realized?
who developed it. A rule of thumb that I have used is that if you have more than
13 result “boxes,” you may have a mess instead of a ToC.
Agreed: If the ToC, no matter how well constructed, is the product of just one
person or a small group, it may not have much support or buy-in, and may not be
Building Robust Theories of Change 163
CJPE 32.2, 155–173 © 2017doi: 10.3138/cjpe.31122
sustainable in the sense of not being used or not lasting long. Broad agreement
among stakeholders would usually suggest a more robust ToC, oen built up
through a participatory approach to building the ToC. And there may be dierent
views on how the intervention is supposed to work. In this case, one may need to
build more than one robust ToC and check each against reality in due course, See,
for example, Hansen and Vedung (2010).
Level of eort: is is a rough check on the plausibility of the intervention.
Does it seem reasonable that the activities of the intervention and their outputs
will be enough for the intervention to make a dierence in the ways expected?
Interventions sometimes have quite ambitious intentions that are expected to be
realized from a quite modest level of eort.
Criteria for the Results
Well dened: e results need to be as well dened as practical as to their
meaning and content, and their measurability. ey should not be subject to dif-
ferent interpretations by dierent readers.
Plausible timing: ere should be an indication of when the results are ex-
pected to come about, and the time frame set out should be realistic, that is, plau-
sible. Setting out realistic timing for when results can be expected is frequently
neglected in developing ToCs, indeed oen completely absent. Unrealistic expec-
tations about timing can point to quite unrealistic interventions. Even less atten-
tion is paid to the trajectory of the expected results, as Woolcock (2009) discusses.
Logical coherence: at is, the step-by-step model from activities/outputs to im-
pact should make sense, based on plausible or at least possible logic and perhaps prior
evidence. e distinction here between plausible and possible logic reects the fact
that dierent ToC models provide dierent levels of detail. A “possible” logic sequence
implies that the causal step is possible but represents a large leap in logic, which may
be due to the level of detail in the ToC or to a causal link at-risk. Remember, the ToC
is a model of expectations, which may of course turn out otherwise. If the ToC is
behaviour-based, such as discussed by Mayne (2015) or Morton (2015) and illustrated
in Figure 1, this signicantly strengthens the logical coherence of the model.
Measurable: e results, or at least key results, should be measurable—there
are reliable and valid measures of the results, and the needed data can be (readily)
collected. Depending on the use being made of the ToC, there may not be a need to
measure all the results set out in a pathway. For example, in a behaviour-based theory
of change model, measuring capacity change can sometimes be a challenge. On the
other hand, measuring behaviour changes is usually much simpler, and may be all
that is needed if the expected behaviour changes have occurred and other aspects of
the model are veried. e ToC analysis should indicate if the result (a) needs to be
measured, (b) might be useful to measure, or (c) do not really need to be measured. It
is useful here to note the likely strength of evidence based on the measures.
Implications for monitoring and evaluation: As part of the analysis, one can also
assess what the implications of each component of the ToC are for monitoring and
evaluation. Implications could be identifying evaluation questions to be addressed;
164 Mayne
© 2017 CJPE 32.2, 155–173 doi: 10.3138/cjpe.31122
issues that need to be carefully watched or explored; issues, results, and/or assump-
tions that should be monitored; and/or identifying data that should be collected.
e analysis would identify specic M&E actions that should be taken to
strengthen monitoring and evaluation.
Criteria for the Assumptions
Well dened: e events and conditions set out in the assumptions need to be
as well dened as practicable as to their meaning, content, and measurability. ey
should not be subject to dierent interpretations by dierent readers.
Logical coherence: Because the assumption should be needed for the eect to
occur, it should be a logical precondition or event.
Justied: e assumptions are justied by a solid argument as necessary or
likely necessary events or conditions for the causal link to work.
Realized: One should expect that the assumptions would be realized. at is,
there is general agreement, strong logic, actions being taken, or prior evidence that
make the assumption plausible. e analysis could identify the at-risk assumptions
that exist to the intervention and corrective or conrming actions that could or need
to be taken to mitigate the risk. Ex ante, this would identify weaknesses in the inter-
vention design and what might be done to strengthen the design. Ex post, it would
identify assumptions that need careful examination in an evaluation. In essence,
here one is assessing the degree of control the intervention has over the assumption.
Sustainable: An assumption may be realized during the period of the inter-
vention, but one normally would hope that the assumption is sustainable aer the
intervention is over. If not, then the assumption is at future risk, as would be the
causal link, the result in question, and indeed the intervention. Where sustainability
is an issue, the intervention might want to undertake some form of corrective action.
Measurable: e assumptions, or at least key ones, are measurable: there are reli-
able and valid indicators, the relevant data can be (readily) collected, and/or there is
adequate prior evidence. e analysis should indicate if the assumption (a) needs to be
measured, (b) might be useful to measure, or (c) does not really need to be measured.
Again, it is useful here to note the likely strength of evidence based on the measures.
e M&E Implications criteria are discussed above.
Criteria for Each Causal Link
Independence: For each causal link, the assumptions should be independent
of each other—that is, be separate events/conditions—and hence be a minimum
set of assumptions, recognizing that the same assumption may be needed for more
than one causal link.
A sucient set: e set of the initial result plus assumptions for the causal
link should be seen as sucient for that link to work, that is, for the cause plus
assumptions to contribute to the eect. e link should be plausible—the link
causal package should be enough to likely bring about the eect.
Strength/status of evidence: is nal criterion is about the likely strength of the
evidence on the causal link occurring (ex ante), or the current status of evidence
Building Robust Theories of Change 165
CJPE 32.2, 155–173 © 2017doi: 10.3138/cjpe.31122
about the link having been realized (ex post), classied as strong, medium, or weak.
Again the analysis would seek to identify intervention design weaknesses or issues
that need exploring in an evaluation. Where evidence appears weak, this might sug-
gest the need for additional monitoring, research, and/or evaluation.
CARRYING OUT TOC ANALYSIS
e actual ToC Analysis needs to carried out in a step-by-step manner. Too of-
ten, a theory of change is developed on the basis of the ideas and beliefs of those
involved without much challenge and analysis. Without structured analysis and
challenge, it is unlikely that a robust theory of change and the implications for
intervention design would emerge. ToCA entails a careful examination of each
element in the theory of change, how the elements t together, and an assessment
of the ToC weaknesses, data needs, and their implications.
ToCA would use the criteria in Table 1 as the basis for analysis, roughly in the
order set out. ere would likely be some interplay among results, assumptions,
and the pathway. e ndings could then be summarized in terms of implications
around the two questions noted earlier.
Step 1: Overall Criteria
e initial analysis is to determine if there is indeed an actual ToC model to work
with, and if the intervention seems at all plausible.
Understandable: If the ToC is hard to understand, such as if pathways are
unclear or there is a proliferation of results, then rethinking and redraing are
needed so that there is something resembling a ToC with impact pathways.
Agreed: If there are dierent views as to how the intervention is expected to
work, more discussion on the ToC is probably warranted. If dierences persist,
then it may be necessary to build more than one ToC and analyze each of them.
Level of eort: If the expectations for results are quite out of line with the level
and nature of the activities being undertaken, there may be a need to rethink the
design of the intervention or to reduce expectations to a more realistic level.
Step 2: Detailed ToC Analysis
e detailed ToC Analysis is best done result level-by-result level in sequence.
at is, using the behaviour-based ToC model, in order:
Getting to Reach: Will the outputs delivered reach the intended target
groups with the right reaction?
Getting from Reach to Capacity Change: Will the outputs delivered and
their reach lead to the intended capacity changes?
Getting from Capacity Change to Behaviour Change: Will the capacity
change lead to the intended Behaviour Changes?
Getting from Behaviour Change to Direct Benets: Will the behaviour
changes lead to the intended Direct Benets?
166 Mayne
© 2017 CJPE 32.2, 155–173 doi: 10.3138/cjpe.31122
Getting from Direct Benets to Well-being Changes: Will the direct ben-
ets lead to the intended Well-being Changes?
If another ToC model is being used, the steps are the same: getting from one
level to the next. For each level, an analysis of results and an analysis of assump-
tions would be done and a summary of ndings set out:
Analysis of Results
Denition: If not well dened, need to further dene terms.
Timing: If not sensible, suggests a structural change to the ToC needed.
Logical coherence: If not OK, suggests a structural change to the ToC needed.
Measurement: Indicate if needed, might be needed, not needed, and through
what means. If strength of evidence is weak and the measurement important,
suggests an issue to be addressed in the M&E Implications.
M&E implications: Brings together the M&E issues. Need to remember that
not all results may need to be measured.
Analysis of Assumptions
Denition: If not well dened, need to redene terms.
Logical coherence: If not OK, suggests the need for structural changes in the ToC.
Justication: If not necessary or not likely necessary, then the assumption
should be dropped.
Realization: If realization is in doubt, then need to identify assumption as
at-risk and set out conrming or corrective actions.
Sustainability: Similarly, if the assumption is found not to be sustainable, a
corrective action may be needed, or, in an ex post case, the issue noted as a lesson
learned for future similar or follow-up interventions.
Measurable: Indicate how measures would be taken and if needed, might be
needed, not needed. If strength of evidence is weak and measurement important,
suggest an issue to be addressed in the M&E Implications.
M&E implications: Need to remember that not everything need be measured.
Assessing the Causal Link
Independence: If assumptions are not independent, consider merging as-
sumptions.
A sucient set: If not a sucient set, additional assumption(s) or more of the
prior result are needed.
Strength/status of evidence: Indicate level of evidence for the link being realized.
Summary of Findings for Getting to a Result
e summary of the analysis can depend on the specic purpose and con-
text, but in general can highlight (a) the changes needed to enhance the
Building Robust Theories of Change 167
CJPE 32.2, 155–173 © 2017doi: 10.3138/cjpe.31122
robustness— structural soundness and plausibility—of the ToC, (b) the level of
evidence there is on results and assumptions, (c) the actions that are needed to
enhance the robustness of the intervention design, and (d) the M&E implications.
Structural Changes Needed
Where the structural criteria for a robust theory of change are not met,
structural changes are needed to the ToC to enhance its robustness—that
is, changes in descriptions used, result statements, coherence, assump-
tions, and/or causal links. Aer any structural changes, we would want
to conclude that the ToC is reasonably sound.
Strength/Status of Evidence
Summary analysis can indicate the strength of evidence for (a) the result
in question, (b) the assumptions associated with the link, and (c) the link
being realized.
Additional Intervention Eort Needed to Enhance Plausibility
If plausibility or sustainability of the ToC/intervention design is questionable
due to at-risk assumptions identied and/or sustainability being questioned,
then conrming or corrective actions are likely needed. We would want the
analysis to conclude that, with the conrming/corrective actions, the inter-
vention design is (or ex post would be) robust. Where the ToC is seriously
contested, more than one ToC may needed to be developed and analyzed.
M&E Actions
To monitor how well implementation is going or to verify the ToC in an
evaluation, it is important to identify what data need collecting and the
likely strength of the resulting evidence.
Conclusion: Overall conclusions for the specic link (component) in the ToC
on robustness, level of evidence, and sustainability.
eories of change are best developed in a participatory manner involving
those designing/implementing the intervention and the evaluator (Mayne, 2015,
pp. 137–138). During this development, of course, the criteria for a robust theory of
change can be kept in mind. In other cases, the ToC Analysis is done on a completed
theory of change, probably (although not necessarily) by the evaluator. e ndings
of the analysis should then be discussed with intervention implementers. is dis-
cussion may bring to light issues that were not, but need to be, included in the ToC,
identify issues about the intervention design that need addressing, and/or identify
data that need to be monitored or that need to be addressed in a planned evaluation.
AN EXAMPLE
To illustrate issues and concepts in ToC Analysis, I examined a previously used case
of an intervention aimed at improving the nutritional diets of children through
providing trainintg and information to mothers (Mayne, 2015). e ToC used there
168 Mayne
© 2017 CJPE 32.2, 155–173 doi: 10.3138/cjpe.31122
is shown in Figure 2, with a few small changes to be consistent with the COM-B
model: the motivation and food availability assumptions have been shown as capac-
ity change assumptions rather than behaviour change assumptions.
e ToC in Figure 2 was used to carry out the ToCA. All the details are
not provided here—but can be found in Mayne (2016b), as the analysis is quite
lengthy. And that is worth a note. ToC Analysis is not a quick and dirty approach:
it takes time, but not a lot, and patience to go through each criterion for each result
and each assumption. But it can be worthwhile. Having developed the original
example, I was not expecting many new insights, but I was wrong!
e ndings from the ToCA are summarized below for each results level.
Getting to Reach and Reaction
Several needed structural changes were identied. e Reach result statement
was not well-dened. What did “mothers with young children reached” mean?
It could mean several things, such as mothers heard about the training, mothers
were asked to participate, or mothers participated in at least the rst session. I
Behaviour Changes
Mother adopt new
feeding practices
External Influences
Lower prices for
food
Other staples
become more
nutritious
Capacity Changes
Mother acquire new
capabilities about
nutrition benefits and
feeding practices
Reach and Reaction
Mothers with young
children
Direct Benefits
Children consume a
more nutritious diet
Capacity Change Assumptions
1. Capabilities -Nutrition benefits
and feeding practices understood
and relevant
2. Opportunities Nutritious food
available and affordable
3. Motivation Mothers want to
improve the health of their children
Reach Assumptions
1. Targeted mothers with young
children reached
2. Approach & material seems
appropriate
Wellbeing Changes
Children’s nutrition
status & health
improves
Activities/Outputs
Training & Informing
on Nutrition Benefits &
Feeding Practices
Time line
Behavioural Change Assumptions
1. Mothers make decisions abou t
children’s food
2. New practices supported by
husbands and mother-in-law
3. Parents see improvements in
children’s health
Direct Benefits Assumptions
1. Practices prove practical
2. No reduction in other
nutritious food intake
Wellbeing Change Assumptions
1. Children have access to health
care
Figure 2. A Nutrition Intervention Theory of Change
Building Robust Theories of Change 169
CJPE 32.2, 155–173 © 2017doi: 10.3138/cjpe.31122
assumed it was the latter case: reach and reaction was asking if mothers at least
started the training and if they had a positive initial reaction. So the result state-
ment needed to be changed.
Further, the rst reach assumption was the same as the reach statement!
Clearly there was a logical coherence problem. A new assumption, in fact two,
were needed: targeted mothers are well identied, and targeted mothers can be
communicated with. To get participation, the intervention needed to know who
and where the targeted mothers were, and needed to be able to get the message to
them about the nutrition training sessions.
If sustainability in the target area was an issue, then there would need to be a plan
of how new mothers beyond the initial reach were to be reached, such as perhaps
building into the training the need to spread the word within their communities.
ere are two M&E implications: namely, the need to track the percentage
of targeted population that initially participated, and to monitor initial reaction
of participants.
Getting from Reach to Capacity Change
Several small structural changes were needed in the wording of the capacity result
and assumptions (see Figure 3, where the changes are underlined).
Assumption 2 about the availability and aordability of nutritious food (oppor-
tunities) is possibly at-risk without more information. In Figure 3, at-risk assump-
tions are bolded. A useful corrective action would be to make local markets aware
of the intervention and the expected increased demand for certain food products.
And a conrming M&E action is needed: the availability and aordability of
nutritious food should be monitored during the life of the project.
Getting from Capacity Change to Behaviour Change
Behaviour change assumptions 1 and 2 overlap somewhat and may be at-risk.
e intervention may need a better understanding about how decisions on food
are made in households, and the sessions oered to households rather than only
mothers.
M&E implications: Household surveys could track adoption of the new prac-
tices and general household support, and identify problems. Perhaps schedule a
survey aer 2 months and a follow-up 1 year later.
Getting from Behaviour Change to Direct Benets
Assumption 2 about substituting other foods is at-risk. A conrming action could
be to include this substitution issue in the nutrition training.
M&E implications: Follow-up household surveys could track childrens di-
etary intake.
Overall, although implied by the timeline, Figure 2 did not set out a clear time
frame for the intervention to have an impact. e level of evidence on realizing
the ToC would be good, using the measures suggested.
170 Mayne
© 2017 CJPE 32.2, 155–173 doi: 10.3138/cjpe.31122
Based on this ToC Analysis, the revised and more robust ToC is shown in
Figure 3. At-risk assumptions are shown in bold, and wording changes are un-
derlined.
CONCLUDING REMARKS
eories of change are the basis for theory-based evaluation approaches, such as
logical analysis (Brousselle & Champagne, 2011; Rey et al., 2012), realist evalua-
tion (Blamey & Mackenzie, 2007; Pawson, 2013), contribution analysis (Mayne,
2012), and process tracing (Schmitt & Beach, 2015). As such, the robustness of the
theory of change used matters. A weak theory of change can only generate weak
ndings. For example, conrming a weak theory of change—one poorly struc-
tured with evident logical gaps—in contribution analysis cannot lead to credible
causal contribution claims.
is article argues the usefulness of building robust theories of change and
structured theory of change analysis, so that evaluation ndings based on these
theories of change are strengthened. ToC Analysis involves assessing a theory of
Behaviour Changes
Mother adopt new
feeding practices
External Influences
Lower prices for food
Other staples become
more nutritious
Capacity Changes
Mother acquire new
capacity about nutrition
benefits and feeding
practices
Reach and Reaction
Targeted mothers
participate
Direct Benefits
Children consume a
more nutritious diet
Capacity Change Assumptions
1. Capabilities -Nutrition benefits
and feeding practices understood
and relevant
2. Opportunities Nutritious food
discussed are known about,
available and affordable
3. Motivation Mothers want to
improve the health of their children
Reach Assumptions
1. Mothers with young children in
the target area are well identified
2.Targeted mothers can be
communicated with
3. Approach & material seems
appropriate
Wellbeing Changes
Children’s nutrition
status & health
improves
Figure 3: A Robust Nutrition Intervention Theory of Change
Activities/Outputs
Training & Informing
on Nutrition Benefits &
Feeding Practices
Time line
Behavioural Change Assumptions
1. Mothers make decisions
about children’s food
2. New practices supported by
husbands and mother-in-law
3. Parents see improvements in
children’s health
Direct Benefits Assumptions
1. Practices prove practical
2. No reduction in other
nutritious food intake
Wellbeing Change Assumptions
1. Children have access to
health care
3 months
4months
6months
1 year
Figure 3. A Robust Nutrition Intervention Theory of Change
Building Robust Theories of Change 171
CJPE 32.2, 155–173 © 2017doi: 10.3138/cjpe.31122
change against a set of criteria (Table 1) for each result, each assumption, and
each causal link, challenging the structure and logic of the theory of change.
e analysis takes some time and discipline to carry out. But it is mainly a desk
review, and overall it entails hours rather than days of work. In my experience, it
inevitably leads to improvements in the theory of change. e results are usually
quite informative, leading to
more robust ToCs,
better intervention designs,
useful M&E actions to help manage the intervention and support future
evaluation, and
ex post, more credible theory-based evaluations.
REFERENCES
Blamey, A., & Mackenzie, M. (2007). Theories of change and realistic evaluation:
Peas in a pod or apples and oranges? Evaluation, 13(4), 439–455. http://dx.doi.
org/10.1177/1356389007082129
Brousselle, A., & Champagne, F. (2011). Program theory evaluation: Logic analysis. Evalu-
ation and Program Planning, 34(1), 69–78. http://dx.doi.org/10.1016/j.evalprogplan.
2010.04.001
Coryn, C. L. S., Noakes, L. A., Westine, C. D., & Schroter, D. C. (2011). A systematic review
of theory-driven evaluation practice from 1990 to 2009. American Journal of Evalua-
tion, 32(2), 199–226. http://dx.doi.org/10.1177/1098214010389321
Davies, R. (2012). Criteria for assessing the evaluability of theories of change. http://
mandenews.blogspot.com/2012/04/criteria-for-assessing-evaluablity-of.html.
Davies, R. (2013). Planning evaluability assessments: A synthesis of the literature with
recommendations (Working Paper 40: DFID). Retrieved from https://www.gov.uk/
government/uploads/system/uploads/attachment_data/le/248656/wp40-planning-
eval-assessments.pdf.
Donaldson, S. I. (2007). Program theory-driven evaluation science: Strategies and application.
Mahwah, NJ: Lawrence Erlbaum. Retrieved from https://www.researchgate.net/pub-
lication/235930890_Program_theory-driven_evaluation_science_Strategies_and_ap-
plications
Funnell, S., & Rogers, P. (2011). Purposeful program theory: Eective use of theories of change
and logic models. San Francisco, CA: Jossey-Bass.
Hansen, M. B., & Vedung, E. (2010). eory-based stakeholder evaluation. American
Journal of Evaluation, 31(3), 295–313. http://dx.doi.org/10.1177/1098214010366174
James, C. (2011). eory of change review. A report commissioned by Comic Relief. Avail-
able at http://mande.co.uk/blog/wp-content/uploads/2012/03/2012-Comic-Relief-
eory-of-Change-Review-FINAL.pdf.
Johnson, N., Mayne, J., Grace, D., & Wyatt, A. (2015). How will training traders contribute
to improved food safety in informal markets for meat and milk?: A theory of change
analysis (IFPRI Discussion Paper 1451). Washington, DC: International Food Policy
172 Mayne
© 2017 CJPE 32.2, 155–173 doi: 10.3138/cjpe.31122
Research Institute (IFPRI). Available at http://ebrary.ifpri.org/cdm/ref/collection/
p15738coll2/id/129293.
Leeuw, F. (2012). Linking theory-based evaluation and contribution analysis:
ree problems and a few solutions. Evaluation, 18(3), 348–363. http://dx.doi.
org/10.1177/1356389012452051
Mayne, J. (2011). Contribution analysis: Addressing cause and eect. In R. Schwartz, K.
Forss, & M. Marra (Eds.), Evaluating the complex (pp. 53–96). New Brunswick, NJ:
Transaction Publishers.
Mayne, J. (2012). Contribution analysis: Coming of age? Evaluation, 18(3), 270–280. http://
dx.doi.org/10.1177/1356389012451663
Mayne, J. (2015). Useful theory of change models. Canadian Journal of Program Evaluation,
30(2), 119–142. http://dx.doi.org/10.3138/cjpe.230
Mayne, J. (2016a). e COM-B theory of change model: Working paper. Retrieved from
https://www.researchgate.net/publication/314086441_The_COM-B_Theory_of_
Change_Model_V3.
Mayne, J. (2016b). An example of ToC analysis. Retrieved from https://www.researchgate.
net/publication/305957815_An_Example_of_ToC_Analysis.
Mayne, J., & Johnson, N. (2015). Using theories of change in the CGIAR Research Program
on Agriculture for Nutrition and Health. Evaluation, 21(4), 407–428. http://dx.doi.
org/10.1177/1356389015605198
Morton, S. (2015). Progressing research impact assessment: A “contributions” approach.
Research Evaluation, 24(4), 405–419. http://dx.doi.org/10.1093/reseval/rvv016
Patton, M. Q. (2008). Utilization-focused evaluation (4th ed.). ousand Oaks, CA: Sage.
Pawson, R. (2013). e science of evaluation: A realist manifesto. Atlanta, GA: Sage.
Peersman, G., Guijt, I., & Pasanen, T. (2015). Evaluability assessment for impact evaluation:
Guidance, checklists and decision support, A Methods Lab Publication. London, UK:
Overseas Development Institute.
Rey, L., Brousselle, A., & Dedobbeleer, N. (2012). Logic analysis: Testing program theory
to better evaluate complex intervention. Canadian Journal of Program Evaluation,
26(3), 61–89. Retrieved from https://evaluationcanada.ca/system/les/cjpe-entries/
26-3-061.pdf
Rogers, P. (2007). eory-based evaluations: Reections ten years on. New Directions for
Evaluation, 114, 63–67. http://dx.doi.org/10.1002/ev.225
Schmitt, J., & Beach, D. (2015). e contribution of process tracing to theory-based
evaluations of complex aid instruments. Evaluation, 21(4), 429–447. http://dx.doi.
org/10.1177/1356389015607739
Tremblay, M.-C., Brousselle, A., Richard, L., & Beaudet, N. (2013). Dening, illustrating
and reecting on logic analysis with an example from a professional development
program. Evaluation and Program Planning, 40, 64–73. http://dx.doi.org/10.1016/j.
evalprogplan.2013.05.004
Trevisan, M., & Walser, T. (2014). Evaluability assessment: Improving evaluation quality and
use. ousand Oaks, CA: Sage.
Building Robust Theories of Change 173
CJPE 32.2, 155–173 © 2017doi: 10.3138/cjpe.31122
Valters, C. (2014). eories of change in international development: Communication, learn-
ing, or accountability? (JSRP Paper 17: e Asia Foundation). Retrieved from http://
www.lse.ac.uk/internationalDevelopment/research/JSRP/downloads/JSRP17.Valters.
pdf.
Vogel, I. (2012). Review of the use of “eory of Change” in international development:
Department for International Development (DFID). Retrieved from http://www.
oxfamblogs.org/fp2p/wp-content/uploads/DFID-ToC-Review_VogelV4.pdf
Wilson-Grau, R., & Britt, H. (2013). Outcome harvesting (FordFoundation, MENA oce).
Retrieved from http://www.outcomemapping.ca/download/wilsongrau_en_Outome
%20Harvesting%20Brief_revised%20Nov%202013.pdf
Woolcock, M. (2009). Toward a plurality of methods in project evaluation: A contextual-
ized approach to understanding impact trajectories and ecacy. Journal of Develop-
ment Eectiveness, 1(1), 1–14. http://dx.doi.org/10.1080/19439340902727719
AUTHOR INFORMATION
John Mayne is an independent advisor on public sector performance. Over the past 13
years he has focused largely on international development evaluation and results-based
management work. He has been working with several government, NGOs, and interna-
tional organizations. He has authored numerous articles and reports, including several on
contribution analysis, and co-edited eight books on program evaluation and performance
monitoring. In 1989 and in 1995, he was awarded the Canadian Evaluation Society Award
for Contribution to Evaluation in Canada. In 2006, he was made a Canadian Evaluation
Society Fellow. Dr. Mayne’s current research interests are on approaches for strengthening
impact evaluation, useful theories of change, and dealing with attribution.
... Scholars should therefore be able to justify why their object of study represents an innovation and not simply an instance of incremental policy change. An explicit description through a theory of change can highlight the (assumed) causal links between policy action and corresponding outcomes (Mayne, 2017). ...
... Once again, what may be(come) substantial in one context could be (or remain) only a small step in another; researchers working on policy innovations should thus be able to explain to what extent they observe a substantial policy innovation and how the step-change evolves. This is the essence of the theory of change for any particular innovation (Mayne, 2017). ...
Chapter
In light of existing unsustainable ways of human life, policy innovations are key to achieve the necessary transition towards sustainable development. This chapter unpacks the concept of policy innovation with a focus on its core characteristics, as well as different types of innovation. It then turns to the sources of innovation such as policy experimentation, how innovations may spread through diffusion processes and ultimately the effects of innovations, which may be detected through careful evaluation. The chapter discusses these conceptual approaches in the context of the UN Sustainable Development goals and what innovations may be expected based on them across different policy sectors. The chapter concludes that while policy innovation may play an important role, one should not subscribe to an uncritical, pro-innovation bias. Future research should study policy innovations in situ, including their characteristics and how they may (not) spread.
... Over time, the role of the program theory of change (ToC) in developing, implementing, and evaluating an intervention has received greater emphasis (Alter & Egan, 1997;Mayne, 2015;Breuer et al, 2015). A ToC describes how change is expected to happen or how change has happened (Mayne, 2017) and guides the evaluator in making choices around when and how to measure these elements of change (Connell & Kubisch, 1998). Most program ToCs specify links between input, output, and outcomes, and can potentially help strengthen the scientific case for attributing the change in outcomes to the activities included in the intervention. ...
Article
This paper used a blended approach that involves multiple techniques to, first, test a set of assumptions around a health behavior change communication intervention theory of change (ToC) and, second, surface some unidentified assumptions involving the local context. The intervention was integrated with women's self-help groups (SHGs) in Uttar Pradesh, India. The key assumption tested in this paper was the linkage between SHG membership, program exposure, and maternal, newborn, and child health practices. Learnings were substantiated through empirical investigations, including structural equation modeling and mediation analysis, as well as 'co-learning' workshops within the community. The workshops aimed to capture and interpret the heterogeneity of local contexts through deep dialogs with the community and program implementers at various levels. Statistical analyses indicated a significant association between the amount of women's program exposure and their health practices. SHG membership was shown to affect maternal health practices; however, it did not have a direct effect on neonatal or child health practices. The 'co-learning' workshops revealed crucial aspects, such as prevailing socio-cultural norms, which prevented pregnant or recently delivered women from participating in SHG meetings. This paper encourages evaluators to work with the community to interpret and co-construct meaning in unpacking the contextual forces that seldom appear in the program ToC.
... To address deficiencies of previous strategic plans more effectively, the CBD has changed course and adopted a ToC approach in the development of the Post-2020 GBF [23]. At a fundamental level, the approach involves identifying beneficiaries and mapping out what needs to happen (actions) to support desired results, putting in place tools and solutions for mainstreaming, creating enabling conditions and adequate means of implementation (e.g., financial resources, human resources, technology) collectively within what is commonly referred to as a 'ToC pathway' [24][25][26]. ...
Article
Full-text available
There is a rapidly closing window of opportunity to stop biodiversity loss and secure the resilience of all life on Earth. In December 2022, Parties to the United Nations (UN) Convention on Biological Diversity (CBD) will meet in Montreal, Canada, to finalize the language and terms of the Post-2020 Global Biodiversity Framework (Post-2020 GBF). The Post-2020 GBF aims to address the shortcomings of the previous Strategic Plan on Biodiversity 2011–2020, by introducing a Theory of Change, that states that biodiversity protection will only be successful if unprecedented, transformative changes are implemented effectively by Parties to the CBD. In this policy perspective, we explore the implications of the Theory of Change chosen to underpin the Post-2020 GBF, specifically that broad social transformation is an outcome that requires actors to be specified. We detail how the health sector is uniquely positioned to be an effective actor and ally in support of the implementation of the Post-2020 GBF. Specifically, we highlight how the core competencies and financial and human resources available in the health sector (including unique knowledge, skill sets, experiences, and established trust) provide a compelling, yet mostly untapped opportunity to help create and sustain the enabling conditions necessary to achieve the goals and targets of the framework. While by no means a panacea for the world’s biodiversity problems, we posit that explicitly omitting the health sector from the Post-2020 GBF substantially weakens the global, collective effort to catalyze the transformative changes required to safeguard biodiversity.
... Barzelay (2007, pp. 523-524) helpfully suggests a two-step assessment procedure that can be used, supplemented with methods drawn from recent work on extrapolation (Khosrowi, 2019;Steel, 2008) and process-related evaluation methods (Mayne, 2015(Mayne, , 2017Wauters & Beach, 2018;Woolcock, 2022). In the first step, a causal lesson is drawn from a source case (here a dramatic policy failure) by analyzing the process leading to the failure (Barzelay, 2007, p. 524) using process tracing methods (e.g., Wauters & Beach, 2018). ...
Article
Full-text available
Why do smart policy makers who try to learn from policy failure end up overgeneralizing these lessons when facing new crises? This article focuses on the policy learning that can come in the wake of perceived policy failure, and the consequences that lesson learning has for diagnosing and tackling subsequent crises. We argue, in contrast to much of the existing literature, that failure can result in policy learning when the policy area is highly complex and there are recognized experts that have epistemic authority in the issue. At the same time, lessons tend to become overgeneralized in which policy makers facing a new crisis extrapolate the lessons learned from the past failure with little consideration of whether the circumstances are similar enough to be applicable. Our argument is assessed empirically using a process-tracing analysis of how the EU responded to the Spanish banking crisis in 2012.
... Barzelay (2007, pp. 523-524) helpfully suggests a two-step assessment procedure that can be used, supplemented with methods drawn from recent work on extrapolation (Khosrowi, 2019;Steel, 2008) and process-related evaluation methods (Mayne, 2015(Mayne, , 2017Wauters & Beach, 2018;Woolcock, 2022). In the first step, a causal lesson is drawn from a source case (here a dramatic policy failure) by analyzing the process leading to the failure (Barzelay, 2007, p. 524) using process tracing methods (e.g., Wauters & Beach, 2018). ...
Article
Why do smart policy makers who try to learn from policy failure end up overgeneralizing these lessons when facing new crises? This article focuses on the policy learning that can come in the wake of perceived policy failure, and the consequences that lesson learning has for diagnosing and tackling subsequent crises. We argue, in contrast to much of the existing literature, that failure can result in policy learning when the policy area is highly complex and there are recognized experts that have epistemic authority in the issue. At the same time, lessons tend to become overgeneralized in which policy makers facing a new crisis extrapolate the lessons learned from the past failure with little consideration of whether the circumstances are similar enough to be applicable. Our argument is assessed empirically using a process‐tracing analysis of how the EU responded to the Spanish banking crisis in 2012. Butz, Adam M., Michael P. Fix, and Joshua L. Mitchell. 2015. “Policy Learning and the Diffusion of Stand‐Your‐Ground Laws.” Politics & Policy 43(3): 347–77. https://doi.org/10.1111/polp.12116. Considine, Mark. 2012. “Applying Design Theory to Public Policy.” Politics & Policy 40(4): 704–24. http://onlinelibrary.wiley.com/doi/10.1111/j.1747‐1346.2012.00372.x. Raile, Eric D., Amber N. Raile, Charles T. Salmon, and Lori Ann Post. 2014. “Defining Public Will.” Politics & Policy 42(1): 103–30. http://onlinelibrary.wiley.com/doi/10.1111/polp.12063. ¿Por qué los responsables políticos inteligentes que tratan de aprender del fracaso de las políticas terminan generalizando en exceso estas lecciones cuando se enfrentan a nuevas crisis? Este artículo se centra en el aprendizaje de políticas que puede surgir a raíz de un fracaso político percibido y las consecuencias que tiene el aprendizaje de lecciones para diagnosticar y abordar crisis posteriores. Argumentamos, en contraste con gran parte de la literatura existente, que el fracaso puede resultar en el aprendizaje de políticas cuando el área de políticas es muy compleja y hay expertos reconocidos que tienen autoridad epistémica en el tema. Al mismo tiempo, las lecciones tienden a generalizarse en exceso y los responsables de formular políticas que enfrentan una nueva crisis extrapolan las lecciones aprendidas del fracaso pasado sin considerar si las circunstancias son lo suficientemente similares como para ser aplicables. Nuestro argumento se evalúa empíricamente utilizando un análisis de seguimiento de procesos de cómo la UE respondió a la crisis bancaria española en 2012. 为何那些试图从政策失败中吸取经验的政策制定者在面临新危机时最终会过度泛化这些经验?本文聚焦于在感知政策失败后可能出现的政策学习,以及经验对诊断和应对后续危机产生的结果。与许多现有文献相比,我们论证认为,当政策领域具有高度复杂性并且存在对该问题具有认知权威的公认专家时,政策失败可能会导致政策学习。与此同时,经验往往过于泛化,即面临新危机的决策者从政策失败中推断出经验,但很少考虑这些经验是否与新危机足够相似而具有适用性。通过对欧盟如何响应 2012 年西班牙银行业危机一事进行过程追踪分析,我们对论点进行了实证评估。
Article
Full-text available
In recent years, for the first time in Italy, impact evaluation has been requested as a standard for social projects implemented by public–private partnerships and, more specifically, a condition for them to be financed through the national educational poverty reduction program. This article presents and discusses the Developmental Outcome Monitoring and Evaluation (DOME) model, a mixed and participatory method for the impact assessment of complex and innovative educational poverty reduction programs.
Article
Full-text available
Introduction Research–practice partnerships (RPPs) are long-term collaborations between research and practice that aim to conduct research that can be used to make practice-based improvements. They intentionally bring together diverse experience in decision making and seek to shift power dynamics so that all partners have a say. The Creating Care Partnerships project aims to explore whether the RPP approach developed within the US educational context can be successfully applied to the English care home context. The project involves a programme of codesign, implementation and evaluation within three case study sites. This protocol set outs the aims, research design and governance of the evaluation. Methods and analysis The evaluation takes a theory-based approach to explore how, why and in what circumstances RPPs in the care home context contribute to enhancing research and research use in local care homes and informing wider improvement efforts. A mixed-methods design will be used for each case study, including semistructured interviews, observations of RPP events and meetings, an online survey, activity diary and review of local data and documents. Data collection will proceed in waves, with the theory of change (ToC) being continually refined and used to guide further data collection and analysis. Insights will be drawn using Contribution Analysis, Realist Evaluation and systems perspectives to assess the contribution made by the case study sites to achieving outcomes and the influence of contextual factors. Economic consequences will be identified through the ToC, using a narrative economic analysis to assess costs, consequences and value for money. Ethics and dissemination The study has undergone ethics review by HRA Research Ethics Committee. It does not pose major ethical issues. A final report will be published and articles will be submitted to international journals.
Technical Report
Full-text available
Con la Iniciativa Volando Juntos, Provita desea promover comportamientos alternativos a la demanda de fauna, que generen hábitos de vida sostenible en las comunidades que conviven con estas especies. A través de una serie de herramientas, deseamos compartir tanto el conocimiento técnico, como la experiencia ganada tras la implementación de nuestras campañas de cambio de comportamiento en dos especies de aves amenazadas por el tráfico ilegal de fauna, el Cardenalito (Spinus cucullatus) y la Cotorra cabeciamarilla (Amazona barbadensis). El objetivo de esta segunda herramienta es describir cómo integrar herramientas de planificación como la Teoría de Cambio, con modelos conductuales para describir y evaluar los mecanismos por los cuales ocurre el cambio conductual deseado.
Article
In the digital age, governments can reach thousands of citizens easily and at a low cost using information and communication technology (ICT). However, little is known about whether such ICT-delivered messages are received by citizens and result in behavioural change. We develop a theory about the barriers that government messages need to overcome if citizens are to receive a message and engage in it. We test the effects of a government health message in a large-scale field experiment among 4,880 representative citizens. The results show that small-scale government information campaigns delivered through ICT can succeed in changing citizen behaviour.
Working Paper
Full-text available
Article
Full-text available
Evaluating complex interventions requires an understanding of the program's logic of action. Logic analysis, a specific type of program theory evaluation based on scientific knowledge, can help identify either the critical conditions for achieving desired outcomes or alternative interventions for that purpose. In this article, we outline the principles of logic analysis and its roots. We then illustrate its use with an actual evaluation case. Finally, we discuss the advantages of conducting logic analysis prior to other types of evaluations. This article will provide evaluators with both theoretical and practical information to help them in conceptualizing their evaluations.
Article
Full-text available
This article focuses on methodological challenges in evaluating complex program aid interventions like budget support. We show that recent innovations in process-tracing methodology can help solve the identified challenges and increase the strength of causal inference made when using case studies in demanding settings. For the specific task of evaluating the governance effectiveness of budget support interventions, we developed a more fine-grained causal mechanism for a sub-set of the comprehensive program theory of budget support. Moreover, based on the informal use of Bayesian logic, we have elaborated on how to increase the conclusiveness of empirical evidence for one part of the theorized causal mechanism. We argue that by establishing an explicit theorized mechanism prior to empirical research and by critically judging our evidence according to an informal Bayesian logic we can remedy some of the problems at hand in much case-study research and increase the inferential leverage in complex within-case evaluation studies.
Article
Full-text available
Theories of change are increasingly being discussed and referenced in development evaluation even while the elements of what a theory of change consist of differ widely among applications. Equally, examples of actual use of theories of change other than as overview illustrations of interventions are rare. In this article, we present generic models of theories of change for both straightforward and more complex interventions. A number of examples of evidence-based theories of change in the area of agriculture research for nutrition and health are discussed, as is the need for different versions of a theory of change for different purposes. We also discuss the use and analysis of these models in the planning, managing and assessing of research-related interventions, illustrating the practical usefulness of well-developed theories of change.
Article
Full-text available
There is an increasing interest in demonstrating the outcomes from research for the purposes of learning, accountability, or to demonstrate the value of research investments. However, assessing the impact of social science research on policy and practice is challenging. The ways in which research is taken up, used, and reused in policy and practice settings means that linking research processes or outputs to wider changes is difficult, and timescales are hard to predict. This article proposes an empirically grounded framework for assessing the impact of research—the Research Contribution Framework. A case study approach was adopted to explore the nature of research impact and how it might be assessed. Findings were used to design, develop, and test a framework to assess the contribution of research to relevant areas of policy and practice and to articulate wider benefits. The framework has been adapted from contribution analysis, using the idea of ‘contribution’ to help explain the ways research is taken up and used to influence policy and practice. The framework allows for a focus on the roles of research users, and examines both processes and outcomes. It is argued that this approach gets round some of the common problems in assessing impact. It provides a method of linking research and knowledge exchange to wider outcomes whilst acknowledging and including contextual factors that help or hinder research impact. It is practical, balancing robustness with feasibility. It is adaptable for a wide range of content, types of impact assessment, and purposes.