ArticlePDF Available

Abstract and Figures

Evaluation practice is vital for the accountability and learning of administrations implementing complex policies. This article explores the relationships between the structures of the evaluation systems and their functions. The findings are based on a comparative analysis of six national systems executing evaluation of the European Union Cohesion Policy. The study identifies three types of evaluation system structure: centralized with a single evaluation unit, decentralized with a coordinating body and decentralized without a coordinating body. These systems differ in terms of the thematic focus of evaluations and the targeted users. Decentralized systems focus on internal users of knowledge and produce mostly operational studies; their primary function is inward-oriented learning about smooth programme implementation. Centralized systems fulfil a more strategic function, recognizing the external audience and external accountability for effects. Points for practitioners Practitioners who design multi-organizational evaluation systems should bear in mind that their structure and functions are interrelated. If both accountability and learning are desired, the evaluation system needs at least a minimum degree of decentralization on the one hand and the presence of an active and independent coordination body on the other.
Content may be subject to copyright.
Structures and functions of
complex evaluation
systems: comparison of six
Central and Eastern
European countries
Tomasz Kupiec
University of Warsaw, Poland
Dominika Wojtowicz
Kozminski University, Poland
Karol Olejniczak
SWPS University of Social Sciences and Humanities, Poland
Abstract
Evaluation practice is vital for the accountability and learning of administrations imple-
menting complex policies. This article explores the relationships between the struc-
tures of the evaluation systems and their functions. The ndings are based on a
comparative analysis of six national systems executing evaluation of the European
UnionCohesionPolicy.Thestudyidenties three types of evaluation system structure:
centralized with a single evaluation unit, decentralized with a coordinating body and
decentralized without a coordinating body. These systems differ in terms of the
thematic focus of evaluations and the targeted users. Decentralized systems focus on
internal users of knowledge and produce mostly operational studies; their primary
function is inward-oriented learning about smooth programme implementation.
Centralized systems full a more strategic function, recognizing the external audience
and external accountability for effects.
Corresponding author:
Tomasz Kupiec, University of Warsaw Centre for European Regional and Local Studies, Krakowskie
Przedmieście 30, 00-927 Warsaw, Poland.
Email: tomasz.kupiec@uw.edu.pl
Article
International
Review of
Administrative
Sciences
International Review of Administrative
Sciences
119
© The Author(s) 2021
Article reuse guidelines:
sagepub.com/journals-permissions
DOI: 10.1177/00208523211026964
journals.sagepub.com/home/ras
Points for practitioners
Practitioners who design multi-organizational evaluation systems should bear in mind
that their structure and functions are interrelated. If both accountability and learning
are desired, the evaluation system needs at least a minimum degree of decentralization
on the one hand and the presence of an active and independent coordination body on
the other.
Keywords
accountability, cohesion policy, evaluation, evaluation systems, evaluation functions,
learning, structures
Introduction
Evaluation, dened as a systematic inquiry of the worth and merit of public interventions
(Fournier, 2005; Patton, 2004), has been considered a crucial element of the policy cycle
and public administration practice for at least several decades (Chelimsky and Shadish,
1997). Its roots can be traced to the United StatesWar on Poverty and Great Society
initiatives in the 1960s (Mark et al., 2006; Rossi et al., 2004). The subsequent waves
of public management reforms, including performance-based management promoted
by the Organisation for Economic Co-operation and Development and the World Bank
(Kusek and Rist, 2004; OECD, 1998), institutionalized evaluation within governments
and made it a ubiquitous practice across public administrations (Furubo et al., 2002;
Stockmann et al., 2020).
In Europe, the European Union (EU) has been the major promoter of evaluation prac-
tice in public administrations at various levels. While some EU Member States have
developed their own evaluation culture, during what is often referred to as the rst and
second evaluation wave (Derlien, 1990), many others, including all those joining the
EU in the twenty-rst century, adopted this practice mainly in response to the require-
ments of EU regulations.
The ultimate and general goal of evaluation is social betterment(Henry and Mark,
2003). In the practice of specic organizations, evaluation translates into various func-
tions (also called purposes), often driven by conicting logic, and therefore leading to
potential tensions among the actors in policy systems (Donaldson et al., 2010). Thus,
it is important for both the theory and practice of evaluation in public administration
to recognize the functions underlying specic evaluation efforts and to identify their
determinants.
The vast majority of evaluation use literature focuses on a single evaluation study as
the unit of analysis (Højlund, 2014; Raimondo, 2018) and misses the wider institutional
context. Only a handful of authors examine evaluation systems in terms of multi-
organizational networks (Olejniczak, 2013; Stockmann et al., 2020), but they do not
explore, in a comparative manner, the relation between structural arrangements and
system functions. Some studies suggest that there might be a relationship between
2International Review of Administrative Sciences 0(0)
functions and the design of evaluation systems. However, these articles take the perspective
of an individual organization and their internal evaluation arrangements. A group of authors
postulate that a broader perspective of evaluation systems is needed, and more attention
should be devoted to studying the arrangement of evaluation systems as multi-organizational
networks (e.g. Kupiec, 2020; Leeuw and Furubo, 2008; Liverani and Lundgren, 2007).
This limited recognition of multi-organizational systems and their structures in evalu-
ation literature is in sharp contrast with the wider body of work on administration and
management. The effects of institutional structures on political and administrative out-
comes lie at the heart of the eld of public administration and policy (Balla et al.,
2015). Management literature also acknowledges the association between functions
(strategy) and structure (Dyas and Thanheiser, 1976; Mintzberg, 1992).
Therefore, there is a gap to be addressed by exploring the relations between structures
of multi-organizational evaluation systems and the functions they full. This study
denes the evaluation system as a set of organizations involved in acquiring or using
evaluative knowledge, while the structure is the relationships between them. Our hypoth-
esis states that the structure of evaluation system correlates with its functions. We do not
intend to establish the direction of this relationship, but we claim that certain structures
may support one function and impede another, thus indicating the dominant orientation
of a specic evaluation system.
To verify this hypothesis, we analyse the evaluation practices of six EU Member
States that are substantial beneciaries of cohesion policy (CP): Bulgaria, Czechia,
Hungary, Poland, Romania and Slovakia. The evaluation practice in CP is considered
to be the most developed among EU policies (Fratesi and Wishlade, 2017). This is
mostly due to the signicant size of the CP budget, the spectrum of activities that it
covers and the long record of making evaluation an integral part of programme manage-
ment. As longitudinal analysis shows, these extensive evaluation practices prescribed by
the European Commission (EC) result in a spectrum of evaluation functions, with some
trade-offs among functions (Batterbury, 2006).
Although the EC introduced the same requirements for monitoring and evaluation in
all Member States (regulation no. 1303/2013), it left open the issue of structural and pro-
cedural arrangements. As a result, national administrations developed country-specic
evaluation systems with distinctive structures and operating procedures. This situation
creates a unique opportunity for a comparative study.
Even though it has not been explicitly stated in the regulations, CP evaluation arrange-
ments seem to be based on the equinality principle, which states that the same evaluation
functions can be performed by various, national-specic systems. This article challenges
this implicit assumption by arguing that national evaluation systems of varying structures
perform different functions, and therefore common CP evaluation goals are not achieved
in all countries to the same degree.
The article commences with a review of evaluation functions in public administration,
summarizing them with a two-dimensional, four-element model of evaluation functions.
This synthesis provides a frame for later empirical study. A discussion on the boundaries
of evaluation systems is also provided in this section, concluding with the concept of a
complex multi-organizational evaluation system. We then move on to explain the
Kupiec et al. 3
research design and methods, including the rationale behind the selection of national
cases, the approach to operationalizing and measuring key variables and sources of
data. The next part is devoted to a comparative analysis of the variations in structures,
activities and functions of six national evaluation systems. The article concludes with
implications for future research and evaluation practice.
The article aims to refocus the evaluation debate onto a new unit of analysis evalu-
ation systems understood as multi-organizational arrangements and provide an initial
typology of system structures. We hope that this will pave the way for a more systematic
inquiry into the relationships between evaluation functions and structures as part of public
administration and policy practice.
Theoretical framework
Evaluations functions
Evaluation is an essential element of a policy process and public administration practice.
Although policy cycle models differ from author to author, they always include an evalu-
ation stage. Evaluation contributes to the improvement of public policies and ultimately
to social betterment (Henry and Mark, 2003). Its importance for policy making, public
administration and, more generally, the public sector is explained in many various, com-
plementing ways. For Ostrom (2007) evaluation is necessary for the efcient and just
management of common pool resources, and Picciotto (2016) argues that from the per-
spective of agency problem, evaluation reduces information asymmetry. Reducing red
tape is also among the highlighted benets (Oh and Lee, 2020). The use, non-use, advan-
tages and limitations of evaluation are discussed in the context of bureaucracies and pol-
itical authorities (Kudo, 2003), local governments (Favoreu et al., 2015) and national
parliaments (Speer et al., 2015).
A rich body of literature describes a wide range of potential evaluation functions also
called purposes (e.g. Batterbury, 2006; Chelimsky and Shadish, 1997; Hanberger, 2011;
Mark et al., 1999; Widmer and Neuenschwander, 2004; Zwaan et al., 2016). The follow-
ing observations emerge from that literature.
First, evaluation functions may be grouped into general categories (Kupiec, 2020):
providing an assessment of performance for the external audience accountability,
oversight,compliance;
inducing a change in behaviour, adjustment of a policy/strategy/programme/practice
learning,improvement (of performance,planning),knowledge creation,building
capacity;
supporting or criticizing a policy, practice or decision based on other (than evaluation)
considerations, building appearances of a learning organization legitimizing,sub-
stantiating,justifying,sanctioning,political ammunition;
involving stakeholders in the evaluation process, ceding power to those normally
excluded from the decision-making process, fostering public debate engaging,
empowering,developing a sense of ownership.
4International Review of Administrative Sciences 0(0)
Second, two purposes accountability and learning are often considered the main ones
(e.g. Van Der Meer and Edelenbos, 2006). Rather than being complementary, these two
aspects stand, to some extent, in opposition to each other. While learning is oriented
inward (to an institution, which implements an intervention subject to evaluation),
accountability is performed for other stakeholders, i.e. supervising bodies, media,
society. Each function requires a different attitude from evaluators, and as Raimondo
(2018) argues, focusing on accountability strengthens norms which then impede learning.
Third, there are some ambiguities in terminology. The terms purposes and functions
are often used interchangeably (e.g. Mark et al., 1999) without specifying their focus-
specic content (i.e. accountability for what actions, learning about what issues, etc.).
Therefore, for our analytical framework, we propose referring to accountability and
learning as evaluation purposes, not functions. Furthermore, the second dimension is
introduced the focus of evaluation activities. It is based on a classic dichotomy from
management literature the strategic vs. operational focus. Combining the purposes
with focus allows four possibilities to be identied. We call them evaluation functions
(Figure 1).
The framework is aligned with the accountability literature distinguishing between
accountability for inputs and process vs. outputs (Dubnick and Frederickson, 2011)
and the organizational learning literature operational improvements in doing things
rightor strategic reection on doing right things(Argyris and Schon, 1995). The
framework also corresponds to the goals of CP evaluation specied in EU regulations:
(1) supporting the management of the programmes (operational and, to a lesser extent,
strategic learning); and (2) assessing their effects (strategic and, to a lesser extent, oper-
ational accountability). The framework provides the basis for dening variables in this
study.
Evaluation systems boundaries and determinants of functions
While there is a rich body of literature on determinants of evaluation use, and specic
types of use, there has been little discussion about the factors conditioning evaluation
functions, especially when restricted to evaluation systems alone. Available sources
suggest one potential factor the design/structure of the evaluation system. Hanberger
(2011), who studied international organizations, suggests that the evaluation functions
depend on the system design. Widmer and Neuenschwander (2004) argue that the organ-
izational design of a system should follow the purpose/function of evaluation but admit
that this relationship is often two-way in the reality of Swiss public agencies.
These empirical observations correspond to some extent with ndings from the man-
agement eld on the association between strategy and structure. The relationship between
these two aspects has been the subject of numerous studies since the 1970s, with the
majority suggesting the impact of strategy on structure (e.g. Dyas and Thanheiser,
1976; Mintzberg, 1992). There is, however, also a substantial body of literature support-
ing the opposite claim, e.g. Hall and Saias (1980) suggest that structure impacts our per-
ception and decisions, while Harris and Ruei (2000) reveal that structure inuences
organizational performance.
Kupiec et al. 5
Building on the ndings discussed above, this article explores the relationships
between the structures and the functions of national CP evaluation systems. What distin-
guishes this study from previous research is the different denition of an evaluation
system. As Williams and Imam (2007) suggest, thinking in terms of evaluation
systems requires dening boundaries deciding what lies within and outside of them.
Hanberger (2011) and Widmer and Neuenschwander (2004) studied single organizations;
therefore, their evaluation systems are embedded within the boundaries of a single organ-
ization. However, this is not the only possible arrangement of or perspective on an evalu-
ation system.
The acquisition of evaluative knowledge is often performed by distinct bodies named
evaluation units. They identify knowledge needs, conduct studies and then feed knowl-
edge to the intended users (Olejniczak et al., 2016). User-units and evaluation units might
be from the same organization but when they are not we are dealing with a multi-
organizational evaluation system.
In complex policy settings, evaluation systems also become complex, and national CP
evaluation systems are model examples. They may consist of a single national evaluation
unit, but they may also contain many interrelated organizations acquiring and using
evaluative knowledge. Not only one evaluation-acquiring organization may serve
several user-organizations, but it is also possible that several organizations each of
them acquiring and using evaluation cooperate in the process or are coordinated by
one central body. Organizations comprising the evaluation system may operate at differ-
ent levels, e.g. national or programme levels. Relations between organizations may
resemble those characteristics of network structures or hierarchies, especially when
several levels are involved. An organization from a higher level often coordinates, facil-
itates and creates a regulatory framework for the subordinate. Lower-level organizations
often feed knowledge to their superiors.
Figure 1. Four potential functions of the evaluation system.
Source: own elaboration based on Olejniczak et al. (2017).
6International Review of Administrative Sciences 0(0)
This variety of possible relations, coordination, and cooperation modes draws our
attention from the perspective of evaluation functions. Therefore, the following hypoth-
esis can be put forward: the structure of the evaluation system correlates with its func-
tions. The evaluation system here is dened as all organizations involved in acquiring
or using evaluative knowledge in the CP context, while the structures are the relationships
between those organizations.
Research design and methods
The research is based on a cross-national comparative design, focusing on the iden-
tication and explanation of potential diversities across CP national evaluation
systems. This approach follows Lijphart (1971), who views the comparative
method as one of the legitimate scientic methods of establishing general empirical
propositions, and Sartori (1991), who sees comparison as an acceptable method for
controlling variables.
Country selection
The study covers CP evaluation systems in six out of 12 Central and Eastern European
Countries (CEECs) that joined the EU between 2004 and 2007: Bulgaria, Czechia,
Hungary, Poland, Romania and Slovakia. The analysis level is a country, and the
unit of analysis is a national CP evaluation system. Poland is the only country in
the studied sample with a CP implementation system consisting of national and
regional level arrangements. The regional level of Polish evaluation units was
excluded from the analysis to ensure comparability between cases. Case selection
was based on the most similar system strategy(Sartori, 1991). Thus, countries
with similar critical variables related to the administration system were selected,
except for the phenomenon that was to be investigated, i.e. CP evaluation system
structure (Figure 2).
All of the studied countries are characterized by a comparable administrative culture
derived from the communist legacy, as they were part of the Soviet Block
(Meyer-Sahling, 2009). After 1990, the course of the socio-political transformation
process and administration development trajectory were also alike in CEECs (Ágh,
Figure 2. Research scheme.
Source: own elaboration.
Kupiec et al. 7
2016). Recent research proves that several common features of the communist-type
administrations have persisted to the present day (e.g. Dimitrov et al., 2006) and
remain reected in the current characteristics and performance of public administration
(Thijs et al., 2017). This methodological choice to pair and compare CEECs with each
other is frequent in public administration studies (e.g. Ágh, 2002).
From the evaluation perspective, resemblances between the selected countries are
even more evident. None of them had developed an evaluation culture before joining
the EU. Hence, the evaluation practice in the CP established after 2004 was solely due
to external pressure from EU regulations. Additionally, all of the studied countries are
among the main beneciaries of CP funds in per capita terms, which places specic pres-
sure on evaluation systems.
Operational variables and data collection
Table 1 summarizes how the variables of structure and dominant function of the evalu-
ation system were operationalized and assigned measures. Data were collected at two
levels: (a) national, with the focus on the structure of the system, regulations and coord-
ination activities; and (b) organization, with the focus on potential variations in activities
and the orientation of evaluation units in different national systems resulting from choices
at the national level. At the organization level, the data were aggregated to provide
average/dominant characteristics for the national system.
The analysed data came from the following sources:
1. A survey among the heads of all CP evaluation units in the studied countries (see
Table 2). The questionnaire covered the unitsorganization and activities performed
and the dominant target audience of executed evaluation reports.
2. In-depth interviews with representatives of coordinating bodies for CP evaluation system,
leading evaluation units, local evaluation experts in each country (see Table 2). The inter-
views covered the activity and arrangements of the evaluation system at the national level.
Table 1. Operational variables and measures.
Operational variable Measure
Structure of evaluation
system
Type of structure Number of organizations
conducting evaluation
Degree of activity of coordination
body
Dominant function of
evaluation system
Type of knowledge provided by
evaluation studies
Proportion of strategic/operational
evaluation studies
Share of impact-oriented evaluation
studies in total
Dominant target audience Relative focus on external/internal
institutions
8International Review of Administrative Sciences 0(0)
3. Desk research of English-language sources, e.g. national guidelines and evaluation
plans, providing general knowledge of evaluation practices in the selected countries.
4. The evaluation database managed by the EC the number of impact evaluations con-
ducted in each country.
5. Six in-depth interviews with DG Regio and DG Employment representatives provid-
ing data triangulation and supplementing the national-based sources.
Descriptive statistics measures were employed to analyse sources (1) and (4). Sources
(2), (3) and (5) were analysed according to the qualitative approach methodology. First,
the main themes were identied with the use of coding. Patterns and relationships were
then specied, which enabled the content and narrative analysis of the gathered material.
Results
Description of the context CP implementation system
The evaluation systems that are the subject of this analysis are an inherent part of the CP
implementation system. This policy is performed based on the shared management
concept, which may be characterized as coordinated action taken by institutions at differ-
ent levels, aimed at setting policy goals and taking concrete actions to achieve them
(Dąbrowski et al., 2014). It means that responsibility and duties are shared between
EU and Member Statesadministrations.
The detailed scope of CP intervention and implementation arrangements in each
country are negotiated between the EU and Member States and formalized in the
so-called Partnership Agreement. The policy is implemented through operational pro-
grammes (OPs) covering specic areas of support. The number of OPs differs, but at
least some of them are implemented in every studied country. Operational programmes
are formulated and implemented by the managing authorities (MA) usually ministries.
Sometimes one institution acts as the MA for more than one OP. The MA may also dele-
gate some of its competences to implementation bodies other ministries or government
agencies.
Table 2. Data sources for analysis: survey and in-depth interviews.
Country
Survey: number of evaluation units surveyed (total
number of evaluation units in the country)
In-depth interviews: number of
persons interviewed
Bulgaria 5 (7) 5
Czechia 11 (12) 3
Hungary 1 (1) 4
Poland 20 (20) 3
Romania 3 (3) 6
Slovakia 13 (15) 3
Total 53 (58) 24
Kupiec et al. 9
Structures and activities of evaluation systems
Although developed under the same EU rules, CP national evaluation systems in the
studied countries differ in terms of structure. The main difference is in the number of
organizations that commission evaluation studies. While this activity is spread across a
number of organizations in some countries we call these systems decentralized, in
the rest, evaluations are commissioned by just one organization and then fed to others
we call these systems centralized. Within the group of decentralized systems, there is
another variation involving a coordination body, which exists and engages in coordinat-
ing activities in just some systems, making their structure less loose and activities more
coordinated.
The most common decentralized system design can be found in Czechia, Slovakia and
Bulgaria. Several evaluation units operate in each of these systems. They are located in
the ministries, which act as MAs for OPs. In these systems, each evaluation unit is
responsible for evaluating a single OP, implemented by the ministry that it is part of.
Poland is an example of even deeper decentralization. Evaluation units operate at the
level of ministries MAs and implementation bodies usually government agencies
responsible for implementing selected priorities under certain OP.
At the other extreme lies Hungary a centralized evaluation system with only one
evaluation unit. It operates in the Prime MinistersOfce and has exclusive competences
regarding the evaluation of seven OPs. The MAs for these programmes are located in dif-
ferent ministries. Hence, the Prime MinistersOfce is the only organization commis-
sioning evaluation, which is then fed to organizations using evaluation ministries.
The nal case, Romania, constitutes another example of a centralized structure. There
is a single unit responsible for evaluation there, operating within the Ministry of Regional
Development, Public Administration and European Funds. However, this unit is divided
into three sub-units two of them responsible for evaluating a single OP and the third
dealing with ve programmes. Implementation of all OPs is the responsibility of units
within the same ministry.
The number of evaluation units corresponds to the level of decentralization of the
system. A single unit in Hungary and one subdivided into three in Romania contrasts
with seven units in Bulgaria, 12 in Czechia, 15 in Slovakia and 18 in Poland.
In decentralized systems, the network of organizations with evaluation units is linked
by what can be termed the central coordination body(CCB). The scope of activities of
the CCB and the extent to which it regulates the operations in the system differ.
Active CCBs operate in Poland, Czechia and Slovakia. They establish working groups
and organize meetings, trainings and postgraduate courses (Poland); they provide
knowledge-sharing activities and evaluation conferences, as well as assisting in public
procurement procedures. They also support the process of designing evaluation plans.
To coordinate and standardize the formulation of evaluation plans, the quality assessment
of completed studies and recommendation follow-up, the Polish CCB issues formal regu-
lation, while the Czech and Slovak CCBs conne themselves to soft guidance and
support. All three Polish, Czech and Slovak CCBs also act as evaluation units focus-
ing on horizontal subjects and CP effectiveness at the Partnership Agreement level.
10 International Review of Administrative Sciences 0(0)
In striking contrast to the three CCBs described above is the Bulgarian CCB. Although
ofcially present in the system, it does not perform any coordinating activities, limiting its
job to merely forwarding communications from the EC and ensuring conformity to EC
requirements. Bulgarian evaluation units are left to themselves when planning and con-
ducting an evaluation or disseminating results. The CCB provides no support or training.
Based on the differences in the number of organizations with evaluation units directly
acquiring evaluation knowledge, and the activity of CCB, we propose a typology of
evaluation system structures consisting of three general types: centralized, decentralized
with active CCB and decentralized with inactive or no CCB (Figure 3).
The analysis identied several differences in the activities and work organization of
evaluation systems relating to their structure. The rst is in the number of conducted
studies. During the programming period, 20072013, around 100 studies were completed
in a centralized system in Hungary and 60 in Romania. At the same time, 180 studies
were accomplished in Slovakia, over 400 in Czechia and 450 in Poland. Bulgaria is
the exception among the decentralized system with only 40 studies. Even allowing for
that, the dominance of decentralized systems in terms of the number of completed
studies is apparent (Figure 4).
Figure 3. Typology of structures of cohesion policy (CP) national evaluation systems.
Source: own elaboration.
Figure 4. Number of completed evaluation studies.
Source: interviews.
Kupiec et al. 11
Evaluation systems also differ in terms of the average share of working time spent on
tasks related to evaluation by the staff of evaluation units. In all of the studied countries,
evaluation units also have other responsibilities. These include analytical activities and
management activities, such as developing programmes or information and communica-
tion tasks. The proportions, however, vary greatly between countries. While other tasks
amount to only 10% of working time in Romania and 25% in Hungary, they consume
around 85% of working time in Bulgaria and 70% in Slovakia. Hence, in the centralized
systems units usually specialize in evaluation and deal almost entirely with tasks related
to evaluation studies. In decentralized systems units usually have a more diversied
prole, and evaluation is just one task among others (Figure 5).
Functions of evaluation systems
This section discusses elements of the practice of CP evaluation systems, which indicate
different functions of the systems. Findings are presented for the three proposed types of
structure (compare Figure 3), allowing observation of the potential correlation between
the systems structure and functions.
Local experts and representatives of CCBs were asked about the proportion of stra-
tegic (focusing on the effects of intervention) and operational evaluation studies (focusing
on the process of implementation) generated in their countries. According to the intervie-
wees from Hungary and Romania, their systems provide a balanced mix of strategic and
operational knowledge. In the case of all decentralized systems, operational focus dom-
inates (Figure 6).
The observations from interviewees are conrmed by the data from the EU library of
CP evaluations. Studies collected there are classied as focused on impact, monitoring/
progress or implementation/process. As presented in Figure 7, impact-oriented studies
Figure 5. Average share of working time spent on evaluation by the staff of evaluation units (%).
Source: survey.
12 International Review of Administrative Sciences 0(0)
(strategic focus in our framework) are in a minority in all countries. However, their share
in the centralized systems of Romania and Hungary is signicantly higher than that in any
decentralized system.
Based on current data, it cannot be determined how exactly the different types of eva-
luations are used in studied countries. However, the EC pays attention almost solely to
impact evaluations generated in Member States, which suggests that in centralized
systems evaluation may serve accountability more often than in decentralized ones.
Figure 6. Focus of evaluation studies perception of local experts. Bulgaria no clear answer
received, but operational evaluations seem to dominate.
Source: interviews.
Figure 7. Share of impact-oriented studies in the total of completed evaluation studies (%).
Source: own elaboration based on EU library of Member States CP evaluations.
Kupiec et al. 13
Our main measure of the evaluation systems dominant purpose is the perceived
intended audience of evaluation units. In Hungary, external audiences other institutions
in the CP implementation system, domestic public institutions dealing with other policies,
and institutions at the EU level are as important as internal users managers of other
units and senior public administration staff in the same institution. This suggests that both
evaluation purposes accountability and learning are equally pursued there. In all other
countries, internal users are more important. The focus on internal users is most apparent
in Polands most decentralized evaluation system and the loose structure lacking an active
CCB in Bulgaria. Therefore, decentralized systems seem to focus on learning purpose
(Figure 8).
Summing up the ndings on the three identied types of national CP evaluation
system structure, there is a pattern in the observed similarities and differences in practice.
Decentralized systems tend to produce more evaluation studies. Evaluation units in those
systems are oriented on internal users, i.e. programme managers and senior public admin-
istration staff from the same organization. The large majority of acquired knowledge is
focused on operations, it concerns the implementation process and procedures. This is
true for both types of decentralized systems with and without an active CCB but
the orientation on internal users is more apparent in the latter. Centralized systems
produce fewer evaluation studies but provide a higher share of knowledge focused on
strategy. Findings from the evaluations are intended for both internal and external
users (e.g. institutions at the EU level). Referring these observations to our theoretical
framework (see Figure 1), it can be stated that decentralized systems are oriented
towards learning about operations and improving the implementation process. In centra-
lized systems, learning about operations is also performed, but there is more attention on
accountability for strategy results achieved (Figure 9).
Figure 8. Target audience of evaluation ndings.
Source: survey.
14 International Review of Administrative Sciences 0(0)
Conclusions
This article has identied the variance in national CP evaluation systemsorganizational
structures and has veried whether this variance correlates with differences in the
systemsfunctions. A comparative analysis has been conducted for six national CP evalu-
ation systems of countries that joined the EU between 2004 and 2007.
The results show that there are visible differences in the structure of evaluation
systems. Three basic types are identied: (1) centralized one evaluation unit serving
a multitude of user-organizations; (2) decentralized with an active CCB several evalu-
ation units spread among different ministries and one central unit performing a coordin-
ating role; and (3) decentralized without an active CCB.
The study shows that the practice of evaluation units functioning in the systems of
these three types differs in a manner indicating focus on different functions.
Decentralized systems are oriented towards learning about operations only, while centra-
lized systems also serve accountability for results.
On the practical side, these ndings provide three insights for the structural arrange-
ments of complex evaluation systems. First, any authority responsible for supervising or
formulating a regulatory framework for evaluation systems (the EC in our case) should
bear in mind that if certain evaluation functions are to be effectively fullled, the appro-
priate system structure must be indicated or required through regulations or guidelines.
Second, all analysed structures have their strengths and limitations. Decentralized systems
respond better to user needs and support learning, but are mostly operational. There are
limited strategic insights for accountability or learning about what works and why.
Figure 9. Relationship between the structure and functions of evaluation systems.
Source: own elaboration.
Kupiec et al. 15
Evaluation units of centralized systems provide more independent impact evaluations suited
to accountability, but their outcomes may not be relevant and useful for potential users within
implementing organizations.
Third, if both accountability and learning are desired, the evaluation system needs
some degree of decentralization on the one hand and the presence of an active CCB
on the other. Decentralized evaluation units located within the same organization as
their potential users are best able to provide relevant and timely operational knowledge.
The CCB: (1) may secure accountability because of its independence; (2) has capacity to
conduct robust impact studies; and (3) is in the best position to guide evaluation units and
combine their ndings into the reliable knowledge needed for strategic learning.
As regards implications for the theory, the study demonstrates a relationship between
the structure and functions of the evaluation system but does not establish causality. This
limitation opens the perspective for further studies. Based on current observations from
the studied evaluation systems, it can be conjectured that it is the structure that inuences
functions and that other contextual factors lie behind decisions on structure. However,
this should be veried empirically.
The issue of factors shaping structure is also open for further exploration. At this point,
we may only speculate that these factors could include the formal and informal charac-
teristics of public administration settings, cultures and traditions (Curristine et al.,
2007), the nature of external pressure on national administrations (Højlund, 2014) or
the power struggle between actors in the system (Martinaitis et al., 2018).
Furthermore, this study opens up the possibility for a more systematic analysis of
diversity in systems performance in terms of independence, quality of ndings and the
specic structuresability to combine individual studies into streams of evidence.
Finally, the current study was limited in geographical and policy terms (six CEECs, CP).
Therefore, the articles typology should be further tested in comparative research on evalu-
ation systems in other countries and specic policies. It could pave the way for a better
understanding of the role of evaluation in public administration and policy practice.
Acknowledgements
The authors thank the representatives of the studied countriesadministrations, who facilitated the
studys execution, and Weronika Orkisz Felcis, who contributed to the data collection process.
Karol Olejniczak acknowledges the Fulbright Program The Senior Award 202021 scholarship,
which allowed completion of the nal version of the article.
Declaration of conicting interests
The authors declared no potential conicts of interest with respect to the research, authorship, and/
or publication of this article.
Funding
The authors disclosed receipt of the following nancial support for the research, authorship, and/or
publication of this article: Empirical data was collected during a project commissioned by the
16 International Review of Administrative Sciences 0(0)
Ministry of Development Funds Poland. The comparative analysis was conducted as part of the
project nanced by the National Science Centre (2019/33/B/HS5/01336).
ORCID iDs
Tomasz Kupiec https://orcid.org/0000-0002-6469-7746
Dominika Wojtowicz https://orcid.org/0000-0002-8433-8935
Karol Olejniczak https://orcid.org/0000-0002-7079-2440
References
Ágh A (2002) Public sector reforms, institutional design and strategy for good governance in east
Central Europe. Studies in East European Thought 53(3): 233255.
Ágh A (2016) The decline of democracy in east-Central Europe. Problems of Post-Communism
63(56): 277287.
Argyris C and Schon DA (1995) Organizational Learning II: Theory, Method, and Practice.
Massachusetts: FT Press.
Balla S, Lodge M and Page E (2015) What makes a classic? Identifying and revisiting the classics
of public policy and administration. In: Balla S, Lodge M and Page E (eds) The Oxford
Handbook of Classics in Public Policy and Administration. Oxford: Oxford University Press,
pp. 111.
Batterbury SCE (2006) Principles and purposes of European Union Cohesion Policy evaluation.
Regional Studies 40(2): 179188.
Chelimsky E and Shadish WR (1997) Evaluation for the 21st Century: A Handbook. London: Sage.
Curristine T, Lonti Z and Joumard I (2007) Improving public sector efciency: Challenges and
opportunities. OECD Journal on Budgeting 7(1): 141.
Dąbrowski M, Bachtler J and Bafoil F (2014) Challenges of multi-level governance and partner-
ship: Drawing lessons from European Union cohesion policy. European Urban and Regional
Studies 21(4): 355363.
Derlien HU (1990) Genesis and structure of evaluation efforts in comparative perspective. In Rist
RC (ed.) Program Evaluation and the Management of Government. New Brunswick, NJ:
Transaction, pp. 147177.
Dimitrov V, Goetz K and Wollmann H (2006) Governing After Communism. Boulder, CO:
Rowman and Littleeld.
Donaldson S, Patton M, Fetterman D, et al. (2010) The 2009 Claremont debates: The promise and
pitfalls of utilization-focused and empowerment evaluation. Journal of MultiDisciplinary
Evaluation 6(13): 1557.
Dubnick M and Frederickson G (2011). Public Accountability: Performance Measurement, the Extended
State, and the Search for Trust. Washington, DC: National Academy of Public Administration.
Dyas GP and Thanheiser H (1976) The Emerging European Enterprise. London: Macmillan.
Favoreu C, Carassus D, Gardey D, et al. (2015) Performance management in the local public sector
in France: An administrative rather than a political model. International Review of
Administrative Sciences 81(4): 672693.
Fournier DM (2005) Evaluation. In: Mathison S (ed.) Encyclopedia of Evaluation. Thousand Oaks,
CA: Sage, pp. 139140.
Fratesi U and Wishlade F (2017) The impact of European Cohesion Policy in different contexts.
Regional Studies 51(6): 817821.
Kupiec et al. 17
Furubo JE, Rist RC and Sandahl R (2002) International Atlas of Evaluation. New Brunswick, NJ:
Transaction.
Hall DJ and Saias MA (1980) Strategy follows structure. Strategic Management Journal 1(2): 149163.
Hanberger A (2011) The real functions of evaluation and response systems. Evaluation 17(4): 327349.
Harris IC and Ruei TW (2000) The strategy/structure debate: An examination of the performance
implications. Journal of Management Studies 37(2): 587604.
Henry GT and Mark MM (2003) Beyond Use: Understanding evaluationsinuence on attitudes
and actions. American Journal of Evaluation 24(3): 293314.
Højlund S (2014) Evaluation use in the organisational context Changing focus to improve theory.
Evaluation 20(1): 2643.
Kudo H (2003) Between the governancemodel and the Policy Evaluation Act: New public man-
agement in Japan. International Review of Administrative Sciences 69(4): 483504.
Kupiec T (2020) Do we learn how to learn? Evolution of functions of evaluation systems The case
of Polish regional administration. Studia Regionalne i Lokalne 81(3): 722.
Kusek JZ and Rist RC (2004) Ten Steps to a Result-based Monitoring and Evaluation System.
Washington, DC: World Bank.
Leeuw FL and Furubo JE (2008) Evaluation systems: What are they and why study them?
Evaluation 14(2): 157169.
Lijphart A (1971) Comparative politics and the comparative method. American Political Science
Review 65(3): 682693.
Liverani A and Lundgren HE (2007) Evaluation systems in development aid agencies: An analysis
of DAC peer reviews 19962004. Evaluation 13(2): 241256.
Mark MM, Greene J and Shaw I (2006) The evaluation of policies, programs, and practices.
In: Shaw I, Greene J and Mark M (eds) The Sage Handbook of Evaluation. London: Sage,
pp. 130.
Mark MM, Henry GT and Julnes G (1999) Toward an integrative framework for evaluation prac-
tice. American Journal of Evaluation 20(2): 177198.
Martinaitis Z
, Christenko A and Krauc
iu
nieneL (2018) Evaluation systems: How do they frame,
generate and use evidence? Evaluation 25(1): 4661.
Meyer-Sahling JH (2009) Varieties of legacies: A critical review of legacy explanations of public
administration reform in east central Europe. International Review of Administrative Sciences
75(3): 509527.
Mintzberg H (1992) Structure in Fives: Designing Effective Organizations. Upper Saddle River,
NJ: Prentice Hall.
OECD (1998) Best practice guidelines for evaluation (PUMA policy brief No. 5). Paris:
Organisation for Economic Co-operation and Development. Available at: http://www.oecd.
org/governance/budgeting/1902965.pdf (accessed 13 August 2020).
Oh Y and Lee K (2020) External control mechanisms and red tape: Testing the roles of external
audit and evaluation on red tape in quasi-governmental organizations. International Review
of Administrative Sciences.
Olejniczak K (2013) Mechanisms shaping evaluation system A case study of Poland 19992010.
EuropeAsia Studies 65(8): 16421666.
Olejniczak K, Raimondo E and Kupiec T (2016) Evaluation units as knowledge brokers: Testing
and calibrating an innovative framework. Evaluation 22(1): 168189.
Olejniczak K, Kupiec T and Newcomer K (2017) Learning from evaluation the knowledge users;
perspective. Evaluation Theory and Practice 5(2): 4974.
Ostrom E (2007) Governing the Commons: The Evolution of Institutions for Collective Action.
New York: Cambridge University Press.
18 International Review of Administrative Sciences 0(0)
Patton MQ (2004) Evaluation research. In: Lewis-Beck MS, Bryman A and Liao TF (eds) The Sage
Encyclopedia of Social Science Research Methods. Thousand Oaks, CA: Sage, pp. 337340.
Picciotto R (2016) Evaluation and bureaucracy: The tricky rectangle. Evaluation 22(4): 424434.
Raimondo E (2018) The power and dysfunctions of evaluation systems in international organiza-
tions. Evaluation 24(1): 2641.
Rossi P, Lipsey MW and Freeman H (2004) Evaluation: A Systematic Approach. London: Sage.
Sartori G (1991) Comparing and miscomparing. Journal of Theoretical Politics 3(3): 243257.
Speer S, Pattyn V and De Peuter B (2015) The growing role of evaluation in parliaments: Holding
governments accountable? International Review of Administrative Sciences 81(1): 3757.
Stockmann R, Meyer W and Taube L (eds) (2020) The Institutionalisation of Evaluation in Europe.
Cham: Palgrave Macmillan.
Thijs N, Hammerschmid G and Palaric E (2017) A Comparative Overview of Public Administration
Characteristics and Performance in EU28. Brussels: European Commission.
Van Der Meer FB and Edelenbos J (2006) Evaluation in multi-actor policy processes:
Accountability, learning and Co-operation. Evaluation 12(2): 201218.
Widmer T and Neuenschwander P (2004) Embedding evaluation in the Swiss federal administra-
tion. Purpose, institutional design and utilization. Evaluation 10(4): 388409.
Williams B and Imam I (2007) Systems Concepts in Evaluation: An Expert Anthology. Point Reyes:
EdgePress of Inverness.
Zwaan P, Van Voorst S and Mastenbroek E (2016) Ex post legislative evaluation in the European
Union: Questioning the usage of evaluations as instruments for accountability. International
Review of Administrative Sciences 82(4): 674693.
Tomasz Kupiec is an assistant professor at The Centre for European Regional and Local
Studies (EUROREG), University of Warsaw, Poland. His main research interest are per-
formance of evaluation systems and use of knowledge in public administration.
Dominika Wojtowicz is a professor and researcher at the Department of Economics of
Kozminski University. Her interests focus on the effectiveness of public interventions
aimed at regional and local development.
Karol Olejniczak is a professor of public policy at SWPS University, Poland, and a
co-founder of research company Evaluation for Government Organizations. His work
focuses on programme evaluation and behavioural policy design.
Kupiec et al. 19
... It addresses the steps involved in defining, gathering, and delivering pertinent information for evaluating decision alternatives (Wanzer, 2021). According to Kupiec et al. (2023), evaluation is a methodical investigation into the value and usefulness of initiatives. Examining the results of policies and programs and guaranteeing organizational learning both rely heavily on evaluation. ...
... Examining the results of policies and programs and guaranteeing organizational learning both rely heavily on evaluation. Three classifications of evaluation systems-centralized, with a single evaluation unit; decentralized, with a coordinating body; and decentralized without a coordinating body-are used to classify evaluation systems (Kupiec et al., 2023). The internal knowledge consumers are the main focus of the decentralized evaluation system. ...
... From the perspective of Kupiec et al., (2023) evaluation is a systematic inquiry of the merit and worth of interventions. Evaluation is a vital tool for ensuring accountability and organizational learning in terms of examining the outcomes of policies and strategies. ...
Article
Full-text available
The purpose of this research was to examine how monitoring and evaluation systems affect small and medium-sized businesses in selected Ghanaian metropolises. The research design used was explanatory. The population for this study comprised owner-managers of SMEs in Ghana captured in the GEA and the AGI databases. From these two sources, a sampling frame was created for those in the selected metropolis. This gave a total population size of 1,189. For this frame, a sample of 423 SMEs was selected comprising SMEs from the Sunyani, Accra, Kumasi, Cape Coast, and Tamale metropolitan areas. The sample size for the study was 423 SMEs. The primary data collection tool was a self-administered questionnaire. Data was collected, cleaned, and coded before being entered into statistical software programs like SPSS and Smart-PLS. For the study's key findings, it was noted that the SMEs that took part in the study had monitoring and evaluation practices in place. The third and fourth objectives showed that monitoring and evaluation systems had a significant effect on both business resilience and business sustainability, respectively. It was discovered that monitoring and evaluation systems had a large and favourable effect on business growth and the digitalisation business respectively, for objectives 5 and 6. The study recommended that managers of SMEs within the selected area improve upon their monitoring and evaluation systems. This will increase the effect it will have on their resilience, growth, sustainability and digitalisation.
... With the emphasis on the organisational dimension of evaluations and the specific role of the ministry of finance, we contribute to scholarship on evaluation use by bridging two hitherto disparate literatures. The first is the small but growing literature on evaluation systems (Rist and Stame, 2006;Leeuw and Furubo, 2008;Raimondo, 2018;Kupiec et al, 2023b), which advocates a shift away from focusing on the influence of individual evaluation studies towards a more holistic emphasis on the very context in which evaluation studies are being produced. Studies on evaluation systems have generally focused on elucidating the processes of institutionalisation through which evaluation systems have come about, as well as analysing how such systems affect the use of evaluations (Højlund, 2014;Andersen, 2020). ...
... As already indicated in 2015 (Jacob et al, 2015), little systematic comparative research across countries exists. However, we know from the few systematic comparative cross-country studies (Jacob and Varone, 2004;Filgueiras and Queiroz, 2021;Kupiec et al, 2023b) that different evaluation styles (De Francesco and Pattyn, 2021) and levels of evaluation maturity (Furubo et al, 2002) exist across countries, and that different institutional design scenarios have been adopted (for example, Jacob et al, 2015;OECD, 2020;Stockmann et al, 2020). While there are several parameters defining the institutional set-up of policy evaluations at country level (Furubo et al, 2002;Jacob et al, 2015;Stockmann et al, 2020), the key feature is arguably the organisation providing strategic direction for evaluation within the executive (that is, criterion 2). ...
Article
Full-text available
Background Despite increasing scholarly interest in the organisation of evaluations within different countries’ political-administrative landscapes, not much attention has hitherto been paid to the consequences of a specific institutional set-up for the function of evaluations within government. Aims and objectives This article investigates how the organisational anchorage of policy evaluations within central administration shapes the function those evaluations primarily serve. Methods We focus on the role of ministries of finance for coordinating countries’ evaluation systems, and study its influence in Denmark and the Netherlands through a combination of document analysis and interviews with centrally placed civil servants. Findings Our analysis shows how the ministries of finance come to influence the evaluation activities of the whole central administration by constituting a specific economic outlook on evaluation, which (1) narrows down the applied evaluation methods and criteria; (2) inserts the ministry of finance as primary evaluation user; and hereby (3) furthers accountability rather than learning as the main function of evaluations within central administration. In both countries, the result is that the ministry of finance’s main role in the evaluation systems favours somewhat defensive qualities, where evaluations are primarily used for control and piecemeal changes in policies, rather than fundamental revisions or reflections on the appropriateness of specific policies. Discussion and conclusions Our findings indicate that the influence of evaluation systems is not only dependent on the degree of institutional anchorage of evaluation activities, but also very much a matter of whom the evaluation systems is centred around.
... Overall overview of policy learning factors in the six cases.External consultants involved in evaluations of CP, but with a small, limited labor market.Source:Authors' elaboration based onBachtler and Wren (2006);Furubo et al. (2002);Kupiec et al. (2021);Pellegrin and Colnot (2020);Polverari (2015). ...
Article
Full-text available
The European Union (EU), especially in the context of Cohesion Policy (CP), has played a crucial role in developing and promoting policy evaluation practices across its Member States. Evaluation systems across the Member States have been established to assess CP investments. Remarkably, the use of evaluation research and its contribution to stimulating policy learning has remained a “black box.” To address this issue, this article aims to develop a novel framework centered around four conditions for evaluation‐based policy learning, namely: (1) policy relevance, (2) resources and organizational settings, (3) quality of evaluation, and (4) evaluation culture. These conditions are retrieved from the existing literature on policy evaluation and applied to the six‐country cases across the EU. The findings suggest how loosening the formal EU evaluation requirements could affect policy learning in the Member States.
... Serangkaian tindakan yang telah dilakukan, baik sosialisasi maupun tidak, sangat penting untuk mengikuti kebijakan dan berkomitmen untuk menerapkan kebijakan melalui program (Taylor & Eberhard, 2020). B.Sumber Daya, ukuran dan kemampuan staff suatu organisasi hubungan formal dan informal antara suatu badan dan badan yang membuat atau melaksanakan keputusan tentang sumber daya manusia, anggaran, dan sarana yang diperlukan untuk menjalankan otonomi daerah yang optimal adalah tingkat pengawasan hirarkis terhadap keputusan subunit dan prosedur dalam badan pelaksana (Kupiec et al., 2023). Hasil observasi peneliti di lapangan bahwa Kepala dinas harus memeriksa situasi di lapangan mengenai tenaga guru kontrak. ...
Article
Full-text available
Program pemerataan tenaga guru kontrak merupakan salah satu Upaya penting dalam mengurangi masalah kekurangan tenaga guru kontrak disekolah negeri milik pemerintah daerah, termasuk di kabupaten kampar. Kecamatan Kampa dipilih sebagai fokus implementasi program ini oleh Dinas Pendidikan Kepemudaan dan Olahraga untuk mengatasi masalah kebutuhan tenaga guru yang signifikan. Implementasi program ini melibatkan berbagai strategi, antara lain penguatan kerja sama antar sektor serta peningkatan edukasi dan sosialisasi kepada tenaga guru kontrak tentang peningkatan kualitas sumber daya manusia, monitoring dan menyesuaikan strategi interverensi sesuai kebutuhan yang terindentifikasi. Tipe penelitian yang digunakan dalam penelitian ini yaitu dengan menggunakan metode deskriptif kualitatif. Hasil dari penelitian ini menunjukan peningkatan kesadaran pemerintah daerah tentang program pemerataan tenaga guru kontrak seperti perubahan perilaku yang lambat tetap menjadi fokus perbaikan dalam menjaga keberlanjutan program ini. Program pemerataan tenaga guru kontrak oleh Dinas pendidikan kepemudaan dan olahraga kabupaten kampar menunjukan progress yang positif dalam upaya masalah kebutuhan tenaga guru kontrak. Untuk memastikan program ini akan berhasil dalam jangka panjang dan meningkatkan kebutuhan tenaga guru kontrak secara keseluruhan, langkah selanjutnya diperlukan.
... Thus, for VTSM functions with multivariate and dependent stakeholders, more systematic decisions will be made. Also, new recommendations will be brought to the durability of safety management systems by making important contributions to how to monitor and manage possible variability [7]. ...
Article
This study analyzes the risks in ship traffic services management using the Functional Resonance Analysis Method (FRAM), a novel approach that focuses on identifying interactions leading to performance variability rather than errors. The research area is the Turkish Straits Vessel Traffic Service Management (VTSM) region, known for frequent risky ship passages. The goal is to ensure safe VTSM, minimize the negative impacts on people, goods, and the environment. The daily routine functions of the VTSM, determined through consultations with Vessel Traffic Operators (VTOs), are explained by following FRAM analysis principles. Qualitative methods, aligned with expert opinions, are used to examine potential performance variabilities and hazard factors. The resulting risk situation for each function is categorized by criticality on a color-coded scale. Solution proposals are provided to manage critical function variability, enhancing the VTSM system's responsiveness and adaptability.
... USA i Kanada to dobrze rozwinięte kultury ewaluacji w oparciu o wewnętrzne czynniki. Polska kultura jest nadal mocno niedojrzała, praktyka ogranicza się do wybranych pól i jest w dużej mierze efektem zewnętrznej presji (Kupiec et al., 2021). To sprawia, że przynajmniej część z przytoczonych wyżej argumentów nie jest adekwatna do naszego kontekstu. ...
... The Polish CP evaluation system is one of the largest in terms of the number of organisations commissioning evaluations and the number of studies conducted (Kupiec et al., 2021). The importance of Polish case stems also from the fact that Poland has been the largest financial beneficiary of the CP for the last 15 years (Bienias & Gapski, 2016), which has drawn the attention of the EU community and raised the importance of CP evaluation conducted in this country. ...
Article
Quality of life (QoL) is a fashionable concept widely used not only in many scientific fields but also in strategic documents, political speeches, and newspaper headlines. Although QoL is a crucial element of political agendas and a goal of public policies, there is little evidence on how this concept has been incorporated in policy analysis and evaluation. This study presents how the QoL has so far been defined, contextualised, and measured in the European Union Cohesion Policy evaluation in Poland, based on the analysis of 1,431 evaluation reports for the period 2004–2020. The findings show that the interest in the QoL is growing, and almost one-third of reports refer to the term. However, the concept is usually not clearly defined, and the number of evaluations with actual questions concerning QoL is limited. Moreover, the methodological approach to QoL measurement is rudimentary, limited to surveys of beneficiaries or end-users.
Article
Full-text available
Artykuł ten jest krytyczną refleksją nad inicjatywą Polskiego Towarzystwa Ewaluacyjnego na rzecz stworzenia ram kwalifikacji ewaluatora. Bezpośrednim celem jest wskazanie warunków sukcesu działań na rzecz profesjonalizacji ewaluacji poprzez certyfikację kompetencji i ocena, na ile inicjatywa PTE te warunki spełniała. Podstawą rozważań był systematyczny przegląd literatury dotyczącej certyfikacji, potwierdzania kwalifikacji ewaluatora i ich wpływu na jakość i wykorzystanie ewaluacji. Przegląd doświadczeń i polemik prowadzonych w środowiskach ewaluatorów innych krajów sugeruje, że skuteczność mechanizmu potwierdzania kwalifikacji ewaluatora wymaga m.in.: silnego, wiarygodnego lidera, zrównoważenia wymogów grandparentingu i decertyfikacji, rozróżnienia w kontekście ram kompetencyjnych co najmniej ról – zamawiającego i wykonującego ewaluację, wskazania przekonującego związku między potwierdzaniem kwalifikacji i jakością ewaluacji oraz zrezygnowaniu z pomysłu certyfikacji na rzecz dobrowolnego potwierdzania kwalifikacji.
Article
Full-text available
Public managers require different types of knowledge to run programs successfully. This includes knowledge about the context, operational know-how, knowledge about the effects, and causal mechanisms. This knowledge comes from different sources, and evaluation studies are just one of them. This article takes the perspective of knowledge users. It explores to what extent evaluation is a useful source of knowledge for public managers of cohesion policy. Findings are based on an extensive study of 116 Polish institutions: surveys with 945 program managers, followed by 78 interviews with key policy actors. The article concludes that: (a) utility of evaluation studies, in comparison to other sources of knowledge, is limited, (b) evaluation reports are used to some extent as a source of knowledge on effects and mechanisms, however, (c) "effects" are shallowly interpreted as smooth money spending, not socio-economic change. In conclusion this article offers practical ideas on what evaluation practitioners could do to make evaluation more useful for knowledge users in policy implementation.
Article
Full-text available
The article examines the evolution of functions of evaluation systems. It is based on the comparison of evaluation studies conducted in eight evaluation systems in Polish regional administration in two periods: 2007-13 & 2014-20. The findings are to some extent contradictory to the expectations built on previous literature on the subject. Although established due to external pressure the systems support accountability as well as learning. Systems do not focus on procedural issues only as the production of strategic knowledge increase in time. Numerous regulations imposed on analysed systems suggest however that the dominant type of evaluation use may be symbolic.
Article
Full-text available
Evaluations do not take place in a vacuum. Evaluation systems are embedded within organizations; they shape and are shaped by organizational norms, processes, and behaviors. In International Organizations, evaluation systems are ubiquitous. Yet, little is known about how they “function,” namely how they are used, how they contribute to organizational performance, and how they influence actors’ behaviors. These are empirical questions that cannot be solved without a robust theoretical grounding, which is currently absent from the existing evaluation literature. This article seeks to bridge some of the identified gaps by weaving together insights from evaluation theory and international organization sociology into a unifying framework of factors. The article then demonstrates how the framework can be used to empirically study the relative power and dysfunction of evaluation systems within International Organizations. A forthcoming connected contribution will illustrate such empirical inquiry through the case of the World Bank’s project-level evaluation system.
Article
Over the past three decades, red tape has been a significant issue in public administration. Several prior studies have discovered the external origins of red tape, noting the roles of external control on increased red tape. Despite the empirical studies, it remains unclear how the particular tools of external controls, such as external audits and evaluations, are associated with employees’ perception of red tape. To answer this question, we measured different external control mechanisms in quasi-governmental organizations in Korea and investigated their effects on red-tape perception. The results show that external audits increase employees’ red-tape perception, whereas external evaluations decrease their perception. However, the positive function of evaluation disappears under a high level of government control because both external audits and evaluations increase red-tape perception when more government control is perceived. Points for practitioners This study sheds theoretical light on the negative impacts of external control in rule management. The impacts of audit and evaluation on red tape are empirically identified in public organizations. Practitioners need to realize that the use of audit and evaluation may produce unintended effects that increase ineffective rules even though such control mechanisms are designed to check mismanagement in quasi-government organizations.
Book
This book examines the progress of institutionalisation of evaluation in European countries from various perspectives. . It describes both prior developments and current states of evaluation in 16 European countries and across the European Union (EU), focussing on three dimensions, namely the political, social and professional systems. These detailed country reports, which have been written by selected researchers and authors from each of the respective countries, lead to a concluding comparison and synthesis. This is the first of four volumes of the compendium The Institutionalisation of Evaluation to be followed by volumes on the Americas, Africa and Australasia. The overall aim is to provide an interdisciplinary audience with cross-country learning to enable them to better understand the institutionalisation of evaluation in different nations, world regions and different sectors. Reinhard Stockmann is Professor for Sociology at Saarland University, Germany. He is Founder and Director of the Center for Evaluation (CEval), Managing Director of two Master of Evaluation programs, Executive Editor of the German Journal of Evaluation (ZfEv) and Founding member of the German Society for Evaluation (DeGEval). He is an evaluation specialist in the field of development cooperation, education, vocational training, environmental policy and foreign cultural policy. Wolfgang Meyer is Assistant Professor at Saarland University, Germany and Adjunct Professor at Uganda Technology and Management University, Kampala, Uganda. He is Vice-Director of the Centre for Evaluation (CEval), Founding Member of the German Evaluation Society (DeGEval), Sociologist with a Focus on Empirical Methodology and Evaluation Specialist in the Field of Environment, Labour Market and Regional Development. Lena Taube was previously a research associate at the Center for Evaluation (CEval) at Saarland University, Germany, where she focused on research regarding the institutionalisation of evaluation as a means to good governance. She is currently working at the German Institute for Development Evaluation (DEval) engaged with evaluating human rights in German development aid.
Article
How does the design of evaluation systems affect the different ways of using the results of evaluations? This article offers a conceptual model that outlines three ‘ideal’ types of evaluation systems. It is a heuristic tool for opening up the ‘black box’ of evaluation systems and assessing their qualitative differences in terms of types of ‘owners’ of evaluations, questions asked, methods deployed, answers provided and avenues for use of evaluative knowledge. We apply the model to study the case of the Lithuanian evaluation system. In contrast to the expectations of some of the previously developed models, it does use evaluation results, and we aim to understand why the generated evidence is more often used in some areas rather than others.