ORGAPET Section A2:
Evaluation Principles and Special Considerations for Evaluating Organic Farming
Policies and Action Plans
Nic
Lampkin
Aberystwyth University, UK
Johannes
Michelsen
University of Southern Denmark
Matthias
Stolze, Hanna Stolz
Research Institute of Organic Agriculture
(FiBL),
CH
Christian
Eichert, Stephan Dabbert
University of Hohenheim,
DE
Version 6, April 2008
A2-1
Introduction
Since the late
1980s, European countries have developed policies for organic farming on a
number of levels. These include
EC Regulation 2092/91 defining organic production;
support for organic production, processing and marketing through agri-environment,
rural development and structural measures; support for research and information
dissemination measures; the development of national and EU action plans for
organic farming; and continuing reforms of the main commodity elements
of the Common Agricultural Policy. Increasingly, an integrated approach to
the implementation of these policies has been developed in the form of action
plans in several countries, and now also at the EU level. The evaluation of
these action plans, and organic farming policies in general, provides the
rationale for ORGAPET.
EU member states
have invested more than €6 billion in policy support for organic farming, with
investment in organic farming agri-environmental measures rising from €250
million annually in 1996, to €650 million annually in 2003 (Lampkin
et al.,
1999; Stolze and Lampkin, 2006), although this is still a tiny fraction of
total public expenditure on agricultural policy.
From a governance
perspective alone, there is a need to ensure that these resources are being
invested appropriately. The impacts and cost effectiveness of these policies
is an issue of increasing importance as the size of the organic sector, and
the consequent demand for resources, increases. There is, in any case, a formal
requirement for ongoing monitoring and evaluation of policies at national
and EU level (e.g. current mid-term review of rural development and structural
programmes). Competing claims on resources are likely to become louder,
and there needs to be clear evidence of benefits to justify their continuing
application to organic farming.
However, the evaluation
of these impacts is not simple because organic farming works on a number
of different levels, with multiple, sometimes conflicting, objectives and
impacts. In addition, if the focus is on a single objective (e.g. hedge-planting), the benefits from supporting a systems approach such as organic farming may be less than can be achieved by more targeted
measures. However, if the total benefits across all relevant objectives are
considered, this may be sufficient to more than justify the costs of
organic support compared with single-objective, single-measure schemes. To
make a full assessment of the effects of such synergies and conflicts in a
complex systems context, current evaluation methodologies need further development.
This document set
outs the basic principles of evaluation as well as the specific issues relating
to organic farming policies that have helped define the scope and
structure of the ORGAP evaluation toolbox.
A2-2
Principles of evaluation
The evaluation
approach used in this project builds on the European Commission’s requirements
for ongoing monitoring and evaluation of policies at both a national and European
level, which feed into the process of reviewing rural development and structural
fund programmes. As part of ongoing concerns about the quality of programme
monitoring and evaluation and the validity of outputs, DG XVI (Regional Policies
and Cohesion) commissioned the MEANS programme (1994-1999)
to develop a coherent set of approaches and methods for future evaluations.
The results (EC, 1999) provide a framework for evaluation, as well as guidance
on developing structures for collecting common indicator sets for monitoring
and subsequent evaluation purposes.
However, it is
appropriate to look in broader terms at principles of evaluation in order
to determine whether the MEANS framework is appropriate and/or adaptations
or alternative approaches are needed.
A2-2.1
Definitions
The following definitions
of policy programme evaluation are based on the literature cited in each
case. Definitions
1 to 4 derive from the context of political science, while definition 5 is
specific to EU structural policy and tailored to the socio-economic evaluation
of agricultural policy programmes (MEANS framework). The sixth definition
is related to programme evaluation.
-
Careful retrospective
assessment of the merit, worth and value of administration, output and outcome
of government interventions, which is intended to play a role in the future,
practical action situation (Vedung, 1997).
-
Judgement of the
value of a (usually) public intervention with reference to criteria and explicit
standards (e.g. its relevance, efficiency, sustainability, equity etc.). The
judgement usually concerns the needs which have to be met by the intervention
and the effects produced by it. The evaluation is based on information which
is specially collected and interpreted to support the judgement. For example:
evaluation of the effectiveness of a programme, cost-benefit evaluation of
a project, evaluation of the validity of a policy and evaluation of the quality
of a service delivered to the public. Certain definitions
of evaluation exclude the judgement dimension and limit evaluation to an assessment,
description or measurement of the intervention’s effects. Other, more restrictive,
definitions limit evaluation to the ex-post estimation of effects. In certain
contexts, evaluation focuses not on a public intervention but on a public
organisation (e.g. evaluation of a university, hospital or EU institution) (Tavistock Institute, 2003).
-
Evaluation as qualified
monitoring is usually a feature of public decision-making systems. As impact
assessment, evaluation is frequently commissioned on specific occasions. Whether
permanent or periodic, monitoring or impact-assessing, however, evaluation
is performed for either accountability, intervention improvement or basic
knowledge advancement (Vedung, 1997).
-
In a
political sense, evaluation is the verification of a programme or an institution by
a group composed for this purpose, which generates a certification on the
basis of expertise, sometimes including recommendation for cessation, continuation
or optimisation of the evaluated scheme (Kromrey, 2001).
-
The evaluation
of public intervention consists of judging its value in relation to explicit
criteria and on the basis of information that has been specially gathered
and analysed. The purpose of policy programme evaluations is to check
the raison d’être of a public intervention,
to confirm both reproducible success stories and failures not to be repeated,
and to report back to citizens. Whereas private sector organisations have
little need to justify or account for their actions since they are judged
by the market, public sector organisations are obliged to evaluate their activities.
The European Union is using a number of different forms of evaluation: managerial
evaluation aimed at improving management, democratic evaluation used for accounting
to citizens and pluralistic evaluation that tries to bring agreement between
the partners involved in public action. Each of these three concepts are to be found
within the experience of evaluating EU policies, with one
of them predominating according to whether the evaluation concerned is ex-ante,
ex-post or mid-term/collateral (EC, 1999).
-
Evaluation is
the systematic investigation of the merit or worth of an object (programme)
for the purpose of reducing uncertainty in decision-making (Mertens, 1998).
In summary, policy evaluation
involves the systematic gathering of information and assessment of a programme
according to specific criteria in order to make judgements about the value
of the programme, thus reducing uncertainty in decision-making about future
actions. The assessment of value may relate to the goals of more than one
specific interest group, including policy-makers, beneficiaries and third
parties, and they may fulfil a range of purposes, from financial control and
accountability to intervention improvement and knowledge advancement.
A2-2.2
Paradigms and conceptions of evaluation
Stockmann (2004)
identifies two different paradigms of evaluation:
-
Evaluation as
an empiric-scientific process which is based on empirical research methods.
Evaluation could therefore be seen as applied social research which has to respect
special research conditions and emphasises the practical use of evaluation
results.
-
The second paradigm
questions the ability of empirical science to identify reality objectively. Moreover,
'reality' is assumed to be socially constructed
from varying and sometimes conflicting perspectives. As stakeholder groups
take different power positions, particular interests can be 'over'-emphasised.
Therefore, the aim of an evaluation is not in giving a description and assessment
as near to reality as possible, but rather in changing reality in favour of
the disadvantaged stakeholder groups. Evaluation in this sense is a transformative
act.
EC (1999, Vol. 1:19-23)
identifies three different conceptual approaches to evaluation: managerial,
democratic and pluralistic.
-
Managerial evaluation
emphasises the efficient use (optimisation) of limited public resources
between competing demands. In this context, the
political dimension is not
important – the evaluator operates to objective, scientific standards, guaranteeing
the independence of the evaluation.
-
The democratic
concept emphasises the need for public authorities to be accountable
to their citizens, involving them directly in the process with judgements
made by elected representatives, through forums such as the Danish public consensus
conferences and open juries in the UK. Evaluation professionals provide technical
advice only.
-
Pluralistic
evaluation (also known as participative or stakeholder evaluation) attempts
in an impartial way to integrate the technical components of the managerial
approach and the political components of the democtratic approach to problem-solving and the involvement of
stakeholders in the evaluation process.
Even though the
discussion of paradigms is not yet concluded, in recent years there is
increasing consensus that evaluations should consider the views and the
needs of stakeholders. The pluralistic approach, to some extent represented
by the MEANS framework and not inconsistent with the second of Stockmann’s paradigms,
reflects this.
A2-2.3
Aims of programme evaluation
Stockmann (2004)
identifies four possible aims of an evaluation (summarised in
Figure
A2-1):
-
Provision
of information and knowledge to the benefit of both the client and the target
group of a programme. Areas of interest are, for example, the way a programme
is running, the needs of the target group, whether the measures reach the
target group, the acceptance of the programme, whether institutions are able
to implement the programme efficiently and effectively, etc. The aim of the
collection of knowledge and of the corresponding assessments (evaluation),
conducted on the basis of evaluation criteria, is to identify control decisions
and to reduce deficits in the programme.
-
Exertion
of direct and indirect monitoring of those who are involved in a programme.
Apart from gathering knowledge about the programme, evaluation also discloses whether
all actors involved fulfil their duties and whether they are competent and
qualified to do so.
-
Creation
of transparency, enabling dialogue between the stakeholders of a programme.
Evaluation thus provides transparency about the success and progression of
the co-operation between stakeholders and decision-makers, providing a basis
for learning processes.
-
Documentation
of achievements over time or legitimisation (sometimes without making use of
results). Evaluations allow the verification of programme input/output and achieved impacts over time. With
ex-post evaluations, the
sustainability of a programme may be verified. Thus, financiers and operational
organisations can be informed about resource efficiency and programme impact
levels. Evaluation results may be used for legitimising
programme operation.
Figure A2-1: Aims of programme evaluation

Source:
Stockmann
(2004)
According to
Stockmann
(2004), evaluation results may be used for:
-
improving the planning
of a programme or measure (ex-ante evaluation)
-
monitoring
implementation processes (ongoing evaluation)
-
ex-post assessment
of the effectiveness and sustainability of an intervention (ex-post evaluation).
Table A2-1 and
Figure A2-2 illustrate this in more detail.
Table A2-1: Dimensions
of evaluation research
Stages
of the
programme process |
Analysis
perspective |
Perception
interests |
Evaluation
concept |
formulating
the
programme/planning stage |
ex-ante |
“analysis
for policy”
“science
for action” |
preformative/formative:
proactive design, process-orientated, constructive |
implementation
stage |
ongoing |
both
possible |
formative/summative:
both possible |
impact
stage |
ex-post |
“analysis
for policy”
“science
for knowledge |
summative:
summarising, making up the balance, result-orientated |
Source:
Stockmann
(2004)
Evaluation
may therefore have a formative (process-orientated, constructive or communication-promoting)
design or a summative (result- oriented, concluding, accounting) character.
Except in the ex-ante
case, evaluations may permit observation of operational processes and the identification
of programme problems. In this context, important questions include: whether the measures
are acceptable to stakeholders; which conflicts of interests are occurring
and whether enough qualified personnel is provided for the implementation
of measures; whether communication and co-ordination with the programme target
group works well; and whether the innovations established by
the programme are leading to programme objectives etc. (Stockmann,
2004).
Figure A2-2: Profile of
tasks in an evaluation
Source:
Stockmann
(2004)
EC (1999, Vol. 1:
24-25) identifies three key aims which correspond, at least in part, with Stockmann’s
perspectives:
-
Verifying that
public action responds to uncovered or insufficiently satisfied needs (legitimisation
– is there a real problem to be addressed and can public intervention be
justified by the failure of the private sector to act? In the EU context,
this also refers to the balance between actions at EU level and considerations
of national sovereignty.)
-
Improving interventions,
particularly through feedback and modification during the period of the
programme.
-
Accountability
(responsibility/liability), reporting to political authorities and citizens
on the results obtained and on the sound use of allocated resources. At
the EU level, community value-added (relative to actions supported at the
national level) is a key element.
Vedung (1997) similarly
identifies control, knowledge and change as key potential aims of evaluations,
with an emphasis on practical application rather than the pursuit of science.
A2-3
Examples of evaluation approaches
A2-3.1
The Stockmann (CEVAL) approach
Stockmann
(2004) describes two relevant aspects of evaluation:
the participative and multi-methodological
aspects. The first aspect is related to the respective roles of persons
concerned in the evaluation, while the second relates to the evaluation methods
used.
A2-3.1.1
The participative aspect
According to
Stockmann
(2004), external evaluators are faced with a range of different stakeholders and
actors and, therefore, a participative aspect is advisable. The participative
aspect of programme evaluation means that evaluators and evaluatees work
together as partners, instead of a relationship consisting of
a controlling and controlled group of persons. The intention of the participative aspect
is to
enhance the validity of evaluation results, because valid assessment
of measures and results is improved through the voluntary and active
co-operation of all those involved. In Stockmann's view, the most effective
method of participative evaluation is achieved where evaluators and evaluatees generate the processes
and criteria of evaluation together. He suggests that they decide
together which actors are to be involved and adjust their ideas alongside
other participants. The substance and implementation of the evaluation should be connected as closely as possible with
the interests and needs of all persons involved. According to Stockmann, such a procedure allows for
continuous adoption of appropriate evaluation instruments, and enables a flexible
reaction to changes of context within the processes of evaluation (Stockmann,
2004).
Based on
Stockmann
(2004), a participative evaluation consists of three stages. During the first
stage, there is emphasis on the methodological knowledge of evaluators.
The second stage is principally focused on an investigation of relevant information delivered
mainly by the evaluatees. Through continuous intermediation
and through workshops with stakeholders, the
integration of all participants in the evaluation process can be guaranteed.
Following data analysis using empirical social research methodology, the third stage of the
evaluation is reached: the review and discussion of findings by both
evaluatees
and evaluators (and, in some cases, with stakeholders). At this stage, development strategies are
worked out which are then implemented by those concerned and their organisations. Through the
establishment of monitoring and evaluation systems, the progress of implementation
may also be observed.
A2-3.1.2
The multi-methodological aspect
According to
Stockmann
(2004), another important aspect of the evaluation concept is the choice of
suitable methods, as well as the precise development of instruments
for data analysis. Lack of time often prevents the use of experimental
design, which would usually be necessary for impact assessment, so a combination
of qualitative and quantitative instruments is advisable. For the analysis
of process-oriented data (e.g. programme steering, programme process, etc.),
qualitative instruments are predominantly suitable, while quantitative investigational
and analytical methods are chosen for the verification of aim achievements,
impacts and causalities. The following methods are used frequently in evaluations:
-
Secondary analysis
of existing data
-
Guideline (semi-structured)
interviews
-
Standardised
(structured) interviews
-
Case studies
The question of which
methods are chosen depends on the respective task and aims of the evaluation but, generally, qualitative and quantitative research
methods should be combined for the benefit of both clients and stakeholders (Stockmann
2004).
A2-3.2
The MEANS approach
The MEANS approach
(EC, 1999) was developed for the European Commission to provide a basis for
assessing, in particular, the socio-economic impacts of rural development and
structural policies. Although the MEANS approach does not formally address
the environmental impacts of policies, which have been addressed separately
by OECD and EU initiatives (e.g. IRENA agri-environmental policy indicators 1, 2, 3, 4 and 6: EEA, 2005), some of the basic MEANS concepts
are applicable also for this purpose.
The MEANS approach
is designed to address two key issues which are also common to the evaluation
of organic action plans (EC, 1999 Vol. 1:27-30):
-
The policies
to be evaluated are intended to achieve multiple goals, and these goals (and
the values attached to them) will differ according to the time frame (short,
medium, or long-term), the political
level (EU, national or local) and the stakeholders involved (politicians, administrators, beneficiaries or
other interest groups). Consequently, evaluations need to consider multiple,
often interlinked, objectives that are difficult to observe and analyse in
detail.
-
The policy
instruments used can be applied in multiple domains (e.g. fisheries, agriculture,
transport, tourism) and a wide range of instruments are available, resulting
in complex interactions between multiple objectives, multiple domains, multiple
instruments and multiple geographic/institutional levels. This makes it difficult
to identify causal links between policies and outcomes.
The MEANS approach,
as set out in the six-volume MEANS collection (EC, 1999), proposes a multi-methodological
approach. It addresses in detail the:
-
setting up of
evaluations including definition of aims and objectives, priorities and
content (Vol. 1);
-
selection and
use of indicators for monitoring and evaluation (Vol. 2), including elements
of stakeholder participation, at least at the level of suppliers and users
of the information, but also potentially including the public (stakeholders/citizens)
in defining key indicators likely to be of interest and relevance to them;
-
range of quantitative
and qualitative evaluation techniques and tools applicable to evaluation
(from SWOT analysis to GIS applications and multi-criteria analysis) (Vol.
3);
-
key areas of
analysis with examples from pilot evaluations (Vol. 4), including impact
mapping (to evaluate different measures by means of common criteria); scoring
scales (to evaluate different measures in a homogeneous manner); cross impact
matrices (for analysing synergy within a programme) and multi-criteria analysis
(for judging the overall success of a programme);
-
transversal
evaluation of programme impacts on key public policy goals, such as environment
and employment, which are not necessarily specific objectives of the programme
to be evaluated (Vol. 5);
-
and includes
a glossary of 300 concepts and technical terms (Vol. 6).
Since the development
of ORGAPET is intended primarily to assist the European Commission in the
evaluation of the European action plan for organic food and farming, it makes
sense to base the evaluation approach on systems already in use by and familiar
to the Commission, and to member states who are required to deliver national/local
evaluations to the Commission with suggested modifications to
address the specific context of organic farming policy (see below). However, it
is already clear that a participative, multi-methodological approach, including
quantitative and qualitative tools, will be needed. The different elements
of ORGAPET will set out how many of the tools and approaches covered by MEANS
can be implemented in the specific context of organic farming. In particular, ORGAPET
Section A5 identifies the key issues
to be considered in planning an evaluation and elaborates on the issues identified
above.
A2-3.3
Evalsed: MEANS updated for 2007-2013
At the end of 2006, the
European Commission published an internet-based framework for the evaluation of
socio-economic development (Evalsed),
with a particular focus on evaluation of regional policy within the 2007-2013
programming period. Evalsed is designed to update the MEANS framework to take account of developments
in both the policy environment and in policy evaluation methodology since MEANS
was originally published.
A2-4
Specific considerations relating to organic farming
policy evaluation
Organic farming
policies have developed progressively in European countries since the 1980s
(Lampkin et al., 1999;
Dabbert et al., 2004;
Stolze and Lampkin, 2006). Initially, policy measures
focused on two main issues:
-
the harmonisation
of organic farming standards, including the development of a legal framework
to protect the interests of consumers and bona fide producers and provide
a basis for the development of the organic food market;
-
the provision
of financial support (typically on a per hectare basis as an agri-environmental
measure) to support producers during conversion to organic production and,
in many cases, the longer-term continuation of organic production, in recognition
of the environmental and other societal benefits but also as a means to encourage
growth in the sector.
These policies
have contributed to the rapid growth in the organic food and farming sectors
since the mid-1990s although, as might be expected with a small and rapidly developing
sector, supply and demand have not grown in tandem. The increase in supply
resulting from area-based support and other supply-push policies has led in
several cases to market disruptions, with over-supply leading to falling prices.
As a result, policy-makers have increasingly turned to demand-pull measures
including promotion and public procurement, as well as marketing and processing
projects, supported to a large extent through EU rural development and structural
programmes. These policies are often integrated in an action plan framework
at national or local (regional) level, in order to address specific local
development bottlenecks as well as to achieve better complementarity between
measures.
The development
of the EU action plan (EC, 2004) was envisaged to provide an overarching vision
and framework for development of policies at national and regional levels.
It recognises the dual role of organic farming – responding to consumer demands
for quality food and providing environmental and other societal benefits.
It also looks at developing opportunities for promotion and research, but
the primary focus is on improvements to the organic regulation and the implementation
of control systems.
The policy mix
which is now in place for organic farming in Europe means that the evaluation
of organic farming policies and action plans faces the same challenges identified
with respect to rural and structural programmes in general (and which themselves provide the
basis for most organic farming policy measures).
However, there
is another aspect to organic food and farming which is critical in the context
of both policy development and evaluation, and that is the role of stakeholders, in particular producers and consumers.
Unlike many agri-environmental and rural development policy measures, the organic farming concept was
not developed by policy-makers and technical experts responding to a specific
policy need. It has evolved since the early 20th century as a global
movement for agricultural change, with committed individuals and producers
working together to define standards and develop systems and practice, in
order to achieve the movement’s goals of environmental protection, animal
welfare, food quality and health, and social justice. At the same time, producers
have had to find solutions to the challenge of maintaining financial viability
while limiting the use of specific output-enhancing technologies and practices
in pursuit of the broader goals. In the absence of direct policy support during
most of the 20th century, organic producers turned to the market
place and the willingness of committed consumers to pay a premium price, in
order to maintain their financial viability.
As the goals of
policy-makers and the organic movement have come closer together, the opportunities
to provide policy support for common societal goals have been recognised,
but the challenge of doing so, consistent with the market focus of the sector,
has not always been met. With increasing policy support, the incorporation
of organic standards in legal frameworks and the involvement of major retailers
and food multinationals in the organic market, there has been increasing concern
about the ‘institutionalisation’ of the organic movement and its ‘takeover’
by public and private corporate interests.
It is thus an issue of evaluations to consider the extent to which policy
interventions change the dynamics of the original organic food and farming sector
and to consider whether the effects are positive or negative. These question can hardly be answered without
a systematic involvement of stakeholders associated with the organic movement – on national, regional or European level.
There have been
a number of recent efforts to focus specifically on evaluation of organic
farming policies at the European level. The EU-funded research project OFCAP
(Organic Farming
and CAP Reform (FAIR3-CT96-1794) looked at policies implemented in the period 1993-1997, with the results reported
in the Organic Farming in Europe – Economics and Policy
series (Volumes 1-10) published by the University of Hohenheim. Of particular
relevance is the overview of policies implemented (Lampkin
et al., 1999) and
the policy impact assessment of Häring (2003), with
Dabbert et al.
(2004) providing an overview of the whole project. More recently,
Häring et
al. (2004) (funded by DG-ENVIRO) provided a first evaluation of the impacts of the Agenda
2000 reforms, in particular the main commodity measures and the rural development
programme, on organic farming. From 2003 to 2007, the EU-funded
EU-CEE-OFP project (Further development
of European organic farming policies, with particular emphasis on EU
enlargement (QLK5-2002-00917) analysed the effectiveness of organic
farming policies (OFPs) in the old EU member states and Switzerland (EU15/CH);
the regional and spatial impacts of existing and potential OFPs on farm structures
and production in EU15/CH; the development of organic farming and the policy
and regulatory environment in the new EU member states from central and eastern
Europe; the development and implementation of organic farming regulations
and markets in the CEE countries; the farm level economic impacts of OFPs,
Agenda 2000 implementation and EU enlargement in selected countries; policy
networks for developing OFPs in selected countries; and the involvement of
policy-makers and stakeholders in identifying parameters for further development
of European OFPs.
In developing approaches
to the evaluation of the EU organic action plan and other organic farming
policies, the ORGAP project needs to build on the experiences and knowledge
developed in these earlier studies, in particular with respect to the application
of evaluation methods, but also with respect to stakeholder involvement. ORGAPET
Section A3 goes into more detail on the programme theory – i.e. the mechanisms
by which policy interventions impact on the development of the organic sector
– to provide a sound technical basis for the conduct of evaluations.
A2-5
Implications for the Organic Action Plan Evaluation
Toolbox (ORGAPET)
The perspectives
outlined above lead to the conclusion that the ORGAP
evaluation toolbox should enable the careful (i.e. systematic) assessment of
policy-driven organic action plans and other
implemented policies,
reflecting the interests not only of policy-makers but also other stakeholders.
Evaluations should not be used primarily to advance scientific knowledge
and understanding (although this may be a by-product) but should aim to find
practical solutions to the further development of action plans and organic
agriculture.
A2-5.1
Objectives of ORGAPET
The ORGAP evaluation
toolbox is intended for use by the European Commission and EU member state
governments and relevant stakeholders to develop and evaluate the European and national organic action
plans. As many of the administrators and members of the action plan stakeholder
groups involved are not expert evaluators or researchers, ORGAPET must be
easy and practical to apply and parsimonious regarding the number of variables
(objectives, measures and indicators) included in the toolbox. ORGAPET must
include a limited number of evaluation procedures, serving different goals
and focusing on different stakeholders.
In order to meet
the different information requirements of ORGAPET, three
levels of information are envisaged:
-
A practical manual aimed at policy-makers and stakeholders engaged in
policy development and in utilising the results of evaluation. At this level,
information on evaluation methods is restricted to that necessary for the interpretation of results.
-
A more detailed overview of each of the main areas of evaluation
aimed at non-expert administrators and stakeholders who are involved in
commissioning or carrying out evaluations. For each section
of ORGAPET, a separate document is provided which aims to set out the key issues that need to be considered.
Some of this information will already be well-known to professional evaluators,
but the overview should still be useful for putting the specific
organic action plan issues that need to be considered in context.
-
The overviews in each section of the toolbox are supported by detailed
annexes which fulfil different supporting roles: they may be examples of
applications of specific techniques including working guidelines and results,
or they may be information or data resources that might be valuable in interpreting
results and making comparisons with other situations.
The objectives
of the toolbox can be defined in more detail by addressing 8 questions
suggested by Vedung (1997) as a basis for planning an evaluation:
-
What is the
aim of the evaluation? (e.g. control/accountability, improved knowledge,
programme modification/change etc.)
-
How is the evaluation
organised? (Who commissions and who carries out the evaluation, and how
are they interrelated?)
-
What is the
programme to be evaluated? (There needs to be a clear description to set
the boundaries for what is to be evaluated)
-
What is the
public management process between input and output? (How has the decision
been implemented, which agency was involved, what was the content, have
target groups responded the way they promised, is there evidence of implementation
failure?)
-
What are the
results (outputs/outcomes) of the programme (The focus should not only be
on intended outcomes –
are there other outputs, e.g. national action
plans resulting from the EU action plan. What further outcomes could be
included?)
-
What are the
factors explaining the results (programme/others)? (What are the effects/impacts
of the policy; what are the causalities involved; how are these influenced
by the general context of society?)
-
What are the
evaluation criteria and standards used for assessment? (What value statements
are involved; what are the goals; what did the stakeholders expect?)
-
How and by whom
is the evaluation to be used? (Intention and reality – possible purposes
include instrumental, political, internal reflection).
Taking each of
these in turn, we can identify both the scope of ORGAPET and the sections of
the toolbox that address them in greater detail.
-
the
aims of organic action plan evaluations for which ORGAPET is developed
are to document policy actions and their impact, to make suggestions for changes
in policy and to promote transparency in the development of action plans,
in order to meet the multiple goals of both policy-makers and organic sector
stakeholders (ORGAPET Section A2 (this document) and
Section
C1 address this);
-
organisation:
action plan evaluations will be commissioned by the European Commission and/or
national/regional governments, with organic stakeholders as participants
and either internal administrators or external researchers/consultants
acting as evaluators. To achieve this ORGAPET needs to have an open and relatively
simplistic approach to evaluation (ORGAPET
Section
A5 addresses the planning of evaluations).
-
programme:
a typology of European and/or national action plans is needed regarding
action plan objectives, programmes and organisation, including? descriptions
of implicit or explicit programme theories (i.e. how and why are distinct goals
transformed to distinct outputs and how are outputs expected to influence
outcomes within given contexts?). A first suggestion is to distinguish between
action plans where programme theories focus a) mainly on market drivers, b)
mainly on public good drivers or c) balancing market and public good approaches.
This distinction will need further refinement on the basis of the variation
among national action plans, but the number of categories should be limited.
The distinction must be based on a characterisation of all objectives and
measures within an action plan in relation to programme theory; this is the
basis for defining outputs, outcome and impacts as well as evaluation criteria
(see ORGAPET Section A3).
-
public management: the focus is on the process of policy development and implementation,
in particular the description of actors participating and their activities.
The policy network analysis conducted as part of the EUCEEOFP project is relevant
in this context, focusing on institutional change involving the farming community,
agricultural policy, the food market and a potential institutional setting
across the three, including a distinction between organic and non-organic
actors (see ORGAPET Section B1 (process) and
Section B2 (synergy and coherence)).
-
results:
data (qualitative or quantitative) with respect to a limited number of
relevant indicators selected as basis for the evaluation (see ORGAPET Sections
C1,
C2,
C3
and C4). Results should cover:
a. output
= direct actions taken by public agencies and/or stakeholders in terms of
support, grants etc.;
b. outcome = effect on organic sector in terms of its objectives;
c. impact
= effect of changes in the organic sector on the (higher level) goals
of policy-makers
Similar
outputs and outcomes may be evaluated differently on the basis of the different
objectives and values of stakeholders.
-
explanations:
did the action plan lead to the expected results (i.e. is the programme theory
supported, rejected or influenced by intervening factors? This is important
for conclusions regarding future actions and would require insights and examples
from research and other sources (see ORGAPET
Section
D1).
-
assessment:
do the evaluation criteria and standards reflect the diversity of views held
among public agencies and organic stakeholders? On the basis of the action
plans and evaluations collected in WP3 and the WP2 workshops, a limited number
of themes regarding action plan objectives, corresponding evaluation criteria
and how they could be assessed from various viewpoints should be formulated.
The emphasis on formulating relatively few but different evaluation criteria
implies acceptance that they may be interpreted differently by various stakeholders
and public agencies. Strategies for dealing with different values among stakeholders
are required (see ORGAPET Section D1).
-
the
use of evaluations will be influenced by the choice of formative or
summative evaluations: formative evaluation implies that evaluators both analyse
the development and implementation of an action plan and intervene in it,
while summative evaluation implies a clear separation of analysis and use
of evaluation in terms of the timing of the evaluation and the involvement
of evaluators in decision-making.
Another aspect of use is the position held by policy makers and stakeholders on the evaluation as such, as well as on the recommendations.
Evaluation may thus be part of a strategy for using the results to involve
stakeholders as much as possible, whether in specific parts or all parts of the evaluation (which is one of the main
reasons for developing participative evaluation).
A2-5.2
Integration of stakeholders
This topic has
been identified as particularly important in the context of organic action
plans. Systematic assessment requires a scientific basis (including relevant
objective criteria and sound evidence) but, at the same time, needs to take
account of subjective values of stakeholders. These values may be determined
with reference to the published aims of the intervention but, in the case
of action plans, many different values or goals may be represented and they
are not necessarily prioritised, and they may reflect the values of governmental
institutions more strongly than they do the direct beneficiaries or other
stakeholders. In the absence of specific value statements, it becomes ambiguous
whose values are being promoted. There is therefore a need to include consideration
of these issues in any evaluation.
The approach chosen
in this project is to involve stakeholders and their values, but stakeholders
have different values that may depend on position in the production chain,
material position, culture and many other aspects. Hence the toolbox to be
developed needs to be able to grasp different sets of values – and any tool
may be interpreted differently by different stakeholders.
Vedung (1997) mentions
two different ways that stakeholders might be integrated in evaluations:
-
the north
American way (Vedung, 1997:
69) – the evaluator decides which stakeholders
to involve and in what capacity – this leads to a strongly differentiated
evaluation, where the evaluator decides how to weigh the views of each stakeholder.
This might involve evaluations where different stakeholders receive different
answers and hence receive different support by means of the evaluation;
-
the Swedish/Scandinavian
way (Vedung, 1997: 76) – the relevant stakeholders are selected by the promoter
of the evaluation and they are all involved directly in the execution of the
evaluation – servicing with manpower and data or just overseeing its
implementation. The aim is to serve the whole spectrum of relevant political
and material interests and to reach a consensus.
The importance
of values in any evaluation, the necessary inclusion of stakeholders in evaluating
organic action plans and the various ways of integrating stakeholders are one important aspect that ORGAPET needs to address. The
role and potential for integration of stakeholders is developed further in ORGAPET
Section A4 while
Section B3 addresses the evaluation of these processes.
A2-6 Conclusions
(with grateful acknowledgement to the Evalsed
Golden Rules!)
Policy evaluation
involves the systematic gathering of information and assessment of a programme
according to specific criteria in order to make judgements about the value
of the programme, reducing uncertainty in decision-making concerning future
actions. The assessment of value may relate to the goals of more than one
specific interest group, including policy-makers, beneficiaries and third
parties, and they may fulfil a range of purposes, from financial control and
accountability to intervention improvement and knowledge advancement.
Evaluation takes place in order to improve programmes, not to undertake
evaluations for their own sake.
Aligning the time cycles of evaluations with the time cycles
of programmes and policies can help ensure evaluations make
their maximum contribution. It is better to deliver an
incomplete or imperfect evaluation on time than to achieve a
10% improvement in evaluation quality and miss the window of
opportunity, when policy-makers and programme managers can
use evaluation results.
Different stakeholders (e.g. policy-makers, professionals,
managers and citizens) have different expectations of
evaluation. If a major stakeholder interest is ignored, this
is likely to weaken an evaluation, either because it will be
poorly designed and/or because its results will lack
credibility. Involving policy-makers and those responsible
for programmes will ensure they take results seriously.
Evaluations must be fully integrated into programme
planning and management. Policy makers should consider
expressing their goals in terms of something measurable and
to ensure availability of such measures. Programme managers need to think of
evaluation as a resource: a source of feedback, a tool for
improving performance, an early warning of problems (and
solutions) and a way of systematising knowledge. Evaluation
is not simply an external imposition. Equally, evaluators
need to take on board the concerns of programme managers
(and their partnerships) and try to take seriously their
need for answers to difficult questions.
Getting good work from the diverse groups which make up
the contemporary evaluation professional community needs
bridge-building and team-building. Bridges need to be built
at national, regional and European levels between the
different traditions of evaluators, social scientists,
economists, policy analysts and management consultants. At a
micro-level, the priority is integration and the combination
of different skills and competences within evaluation teams.
Evaluation is not only about looking back to rate success
or failure and allocate blame. It has a contribution to make
at every stage in the programme cycle. In particular,
evaluation can strengthen programmes
at the earliest stage by helping to unpick intervention logics and reveal
weaknesses in programme design, allowing remedial action to
be taken early.
It is not appropriate to gather large quantities of data
in the belief that these will eventually provide answers to
all evaluation questions. Data dredging is nearly always
inefficient. This does not mean that data systems are not
essential: they must be put in place at an early stage.
However, by being clear about assumptions, by drawing on
available theory and being clear about the type of
evaluation that is needed, evaluations can be more focused
and offer a higher yield for the resources expended.
The policy context is an important framework within which
evaluations need to be located. Of course, policies change
or are being constantly restated in different terms and with
subtly changing priorities. However, it is always necessary
to keep one eye on policy debates and decisions in order
to ensure that evaluations are sensitised to policy
priorities. The broader criteria that need to be designed into evaluations usually derive from the wider policy
framework.
While all stakeholders are important, particular
prominence needs to be given to one important and often
neglected group: the intended beneficiaries of the programme
interventions. Incorporating the voice of these intended
beneficiaries in evaluations implies more than asking their
opinions. It also implies incorporating their criteria and
judgements into an evaluation and accepting that their
experience and benefits are the justification for programme
interventions. This is consistent with the logic of
bottom-up, participative and decentralised approaches that
are so common now in socio-economic development. It is also
why responsive and participatory methods have become such an
important part of the evaluators toolkit.
Organic action plan and organic farming policy
evaluations need to take account also of the complex systems and multiple
objective nature of the organic approach, with due attention to synergies
and conflicts between objectives, and the different emphases that will be
placed on these by different stakeholders.
Finally, we live in an imperfect world where resources
are limited, administrators are not always efficient, co-ordination is
imperfect, knowledge is patchy and data is often not available. It is
nonetheless worth taking small steps, working with what is available and
increasing, even marginally, the efficiency and legitimacy of organic action
plans and policies. Even modest outputs can make a big difference especially
when this is set within a longer-term vision to build capacity and allow for
more ambitious evaluations in the future.
A2-7
References
Dabbert, S.; A. M. Häring and R. Zanoli (2004) Organic Farming: Policies and Prospects.
Zed Books, London.
EC
(1999) Evaluating Socio-economic Programmes. MEANS Collection Vols.
1-6. Office for Official Publications of the European Communities, Luxembourg.
EC (2004)
European
Action Plan for Organic Food and Farming, COM(2004)415 final, Commission
of the European Communities, Brussels.
EEA (2005)
IRENA:
Indicator Fact Sheets and Data Sheets. CIRCA v3.2 [on-line],
European Environment Agency, Copenhagen.
Häring,
A. M. (2003)
An Interactive Approach to Policy Impact Assessment for Organic
Farms in Europe.
Organic Farming in Europe: Economics and Policy,
Vol. 10, University of Hohenheim, Stuttgart.
Häring, A. M.,
S. Dabbert, J. Aurbacher, B. Bichler, C. Eichert, D. Gambelli, N. Lampkin,
F. Offermann, S. Olmos, J. Tuson and R. Zanoli (2004)
Organic Farming and
Measures of European Agricultural Policy.
Organic Farming in Europe: Economics and Policy, Vol. 11, University of Hohenheim, Stuttgart.
Kromrey, H. (2001)
Evaluation - einvielschichtiges Konzept. Begriffe und Methodik von Evaluierung
und Evaluationsforschung. Empfehlungen fuer die Praxis. In: SuB Sozialwissenschaften
und Berufspraxis. 24(2):105-132.
Lampkin,
N., C. Foster, S. Padel and P. Midmore (1999)
The Policy and Regulatory
Environment for Organic Farming in Europe.
Organic Farming in Europe: Economics and Policy, Vols. 1 and 2, University of Hohenheim, Stuttgart.
Mertens, D. (1998)
Research methods in education and psychology: Integrating diversity with quantitative
and qualitative approaches. Sage, Thousand Oaks.
Stolze, M. and
N. Lampkin (2006)
European organic farming policies: an overview.
In: Proceedings of the European Joint Organic Congress, 30-31 May 2006.
(Andreasen
CB et al. (eds.)), Danish Research Centre for Organic Food and Farming, Foulum.
Stockmann, R. (2004)
Was ist eine gute Evaluation? CEVAL, Arbeitspapier Nr. 9.
Centrum für Evaluation, Saarbrücken.
Tavistock Institute with GHK and
IRS (2003) The Evaluation of Socio-Economic Development: The Guide,
Tavistock Institute, London [on-line], http://www.evalsed.com/
accessed 21/08/2006. (NB. The original website accessed is now superseded by
the EU Commission's Evalsed).
Vedung, E. (1997)
Public Policy and Program Evaluation. Transaction Publishers, New Brunswick,
New Jersey.
A2-8
Annexes
Annex A2-1 Impact Assessment Guidelines (EU Commission) (SEC(2005) 791)