
EN: Methodological bases and approach
The EU has developed and formalised a methodology for evaluating its external assistance in which priority is given to results and impacts. This page presents the methodological guidelines designed to facilitate the move towards an evaluation practice focused on programmes and strategies in EC Development Cooperation.
-
This first chapter presents the methodological bases of the evaluation approach:
Methodological bases or what, when and why to evaluate and who is a part of the evaluation:
Subject, timing of the evaluation and intervention cycle, utilisation and roles, as well as specificity of the evaluation compared to monitoring and auditing are described in details.
Evaluation methods or how to evaluate
Intervention strategy, evaluation questions, judgement references, methodological design, data collection, analysis, judgement and quality assurance are presented and explained.
Evaluation tools
The rationale and the use of each evaluation tool are systematically presented.
The second and third chapters of these guidelines contain respectively:
-
The detailed guidelines for Project and programme evaluations
-
The detailed guidelines for Geographic, thematic and other evaluation
The way development aid is provided is constantly evolving. To reflect this changing context, DEVCO develops specific methodological tools, often in cooperation with international institutions such as the OECD and the World Bank.
In this context, three specific methodological approaches to evaluate budget support programmes, capacity development interventions and gender as a cross cutting issue have been developed:
Disseminating Evaluations
Creative Communications for Evaluation Dissemination
This space is created to help you promote evaluation findings in a way that is attractive and communicative by exploring different options, guides and good practice examples at the EU level and externally.
An INTPA/ESS study on EU evaluation dissemination suggests that the most innovative ways to communicate evaluation results, besides organising seminars or webinars, include infographics, briefs, videos, blogs and podcasts, among others.
Please find below some ‘Good Practice Examples’ from different EU offices, findings of the INTPA/ESS study in a dynamic report, and How-To Guides on creating different Dissemination Knowledge Products.
Background of the study: In the first semester of 2020, the Evaluation Support Service of DG DEVCO conducted a study to examine and document the current situation regarding how the knowledge generated from evaluation of the international development and co-operation interventions is being translated and disseminated to a wide variety of stakeholders to ensure ‘evaluations influence change’, in line with DG DEVCO’s Evaluation Policy (Evaluation Matters, 2014). Related to this, the study captures examples of good dissemination practice within DG DEVCO and non-EU institutions such as the Asian Development Bank, African Development Bank, International Fund for Agricultural Development, World Food Programme, Global Environment Facility, Green Climate Fund, InsighShare, UN Women, UNICEF, UNFPA, Y Care International and Oxfam.
A detailed report on the findings of the study and its recommendations is also available on request. Please write to the ESS team to have a copy of the full report.
Further support: Should you need any support in the management of your evaluations, including advice on ways to best disseminate their results, the ESS helpdesk remains at your full disposal and can be reached at helpdesk@evaluationsupport.eu. Please do not hesitate to contact them and share your examples of evaluation dissemination: we would be happy to publish them in this C4D space!
Disseminating knowledge generated by evaluations. Some examples from the EU.
This space has been created to showcase good-practice examples of Evaluation Dissemination Knowledge Products (EDKPs) by different EU Delegations and Headquarters.
These knowledge products are produced to further encourage the use of knowledge generated through evaluations by key decision-makers in the EU and overseas, and to support our aim to be more transparent and accountable to the wider EU public. These knowledge products are created in easily digestible and simple language and the full reports can be provided upon request.
For EU staff members: we would like to invite you to help us populate this section by sharing your evaluation dissemination knowledge products with the Helpdesk@evaluationsupport.eu
Examples of these knowledge products can be accessed by clicking on the different categories shown below.
Briefs
Here you will find examples of evaluation reports that were summarised in briefs, among which is the multi-country study “La voix des jeunes au Sahel”, with a global brief and six country-specific editions. Full evaluation reports can be provided upon request.
Please access the briefs by clicking on the thumbnails below.
Infographics
Here you will find examples of evaluation reports that were summarised in an infographic. Full evaluation reports can be provided upon request.
Please access the infographics by clicking on the thumbnails below.
Evaluation of the EU's regional development cooperation with Latin America (2009-2017) (EN) | ||
|
Reader-friendly reports
Here you will find examples of evaluation reports that were edited in a communicative and reader-friendly manner. Full evaluation reports can be provided upon request.
Please access the reports by clicking on the thumbnails below.
Ex-post evaluation of the CBTPSD project in Malawi. (EN) | Evaluation of the EU SRSP's PEACE Programme. (EN) | Analysis of the PARAQ Programme. (FR) |
|
Video
Here you will find examples of EU evaluations of which the findings have been summarised in a short, communicative video. Full evaluation reports can be provided upon request.
Please access the videos by clicking on the thumbnails below.
An infrastructure project in Ivory Coast, 2019 (FR) | ||
|
How-to Guides on Evaluation Dissemination
Through analysing a large number of evaluation dissemination products produced recently and information collated from interviews with relevant institutions, this study produced a series of five “How-to guidelines” on the most frequently used dissemination products (Videos, Infographics, Evaluation Briefs, Podcasts and Blogs) to improve and support evaluation dissemination practice. These How-to-Guides can be accessed by clicking on the following:
How-to Guide on Evaluation Dissemination Seminars
A dissemination seminar is a way to share evaluation findings during an online or face-to-face meeting with key stakeholders or a wider audience. It is a good way to get feedback before the final report is written and ensure that findings are supported by relevant parties.
These guidelines support your work with practical tips on organising a seminar to communicate the learning from your evaluation.
You may access the document by clicking on the image below.
How-to guide on producing blogs for evaluation dissemination
A blog provides an accessible and user-friendly format to communicate your evaluation results. You can share findings from the evaluation as a whole, or specific aspects such as key learning, recommendations or experience of using different methods and tools. All of these may incite your audience to read the full report.
The DEVCO/ESS study to understand the different ways in which evaluation results are communicated reveals that 7 out of the 18 institutions consulted use blogs to disseminate their evaluations. Blogs do not require particular preparation in terms of budgeting and planning, but there are some writing principles to respect.
These guidelines provide you with practical tips, examples, and references to additional resources to get you started on evaluation blogs. Click on the document below and let us know how you get on with evaluating your evaluation results through blog posts.
You may access the document by clicking on the image below.
How-to guide on producing briefs for evaluation dissemination
Evaluation brief is a way to share key findings on a particular thematic area through an attractive 2-4 page summary. The DEVCO/ESS study to understand the different ways in which evaluation results are communicated reveals that 9 out of the 18 institutions consulted use briefs to disseminate their evaluations. They are relatively easy to make and do not require important budget or preparation. Critical, though, is to make sure you provide enough context and adapt the language for your audience to understand the summary.
These guidelines support your work with practical tips, examples, and references to additional resources to communicate the learning from your evaluation.
You may access the document by clicking on the image below.
How-to guide on producing infographics for evaluation dissemination
Infographics provide an increasingly popular way of sharing your evaluation findings. The DEVCO/ESS study to understand the different ways in which evaluation results are communicated reveals that 11 out of the 18 institutions consulted use infographics to disseminate their evaluations.
Infographics are relatively low cost and do not require particular planning at the start of the evaluation process. What they do require is the definition of clear key messages that convey a story.
These guidelines provide you with practical tips, examples, and references to additional resources to communicate the
learning from your evaluation.
You may access the document by clicking on the image below.
How-to guide on producing podcasts for evaluation dissemination
Podcasts are an increasingly popular way to communicate evaluation results. The DEVCO/ESS study to understand the different ways in which evaluation results are communicated reveals that 6 out of the 18 institutions consulted use podcasts to disseminate their evaluations. Podcasts can be engaging; they are easy to share and provide convenient access to knowledge.
These guidelines provide you with practical tips, examples, and references to additional resources to get you started on evaluation podcasts. Have a look and let us know how you get on with promoting your evaluation results through the power Podcasts.
You may access the document by clicking on the image below.
How-to guide on producing video for evaluation dissemination
Evaluation Video for Dissemination
Video is an engaging way of sharing your evaluation findings. The DEVCO/ESS study to
understand the different ways in which evaluation results are communicated reveals
that 12 out of the 17 institutions consulted use video to disseminate their evaluations. It
requires preparation in terms of budgeting and planning, as well as in terms of reflecting on
the story you want to tell. Whose story will it be? Is it an account of programme success
or an evaluative journey? These guidelines provide you with practical tips, examples, and
references to additional resources to communicate the learning from your evaluation.
Evaluation Dissemination - Dynamic report
Here you can download and read the dynamic interactive PDF report that has been produced to disseminate the findings of the study on evaluation dissemination at DG DEVCO, EU Delegations and non-EU organisations around the world such as the Asian Development Bank, African Development Bank, International Fund for Agricultural Development, World Food Programme, Global Environment Facility, Green Climate Fund, UN Women, UNICEF, UNFPA, Y Care International and Oxfam.
Please note that this Dynamic Interactive Report has been produced from the ESS study, ‘A Report on the current practices of evaluation dissemination within the European Commission DG INTPA, and other non-EU institutions’. This report can be accessed via this weblink.
Should you need any support in the management of your evaluations, including advice on ways to best disseminate their results, the ESS helpdesk remains at your full disposal and can be reached at helpdesk@evaluationsupport.eu. Please do not hesitate to contact them and share your examples of evaluation dissemination: we would be happy to publish them in this C4D space!
Methodological bases
This section presents the following evaluation's methodological bases :
- The subject or what is evaluated (scope)
- The timing or when to evaluate (evaluation and intervention cycle)
- The utilisation or why to evaluate (users, type of use)
- The roles or who is part of the evaluation (manager, reference group, external team, stakeholders)
- Specificity of the evaluation compared to monitoring and auditing
Subject
.
This section is structured as follows:
- Evaluation scope
- Evaluation of projects, programmes and strategies
- Sectors, themes and cross-cutting issues
.
Timing of the evaluation and intervention cycle
.
This section is structured as follows:
.
TIMING OF THE EVALUATION |
. What does this mean?An evaluation can be performed before, during or after the evaluated intervention. Depending on the timing, its purpose and use will differ.
What is the point?
|
Different times of an evaluationEx anteAn ex ante evaluation is performed before adopting or implementing the intervention. It gives support to the intervention design and contributes to ensuring the design quality. It is concerned with the following points:
It aims at having a direct influence on the decisions upstream from the implementation, to the extent that it transposes lessons from past experiences into the framework of the new intervention. Mid-term or finalAn evaluation during or at the end of the implementation is intended to draw lessons from the first years of the intervention implementation and to adjust the contents of the ongoing intervention in relation to realities in the field and/or contextual developments. It often includes a report on outputs and an analysis of the first results and impacts achieved. It aims at improving the intervention under way and its conclusions may be supported by observations in the field. Ex postThe ex post evaluation is performed right after or a long time after completion of implementation. It is mainly concerned with checking achieved impacts, identifying and judging unexpected impacts and assessing the sustainability of the intervention's benefits. It enables to detect the real changes in the field and, if the changes occur soon enough, they can be analysed to estimate those that are attributable to the intervention. The ex post evaluation often aims to report to the institutions that have allocated the resources. Likewise, it helps to transfer acquired experiences to other countries or sectors. |
Recommendations
. |
EVALUATION AND THE INTERVENTION CYCLE |
. What does this mean?The evaluation can be scheduled:
Is there a best time to evaluate or should three evaluations be performed for each cycle? What is the purpose?Many interventions are characterised by successive cycles and show a relative continuity between cycles. A new cycle can start before the intervention of the previous cycle has yielded all its effects. In light of this, several cycles need to be considered when choosing the timing of the evaluation, in order to:
How to take the cycles into account?For defining the purposes of the evaluation:
For choosing evaluation questions
RecommendationsWhere possible, ensure that the evaluation ends about a year before the end of the cycle. At that stage it will still be possible and useful to adjust the implementation because it always lasts for a year or two after the end of the cycle. It will also be possible to assist in planning the intervention strategy of the following cycle. Finally, it is not too late to perform a sound analysis of the impacts of the previous cycle's intervention. |
Utilisation
.
This section is structured as follows:
- Users of an evaluation
- Types of use
- Evaluation and decision making
- Evaluation and knowledge transfer
- Dissemination of the evaluation
- Documents presenting the evaluation
- Presenting an evaluation in a seminar
.
USERS OF AN EVALUATION |
. Who do we mean?Evaluation is intended for a variety of users:
What is the purpose?To optimise the usefulness of the evaluation for the various partners, and especially:
|
Policy-makers and designersPolicy-makers and designers use the evaluation to prepare the launching of new interventions, the reform of existing interventions, the choice of strategic orientations, and decisions on allocation of budgetary, human and other resources, etc. They need information that:
They are interested in strategic issues, external coherence and global impacts, which constitute the ultimate goal of the intervention. |
Managers, partners and operatorsThe managers are responsible for the implementation and monitoring of the intervention, from headquarters to the field. The actors closest to the public are the operators. Field level operators may either belong to the EC or to partner organisations sharing the responsibility of implementation. They use evaluation findings about the results of their action as a feedback. They need information that arrives as early as possible to remedy problems or validate changes. They are able to interpret complex and technical messages. They are interested in the direct results of the intervention, in the needs and behaviour of the targeted group, and in interactions between partners. |
Other actorsThe institutions that funded an intervention expect accountability. This applies to Parliament or the Council of Ministers, but also to all the co-funders. The taxpayers and citizens are also addressees of an evaluation. The public authorities that conduct related or similar interventions are potential users of the evaluation, especially in the form of transfer of lessons learned. The same applies to the expert networks concerned by the intervention. Finally, an evaluation is likely to be used by the actors in civil society, especially those representing the interests of the beneficiary groups. |
Recommendations
. |
TYPES OF USE |
. What does this mean?An evaluation can be used:
These three types of use are not mutually exclusive. It is necessary to understand in order to judge, and to judge in order to decide. What is the purpose?To optimise the usefulness of an evaluation, that is:
Assisting decision makingThe evaluation may be undertaken for the benefit of those who have to decide or to negotiate an adjustment or reform of the evaluated intervention. In that case it is used to adjust the implementation, to design the next intervention cycle or to redefine political orientations. To facilitate this type of use, also called feedback, the evaluation questions must be asked in relation to the decision-makers' expectations and to their planned decision-making agenda at the time the report is submitted. The evaluation can aid decision-making in two different ways:
Evaluations may assist decision-making in different ways, depending on the context of the decision:
Assisting the formulation of judgementsThe evaluation may help users to shape their opinion on the merits of the intervention. The formulation of an overall assessment is particularly useful for accountability purposes. In this case, the evaluation examines the merits of the intervention in relation to the different points of view (summative evaluation). It answers questions that are important for the funding institutions. The report is accessible to the general public. The independence of the evaluation and the transparency of the judgement are highlighted. In this instance, particular attention is paid to the definition of judgement criteria (also called "reasoned assessment criteria"). Yet the judgement itself is definite only when the final report is submitted and its conclusions are discussed. Using the evaluation for accountability purposes therefore means having to wait for the end of the process. Knowing and understandingApart from assisting in making decisions and formulating judgements, which are the two main forms of use, the evaluation may also enable users to learn from the intervention, to better understand what works and what does not, and to accumulate knowledge.. Indirectly, it contributes to transferring knowledge thereby acquired, to the benefit of professional networks that may not have a direct link with the evaluated intervention. Unlike feedback, which directly concerns those responsible for the evaluated intervention, the transfer of lessons is an indirect process that takes place through networks of experts both within and outside the European Commission. Capitalising on knowledge often starts during the evaluation process, through the experts who belong to the evaluation team or reference group. However, the transfer of lessons learnt may only occur after the final report has been delivered. A key step in this perspective is presentation of the evaluation in specialised networks, in the form of seminars or technical articles. |
RecommendationsFrom the outset, the evaluation manager should prioritise one or more types of use, and then optimise the evaluation process in order to make it as "user friendly" as possible, for instance in adjusting:
. |
EVALUATION AND DECISION MAKING |
. What is this about?Evaluation provides feedback and thus facilitates decision-making, for instance in adjusting the implementation of the intervention, designing the next cycle, or helping to redefine political priorities. In this context the formulation and follow-up of recommendations are key steps in the process. What is the purpose?Facilitating future decision-making is generally the main type of use of evaluation. Decision-makers' needs have, therefore, to be taken into account throughout the evaluation process to increase the chances of the evaluation being useful to them. |
What is the link between evaluation and decision-making?Some evaluations are designed primarily to provide information for management decisions or for reform of the evaluated intervention. They are intended for operational actors in the field, management services, and the authorities responsible for the intervention or their partners. In this perspective, mid-term evaluation is to be preferred and careful attention needs to be paid to the formulation and follow-up of recommendations. These evaluations are referred to as formative. Other evaluations are designed primarily to learn lessons from the experience and to serve decision-making in other contexts. In this perspective, ex post evaluation is the most appropriate and careful attention must be paid to the formulation and transferability of the lessons learned. These evaluations are referred to as summative. Yet one has to be realistic: decision-makers will not necessarily follow the recommendations and lessons. The decision-making process almost always involves many actors and multiple factors, evaluation being only one of them. |
Advice for performing a decision-making oriented evaluation
. |
EVALUATION AND KNOWLEDGE TRANSFER |
. What is this about?Evaluation is a learning-intensive exercise in so far as lessons learned from experience can be capitalised on, and knowledge acquired can be transferred and reused by actors who have no direct link with the evaluated intervention. In this context the identification of good practices and transferable lessons is a key step. What is the purpose of knowledge transfer?
|
Draft the report to promote the accumulation of knowledgeThe only pages of a voluminous evaluation report that are oriented towards knowledge transfer are those concerning good practices and lessons learned. It is therefore important to identify them and to draft them with that in mind. Identifying a good practice means judging that it produced the expected effects in a particularly effective, efficient or sustainable way. Identifying a transferable lesson means judging that a given practice generally succeeds (or fails) to produce the expected effects in a given context. The summary of the evaluation report highlights the main lessons learned with a view to facilitating their accumulation and transfer. For each lesson, references to the body of the report serve to specify:
Mobilise available knowledgeCertain public institutions or professional networks accumulate knowledge in Intranet or Internet databases. Knowledge may also be accumulated informally through expert networks that build on lessons learned in a particular sector or on a particular topic. In the latter case, the members of the network try to validate the knowledge before capitalising on it, through meta-evaluations or expert panels. Not only does evaluation contribute towards the accumulation of knowledge, it also facilitates the circulation of the acquired knowledge. The evaluation team mobilises available expertise and documentation to provide an initial partial answer to the questions asked. If the transferable lessons have been learned through other evaluations and properly capitalised, the evaluation team identifies and reuses them. . |
DISSEMINATION OF THE EVALUATION |
. What does this mean?Dissemination concerns the final evaluation report, as well as all other means of publicising the conclusions, the lessons learned and the recommendations. Dissemination activities target the services of the Commission, European Institutions, external partners, networks of experts, the media and the wider public. What is the purpose?The dissemination process promotes the use of the evaluation if it is done well. It serves to:
|
Which measures need to be taken?
|
Which channels for dissemination?The evaluation report is disseminated on the Internet and is thus accessible to all audiences. More active dissemination is also undertaken for specific audiences:
|
Recommendations
. |
DOCUMENTS PRESENTING THE EVALUATION |
. What does this mean?Since full-length evaluation reports are available on the Internet, it is advisable to disseminate one or more shorter documents that are suited to the different audiences and present the evaluation in an attractive and accessible way. What is the purpose?These documents are intended to:
Formats and uses of the different documentsThe executive summary is one of the starting points for drafting presentations. It is usually three pages long and no particular audience is targeted in the content. Beyond the executive summary, various documents are or may be produced:
Advice for drafting a presentation document- Choose the orientation of the document
- Describe the evaluated intervention
- Describe the evaluation
- Main messages
- MethodDescribe the methodological design in a non-technical way and make a statement about the soundness of the main messages. If one aspect of the method is of particular interest to the targeted audience, add a more technical boxed-in section. Recommendations
. |
PRESENTING AN EVALUATION IN A SEMINAR |
. What is the purpose?In an environment with an over-abundance of written information, an effective way of reaching potential users consists in presenting the evaluation results in meetings or seminars. An oral presentation of the evaluation helps to:
Practical adviceA short presentation (10 to 20 minutes) is enough for the main points but more time needs to be left for questions (20 to 40 minutes). The presentation covers the following points: - The evaluated intervention
- The evaluation
- Messages resulting from the evaluationA few particularly important messages from the point of view of the people participating in the meeting or seminar (key data, findings, conclusions, transferable lessons and/or recommendations). - Strengths and weaknesses of the messagesExplanation of the methodology employed and reasons for which a particular message is sound (valid) or fragile. Recommendations for using messages if they are fragile. Recommendations
|
Roles
.
This section is structured as follows:
- The distribution of roles
- The evaluation manager
- The reference group
- The external evaluation team
- The stakeholders
.
This Evaluation methodological approach is generally intended to Evaluation managers and to external evaluation teams. This is why we have visually distingueshed their respective tasks, as follows : | |||||||||||||||||||||||||||||||||
|
|||||||||||||||||||||||||||||||||
These icons are used in main sections of guidelines for Project and programme evaluations and Strategic, geographic and other complex evaluations. | |||||||||||||||||||||||||||||||||
. |
|||||||||||||||||||||||||||||||||
THE DISTRIBUTION OF ROLES |
|||||||||||||||||||||||||||||||||
. What does this mean?The three key players are the evaluation manager, the reference group and the external evaluation team. The distribution of roles is specific to each phase and each stage of the evaluation. All the actors in the evaluation process have to know what their role and responsibilities are, for example: Who is responsible for the recommendations? Who has to assess the quality of the report? What is the purpose?
Who does what?The three key players in an evaluation are the evaluation manager, the reference group and the external evaluation team. Their roles differ, depending on the following phases and steps: - Decision to evaluateThe decision is taken by the authorities responsible for the intervention and is reported in a document such as:
The decision refers to an evaluation framework that defines the rules of the game. If necessary, it completes this framework. - Preparation (phase 0)Constitution of the reference group:
Drafting of the Terms of Reference:
Selection or approval of the external evaluation team:
- Desk (1)Clarification of the intervention logic:
Choice and drafting of evaluation questions:
Definition of the judgement criteria (also called reasoned assessment):
Choice of indicators:
Design of the method (data sources, tools and work programme):
Collecting available data at the head office of the Commission and/or the Delegation (s):
- Field (2)Data collection:
- Synthesis (3)Report, including the conclusions and recommendations:
Verification of the quality of the report:
- Dissemination and follow up of recommendations (4)Internal and external dissemination:
Use of the conclusions and recommendations:
. |
|||||||||||||||||||||||||||||||||
THE EVALUATION MANAGER |
|||||||||||||||||||||||||||||||||
. Who is this?The evaluation manager is a member of the service in charge of the evaluation. He/she manages the entire evaluation process, from A to Z on behalf of the commissioning body. Why appoint an evaluation manager?
What does he / she do?
Recommendations
. |
|||||||||||||||||||||||||||||||||
THE REFERENCE GROUP |
|||||||||||||||||||||||||||||||||
. What is this?A small group gathering the services concerned by the intervention and possibly other stakeholders. One or more experts may also be invited to participate. The reference group is an interface between the evaluation manager and the external evaluation team. Its members help the external evaluation team to identify data sources and to access them. They validate the evaluation questions, and discuss the conclusions and recommendations. The reference group allows the variety of points of view on the evaluated intervention to be expressed. What is its role?Reference group members:
Whom to invite and how?Different types of actor can be invited to participate in the reference group. If the evaluation is managed at headquarters, membership includes:
If the evaluation is managed in the partner country, membership may extend to:
Experience has shown that it is preferable for the group to be limited in size (10 to 12 members) if it is to function effectively. After identifying the services, institutions or organisations to invite to the reference group, the head of the service managing the evaluation sends them an invitation to:
|
|||||||||||||||||||||||||||||||||
When should it be asked to meet and how should it be involved?The reference group is generally involved in the following way:
|
|||||||||||||||||||||||||||||||||
Opening the evaluation processIn addition to inputs made by the reference group, the evaluation may benefit from a wider variety of views through other channels like:
|
Experts invited to the reference group |
What does this mean?One or more members of the reference group may be invited as a resource person, for their expertise. Why invite them?
Who should be invited?
|
Charging a reference group meeting |
What to do?
|
What to do in case of disagreement?In case of lasting disagreement on the evaluation method, the service in charge of the evaluation arbitrates. If necessary, the decision is taken by the head of the Unit. . |
THE EXTERNAL EVALUATION TEAM |
. What does this mean?The external evaluation team is responsible for data collection and analyses and for the evaluation report, including the formulation of judgements (also called reasoned assessment) and the drafting of conclusions and recommendations. It interacts with the reference group and the evaluation manager. It provides evaluation services to the commissioning body under contract. The members of the external evaluation team are independent of the organisations that participated in the design and implementation of the evaluated intervention. They belong to national or international research organisations or consultancy firms, or are experts working independently. Where possible, some members of the team are from the country or countries targeted by the evaluated intervention. Why contract an external evaluation team?
|
How to select the team?The external evaluation team is selected on the basis of proposals via different procedures, e.g. pluri-annual evaluation service contract or specific call for tenders. The evaluation manager engages the team after consulting members of the reference group. The following criteria are used as guidelines:
The evaluation manager sets the relative weight of criteria and specifies it in the terms of reference. Recommendations
|
Evaluation capacity |
What does this mean?Having evaluation capacity implies sufficient knowledge and experience to:
Evaluation capacity is required both:
How to assess professional experience?An evaluation professional meets most of the following criteria:
. |
THE STAKEHOLDERS |
. What does this mean?The stakeholders are individuals, groups or organisations that have responsibilities and/or direct or indirect interests in the evaluated intervention. They may be affected by the intervention or not. Why identify them?
How to take them into account?
Recommendations
|
Specificity of the evaluation compared to monitoring and auditing
.
This section is structured as follows:
- What does this mean?
- What is the purpose?
- What distinguishes the three exercises?
- How to ensure complementarity between the three exercises?
- Recommendations and clarification
.
What does this mean? |
. Monitoring, auditing and evaluation are all exercises in sound public management. They are similar and complementary but cover different issues. All three contain value judgements but the judgement references are not the same. . |
What is the purpose? |
.
. |
What distinguishes the three exercises? |
. Subject of study
Judgement criteria
Time scale
Use
The required professional qualitiesEach exercise requires an appropriate professional capacity and distinct core competencies:
. |
How to ensure complementarity between the three exercises? |
. A good monitoring system considerably facilitates the evaluation. When an evaluation deals with intended effects, the first thing to do is to check whether these effects have been monitored, which indicators have been used, and whether baseline data have been recorded. In turn, evaluation reports often include suggestions for improving the monitoring system. Auditing and monitoring may detect apparent success or failure and recommend that the evaluation deepens the causes and effects analysis, clarifies the judgment criteria, or validates a good practice. An evaluation can show that the right application of a legal or professional standard is a factor of either effectiveness or ineffectiveness, and therefore provides feedback into auditing or monitoring practices. . |
Recommendations and clarification |
.
|
Evaluation methods
This section is structured as follows :
Intervention strategy
.
This section is structured as follows:
.
RATIONALE OF THE INTERVENTION |
. What is this?The rationale of an intervention is to address the needs, problems or issues that are considered to be priorities in a given context, and that cannot be addressed more effectively in another way. It is in the programming stage that the rationale of an intervention must be justified. At the evaluation stage it is enough to note the main points or to redefine them if the programming documents lack precision. The evaluation identifies the reasons for which the priorities have been chosen, for example: priorities of the policy in which the intervention takes place, urgency of the needs to be satisfied, and comparative advantages compared to alternative options. What is the purpose?
How to clarify it?. - Reviewing the problems and responsesThe evaluation report succinctly sets out the following:
The evaluation team searches for this information in the official documents that instituted the intervention and in the preparatory design work. It completes its understanding with interviews with the key informants. Certain evaluation questions may concern problems addressed by the intervention, for example:
- Reviewing the rationaleWhere relevant, the evaluation report highlights the following:
The evaluation team looks for this information in the official documents that instituted the intervention. It completes its understanding of the rationale by interviews with key informants. Certain evaluation questions may concern the rationale of the intervention, for example:
Recommendations
|
. |
INTERVENTION LOGIC |
. What is this?All the activities and expected effects (outputs, results and impacts) of an intervention, as well as the assumptions that explain how the activities will lead to the effects in the context of the intervention. The intervention logic may be "faithful" to the programming documents and to the documents establishing the policy to which the intervention is related. In this case, the expected effects are inferred from the stated objectives in the official documents. When the intervention logic is reconstructed during the evaluation, implicitly expected effects that were not mentioned in the initial documents may be taken into account. The fact that this is no longer a "faithful logic" must then be mentioned. The "faithful" approach is relevant when the objectives are expressed precisely and in a verifiable way. The other option is preferable if objectives are too vague or ambiguous. The intervention logic often evolves over time. In such cases, it is justified to reconstruct it several times, for successive periods. The intervention logic is a useful simplification of reality, but one has to bear in mind the complexity of the real world. In addition to the reconstruction of the intervention logic, it is useful to identify the main external factors that condition or limit the implementation and effects. One also has to remember that real causal explanations are often more complex than initial assumptions. What is the purpose?
|
How can it be reconstructed?- Collect and analyse reference documents:
|
The most common presentations. - Logical frameworkThis technique consists in producing a presentation in the form of a matrix that specifies the objectives (first column), how they should be verified (second column: indicators, and third column: sources), and the main assumptions on the external factors (fourth column). - Objectives diagramThis technique consists in the identification of officially stated objectives and a graphical presentation of the logical relations between objectives, from the most operational to the most global. The intervention logic is represented in the form of boxes and arrows. - Diagram of expected effectsThis technique is similar to the diagram of objectives since it also builds upon officially stated objectives. However, the objectives are translated into expected effects before being presented as a diagram. By translating objectives into expected effects, more concrete and easily verifiable concepts can be worked on. |
Recommendations
|
. |
OBJECTIVES |
. What are these?Objectives are the initial and formal statements on what is to be achieved by the intervention. They may relate to outputs, direct results or impacts. Objectives may be expressed in a verifiable way, which specifies clearly the expected achievements. They may be expressed in more political or rhetoric way, making difficult to connect them to a precise achievement. The evaluation process may uncover implicit objectives, i.e. cause-and-effect assumptions that necessarily explain how an objective may be achieved. What is the point?Identifying and clarifying objectives is a necessary step towards:
How to clarify them?- Examining the clarity of objectivesThe evaluation identifies the objectives and, for each one, sets out the following:
The evaluation team searches for the objectives in the official documents that instituted the intervention and in the documents framing it at a higher political or strategic level. When the statement of an objective contains ambiguities, the evaluation team proposes clarifications based on interviews with key informants. - Examining the hierarchy of objectivesThe evaluation ranks the objectives of the intervention:
The reconstruction of the intervention logic helps to differentiate the levels of objectives as well as the effects, direct finding s and impacts corresponding to them. Some evaluation questions may concern the articulation of objectives, e.g.:
|
RecommendationsIf the intervention is framed by numerous documents on a political or strategic level, or if these documents also contain many objectives, the evaluation team is faced with an over-abundance of objectives. Rather than presenting them all, the evaluation team underlines this fact and concentrates on the most important ones in the context of the evaluated intervention. There is a continuum between more or less explicit objectives:
All these elements warrant explanation in the report, but the status of each one must be specified. |
. |
RELATED POLICIES |
. What does this mean?The idea is to identify the main related interventions and to situate the evaluated intervention in relation to them. What is the purpose?
How to proceed?The evaluation report briefly sets out the following:
This analysis offers a brief insight of the main related policies, if possible with their starting and finishing dates, the tools used, the groups and objectives aimed for, and the resources allocated. Certain evaluation questions may concern related policies for a more in-depth examination, for example:
|
Evaluation questions
This section is structured as follows:
Focus the evaluation on key questions
.
Structure de la section:
WHAT IS ABOUT THE PURPOSE? |
||||||||||||||||||||
. There are technical limitations that make it impossible to answer multiple questions or, more precisely, to provide quality answers to an excessive number of questions. This guide recommends a maximum of ten questions. How to choose the questions?. - Identify questionsA first version of the evaluation questions is proposed on the basis of:
In a second version, the list and wording of the evaluation questions also take into account:
- Assess the potential usefulness of answersAssuming that a question will be properly answered, it is necessary to assess the potential usefulness of the answer, by considering the following points:
If the choice of questions has to be discussed in a meeting, it may be useful to classify them in three categories of potential utility: higher, medium, lower. - Check that nothing important has been overlookedExperience has shown that it is most harmful to the quality of the evaluation if the following type of question is left out:
- Assess the feasibility of questionsThe feasibility (evaluability) of a question should be examined, but always after its usefulness. For this purpose the following should be consulted:
If the choice of questions has to be discussed in a meeting, it may be useful to classify them in three categories:
If a question is potentially very useful but difficult to answer, check whether a similar question would not be easier and equally useful. For example, if a question concerns a relatively distant or global impact, its feasibility could probably be improved by focusing on the immediately preceding impact in the intervention logic. - Discuss the choice of questionsThe choice of questions is discussed at the inception meeting. The selection is more likely to be successful if potential users have been consulted and have agreed on the selected questions, and if no legitimate point of view has been censored. . |
||||||||||||||||||||
REASONS FOR SELECTING A QUESTION |
||||||||||||||||||||
. Because someone raised itSomeone who proposes a question tends to cooperate in answering it and in accepting the conclusions. It is therefore preferable to select questions clearly requested by the actors concerned, for example:
An actor may ask a question primarily with the intention of influencing or even obstructing the action of another actor. The potential usefulness of this type of question has to be examined carefully. Because it is usefulA question is particularly useful if:
Because the answer is not knownA question is useless if:
It may nevertheless be useful to ask the question again if the answer requires verification. Assessing the overall intervention through a limited number of questionsFocusing on questions does not prevent one from drawing conclusions on the intervention as a whole. On the contrary, it makes it possible to formulate an overall assessment which builds upon professional data collection and analysis, and avoids the risk of being superficial and impressionistic. This can be explained with the analogy of oil exploration. One cannot discover oil by just looking at the surface of the earth. Oil seekers need to explore the underground. The same applies to an intervention being evaluated. The surface of things is visible through reporting, monitoring information, and change in indicators, but what needs to be discovered remains invisible, e.g. the EC's contribution to changes, sustainability, etc. Evaluation questions can be compared with the oil seekers' exploratory wells. Each evaluation question provides a narrow but in-depth view into what is usually invisible. By synthesising what has been learnt by answering the questions, it becomes possible to provide an overall assessment of the intervention. The process can be compared to that of mapping oil fields after an exploratory drilling campaign. |
||||||||||||||||||||
QUESTIONS AND COMPLEXITY OF EVALUATION |
||||||||||||||||||||
Why work with a limited number of questions?Focusing an evaluation on a few key questions is all the more necessary when the intervention concerned is multidimensional and when the evaluation itself is multidimensional. In that case, if one wanted to evaluate all the dimensions of the aid and all the dimensions of the evaluation, the work would be extremely costly or very superficial. It is therefore necessary to make choices. Multidimensional interventionsAn intervention is multidimensional if it concerns several sectors, applies several instruments, and targets several objectives, population groups and/or regions. .
. Multidimensional evaluationsAn evaluation is multidimensional if it refers to several families of evaluation criteria and covers several cross-cutting issues and/or neighbouring policies. .
|
Preparing an evaluation question
.
This section is structured as follows :
- Specify the nature of the expected use
- Ensure that the question concerns evaluation
- Specify the scope of the question
- Inferring a question from the Intervention logic
- Specifying the evaluation criterion
- Should question be open or closed?
.
Why is it important? |
. The questions serve to concentrate the work on a limited number of points in order to ensure that the conclusions are useful and of a high quality. They therefore have to be carefully prepared and worded with precision. Ensure that the answer to the question will be usefulAs far as possible, the evaluation questions are proposed together with a comments on the following points:
If there is uncertainty on the usefulness of the question, it is better to exclude it and to concentrate the evaluation on other more useful questions. . |
SPECIFY THE NATURE OF THE EXPECTED USE |
. |
Question for knowing and/or understanding |
A question for knowing, understanding and/or estimating the effects of the intervention, for example:
Questions of this nature reveal new aspects of the intervention. They help to understand the effects and impact mechanisms, and raise the level of knowledge. |
Question for judging |
A question for formulating or helping to formulate a judgement on the evaluated intervention, for example:
Questions of this nature allow users to judge the merits of the intervention and to recognise good and unsatisfactory practices. They use the evaluation to communicate on the intervention, positively or negatively. |
Question for deciding |
A question for showing how the intervention can be improved, for example:
Questions of this nature lead to recommendations based on lessons from experience. The answers serve to prepare reforms or adjustments. |
Choosing one of three options |
The three types of question are not exclusive. On the contrary, there is a progression in the nature of the questions:
If all the questions of the same evaluation have no purpose other than furthering knowledge and understanding, the exercise is more a study or a piece of research than an evaluation. The nature of use has different levels: to decide, one has to have judged, and to judge one has to have understood. It is therefore enough to draft the question in relation to the highest level of use (question for decision-making or judging). Uses on a lower level can be considered as sub-questions. For example the question might be:
and a sub-question would be:
|
Recommendations |
. |
ENSURE THAT THE QUESTION CONCERNS EVALUATION |
. |
Before drafting a question, ensure that it does not concern audit or monitoring |
If a question concerns audit or monitoring, there are two options:
|
Avoid auditing and monitoring questions |
Evaluation, auditing and monitoring do not serve the same purpose. Consequently, the questions asked are not the same in each exercise. It is important to check whether a question is relevant to evaluation and, where necessary, to amend it so that it does not primarily concern auditing nor monitoring. Amending an auditing question
This question is limited to the verification of the legality and regularity of the implementation of a project, which is a matter of auditing. In an evaluation it would be relevant rather to ask whether the application of regulations was a particular factor of effectiveness or ineffectiveness.
Amending a monitoring question
This question concerns only the programme outputs, which is a matter of monitoring. In an evaluation it would be relevant to ask whether the quality of outputs is a particular factor of efficiency of inefficiency.
. |
SPECIFY THE SCOPE OF THE QUESTION |
. |
What is this about? |
Generally, an evaluation question concerns both the effects of the intervention and the intervention itself:
|
What is the purpose? |
|
Questions on the intervention design |
These questions may refer to the design of the intervention as a whole or to a particular step in the design process, such as:
Have the procedures of dialogue with the actors favoured ownership of the strategy and increased the chances of sustainable impacts? |
Questions on the implementation |
These questions concern the entire implementation of the intervention or a particular aspect of the implementation process such as:
To what extent do administrative funding and project management procedures facilitate or hinder the adaptation of aid to beneficiaries' needs? To what extent have phasing out procedures favoured the sustainability of impacts as regards food safety? |
Questions on the modalities of aid |
These questions concern modalities such as:
Has the funding modality opted for made it possible to obtain better effects in terms of food safety? To what extent have interventions in the field of transport been conducted in the form of sector-specific programmes, and what difference does this make from the point of view of effects generated? |
Questions on the intervention as a whole |
Finally, certain questions are drafted broadly and concern the intervention as a whole, including its design and implementation, for example: "To what extent has the design and implementation of the intervention helped to produce effect X?".
If the question concerns the entire design and implementation, we choose to focus it on a precise, immediate effect. A question concerning the entire intervention and all its effects would probably be unevaluable. . |
INFERING A QUESTION FROM THE INTERVENTION LOGIC |
. |
What is this about? |
The intervention logic specifies the expected effects. Most evaluation questions concern one (or several) effects, which have to be specified. |
What is the purpose? |
To focus the question on the effects that are considered to be the most important or the least known. To find the right balance between:
|
Effect, need or problem |
Where possible and relevant, the question specifies the effect concerned. This is easy for effectiveness questions such as:
It is also preferable to specify the effect concerned in more complex cases, for example:
As shown above, many questions can be related to the intervention logic either directly (effectiveness) or indirectly (sustainability, efficiency, relevance). The only real exception is the question on unexpected effects. Effect
This question concerns an expected effect: better access to basic services. It is inferred directly from the intervention logic. Need
This question concerns a need that must be satisfied: better food for local communities dependent on fishing. It may be indirectly related to the intervention logic in so far as satisfaction of the need corresponds to an objective. Problem
This question concerns a problem. Is it realistic to want to improve the managerial capacity of local authorities in the current context of the country? The question is indirectly related to the intervention logic in that it concerns an objective. |
More or less extensive effects |
The question specifies whether it concerns:
Questions on sets of effects are of interest to policy-makers and strategic decision-makers but are generally more difficult to answer. Operators and field level managers are more interested in questions on precise effects. These questions are also easier to answer. If the question concerns all effects, it is focused on a specific aspect of intervention design or implementation. A question that covered the entire intervention and all its effects would probably not be evaluable. All the effects
The scope of this question is very broad: all the impacts of European Aid in a country. Group of effects
This question concerns a group of logically related effects. A particular effect
This question concerns a precise effect, for a clearly defined public. |
Close or distant effects |
The wording of the question indicates whether one is interested in:
Questions on the most distant impacts are of interest to policy-makers and strategic decision-makers but are generally more difficult to answer. Operators and field level managers are more interested in questions on direct results, on the uptake by the targeted group, on their needs, or on the closest impacts. These questions are also easier to answer. Reaching beneficiaries
This question concerns an initial result. Result
This question concerns a short-term result for direct beneficiaries. Intermediate impact
This question concerns an intermediate effect (improvement of access to services). A more direct effect would have been, for example: greater priority given to the territories neglected by local authorities. A more distant effect would have been, for example: reduction of regional disparities in access to basic services. Global impact
This question concerns a set of distant effects. A closer effect would have been, for example, local authorities' adoption of practices that involve actors in project design. . |
SPECIFYING THE EVALUATION CRITERION |
. |
What is this about? |
The questions are classified in different families that correspond to different "viewpoints" on what is being evaluated. Seven of these viewpoints, also called evaluation criteria, are to be considered: relevance, effectiveness, efficiency, sustainability, impact, coherence/complementarity, and Community value added. A question is drafted in relation to an evaluation criterion unless it is intended only for knowledge and understanding. Questions of efficiency and sustainability are asked more rarely, partly because they are difficult to answer. Yet they are often more useful. |
What is the purpose? |
|
Focus on one evaluation criterion for each question |
Questions relating to several evaluation criteria are complex, for example: Has the intervention produced effect X satisfactorily compared to the objectives (effectiveness) and were the objectives in phase with the targeted beneficiaries' needs (relevance)?
With this type of question it is probably necessary to implement two different evaluation methods: one to judge whether the expected effect was obtained and the other to identify the target group's main needs and judge whether the intervention satisfies them. The two methods use different survey and analysis techniques. Applying them both would probably entail doubling the costs or halving the quality of the answer. In this example it would be better to draft two different questions, to decide which one is more useful, and then to concentrate the resources on that question. |
Exception |
In some cases it is not possible to consider an evaluation criterion independently from another one, for example: Is effect X likely to continue after the end of financial support (sustainability)? In this example, we cannot answer the question without first having asked a sub-question: Has effect X been obtained (effectiveness)?
In this case it is preferable to ask only one final question (sustainable effect). In preparing the answer, the evaluation team notes that a sub-question will be: Has effect X been obtained? |
The families of evaluation criteria |
This page proposes a typology of seven families of criteria. The first five correspond to the traditional practice of evaluation of development aid formalised by the OECD (DAC). The following two apply to all EC policies. RelevanceThe extent to which the objectives of the development intervention are consistent with beneficiaries' requirements, country needs, global priorities and partners' and donor's policies. Note: Retrospectively, the question of relevance often becomes a question as to whether the objectives of an intervention or its design are still appropriate given changed circumstances. Examples
EffectivenessThe extent to which the development intervention's objectives were achieved, or are expected to be achieved, taking into account their relative importance. Note: Also used as an aggregate measure of (or judgment about) the merit or worth of an activity, i.e. the extent to which an intervention has attained, or is expected to attain, its major relevant objectives efficiently in a sustainable fashion and with a positive institutional development impact. Related term: efficacy. Examples:
EfficiencyA measure of how economically resources/inputs (funds, expertise, time, etc.) are converted to results. Examples:
SustainabilityThe continuation of benefits from a development intervention after major development assistance has been completed. The probability of continued long-term benefits. The resilience to risk of the net benefit flows over time. Examples:
ImpactPositive and negative, primary and secondary long-term effects produced by a development intervention, directly or indirectly, intended or unintended. Example:
Coherence/complementarityThis criterion may have several dimensions: 1) Coherence within the Commission's development programme Example: Can it be said that the activities and outputs logically allow the objectives to be achieved? Are there contradictions between the different levels of objective? Are there duplications between the activities? 2) Coherence/complementarity with the partner country's policies and with other donors' interventions Example: Can it be said that there is no overlap between the intervention considered and other interventions in the partner country and/or other donors' interventions, particularly Member States? 3) Coherence/complementarity with the other Community policies Example: Is there convergence between the objectives of the intervention and those of the other Community policies (trade, agriculture, fishing, etc.)? Community value addedThe extent to which the development intervention adds benefits to what would have resulted from Member States' interventions only in the partner country. Examples
Use of the term criterion: A warning!This document concerns evaluation criteria, that is, the main ways of judging the intervention. In order to formulate fully transparent value judgements, the approach needs to be refined into evaluation questions, and then into judgement criteria. The word "criterion" is also used with a third meaning in the framework of quality assurance. In that case it concerns the quality assessment criteria of the evaluation. . |
SHOULD QUESTIONS BE OPEN OR CLOSED? |
. |
What does this mean? |
Generally questions requiring a 'yes or no' answer are avoided for two reasons:
|
Open questions requiring qualified answers |
Examples:
This type of wording is more appropriate if the question is intended to acquire knowledge or understanding, or to aid decision-making. In such cases the users expect qualified answers. |
Closed questions requiring "yes or no" answers |
Examples:
This type of question is more appropriate for accountability purposes, in a context where the objectives were set with precision. A closed question can also be useful if the intention is to validate a procedure or innovative instrument or to confirm a good practice. |
Examples
.
This section is structured as follows:
.
EVALUATION AT COUNTRY LEVEL |
. For an evaluation at country level the questions concern, for example:
The questions below are examples. Many are drawn from recent country evaluations undertaken by the Evaluation Unit. Questions on relevance
|
Questions on effectiveness- Questions on sector-specific effectivenessThis site proposes a method for choosing questions specific to a particular aid sector, based on the analysis of the intervention logic. The method generates the following standard type of question:
These are questions on effectiveness (attaining a specific objective or obtaining an expected impact). Each standard question can be reformulated in many ways: ...by specifying the scope for example:
...or by specifying the effect concerned, for example:
...or by changing the evaluation criterion, for example:
|
- Generic questions on effectiveness
Questions on efficiencyQuestions on efficiency (also called cost-effectiveness) may be specific to an aid sector and are formulated as follows:
It is also useful to ask generic questions on efficiency:
|
Questions on sustainabilityQuestions on sustainability are generally specific to an aid sector and are formulated in one of the following two ways:
|
Questions on coherence and complementarity
Questions on cross-cutting issuesThe cross-cutting issues mentioned most frequently in recent evaluations are:
Each of the following questions may concern one or more cross-cutting issues:
Questions on instruments and implementation
|
Example of a recent evaluationThe following example presents nine questions chosen in the case of a recent country evaluation (with several important simplifications).
. |
EVALUATION AT REGIONAL LEVEL |
. For an evaluation at regional level the questions concern, for example:
The questions below are examples. Many are drawn from recent country evaluations undertaken by the Evaluation Unit. Questions on relevance
|
Questions on effectivenessThis site proposes a method for choosing questions peculiar to a particular aid sector, based on the analysis of the intervention logic. The method generates the following standard type of question:
These are questions on effectiveness (achieving a specific objective or obtaining an expected effect). Each standard question can be formulated in many other ways: ...by specifying the scope for example:
or by specifying the effect concerned, for example:
...or by changing the evaluation criterion, for example:
It is also useful to ask questions on generic effectiveness, that is, cutting across an entire intervention logic, for example:
Questions on efficiencyQuestions on efficiency (also called cost-effectiveness) may be specific to an aid sector and may be in the following form:
It is also useful to ask generic questions on efficiency, for example:
|
Questions on sustainabilityQuestions on sustainability are generally specific to an aid sector and are formulated as follows:
|
Questions on coherence/complementarity
Questions on cross-cutting issuesThe cross-cutting issues most frequently mentioned in recent evaluations are:
Each of the following questions may concern one or more cross-cutting issues:
Questions on instruments and implementation
|
Example of a recent evaluationThe following example presents the nine questions chosen in the case of a recent evaluation in the MERCOSUR region (with several important simplifications):
. |
SECTOR OR THEMATIC GLOBAL EVALUATION |
. For an sector or thematic global evaluation the questions concern, for example:
The following questions are examples, many of which are drawn from recent country evaluations undertaken by the Evaluation Unit. Questions on relevance
|
Questions on effectivenessThis site proposes a method for choosing questions from the analysis of the intervention logic. The method is generally based on the analysis of one or more strategic documents that define a coherent set of objectives. Such documents may not exist in the case of thematic or sector evaluation on a global scale, but that does not preclude the more informal reconstruction of the intervention logic. It is then possible to formulate questions with the following standard form:
These are questions on effectiveness (attaining a specific objective or one of the expected effects). Each standard question can be reformulated in many ways: ...by specifying the scope for example:
...or by specifying the effect concerned, for example:
or by changing the evaluation criterion, for example:
It is also interesting to ask generic effectiveness questions, that is, questions cutting across the entire intervention logic, for example:
Questions on efficiencyQuestions on efficiency (or cost-effectiveness) can be specific to the sector/theme. They are formulated as follows:
|
Questions on sustainabilityQuestions on sustainability are generally formulated in one of the following two ways:
|
Questions on coherence/complementarity
Questions on cross-cutting issuesThe cross-cutting issues mentioned most frequently in recent evaluations are:
Each of the following questions can concern one or more cross-cutting issues:
Questions on instruments and implementation
|
Example of a recent evaluationThe following example presents the nine questions chosen in the global evaluation of the 'Commerce' sector:
|
Judgement references
.
This section is structured as follows:
.
JUDGEMENT CRITERION |
|||
. What does this mean?A judgement criterion specifies an aspect of the evaluated intervention that will allow its merits or success to be assessed. Whilst "judgement criterion" is the appropriate word, an acceptable alternative is "reasoned assessment criterion". The criterion is used to answer an evaluation question. One or more judgement criteria are derived from each question. What is the purpose?
How can a judgement criterion be clarified on the basis of a question?All the evaluation questions relate to one or more judgement criteria, unless they are designed only to further knowledge or understanding about the intervention or its effects. The following is an example of a question:
Like most evaluative questions, it has two parts:
The judgement criteria develop and specify the second part of the question, for example:
The judgement criteria derive from the question, for instance in the case of the first criterion:
To be used in practice, each judgement criterion has to be accompanied by a target level and one or more indicator(s). |
|||
Recommendations
Be careful not to confuse conceptsOn this site, the word criterion is used for three different concepts:
According to the EC, the value added of an evaluation is the formulation of value judgements on the basis of evidence and explicit judgement criteria. When dealing with organisations which are not familiar with evaluation, it may be wise not to use the word "judgement", which may induce resistance. An acceptable alternative is "assessment", or preferably "reasoned assessment". . |
|||
TARGET |
|||
. What does this mean?The concept of a 'target' is widely used in the context of public management for setting a verifiable objective or a level of performance to be achieved. In an evaluation context it is used in a much wider sense since the evaluated intervention may have to be judged against targets that were not set in advance but that are specifically identified, such as a benchmark, a success threshold or a comparable good practice. What is the purpose?
How can they be determined?- By reference to an objective defined in a verifiable wayThe target may appear in one of the intervention objectives, that is, as long as they have been established in a verifiable way. In this particular case, the same indicator helps to define the objective, to make the judgment criterion operational and to determine the target.
|
|||
- In relation to comparable good practices outside the interventionIn this case, the target is established at the outset of the evaluation. It is not related to an objective or a performance framework existing prior to the evaluation.
The procedure is as follows:
|
|||
- Compared to best practices identified within the interventionThe target can be found within the evaluated intervention itself during the synthesis phase, provided that specific practices can be considered as good as regards the judgement criteria under consideration. In this case, the good practices will serve as benchmarks to judge the others. Of course, it is advisable to check that the contextual conditions are close enough so as to allow for comparison.
|
|||
When should they be determined?- Earlier or later in the evaluation processIf the target is derived from a verifiable objective or a performance framework, then it can be determined at the very first stage of the evaluation process. If the target is derived from an outside benchmark, then it should be identified during the early stages of the evaluation. However, the process may involve the gathering of secondary data with a view to specifying the benchmark, as well as a careful examination of comparability. This means that the target will not be completely defined in the first phase of the evaluation. If the target is to be derived from the best practices discovered within the intervention by the evaluation team, it will be determined in the synthesis phase. - After choosing the judgement criterionDetermining the target takes place in a three-step process:
Evaluation targets and othersWhen the evaluation question pertains to an intended result or impact, the target level is usually derived from a verifiable objective or borrowed from a performance assessment framework. Performance monitoring may however be of little or no help in the instance of evaluation questions relating to cross-cutting issues, sustainability factors, unintended effects, evolving needs and problems, coherence, etc. . |
|||
INDICATOR |
|||
. What does this mean?The evaluation team may use any kind of reliable data to assess whether an intervention has been successful or not in relation to a judgement criterion and a target. Data may be collected in a structured way by using indicators. Indicators specify precisely which data are to be collected. An indicator may be quantitative or qualitative. In the latter case the scoring technique may be used. Unstructured data are also collected during the evaluation, either incidentally, or because tools such as case studies are used. This kind of evidence may be sound enough to be a basis for conclusions, but it is not an indicator. What is the purpose?
|
|||
Evaluation indicatorsThe main evaluation indicators are those related to judgement criteria, that specify the data needed to make a judgement based on those criteria. An indicator can be constructed specifically for an evaluation (ad hoc indicator) and measured during a survey, for example. It may also be drawn from monitoring databases, a performance assessment framework, or statistical sources. A qualitative indicator (or descriptor) takes the form of a statement that has to be verified during the data collection (e.g. parents' opinion is that their children have the possibility of attending a primary school class with a qualified and experienced teacher). A quantitative indicator is based on a counting process (e.g. number of qualified and experienced teachers). The basic indicator directly results from the counting process. It may be used for computing more elaborate indicators (ratios, rates) such as cost per pupil or number of qualified and experienced teachers per 1,000 children of primary-school age. Indicators may belong to different categories: inputs, outputs, results or impacts. |
|||
Evaluation indicators and othersWhen an evaluation question pertains to an intended result or impact, it is worth checking whether this result or impact has been subject to performance monitoring. In such cases, the evaluation team uses the corresponding indicators and data, which is a considerable help, especially if baseline data have been recorded. Performance monitoring may, however, be of little or no help in the instance of evaluation questions relating to cross-cutting issues, sustainability factors, unintended effects, evolving needs or problems, coherence, etc. Quality of an indicatorAn indicator measures or qualifies with precision the judgement criterion or variable under observation (construct validity). If necessary, several less precise indicators (proxies) may be used to enhance validity. It provides straightforward information that is easy to communicate and is understood in the same way by the information supplier and the user. It is precise, that is, associated with a definition containing no ambiguity. It is sensitive, that is, it generates data which vary significantly when a change appears in what is being observed. Performance indicators and targets are often expected to be SMART, i.e. Specific, Measurable, Attainable, Realistic and Timely. The quality of an evaluation indicator is assessed differently. |
|||
Indicators and effects: a warning!The indicator used to evaluate an effect is not in itself a measurement or evidence of that effect. The indicator only informs on changes, which may either result from the intervention (effect) or from other causes. The evaluation team always has to analyse or interpret the indicator in order to assess the effect. |
|||
Categories of indicators |
|||
- Indicators and the intervention cycleIndicators are used throughout the intervention cycle. They are first used to analyse the context; then, for the choice and validation of the intervention strategy, afterwards for monitoring outputs and results and, finally, for the evaluation. Indicators and intervention designContext indicators may be used to support the identification of the needs, problems and challenges which justify the intervention. As far as possible, objectives and targets are defined in a measurable way by using indicators. Indicators, monitoring and performance assessmentMonitoring systems and performance assessment frameworks also use indicators which derive from the diagram of expected effects (also called results chain). Monitoring indicators primarily relate to inputs and outputs. Performance indicators primarily focus on intended results and impacts. The EC's Result Oriented Monitoring (ROM) does not rely that much on indicators. It delivers systematic assessments of external aid projects in the form of ratings with a view to intended results and impacts. Indicators and evaluationEvaluation indicators are used to help answering specific evaluation questions. Depending on the question, they may relate to the needs, problems and challenges which have justified the intervention, or to the achievement of intended outputs, results and impacts, or to anything else. - Global and specific indicatorsGlobal or contextual indicators apply to an entire territory, population or group, without any distinction between those who have been reached by the intervention and those who have not. They are mainly taken from statistical data. This site offers help to look for contextual indicators. Specific indicators concern only a group or territory that has actually been reached. With specific indicators, changes among those affected by the intervention can be monitored. Most of these indicators are produced through surveys and management databases. - Indicators and intervention logicInput indicatorsInput indicators provide information on financial, human, material, organisational or regulatory resources mobilised during the implementation of the intervention. Most input indicators are quantified on a regular basis by the management and monitoring systems (providing that they are operational). Output indicatorsOutput indicators provide information on the operators' activity, especially on the products and services that they deliver and for which they are responsible. To put it simply, one could say that outputs correspond to what is bought with public money. Result indicatorsResult indicators provide information on the immediate effects of the intervention for its direct addressees. An effect is immediate if the operator notices it easily while he/she is in contact with an addressee. Because they are easily recognised by the operators, direct result indicators can be quantified exhaustively by the monitoring system. Impact indicatorsImpact indicators provide information on the long-term direct and indirect consequences of the intervention. A first category concerns the consequences that appear or last in the medium or long term for the direct beneficiaries. A second category of impacts concerns people or actors that are not direct beneficiaries. Impact indicators cannot be produced in general from management information. They require statistical data or surveys specially conducted during the evaluation process. |
|||
Indicators derived from scoring |
|||
What does this mean?Scoring (or rating) produces figures that synthesise a set of qualitative data and or opinions. Scoring is guided by a scoring grid (or scorecard) with varying degrees of detail. From an evaluation point of view, both words scoring and rating can be used. What is the point?Scoring allows the production of structured and comparable data on judgement criteria that do not lend themselves to a measurement using quantitative indicators. How to construct a scoring grid
How to use the scoring gridScoring grids usually apply to projects or components of the intervention and allow for comparing these. The evaluation team puts together all the data it has on the project or intervention to be assessed. It then chooses the level (or descriptor) in the scoring grid that corresponds best (or the least badly) to this information. The score results from this choice. |
|||
RecommendationsThe more detailed the scoring grid the less subjective the score will be and the more comparable the scores allocated by two different evaluators will be. . |
|||
FROM QUESTIONS TO INDICATORS |
|||
.
From questions to judgement criteriaThe judgement criterion (also called reasoned assessment criterion) specifies an aspect of the evaluated intervention that will allow its merits or worth to be assessed in order to answer the evaluation question, . For instance:
The judgement criterion gives a clear indication of what is positive or negative, for example: "enhancing the expected effects" is preferable to "taking potential effects into account".
A more precise judgement criterion than the questionThe question is drafted in a non-technical way with wording that is easily understood by all, even if it lacks precision. The judgement criterion focuses the question on the most essential points for the judgement. Yet the judgement criterion does not need to be totally precise. In the first example the term "satisfactory quality" can be specified elsewhere (at the indicator stage). Not too many criteriaIt is often possible to define many judgement criteria for the same question, but this would complicate the data collection and make the answer less clear. In the example below, the question is treated with three judgement criteria (multicriteria approach):
A judgement criterion corresponding to the questionThe judgement criterion should not betray the question. In the following example, two judgement criteria are considered for answering the same question:
The first judgement criterion is faithful to the question, while the second is less so in so far as it concerns the success in primary education, whereas the question concerns only the access to it. The question may have been badly worded, in which case it may be amended if there is still time. Also specify the scope of the questionMost questions have a scope (what is judged) and a judgement criterion (the way of judging). In addition to the judgement criterion, it is therefore often desirable to specify the scope of the question, for example: "European aid granted over the past X years", "design of programme X", "the principle of decentralisation adopted to implement action X". Also specify the type of cause-and-effect analysisSome questions imply a cause-and-effect analysis prior to the judgement. It may therefore also be useful to specify the type of analysis required by means of terms such as "has European aid led to", "has it contributed to", "is it likely to". |
|||
From judgement criteria to indicatorsAn indicator describes in detail the information required to answer the question according to the judgement criterion chosen, for example:
Not too many indicatorsIt is possible to define many indicators for the same judgement criterion. Relying upon several indicators allows for cross-checking and strengthens the evidence base on which the question is answered. However, an excessive number of indicators involves a heavy data collection workload without necessarily improving the soundness of the answer to the question. In the examples below three indicators are applied to a judgement criterion ("capacity of the primary school system to enrol pupils from ethnic minority X with satisfactory quality"):
Indicator corresponding to the judgement criterionThe indicator should not betray the judgement criterion. Two indicators are considered below:
The first indicator corresponds faithfully in so far as it describes an essential aspect of the judgement criterion. The second indicator is less faithful because it fails to reflect the concept of "satisfactory quality". Its construct validity is not good. Unambiguous indicatorsAn indicator must be defined without any ambiguity and understood in the same way by all the members of the evaluation team. For instance, in the above examples it is necessary to specify what a "qualified and experienced teacher" is. This can be done with reference to an existing definition, or else a definition can be formulated as precisely as possible until there is no more ambiguity whatsoever. Indicators independent from the observation fieldThe same indicator should be able to serve to collect data in several contexts, for example:
In this case the same indicator is applied in both types of area and serves as a comparison, on the basis of which a judgement is formulated. Quantitative and qualitative indicatorsThe following two examples present an alternative between a quantitative indicator and a qualitative indicator for treating the same judgement criterion:
An indicator is preferably associated with a targetThe target indicates which comparison should be made in order to answer the question, for example: "In the areas where ethnic minority X concentrates, the indicator is at least as good as in the entire country in average". The target and the indicator are often specified interactively in successive steps. It is important not to digress from the judgement criterion during this process. When the evaluation question pertains to an intended result or impact, the target is usually derived from a verifiable objective or borrowed from a performance assessment framework. Feasible indicatorsThe indicator makes it possible to focus and structure data collection but serves no purpose as long as data does not exist. To ensure the feasibility of an indicator, it is necessary to indicate the source of the information to use, for example:
If no source is available or feasible, the indicator should be changed. If no feasible indicator can be found, excluding the question should be envisaged. |
|||
Example of a country evaluation |
|||
Question
Judgement criterion
The judgement criterion is derived from the question in the following way:
Indicator
The indicator is derived from the judgement criterion in the following way:
|
|||
Example of a sector evaluation |
|||
Question
The question refers to a family of evaluation criteria: effectiveness. Judgement criterion
The judgement criterion is derived from the question in the following way:
Indicator
The indicator derives from the judgement criterion in the following way:
|
Methodological design
.
This section is structured as follows:
- What does it mean?
- Design table per question
- Optimising the overall design
- Developing a tool
- Feasability (evaluability) of a question
.
WHAT DOES IT MEAN? |
||||||||||||||||||||||
. What is the purpose?To set up the method that will allow the external evaluation team to answer the questions and to come to an overall assessment. In addition to selected questions, judgement criteria, indicators and targets, the evaluation method includes:
|
||||||||||||||||||||||
When does it take place?The evaluation team starts designing a provisional method as early as from the drafting of its proposal in order to draw up cost estimates. A key assumption at this stage is the extent to which the evaluation will rely on secondary data or will involve specific data collection work in the field. The main frame of the method is then established during the inception stage, in line with the evaluation questions, judgement criteria, indicators, data collection tools and analysis strategy.
The method is refined and finalised before the field phase and fully described in the first phase report (desk).
. |
||||||||||||||||||||||
DESIGN TABLE PER QUESTION |
||||||||||||||||||||||
. What does this mean?The design table explains how an evaluation question will be answered, including the chain of reasoning which connects data, findings and conclusions. |
||||||||||||||||||||||
When is it constructed?A design table is developed for each question and progressively refined in successive versions:
|
||||||||||||||||||||||
Opening the tableThe first lines of the table summarise the steps which have already been taken (see from question to indicator), i.e.
|
||||||||||||||||||||||
Sub-questionsWhat does this mean?Together with the judgement criteria, indicators and targets, the sub-questions are meant to explain how the evaluation team will answer the question, for instance:
Who does what, and when?A set of evaluation questions is drawn up at the inception stage, together with sub-questions.
What is the purpose?The sub-questions describe the chain of reasoning through which the evaluation team plans to answer the question. The intended reasoning is indicative but it is worth clarifying it in advance because:
It may furthermore be worth suggesting provisional sub-questions together with the draft set of key questions. Members of the reference group will then realise that many of their expectations will be satisfied through answering the sub-questions. It will help them to accept the principle of a limited list of well focused evaluation questions. Sub-questions pertaining to indicatorsThese sub-questions may pertain to:
(2) changes in the indicators, for instance:
As seen in the examples above, the sub-questions may be quantitative or qualitative. |
||||||||||||||||||||||
Question:
Sub-question:
In this example, the sub-question is simply meant to show how the indicator (number of experienced and qualified teachers) will be applied. |
||||||||||||||||||||||
Sub-questions pertaining to analysisThese sub-questions are written with a view to:
(4) Challenging assumptions about the success of the intervention and substantiating a negative answer to the question, for instance:
|
||||||||||||||||||||||
Question:
Sub-question:
In this example the sub-question relates to a specific short-term result (promotion of decentralisation through policy dialogue) which is a driver to the wider impact highlighted in the question (capacity building). |
||||||||||||||||||||||
Sub-questions pertaining to judgementThese sub-questions are meant to assist in the formulation of conclusions involving explicit value judgements. They are written with a view to:
(6) applying or developing the targets in the specific context of the intervention, for instance:
|
||||||||||||||||||||||
Question:
Sub-question:
|
||||||||||||||||||||||
Analysis strategyFour strategies can be considered:
The three last strategies are appropriate for the questions which require a cause-and-effect analysis. The first one is appropriate in other instances. |
||||||||||||||||||||||
Investigation areasThe evaluation team may consider collecting and analysing data at the level of the intervention as a whole, or investigating some areas more specifically, for instance:
|
||||||||||||||||||||||
Example |
||||||||||||||||||||||
. |
||||||||||||||||||||||
OPTIMISING THE OVERALL DESIGN |
||||||||||||||||||||||
. What is the purpose?To finalise the overall evaluation method in a way which cuts across the evaluation questions and which makes a good enough mix of evaluation tools, considering the available time and resources. |
||||||||||||||||||||||
When does it take place?The evaluation method is designed through an iterative process at three levels:
Several iterations may be needed in order to allow the evaluation team to optimise the overall design whilst ensuring a high quality to each question. |
||||||||||||||||||||||
Selecting toolsIn parallel with the design tables, which are established on a question-by-question basis, the evaluation team designs its overall evaluation method which covers all questions and includes a number of tools such as:
These examples are all but limitative. The evaluation team may even have to develop a completely new tool, where relevant. |
||||||||||||||||||||||
Combining tools and questionsThe evaluation team draws up the list of all evaluation tools suggested in the design tables. Each tool is then considered from the viewpoint of its capacity to help answering several questions and sub-questions, for instance:
An image of this process is given by the matrix below in which the first tool helps answering sub-questions Aa, Cf and Eb, etc. ![]() This image suggests that the set of tools will provide the evaluation team with possibilities of cross-checking data on sub-questions Aa and Eb. Tools and questions are combined in order to optimise such possibilities of cross-checking. |
||||||||||||||||||||||
Verifying the adequacy of toolsThe evaluation team confirms that each tool envisaged is adequate in the sense that:
The evaluation team explains and justifies its main technical choices, with alternative options, pros and cons, and associated risks. |
||||||||||||||||||||||
Articulating toolsThe evaluation team assesses whether the selected tools may reinforce one another, for instance:
|
||||||||||||||||||||||
Preparing the overall assessmentIn the process of progressively optimising its design, the evaluation team examines all the design tables in a cross-cutting manner with a view to preparing its final synthesis, i.e. an overall assessment that draws upon the answers to all evaluation questions.
|
||||||||||||||||||||||
Allocating resourcesIn the process of designing its method, the evaluation team tries to adequately share its limited resources between questions and sub-questions. Some questions deserve to be addressed with costly tools such as questionnaire surveys of end users, several case studies, focus groups, etc. Other questions should rather be answered on the basis of a documentary analysis only and a few interviews with EC and Government officials.
A question is rarely unevaluable in the absolute. It is more likely to be an accumulation of difficulties and constraints that leads to feasibility problems. At the earliest stages of the evaluation, it is often possible to amend a difficult question so as to make it more evaluable. This can be done, for example, by limiting the scope of the question or choosing to apply it to a less distant effect or to a probable effect if the real effect is not yet observable. Once a difficult question has been validated, the evaluation team has to design an appropriate method, and to allocate adequate resources. |
||||||||||||||||||||||
Cost and time constraintsSuccessive versions of the method are designed within the evaluation team until the following constraints are matched:
. |
||||||||||||||||||||||
DEVELOPING A TOOL |
||||||||||||||||||||||
. No evaluation method without toolsEvaluation tools are needed in order to collect primary and secondary data, to analyse data, and to formulate value judgements (or reasoned assessments). The external evaluation team may carry out some of these tasks without evaluation tools, but several tools are always needed for answering each evaluation question. |
||||||||||||||||||||||
How are tools chosen?Tools are chosen during the iterative process of designing the overall evaluation method, with an aim to:
|
||||||||||||||||||||||
When is it developed?All evaluation tools are developed progressively and finalised before the beginning of the field phase, although some tools need to be implemented, and therefore developed earlier in the process, e.g. interviews with key stakeholders at the inception stage, analysis of management databases at the desk phase, etc. |
||||||||||||||||||||||
Developing a toolWhilst the set of evaluation tools is to be selected as a part of the overall evaluation design, each tool is to be developed in a separate way. An example of developed tool is provided on this site. |
||||||||||||||||||||||
Questions and sub-questionsThe evaluation team lists the questions and sub-questions that have to be addressed by the tool. It refers to the design tables. |
||||||||||||||||||||||
Technical specificationsThe evaluation team develops the technical specifications of the tool through a preparatory stage. Technical specifications depend on the type of tool. They cover issues like:
Caution! - When developing a questionnaire or an interview guide, the evaluation team should not proceed by copying and pasting evaluation sub-questions. If evaluation questions and sub-questions are naïvely passed to informants, there are considerable risks of biases. |
||||||||||||||||||||||
Risk managementThe evaluation team assesses the main risks with data collection and analysis, as well as potential biases. As far as relevant, it prepares second best solutions in case the tool cannot be applied satisfactorily. The following lines provide examples of risks associated with evaluation tools, and examples of second best solutions:
This list is not limitative. |
||||||||||||||||||||||
Mode of reportingThe outputs vary from one tool to another. They may take the form of tables, lists of quotations, lists of verbatims, monographs, etc. The evaluation team decides on how the outputs will be reported, for instance:
The evaluation team also decides on how to report about the implementation of the tool and the associated limitations if any, e.g.
|
||||||||||||||||||||||
ResponsibilitiesThe tasks and roles are shared among the evaluation team members, e.g.
|
||||||||||||||||||||||
QualityQuality criteria are precisely defined. Depending on the tool they may cover issues like:
This site proposes a series of quality check-lists for frequently used tools. |
||||||||||||||||||||||
Time schedule and resourcesIn some instances, it may be necessary to specify practicalities like:
|
||||||||||||||||||||||
Example |
||||||||||||||||||||||
. |
||||||||||||||||||||||
FEASABILITY (EVALUABILITY) OF A QUESTION |
||||||||||||||||||||||
. What is this?Certain questions are easy to answer while others may raise evaluability problems. It is therefore necessary to assess the feasibility of evaluation questions from the outset. What is the purpose?
What has to be taken into account?To establish whether a question is evaluable, we check:
|
||||||||||||||||||||||
What are the most frequent limitations?A highly innovative activityIf the question concerns an innovative instrument, activity or procedure, the following difficulties may arise:
A very recent activityIf the question concerns a recently implemented activity:
Managerial weaknessesIf the question concerns activities in which there are or were managerial weaknesses, the following difficulties may be encountered:
In case of a strong suspicion of illicit or illegal practices, it is preferable to postpone the evaluation question for later and to start with an audit. A scope that is too complexIf the question concerns a multi-dimensional or multi-sectoral scope, the following difficulties may be encountered:
A far-reaching impactIf the question concerns a far-reaching impact which is connected to the evaluated activity through a long chain of causes and effects, then the following difficulties may be encountered:
An intervention that is too marginalIf the question concerns a very small activity compared to other neighbouring policies or to the context, the following difficulties may be encountered:
|
||||||||||||||||||||||
RecommendationsA question is rarely unevaluable in the absolute. It is more likely to be an accumulation of difficulties and constraints that makes the question difficult. |
Data collection
.
This section is structured as follows:
.
For the purpose of answering questions, the evaluation team collects data that are already available (secondary data) and applies data collection tools with a view to obtaining new information (primary data). . |
||||||||||||||
USING AVAILABLE DATA |
||||||||||||||
. What are they?To avoid duplicating efforts, running up unnecessary costs and tiring the informants, it is recommended wherever possible to rely on existing information (secondary data), such as administrative data, research, previous evaluations, management and monitoring databases, statistics. This information can be obtained at a lower cost. It can help to provide partial answers to some of the questions asked. Why is it important?
Main channels for identifying and gathering secondary data
|
||||||||||||||
Reliability problemsBefore making use of secondary data, particular attention should be paid to the following points:
|
||||||||||||||
Recommendations
. |
||||||||||||||
MOST FREQUENT DIFFICULTIES WITH DATA COLLECTION |
||||||||||||||
. What are they? How to cope with them?Even if the data collection programme has been wisely prepared, the evaluation team often encounters problems during its field work. The most frequent difficulties occur with: - Access to informants
- Cultural gap
- Lack or weakness of data
|
||||||||||||||
Recommendations
. |
||||||||||||||
RELIABILITY OF COLLECTED DATA |
||||||||||||||
. What are the risks?While gathering information, the evaluation team faces various risks of biases which may undermine the reliability of collected data. Why should biases be considered carefully?
|
||||||||||||||
Most frequent biases- Confirmation biasThis risk is a threat to all data collection approaches. It results from a tendency to seek out evidence that is consistent with the intervention logic, rather than evidence that could disprove it. When subject to this bias, the evaluation team and informants tend to focus on intended effects and systematically to overlook external factors, unintended effects, negative effects, interactions with other policies, outside stakeholders, alternative implementation options, etc. This bias is avoided by relying on independent and professional evaluators. - Self-censorshipIn some instances, informants may be reluctant to freely answer questions, simply because they feel at risk. They tend to rigidly express the views of their institution or their hierarchy. This bias is combated by guaranteeing confidentiality and anonymity in the treatment of answers. The interviewer should also insist on factual questions and avoid collecting opinions. - Informants' strategyThose who have stakes in the intervention may distort the information they provide, with the aim of obtaining evaluation conclusions closer to their views. This bias will be reduced if the whole range of stakeholders is included in the data collection work plan and if various sources of information are cross-checked. - Unrepresentative sampleThis bias may be a matter of concern if the evaluation team generates quantitative data through a questionnaire survey. It should also be considered when using secondary data obtained from a questionnaire survey. In this instance, the evaluation team should verify that the sample of surveyed informants is large enough and representative of the population as a whole. - Question induced answersThis bias and the following ones are frequent in interviews and questionnaires. The way in which questions are asked by interviewers or the interviewer's reaction to answers can generate a bias which is either positive or negative. Even the order of the questions in a questionnaire may change the substance of the answers. This bias will be limited by having questionnaires designed and tested by experienced professionals. - Empathy biasInterviewees may not have a pre-determined opinion about the questions put to them. They try to make up their mind in a few seconds when responding to the interviewer or to the questionnaire. While doing so, they may be strongly influenced by the context. Especially in the case of interviews, the evaluation team has to create a friendly (empathetic) atmosphere, at least for the sake of achieving a high rate of answers and fast completion of the survey. The combination of the two introduces a systematic positive bias in the answers, which tends to overestimate the benefits of the intervention and to underestimate the role of external factors. This bias is prevented by relying on properly trained interviewers. - Sample selection biasPeople who agree to be interviewed may not be representative of the overall target audience. |
||||||||||||||
Recommendations
|
Analysis
.
This section is structured as follows:
- Key elements
- Analysis strategy
- Counterfactual
- External factors
- Exploratory and confirmatory analysis
- Validity
.
KEY ELEMENTS |
|||||||||
. An analysis is required to convert data into findings, which themselves call for a judgement in order to be converted into conclusions. The analysis is carried out on a question-by-question basis, in the framework of an overall design cutting across all questions. |
|||||||||
Data, evidence and findings |
|||||||||
Any piece of qualitative or quantitative information that has been collected by the evaluation team is called data, for instance:
A piece of information is qualified as evidence as soon as the evaluation team assesses it as reliable enough, for instance:
Findings establish a fact derived from evidence through an analysis, for instance:
Some findings are specific in that they include cause-and-effect statements, for instance:
Findings do not include value judgements, which are embedded in conclusions only, as shown below:
|
|||||||||
Strategy of analysis |
|||||||||
Four strategies can be considered:
The first strategy is the lightest one and may fit virtually all types of questions, for instance:
The three last strategies are better at answering cause-and-effect questions, for instance:
The choice of the analysis strategy is part of the methodological design. It depends on the extent to which the question raises feasibility problems. It is made explicit in the design table. |
|||||||||
Data processing |
|||||||||
The first stage of analysis consists in processing information with a view to measuring or qualifying an indicator, or to answering a sub-question. Data are processed through operations such as cross-checking, comparison, clustering, listing, etc.
Provisional findings emerge at this stage of the analysis. Further stages aim to deepen and to strengthen the findings. |
|||||||||
Exploration |
|||||||||
The exploratory analysis aims to improve the understanding of all or part of the evaluated area, especially when knowledge is insufficient and expertise is weak, or when surprising evidence does not fit available explanations. The exploratory analysis delves deeper and more systematically into the collected data in order to discover new plausible explanations such as:
The exploratory stage may not be needed for all questions. When such an analysis is carried out, brainstorming techniques are appropriate. The idea is to develop new plausible explanations, not to assert them. |
|||||||||
Explanation |
|||||||||
This next stage ensures that a sufficient understanding has been reached in terms of:
Depending on the context and the question, the explanation builds upon one or several of the following bases:
A satisfactory explanation (also called explanatory model) is needed for finalising the analysis. |
|||||||||
Confirmation |
|||||||||
The last stage of the analysis is devoted to confirming the provisional findings through a valid and credible chain of arguments. This is the role of the confirmatory analysis. To have a finding confirmed, the evaluation undertakes a systematic self-criticism by all possible means, e.g. statistical tests, search for biases in data and analyses, check for contradictions across sources and analyses. . |
|||||||||
ANALYSIS STRATEGY |
|||||||||
. |
|||||||||
Cause-and-effect analysis |
|||||||||
What does this mean?Approach through which the evaluation team asserts the existence of a cause-and-effect link, and/or assesses the magnitude of an effect. |
|||||||||
Attribution or contribution- Attribution analysisAttribution analysis aims to assess the proportion of observed change which can really be attributed to the evaluated intervention. It involves building a counterfactual scenario. |
|||||||||
- Contribution analysisContribution analysis aims to demonstrate whether or not the evaluated intervention is one of the causes of observed change. It may also rank the evaluated intervention among the various causes explaining the observed change. Contribution analysis relies upon chains of logical arguments that are verified through a careful confirmatory analysis.
|
|||||||||
Analytical approaches- CounterfactualThe approach is summarised in the diagram below: ![]() The "policy-on" line shows the observed change, measured with an impact indicator, between the beginning of the evaluated period (baseline) and the date of the evaluation. For instance: local employment has increased, as has literacy. The impact accounts for only the share of this change that is attributable to the intervention. |
|||||||||
- Case studiesAnother analytical approach relies on case studies. It builds upon an in-depth inquiry into one or several real life cases selected in order to learn about the intervention as a whole. Each case study monograph describes observed changes in full detail. A good case study also describes the context in detail and all significant factors which may explain why the changes occurred or did not occur. - Causal statementsThe approach builds upon documents, interviews, questionnaires and/or focus groups. It consists in collecting stakeholders' views about causes and effects. Statements by various categories of stakeholders are then cross-checked (triangulated) until a satisfactory interpretation is reached. A panel of experts may be called to help in this process. |
|||||||||
- Meta-analysisThis approach builds upon available documents, for instance:
In performing meta-analyses, the evaluation team needs to (1) assess the quality of information provided by the reviewed documents, and (2) assess the transferability to the context of the evaluation underway. |
|||||||||
- GeneralisationThe first two approaches (counterfactual and case studies) have the best potential for obtaining findings that can be generalised (see external validity), although in a different way. Findings can be said to be of general value when all major external factors are known and their role is understood. Counterfactual approaches build upon explanatory assumptions about major external factors, and strive to control such factors through statistical comparisons involving large samples. Case studies strive to control external factors through an in-depth understanding of cause-and-effect mechanisms.
|
|||||||||
RecommendationThe evaluation team should be left with the choice of its analysis strategy and analytical approach. |
|||||||||
Cause-and-effect questions |
|||||||||
What does this mean?Cause-and-effect questions pertain to the effects of the evaluated intervention. They are written as follows:
These questions call for an observation of change, and then an attribution of observed change to the intervention, or an analysis of the intervention's contribution to observed changes. Causality and evaluation criteriaEffectiveness and impact questions tend to be cause-and-effect questions in the sense that they link the evaluated intervention (the cause) to its effects.
The latter example involves causes and effects, but only in a prospective and logical manner. The evaluation team is not expected to assert the existence of cause-and-effect links and/or to assess the magnitude of actual effects.
Questions pertaining to the EC value added may be cause-and-effect questions if the evaluation team attempts to assert the existence or the magnitude of an additional impact, due to the fact that the intervention took place at European level. Caution! Questions which do not require a cause and effect analysis do nevertheless call for a fully-fledged analysis covering all or part of data processing, exploration, explanation and confirmation. . |
|||||||||
Counterfactual |
|||||||||
. What does this mean?The counterfactual, or counterfactual scenario, is an estimate of what would have occurred in the absence of the evaluated intervention.
|
|||||||||
What is the purpose?By subtracting the counterfactual from the observed change (factual), the evaluation team can assess the effect of the intervention, e.g. effect on literacy, effect on individual income, effect on economic growth, etc. |
|||||||||
Comparison groupOne of the main approaches to counterfactuals consists in identifying a comparison group which resembles beneficiaries in all respects, except for the fact that it is unaffected by the intervention. The quality of the counterfactual depends heavily on the comparability of beneficiaries and non-beneficiaries. Four approaches may be considered for that purpose. Randomised control groupThis approach, also called experimental design, consists in recruiting and surveying two statistically comparable groups. Several hundred potential participants are identified and asked to participate or not in the intervention, on a random basis. The approach is fairly demanding in terms of preconditions, time and human resources. When the approach is workable and properly implemented, most external factors (ideally all) are neutralised by statistical rules, and the only remaining difference is participation in the intervention. |
|||||||||
Adjusted comparison groupIn this approach a group of non-participants is recruited and surveyed, for instance people who have applied to participate but who have been rejected for one reason or another. This approach is also called quasi-experimental design. In order to allow for a proper comparison, the structure of the comparison group needs to be adjusted until it is similar enough to that of participants as regards key factors like age, income, or gender. Such factors are identified in advance in an explanatory model. The structure of the comparison group (e.g. per age, income and gender) is adjusted by over- or under-weighting appropriate members until both structures are similar. |
|||||||||
Matching pairsIn this approach a sample of non-participants is associated with a sample of beneficiaries on an individual basis. For each beneficiary (e.g. a supported farmer), a matching non-participant is found with a similar profile in terms of key factors which need to be controlled (e.g. age, size of farm, type of farming). This approach often has the highest degree of feasibility and may be considered when other approaches are unpractical. |
|||||||||
Generic comparisonThe counterfactual may be constructed in abstracto by using statistical databases. The evaluation team starts with an observation of a group of participants. For each participant, the observed change is compared to what would have occurred for an "average" individual with the same profile, as derived from an analysis of statistical databases, most often at national level. |
|||||||||
Comparative approachesDifferent forms of comparison exist, each with pros and cons, and varying degrees of validity.
|
|||||||||
Strengths and weaknesses in practiceA well-designed comparison group provides a convincing estimate of the counterfactual, and therefore a credible base for attributing a share of the observed changes to the intervention. A limitation with this approach stems from the need to identify key external factors to be controlled. The analysis may be totally flawed if an important external factor has been overlooked or ignored. Another shortcoming stems from the need to rely upon large enough samples in order to ensure statistical validity. It is not always easy to predict the sample size which will ensure validity, and it is not infrequent to arrive at no conclusion after several weeks of a costly survey. |
|||||||||
ModellingThe principle is to run a model which correctly simulates what did actually occur in reality (the observed change), and then to run the model again with a set of assumptions representing a "without intervention" scenario. In order to be used in an evaluation, a model must include all relevant causes and effects which are to be analysed. These are at least the following:
Complex models (e.g. macro-economic ones) may include hundreds of causes, hundreds of effects, hundreds of mathematical relations, hundreds of adjustable parameters, and complex cause-and-effect mechanisms such as causality loops. When using a model, the evaluation team proceeds in three steps:
Modelling techniques are fairly demanding in terms of data and expertise. The workload required for building a model is generally not proportionate to the resources available to an evaluation. The consequence is that the modelling approach is workable only when an appropriate model and the corresponding expertise already exist. . |
|||||||||
EXTERNAL FACTORS |
|||||||||
. |
|||||||||
What are they? |
|||||||||
Factors which are embedded in the context of the intervention and which hinder or amplify the intended changes while being independent from the intervention itself. External factors are also called contextual, exogenous or confounding factors. |
|||||||||
Why are they important? |
|||||||||
|
|||||||||
Typical examples |
|||||||||
Factors explaining participation in the intervention:
Factors explaining the achievement of specific impacts:
Factors explaining global impact
When dealing with such external factors, the evaluation may usefully consult the contextual indicators that are available on the web. |
|||||||||
How can they be identified? |
|||||||||
In a given evaluation, external factors are potentially numerous and it is crucial to highlight the most important ones. The following approaches may help:
Identifying external factors is one of the main purposes of the exploratory analysis. |
|||||||||
Recommandations |
|||||||||
Do not try to identify all possible external factors when clarifying the intervention logic in the structuring phase of the evaluation. They are simply too numerous. This task should be undertaken only when working on a given evaluation question, and only if the question involves a cause-and-effect analysis. . |
|||||||||
EXPLORATORY AND CONFIRMATORY ANALYSIS |
|||||||||
. |
|||||||||
Exploratory analysis |
|||||||||
What does this mean?If necessary, the evaluation team delves into the collected data in order to discover new plausible explanations such as:
What is the purpose?
How to carry out the exploratory analysisThe analysis explores the set of data (quantitative and qualitative) with a view to identifying structures, differences, contrasts, similarities and correlations. For example, the analysis involves:
The approach is systematic and open-minded. Brainstorming techniques are appropriate. Ideas emerge through the first documentary analyses, interviews, and meetings. The exploration may continue through the field phase. |
Confirmatory analysis |
What does this mean?Provisional findings progressively emerge during the first phases of the evaluation team's work. They need to be confirmed by sound and credible controls. That is the role of the confirmatory analysis. |
What is the purpose?
How is a confirmatory analysis performed?For a finding to be confirmed, it is systematically criticised by all possible means, e.g.:
Recommendations
. |
VALIDITY |
. |
What does this mean? |
Validity is achieved when:
|
What is the purpose? |
A lack of validity may expose the evaluation to severe criticism from those stakeholders who are dissatisfied with the conclusions and recommendations, and who will point out any weaknesses they may have found in the reasoning. Validity is part of the quality criteria. It should be given an even higher level of attention when the intended users include external stakeholders with conflicting interests. |
External validity |
Quality of an evaluation method which makes it possible to obtain findings that can be generalised to other groups, areas, periods, etc. External validity is fully achieved when the evaluation team can make it clear that a similar intervention implemented in another context would have the same effects under given conditions. Only strong external validity allows one to transfer lessons learned. External validity is also sought when the evaluation aims at identifying and validating good practice. |
Internal validity |
This is the quality of an evaluation method which, as far as possible, limits biases imputable to data collection and analysis. Internal validity is fully achieved when the evaluation team provides indisputable arguments showing that the findings derive from collected facts and statements. Internal validity is a major issue in the particular case of cause-and-effect questions. When striving to demonstrate the existence and/or to assess the magnitude of an effect, the evaluation team is exposed to risks such as:
|
Construct validity |
This is the quality of an evaluation method which faithfully reflects the changes or needs that are to be evaluated. Construct validity is fully achieved when key concepts are clearly defined and when indicators reflect what they are meant to. Construct validity is threatened if the evaluation team does not fully master the process of shifting from questions to indicators. Construct validity is also at risk when the evaluation team uses indirect evidence like proxies. |
Recommandations |
|
Judgment
.
This section is structured as follows:
.
CONCLUSIONS AND LESSONS |
. What does this mean?Conclusions provide clear answers to the evaluation questions. They incorporate value judgements. Lessons are transferable conclusions to subsequent cycles of the same intervention or of other interventions. What is the purpose?
How should they be formulated?Conclusions and lessons stem from the preceding steps as follows: Conclusions answer the questionsThe questions asked at the beginning of the evaluation find their answers by means of the conclusions. A conclusion may answer several questions and several conclusions may answer a single question. Provided that all the questions asked have been answered, the evaluation team can present additional conclusions to take into account unexpected and important information and results. The conclusions follow from data and findingsUpon writing a conclusion, what is being judged is one aspect of the intervention, for example: a strategic guideline (Is it relevant?), a practice (Is it efficient?), an expected effect (Was it obtained?), or an unexpected one (Is it positive?). Thus, conclusions stem from collected data and evidence, from analysis and interpretations performed, from findings and new knowledge generated. Conclusions are based on judgement criteriaTo formulate its conclusions, the evaluation team applies the judgement criteria (also called "reasoned assessment criteria") that were agreed upon in the first phase (desk) of the evaluation. Data collection and analysis are structured according to these criteria. As long as this is possible, the findings are compared against targets.
At the stage of the draft final report, the evaluation team may have to refine its judgement criteria and targets. In such a case, the issue is discussed with the reference group. A lesson is a transferable conclusionA lesson is a conclusion that can be transferred to subsequent cycles of the same intervention or to other interventions.
|
How should they be presented?One chapter of the report introduces the conclusions relative to each question, as well as the conclusions that emerge from points not raised by the questions. The conclusions are organised in clusters in the chapter in order to provide an overview of the assessed subject. The chapter does not follow the order of the questions or that of the evaluation criteria (effectiveness, efficiency, coherence, etc.) It features references to the sections of the report or to annexes showing how the conclusions derive from data, interpretations, analysis and judgement criteria. The report includes a self-assessment of the methodological limits that may restrain the range or use of certain conclusions. A paragraph or sub-chapter picks up the 3 or 4 major conclusions organised by order of importance, while avoiding being repetitive. This practice allows to better communicate the evaluation messages that are addressed to policy makers within the Commission. The conclusion chapter features not only the successes observed but also the issues requiring further thought on modifications or a different course of action. Suggestions
. |
RECOMMENDATIONS |
. What is this?The recommendations are derived from conclusions.They are intended to improve or reform the intervention in the framework of the cycle under way, or to prepare the design of a new intervention for the next cycle. What is the purpose?
How to draft and present themThe recommendations must be related to the conclusions without replicating them. A recommendation derives directly from one or more conclusions. The recommendations must be clustered and prioritised. The report mentions the addressees of the recommendations, e.g. EC Delegation, services in charge of designing the next intervention, etc. The recommendations are useful, operational and feasible, and the conditions of implementation are specified. Wherever possible and relevant, the main recommendations are presented in the form of options with the conditions related to each option, as well as the predictable consequences of the implementation of each option. The recommendations are presented in a specific chapter. This chapter highlights the recommendations derived from the three or four main conclusions. |
How to promote themThe recommendations are valuable as far as they are considered and, if possible, taken up by their addressees. To promote their take-up, the manager drafts a fiche contradictoire in order to:
Advice
. |
ETHICAL PRINCIPLES |
. What is this?The conclusions include value judgements on the merits and worth of the intervention. This dimension of the evaluation exercise is particularly sensitive and the evaluation team has therefore to respect specific ethical principles. What is the purpose?
What are the main principles?- Responsibility for the judgementThe conclusions are primarily a response to questions. Members of the group are partially responsible for the judgement in so far as they orientate it through the evaluation questions they validate. The external evaluation team also intervenes in the preparation of the judgement by making proposals to define the questions, clarify the judgement criteria and set the targets. In the synthesis phase, the evaluation team applies the judgement criteria agreed on, as faithfully as possible, and produces its own conclusions. The conclusions are discussed within the reference group but remain the entire responsibility of the evaluation team. As part of the quality assurance process, the evaluation manager can require sounder justification of a judgement, or better application of an agreed judgement criterion. By contrast, he or she cannot require the removal or amendment of a conclusion if it is methodologically sound. - Legitimacy of the judgementThe questions and criteria take into account the needs and point of view of the public institution that initiated the evaluation. The members of the reference group contribute different points of view, which reinforces the legitimacy of the evaluation. During the desk phase the evaluation team holds interviews, which may enable it to identify other points that were not expressed by the reference group members. It makes them known in reference group meetings and may take them into account in the judgement criteria. More generally, the evaluation team has a responsibility to bring to light important findings and judgement criteria which have arisen during the evaluation process, even if they are not covered by the evaluation questions, provided that such points are legitimate. A point of view is legitimate if:
- Impartiality of the judgementThe impartiality of the judgement concerns the entire evaluation, that is, the choice of questions and judgement criteria, the determination of targets and the formulation of conclusions. The entire process is exposed to risks of partiality, for example:
When there are differences in the way of judging, in the judgement criteria or in the target levels, impartiality consists in:
In case of divergence, a solution may consist in judging in relation to several criteria and/or formulating several conclusions that correspond to different points of view. This solution has the drawback of diluting the conclusions and thus of making the evaluation less conclusive. It is often preferable to make choices and to explain them transparently. - Protection of peopleThe conclusions concern the merits of the evaluated intervention, not the people who implement it or benefit from it. Individuals' names are cited only when this enhances the credibility of the evaluation. The evaluation team must respect people's right to provide information in confidence and ensure that sensitive data cannot be traced to its source. Before citing a person or organisation, the evaluation team or any other evaluation actor anticipates and avoids the risks involved for that person or organisation. Evaluations sometimes uncover evidence of wrongdoing. Bad professional practices are never reported in a traceable way. However, the evaluation team member who encounters illegal or criminal acts deals with them as any other citizen should do. In the later case, the issue should be discussed with the evaluation manager. |
Quality assurance
.
This section is structured as follows:
.
KEY ELEMENTS |
. What is this?Establishing quality check-points at the main phases of the process, defining the responsibilities of quality assurance, and establishing the rules of the game in case of quality defects. Quality assurance is organised on a step-by-step basis: successive aspects of quality are secured at each step, thus creating a sound base for the subsequent step. What is the purpose?
Rules of the gameThe rules of quality assurance are specified in the terms of reference and pointed out when the external evaluation team is engaged. These rules concern:
The quality assurance process benefits from the contribution of all actors whilst limiting the potential conflicts that might arise between them, e.g.
Key stepsInception reportIn the inception report stage, quality assessment is used to ensure that the evaluation team has clearly understood the scope and aims of the evaluation, as well as the logic of the evaluated intervention and the questions to answer. First phase report (desk)In the first phase report stage (desk), quality assessment is used to check the quality of the documentary analysis as well as the appropriateness of the proposed method for the field phase, including the indicators, data sources and analytical approach. Final reportIn the final report stage, quality assessment checks whether the evaluation provides valid and impartial answers to the questions asked, and whether the form of the report is suited to the targeted users' needs. At this stage the quality assessment concerns first the draft version of the report and then the final version. |
Key playersIt is expected that all key players involved in the quality assurance process have a sufficient evaluation capacity, except reference group members. If a key player is not qualified enough, he/she ensures that appropriate support is provided. Evaluation team leaderThe evaluation team leader is primarily concerned with preventing major risks that threaten quality, e.g. overlooking a major question, not consulting an important stakeholder, elaborating upon unreliable data, judging in a partial way. He/she also ensures that each report is submitted to a detailed quality check before it is released to the evaluation manager. Quality assessor(s)The company or consortium in charge of the evaluation contract appoints one or more quality assessors who carefully check each report for quality before it is released to the evaluation manager. The quality assessor should be experienced in evaluation and should not belong to the evaluation team. Evaluation managerThe evaluation manager has the ultimate responsibility for assessing whether or not the methodological quality of a report is sufficient for allowing the next step to be taken. In this process he/she takes advice from the reference group members, and relies upon technical support as far as necessary. Reference groupReference group members receive all draft reports for comments. They may comment on the factual bases substantiating the evaluation, which is a contribution to quality assurance. When their comments apply to the content of conclusions, they are passed on to the evaluation team which freely decides on whether or not to take them into account. . |
APPROVAL OF DELIVERIES |
. What is this?Formal recognition by the evaluation manager that a delivered document has an adequate content and meets the applicable quality criteria in methodological terms. What are the approval steps?Generally, a document is approved in the following way:
|
Which documents are approved and by whom?The following documents require approval:
Approval is generally performed by the evaluation manager and may be confirmed by his/her superior. The reference group members validate the set of evaluation questions, i.e. group members formally accept, or at least do not reject, the set of evaluation questions. If necessary, divergences are resolved by the evaluation manager. In that case dissenting views are noted in subsequent reports. |
Rules of approval and independenceThe approval process does not contradict the external evaluation team's independence. In this respect the requests for amendments made during the approval exercise need to be differentiated:
. |
QUALITY ASSURANCE SUPPORT |
. What is this?Assessing the quality of an evaluation is a multifaceted task. The commissioning body and the stakeholders are in the best position to assess whether their needs have been treated impartially. The experts and those who implement the intervention are in a good position for assessing the relevance of data collected and the fairness of interpretations. It is the evaluation manager's role to assure that the evaluation has sound methodological and factual bases. The manager receives a quality assurance support if he/she lacks training or experience. Who can play this role?Within the EC headquarter Evaluation Unit, managers have the required training and skills to assess the methodological quality of reports. All evaluations are monitored by two members of the Unit, both of whom participate in quality assessment. In the case of devolved evaluations, the manager uses the quality criteria proposed by this site as well as the indications given to fill out the quality assessment grid, if necessary with the help of the Evaluation Unit. The Evaluation Unit can perform cross-cutting quality assessments autonomously. |
Evaluation tools
This section presents the following evaluation tools:
- Toolbox (overview of main tools)
- Problem diagram
- Objectives diagram and the effect diagram
- Decision diagram
- Interview
- Focus group
- Survey
- Expert panel
- Case study
- Context indicators
- SWOT (Strenghts, Weakness, Opportunities, Threats)
- Multi-criteria analysis
- Cost effective analysis
- Cultural and social analysis
Toolbox
.
This section is structured as follows:
- What types of tools
- When to use which tools
- Which combination of tools
- What specific constraints and requirements
- Check lists for tool's implementation
- Toolbox and direction for use
.
What type of tools? |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
. The evaluation's four tasksEvaluations are usually divided into four tasks which are not strictly sequential: the evaluation's organisation (based on the evaluation questions wording leading to the overall assessment), the collection of quantitative and qualitative information, the analysis of the information collected, and the evaluation questions assessment leading to the formulation of conclusions and recommendations.
The elaboration of an evaluation methodology based on the evaluation questions (which are designed for the development of an overall assessment) is crucial in the selection of tools used for each of the evaluation's four stages. |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Importance of the documentary stageAlthough no tool presented in this methodology is specifically dedicated to this purpose, the collection of information from the European Commission services and on-site (particularly the information collection using CRIS database) is a component of the evaluation process.
|
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Presentation of the toolsThirteen tools have been developed. They are usually familiar to the evaluators. Their specificities are described here. Objectives diagram and impact diagramThe objectives diagram displays the classification of the objectives to be achieved for the strategy implementation, from the European Union's global objective to activities carried out for operational programmes. The impact diagram displays the classification of activities, outcomes and expected impacts. The expected impacts are the objectives in terms of results. Problem diagramProjects and programmes in development assistance aim at satisfying priority needs through the resolution of a range of issues. It is theoretically possible to construct a diagram taking the shape of a tree, with the trunk (the core problem), roots (the causes) and branches (the consequences and impacts). Decision diagramThe decision diagram displays the process during which the strategic objectives and the overall co-operation policies with developing countries, which are defined by the European Union's assistance agreements, are converted into short-term and medium-term bilateral co-operation decisions. InterviewThe interview collects information from stakeholders and beneficiaries throughout the evaluation stages: facts and verification of facts, opinions and points of view, stakeholders analyses and suggestions. Focus groupThe focus group is a means of discussing information, opinions and judgements already collected. The tool explains why opinions have been expressed (and the analyses supporting them) and checks their consistency. Focus groups are frequently used to collect the beneficiaries opinions concerning their participation in a programme and what they had drawn from it (positive and negative aspects). They are an alternative to interviews. Whatever their usage, the focus group's specificity is a collection of opinions which have been moderated by an in-depth discussion rather than a collection of spontaneous opinions. SurveyThe survey collects comparable answers from a sample of the population. When the sample is representative, the survey displays statistical measures which can be useful for quantified indicators. Case studyCase studies are the preferred evaluation tool when "how" and "why" questions are being posed, because they allow a detailed examination of the actual elements in line with the evaluation goals. In contexts allowing or requiring it, the case(s) can be selected to yield general conclusions for the overall evaluation. Expert panelThe expert panel is a group of independent specialists, recognised in at least one of the fields addressed by the programme under evaluation. The panel yields a collective assessment which is nuanced, argued and supported by the knowledge and experience of the experts. SWOTSWOT analysis combines the study of the strengths and weaknesses of an organisation, a geographical area, or a sector, with the study of the opportunities and threats to their environment. Frequently used in ex ante evaluations, it can also be used in ex post evaluations to assess the orientations taken. Context indicatorsThe tool ranks a country through the comparison of its context indicators with that of other countries. A context indicator is a datum which produces simple and reliable information describing a variable relative to the context. The tool evaluates development dynamics through the comparison of the level and evolution of a country's main indicators with that of other countries with similar contexts. Multicriteria analysisIn ex ante situation, multicriteria analysis is a decision-making assistance tool. In ex post evaluations, it usefully contributes to the formulation of a judgement based on a range of heterogeneous criteria. Cost-effectiveness analysisThe tool identifies the economically most efficient way to fulfil an objective. It compares the efficiency of projects or programmes with comparable impacts. It usefully contributes to the formulation or validation of a judgement on the selection of the most efficient projects and programmes. Cultural and social analysisIn country evaluations, the cultural and social analysis identifies the constitutive components of social, ethnic, religious and interest groups. It also highlights all the values shared in a society as well as its internal divisions. |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Absence of statistical toolsIn the context of assistance to developing countries, statistical data are often difficult to collect and their relevance is limited by the delays in their publication and a weak reliability. The available data are general and descriptive, and only allow for straightforward analysis. Thus, it is common practice to develop tools based on easily available data. Rules for the methodology's applicationGenerally speaking, the feasibility of field work or the limitations of such a task should be checked. Indeed, specific contexts within the country where the study is to be carried out, such as logistical and implementation costs constraints, can constrain the choice of tools.
In short, the evaluation team should use several tools and choose the most efficient combination of tools for the specific context of the evaluation. . |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
When to use which tools? |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
. The tools and the evaluation's four functionsA first selection can be made when the tools are classified within the four tasks of the evaluation. Organisation toolsThe organisation of the evaluation can usefully be supported by a series of tools called organisation tools. In homogeneous project or programme evaluations, the normal organisation tool is the logical framework which describes the objectives (overall and specific) of the intervention, the issues to which the evaluation answer and the expected outcomes. A problem diagram can usefully complement the logical framework. Evaluations where the scope includes a range of heterogeneous objectives and activities (such as geographic evaluations), theoretically require three tools, the objectives diagram and the impact diagram being the core organisation tools. In complex programmes or strategies, the problem diagram brings precision on the relevance of objectives, identifies the goals and issues of the interventions, as well as the problems neglected by the evaluation's objectives. The decision diagram complements the objectives diagram with information about the reasons for the programme's orientations, and especially the basis for the rejection and negotiation of options. These tools are also useful for the analysis and judgement stages (polyvalent nature of the tools). Collection, analysis and judgement toolsDuring these three evaluation stages, a large range of tools is available, complementary to each other and/or polyvalent. As a consequence, numerous factors are taken into account to make an efficient choice. Although no rule securing an optimal choice exists, a logical process can facilitate the development of a homogeneous methodology which will provide well-grounded answers to the evaluation questions. |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Polyvalent nature of the toolsThe resources allocated for an evaluation are not sufficient for the implementation of all the tools mentioned previously. Choices must be made with respect to the evaluation's priorities and the maximisation of the use of resources. The wider the scope and questioning, the greater the risk of dispersion, which means that the evaluation team must ensure that their observations and analyses provide answers to the most essential issues of the evaluation. The evaluation team should remember that several tools are usually relevant for the same evaluation stage and can be used to confirm the information, and that the tools are often polyvalent and can answer several questions at the same time. Selection criteria for the toolsBesides the tools specific functions and their ability to be implemented in one of the four stages of the evaluation, other selection criteria should be examined.
The table below grades each tool using the five criteria and indicates those demanding particular attention. Each criteria is awarded a grade from 1 to 3. Grade 1 means that the criteria does not constitute a particular problem for the tool; grade 3 means that it is recommended that the feasibility of the tool's implementation should be checked with regards to this criteria. The tools' specific requirements
A grade 3 does not mean that the tool should not be used. Indeed, the priority should be given to the tools providing the best answers to the evaluation questions, and then, the evaluator can check the possibility of using them in the context and with the available resources of the evaluation. Development of a homogeneous methodology
In essence, the methodology should be constructed with a range of available tools and take into account their advantages and limitations, the conditions for their implementation in the context of development assistance evaluations, the prerequisites for their implementation and the limits of their findings due to the context. Analysis of the toolsObjectives diagram and impact diagram
Problem diagram
Decison diagram
Interview
Focus group
Survey
Case study
Expert panel
SWOT
Context indicators
Cultural and social analysis
Multicriteria analysis
Cost-effectiveness analysis
. |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Which combination of tools? |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
. The combination rationaleSome tools require the implementation of other tools prior to their use. Such is the case when a tool yields useful information for the implementation of another tool, or when bringing a different viewpoint to the analysis, it strengthnens or nuances the conclusions reached which another tool. Before taking the decision to implement complex tools (such as judgement tools), the evaluator should check whether preliminary information which is capable of improving their performance are available, and find the optimum tool capable of yielding such information. Tools required by other tools ![]() The table is indicative and other combinations can be developed in particular contexts. Usually, collection tools (interviews, focus group and surveys) are the most frequently combined with analysis and judgement tools, because the latter require specific information for their implementation. Thus, special care should be given to analysis and judgement tools, so as to ensure the homogeneity of the methodology and the maximum performance of the tools throughout the evaluation. |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
The tools' frequency of useThe table shows that the interview is the tool the most frequently used by other tools, including collection tools (which is not surprising). Conversely, case studies are seldom used and belong to the category of meta-tools requiring the support of all the collection tools and, if needs be, analysis tools.
Example of combinationTesting of two country evaluation tools - the survey and the focus group - during the evaluation in Benin.
Four focus group investigations:
The last 2 focus group investigations were set up to complement the outcomes of the 2 focus groups with beneficiaries and that of the surveys. The goal was to confront different perspectives on a same situation where a few dysfunctional elements could be raised. . |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
What specific constraints and requirements? |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
. Limitations and risks depend on the constraints and requirements specific to each of the available tools, whatever their categories. Possible constraintsAccess to information and sources of informationMost of the tools require relatively straightforward access to baseline documentation (Objectives diagram and impact diagram, Problem diagram, Decision diagram, Case study, Context indicators). Some tools strongly depend on the representative nature of the interlocutors and stakeholders, on their availability and co-operation, such as the Interview, the Focus group and the Survey. Quality of the collected informationThe quality of the collected information is crucial, and can be influenced by:
Knowledge of specific techniquesA good knowledge of the tool is needed for all the tools, and especially for: the Survey, the Cultural and social analysis, SWOT, Cost-effectiveness analysis, Multicriteria analysis. |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
PrerequisitesHuman resourcesThe availability of acknowledged experts must be guaranteed whatsoever. It is a key condition for: the Cultural and social analysis, the Expert panel, SWOT, Cost-effectiveness analysis, Multicriteria analysis and Focus group. CostsImplementation costs should be examined particularly for the following tools: Cost-effectiveness analysis, Multicriteria analysis, Survey, Case study. Time spanAn appropriate time schedule is important for all tools, and particularly for the implementation of: Cost-effectiveness analysis, Multicriteria analysis, Survey, Case study and Interview. Impact of preparatory stagesWhen a tool requires a preparatory information collection, its time span, cost and human resources should not be under-estimated. If needs be, access to sites and beneficiaries availability must be checked (Case study). . |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Check list for the tool's implementation |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
. Check list for the evaluation team
Check list for contract managers
|
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Toolbox and directions for use |
|||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
The aim of the evaluation is to produce operational recommendations supported by solid conclusions which are based on clear judgement criteria, solid and concrete information and rational argumentation.
|
Problem diagram
SUMMARY |
||||||||
. |
||||||||
Why are this tools used in evaluation? |
||||||||
. The analysis of problems is a means to test the validity of the objectives of a project, a programme or a strategy. As a programme aims at solving a range of problems, the evaluation should be concerned with the validity of its analysis. The evaluators should therefore check:
. Programmes and strategies depend on the analysis of the situation. This analysis presents the primary data relative to the problems addressed by the programmes and strategies, and the information about the context (economic, social, cultural, institutional, environmental) in which they will be implemented. This crucial analysis identifies:
Therefore, the evaluation team can be asked to assess the quality of the analysis, and the conformity between the analysis of the situation and the adopted strategy (or programme). The problem diagram, as an ex post construction, can be one of the tools used to check the coherence and the relevance of the analysis in respect of the main contextual problems. . |
||||||||
What are the possible uses of these diagrams? |
||||||||
. When it is impossible to directly establish objectives diagrams, problem diagrams play an essential role in the organisation stage of the evaluation. Problem diagrams present a summarised vision of the situation to be improved, partially at least, by the strategy. The classified objectives included in the strategy should be deduced from the problems diagrams. .
. The reconstruction of the problem diagram includes a step which differentiates between context problems and intervention problems. As a consequence, the diagram resulting from this selection should be completely convertible into a logically reconstructed objectives diagram, i.e. the higher-order overall objective corresponds to the core problem, and each row of subordinated objectives to its equivalent row of problems. . |
||||||||
How is the problem diagram constructed? |
||||||||
. Stage 1: How to identify the problems?Record the references to teh problems found in the documentation: quotations from the evaluation's baseline documents are used to identify problems which may not be systematically depicted as such in the texts. Sometimes, problems are identified as assistance objectives or impacts targeted by these objectives. Problems are thus expressed as:
The problems directly targeted by the intervention may not be explicitly identified. Main problems and context problems can be intermingled, which complicates the construction of the diagram. The evaluator will only be able to differentiate the two types of problems after the completion of the diagram. The crucial stage of the process is the identification of the core problem among the variety of the selected problems. Three situations can be encountered:
In the last situation, the evaluator plays the role of the planning manager, and decides which of the problems will be the core problem. He/She may:
The selection of the core problem should be conducted concurrently with the classification of the problems into levels. Indeed, the selection of the core problem should be supported and justified by the coherence of the whole diagram. Stage 2: How to construct a problem diagram?
Usage highlights the fact that these two types of problems/causes are the easiest to identify, whereas intermediary causes are the hardest to determine and classify. Thus, it is recommended that the development of the diagram starts with its extremities at the same time.
Test the temporary diagramWhere possible, the authors of the documentation referred to above should test the diagram in order to validate the classification of the problems by rows and links. The aim is to check that the evaluator's interpretation reflects the authors' intention correctly. If the authors are not available or, in order to complement their original contribution, the evaluator should consider asking for the participation of other actors responsible for the drafting process. Contacting the authors is usually possible when the documentation is recent and the authors are still in their position or contactable in one of the services. This task is challenging when the documentation is old and its authors are not easily identifiable or contactable. Develop the final version of the problem diagramThe final version takes into account the opinions collected during the testing of the temporary diagram. It is an accurate account of the initial intentions logically reconstructed of the European Commission, taken from official documentation. . |
||||||||
What are the preconditions for its use? |
||||||||
. The need for human resources varies with the tasks to be achieved and the situation encountered: reconstruction of the problem diagram with the analysis of the situation displayed in the documentation; retrospective reconstruction of the problem diagram as it should have been established to support the relevance of the strategy or programme under assessment. Most of the graphic problems can be solved with software like MS PowerPoint. The evaluator can also benefit from specific software designed to assist decision-making, such as MS Visio, TeamUp-PCM for Windows, and Palisade PrecisionTree.
. |
||||||||
What are the advantages and limitations of the tool? |
||||||||
.
|
Detailed presentation
.
This section is structured as follows:
- What is a problem diagram?
- Why and when?
- How should the problem diagram be constructed?
- In country evaluation
- Examples
- Bibliography
.
WHAT IS A PROBLEM DIAGRAM? |
|||||||||||||
. WHAT DOES A PROBLEM DIAGRAM REPRESENT?What is the definition of a problem?In development assistance, projects and programmes have two targets: to satisfy priority needs or achieve a specific goal, and in this view to solve a range of hindering problems. Generally, the success of a policy or strategy depends on the resolution of difficult issues for the partners and potential beneficiaries. Problems seldom appear separately. Cause-and-effect relationships often link them and constitute a system, which can be presented as a tree-like diagram. The importance of problems varies. It is thus theoretically possible to identify a core (or central) problem and derive from it a range of causes and effects. Chart of the problem diagramWhen a core problem is identified, the diagram takes the shape of a complete tree, provided with a trunk (the core problem), roots (the causes) and branches (the consequences and impacts). Standard charts illustrate the core problem in the centre of the diagram, causes at the bottom, and consequences at the top. The diagram is read from the bottom upward. Standard chart: the complete problem tree
![]()
A strategy, programme or project is all the more effective when it studies the fundamental causes of the problem(s) to be solved. Thus, once the core problem is identified, the diagram only represents the causes and is displayed as an inverted tree. The diagram below illustrates a simple case, where three levels of problems are identified: fundamental causes, intermediary causes and the core problem. Simplified problem diagram ![]()
In the context of the European Commission's development assistance, the problem diagram illustrates the core problem with its immediate and in-depth causes. It may also depict the core problem's main consequences which are addressed by the assistance strategy (defined in the strategy or policy papers and the corresponding programmes). |
|||||||||||||
WHERE DOES THE PROBLEM DIAGRAM COME FROM?The problem diagram comes from objective-based management whose use expanded during the 1960s in the American economic sphere and in Western economies. Also called cause-and-effect diagram, it has been developed for a variety of fields, such as medical diagnosis, prevention and research on the causes of accidents, and total quality management. It is used as a country assistance decision tool. It is linked to the objectives tree in the context of Goal Oriented Project Planning (GOPP, originally called ZOPP in German for Zielorientierte Projektplanung) which derives from the logical framework developed by GTZ at the start of the 1980's. Used as a problem tree, it is:
Following GTZ's initiative, various public donors and operators (agencies and NGOs) have used the problem tree as a component of project planning and management. However, its application as a participatory and ownership tool does not always play the crucial role defined by the ZOPP designers. |
|||||||||||||
WHAT SHAPE CAN IT TAKE?Tree shape or complex diagramComplex and simplified problem trees are considered to be standard charts. They are subject to the following basic rules:
This simplification of reality is understandable in the context of the objectives-oriented participative planning process. It becomes a handicap in an evaluation because it avoids the complexity of reality and the rational limitations inherent in strategy and planning papers. Thus, the standard chart has been abandoned and the term "tree" replaced by "diagram". Complex problem diagram ![]()
This diagram represents a complex problem system.
Examples of more complex problem systems can be found, having two core problems and subdivisions of intermediate causes. In any case, such problem diagrams do not illustrate feedback links. Vertical or horizontal diagramDiagrams can either be vertically or horizontally oriented. In their vertical shape, they read from top to bottom, or from bottom to top. Usually, the core problem is located in the upper part of the diagram. In their horizontal shape, diagrams read from right to left, or left to right. Other shapesOther shapes have been developed, such as the fishbone diagram (also called Ishikawa diagram). Dedicated to the study of the cause resulting from a selected impact, it is also used in quality control process, medical diagnoses and accident prevention. |
|||||||||||||
WHAT IS A LOGICAL LINK?DefinitionProblems and causes illustrated in the diagram are related to each other with horizontal or vertical links. These links are called "logical" when expressing an inference relationship (induction or deduction) which has been validated by experience. They highlight the fact that:
The logic of the links between problems (causes) is essentially based on experience and development theories (originating from this experience). Experiments and theories usually provide good indicators for the determination of subordinate problems (or causes) in which a particular problem (or cause) in a certain row has its origin. However, in terms of development, experiments do not always result in the same conclusions and, as a consequence, should not be considered universal. This can be explained by the dependency of the problems on various factors, which are often not well-known. Even when experts agree on the relevance of a core problem, they may disagree on the definition of intermediate and in-depth causes. The content of core problems, such as poverty alleviation, is continuously debated and evolving. Verification of the diagram's logicThe logical links between problems (i.e. the problem diagram's logic) can be simultaneously checked with:
These two groups of actors can validate the diagram's logic in whole or in part. Conversely, a link which is contradicted by the experience of the experts and operators can be classed as illogical. The borderline between the logical and illogical nature of a situation is determined when there is a doubt, and contradictory or ambiguous experiments. . |
|||||||||||||
WHY AND WHEN? |
|||||||||||||
. WHEN SHOULD A PROBLEM DIAGRAM BE USED?Evaluation of standard projects and programmesAs a programme aims at solving a range of problems affecting potential beneficiaries (individuals or groups of beneficiaries), the evaluation should be concerned with the validity of its analysis. When a logical framework (ZOPP, Project Cycle Management) supports the establishment of the programme, its drafting usually includes the analysis of the problems. The evaluators should therefore check:
The assessment of the analysis of the problem and its corresponding diagram is a means to test the validity of the objectives system and its corresponding diagram, and to draw conclusions from them. If the programme does not have a logical framework, at least the problems to be addressed should have been analysed. In this context, it may be useful to develop a rational diagram based on a study of the definition of the programme and programming documentation. Evaluation of complex strategies and programmesIn complex strategies and programmes, the evaluation usually deals with a range of activities (projects and programmes) which lack explicit justification within a logical framework. The evaluation team may not be provided with an explicit and logically structured presentation of the objectives targeted by the donor, nor with thorough analysis of the difficulties which are supposed to be solved as a result of adopting such objectives. The available documentation is not always clear about how the various types of problems have been taken into account. The evaluation team must therefore reflect upon questions such as:
In addition, the methodology which was used to analyse the situation and select the problems should be presented, which is not always done prior to the evaluation. Answers to the following questions should be considered as part of the evaluation:
Evaluation of the analysis of the initial situationProgrammes and strategies depend on the analysis of the situation. This analysis presents the primary data relative to the problems addressed by the programmes and strategies, and the information about the context (economic, social, cultural, institutional, environmental) in which they will be implemented. This crucial analysis identifies:
Therefore, the evaluation team can be asked to assess the quality of the analysis, and the conformity between the analysis of the situation and the adopted strategy (or programme). The problem diagram, as an ex post construction, can be one of the tools used to check the coherence and the relevance of the analysis in respect of the main contextual problems. The evaluation team should not underestimate the amount of work, nor the level and variety of competences required (Human resources and working arrangement) to construct the problem diagram. Indeed, the evaluation team should not simply present a diagram with the problems developed in the documentation, but also reconstruct the problem diagram which should have been prepared during the development of the objectives diagram (and therefore, during the drafting of the strategy) |
|||||||||||||
PROBLEM OR OBJECTIVE: WHICH COMES FIRST?In project cycleIn the planning process and project management field (ZOPP, PCM), specialists place the preparation of the problem diagram before the objectives diagram. The typical sequence can be pictured as followed: Analysis of the problems in a simplified sequence of the project cycle ![]()
Problems analysis may be undertaken before stakeholder analysis, but should always precede the objectives analysis. Indeed, in Project Cycle Management (PCM), objectives are always deduced from the problem analysis. In the strategy policy co-ordination cycleIf the definition of the donor's activities in bilateral co-operation does not systematically result from the analysis of the problems, it will target one or more overall objectives (strategic and philosophical) which are explicitly or implicitly mentioned in the documentation. The donor (alone or in partnership) defines the main fields of the development assistance in accordance with these objectives. The chart below illustrates the logical position of the problem analysis within the strategy policy co-ordination cycle and, particularly, in relation with the definition of the objectives. Analysis of the problems in a simplified sequence of the strategy policy co-ordination cycle ![]()
The problem analysis and the resulting diagram derive from the definition of the intervention scope, and thus from the overall objectives of the donor (the European Commission). They also form the basis for the definition of specific objectives of the development assistance, which is to be provided during the period covered by the strategy. |
|||||||||||||
WHAT ARE ITS ADVANTAGES AND LIMITATIONS?Its advantagesPresentation of the problemsThe diagram presents the various problems and their relationship with the core problem through a system of rows. It displays the causal logical links between them or, conversely, the poor logic of these links. Main problems and contextual problemsIn the analysis of the situation, the diagram distinguishes the problems relating to the activity's context from the problems to be solved by the strategy and the planning. As a consequence, its construction requires the highlighting of priorities of the development assistance, and explains why certain problems are considered as important features of the strategy while others are not. Definition of the objectivesThe problem diagram enables the evaluator to:
It contributes to the organisation of the evaluation around a crucial question, which should be systematically answered: To what extent have the objectives been achieved?
Its limitationsThe diagram rationally organises the main difficulties on the basis of the analysis of the situation. It may entail the following limitations. Access to informationThe lack of data or difficult access to information may lower the quality of the analysis of the situation. The evaluator should therefore reflect upon the subsequent questions:
Quality of the analysisThe methodology for the analysis does not guarantee high quality of the data, for neither the methodology, nor the sources of information are usually mentioned in the strategic, political and programming papers. The evaluator must therefore systematically enlarge the assessment to include the sources of the analysis displayed in these documents. The methodology used, the nature of the main sources of information and the identity of the authors should also be noted. Main problem / contextual problemThe determination of the priorities for each problem results in their classification into two categories: contextual problems and main problems. Two questions must be addressed:
Selection of the main problemThe selection of the main problem, which is crucial for the construction of the diagram, is particularly challenging when the objectives of the activities are general and the whole range of the country's problems (or the region's) are to be considered. The documentation may show two main problems which lack links between them, or may appear insufficient for the determination of a single core problem. Tree-like illustrationThe standard setting of the challenges into a tree diagram illustrates a straightforward classification (which types of problem diagrams can be used?), which does not always highlight the complexity of the situation and the interactions between issues. Indeed, the construction of diagrams depends on graphical conventions, such as:
More sophisticated representations should therefore be tried. Knowledge of the situation in the country or the regionThe evaluation team may lack a sufficient understanding of the country or the region under consideration to assess the relevance of the analysis undertaken, the determination and the logical setting of the main problem. |
|||||||||||||
WITH WHICH TOOLS CAN IT BE COMBINED?Objectives diagramThe problem diagram is usually associated with the objectives analysis (and as a consequence, with the objectives diagram) and the development of the logical framework. The problem diagram is used for the construction of the objectives diagram, often through a simple transposition. Decision diagramThe analysis of the programme's development process (decision diagram) can be used to explain the reasons underpinning the determination of the main problems and the core problem. In intermediary evaluations of programmes, the evaluator may be asked to construct a decision and an objectives diagram, in addition to the problem diagram itself. Effect diagramIn an ex post evaluation or in an evaluation with ex post components, the initial analysis of the situation and the selection of the main problems can be compared with the analysis of the situation undertaken at the end of the period assessed. They can also be compared to an effect diagram resulting from the strategy's implementation, in order to assess the contribution of the development assistance to the resolution of problems. Other analytical toolsIn the context of an ex post reconstitution of the events, the problem diagram may be supported by other analytical tools, such as:
|
|||||||||||||
WHAT ARE THE PRECONDITIONS FOR ITS USE?Available informationWhen the problem diagram is found in the basic documentation of the programme, or when it can be directly deduced from a logical framework, it must be presented with a study of the choices underpinning its construction. This requires the consultation of the documentation which has supported its construction, and the interviewing of the actors directly involved in its development. When key actors and documentation are not availableSubsequent to the consultation with specialists, if required, the evaluator may present conclusions about the relevance of the problems system (illustrated in the diagram). When the diagram has been developed on the basis of the documentation at the disposal of the evaluator, the latter should:
If neither the archives documentation, nor the key actors are available, the evaluator must explain the hypotheses at the basis of the diagram's construction, by providing the reader with precise sources (quotations and comprehensive references of the documentation) underpinning the hypotheses. . |
|||||||||||||
HOW SHOULD THE PROBLEM DIAGRAM BE CONSTRUCTED? |
|||||||||||||
. STAGE 1: HOW TO IDENTIFY THE PROBLEMSRecord the references to the problems found in the documentationQuotations from the evaluation's baseline documents are used to identify problems which may not be systematically depicted as such in the texts. Sometimes, problems are identified as assistance objectives or impacts targeted by these objectives. Problems are thus expressed as:
What is to be recorded in the documentation? ![]() ![]() The preliminary analysis of quotations can reveal:
A thorough selection should therefore be undertaken. Distinguish context problems from main problemsAmong the various problems, the documentation often identifies problems which should be solved by the intervention. The evaluation team should therefore identify the problems related to the intervention in the analysis of the situation, in order to incorporate them into the diagram. The remaining problems should be considered part of the context. The problems directly targeted by the intervention may not be explicitly identified. Main problems and context problems can be intermingled, which complicates the construction of the diagram. The evaluator will only be able to differentiate the two types of problems after the completion of the diagram. Whatever the situation, the evaluator should present the list of all the problems under consideration in the annexes, split into two categories:
|
|||||||||||||
Identify the core problemThe crucial stage of the process is the identification of the core problem among the variety of the selected problems. Three situations can be encountered. The core problem is mentioned in the documentation The core problem is not clearly mentioned, whereas the overall objective is These two situations do not need further developments: the evaluator records the core problem explicitly or implicitly mentioned in the documentation, whatever his/her opinion about the problem selected. Neither the core problem, nor the overall objective are explicitly mentioned
Whatever the procedure, the selection of the core problem should be conducted concurrently with the classification of the problems into levels. Indeed, the selection of the core problem should be supported and justified by the coherence of the whole diagram. Proceed to a first classification of the problems by rowsSeveral situations must be considered:
Each of the problem's classification by row must be justified (with interpretations presented as explicit assumptions, if needed). One of the main difficulties in the graphical illustration of problems is to distinguish between "short-cuts" and inconsistencies. In a "short-cut", the succession of causal links is pictured as a causal link between the first and the last problem of the causal chain. For example, the chain illustrating the impact of inadequacies of the road infrastructure on poverty can be illustrated as follows: ![]()
In the same document, quotations can be illustrated as follows: ![]() Or: ![]() This example does not portray inconsistencies but stylistic short-cuts.
|
|||||||||||||
STAGE 2: HOW TO CONSTRUCT A PROBLEM DIAGRAMEstablish a temporary diagramThe diagram may intend to faithfully reflect the problems system as stated in the strategic and planning documents with missing elements and inconsistencies. If no faithfully constructed diagram is explicitly requested, the evaluators will consider it as a step to the construction of a logically reconstructed problem diagram. Stages for the drafting of temporary diagram ![]()
Should the diagram's construction start with the core problem or the in-depth causes? Usage highlights the fact that these two types of problems/causes are the easiest to identify, whereas intermediary causes are the hardest to determine and classify. Thus, it is recommended that the development of the diagram starts with its extremities at the same time. Identify the authors who have analysed the situationBefore testing the temporary diagram, the evaluator should identify the main and secondary authors who have contributed to the analysis of the situation (writing of a chapter, sector-based contributions, participation in working sessions, etc.). It may be useful to check the impact of this analysis on the definition of the strategy and programme operational objectives. Test the temporary diagramWhere possible, the authors of the documentation referred to above should test the diagram in order to validate the classification of the problems by rows and links. The aim is to check that the evaluator's interpretation reflects the authors' intention correctly. If the authors are not available or, in order to complement their original contribution, the evaluator should consider asking for the participation of other actors responsible for the drafting process. Contacting the authors is usually possible when the documentation is recent and the authors are still in their position or contactable in one of the services. This task is challenging when the documentation is old and its authors are not easily identifiable or contactable. Develop the final version of the problem diagramThe final version takes into account the opinions collected during the testing of the temporary diagram. It is an accurate account of the initial intentions logically reconstructed of the European Commission, taken from official documentation. |
|||||||||||||
HOW MANY DIAGRAMS MUST BE DEVELOPED?A single diagramIf the analysis of the situation does not radically evolve during the evaluation and if the activities focus on a single core problem, one diagram is enough. Several diagrams
|
|||||||||||||
HOW SHOULD PROBLEM DIAGRAMS BE USED?When it is impossible to directly establish objectives diagrams, problem diagrams play an essential role in the organisation stage of the evaluation. Conversion into an objectives diagramProblem diagrams present a summarised vision of the situation to be improved, partially at least, by the strategy. The classified objectives included in the strategy should be deduced from the problems diagrams. Planning stageDuring the planning stage (of a project or a programme), the conversion of the problem diagram into an objectives diagram is not automatic, because each problem integrated into the diagram does not have the same priority. Evaluation stageDuring the evaluation stage, the reconstruction of the problem diagram includes a step which differentiates between context problems and intervention problems. As a consequence, the diagram resulting from this selection should be completely convertible into a logically reconstructed objectives diagram, i.e. the higher-order overall objective corresponds to the core problem, and each row of subordinated objectives to its equivalent row of problems. Conversion of the problem diagram into an objectives diagram Extract from 'A summary of the theory behind the LFA method', The Logical Framework Approach (LFA), SIDA - Swedish International Development Co-operation Agency, Draft, June 2002. ![]()
The objectives diagram displays the classification of objectives, from the higher-order objective to the projects and programmes that are designed to achieve the overall objective. The diagram can provoke comments about:
These comments must be presented in the reports and notes of the evaluation organisation stage. They must also appear as an annex in the final report. Definition of the themes for evaluation questionsThe previous analyses raise questions about:
These questions lead to the determination of a range of themes which could be investigated during the subsequent stages of the evaluation, particularly through evaluation questions. However, evaluation questions cannot be automatically deduced from these analyses. Other questions may emerge during the organisation stage, formulated by the strategy implementing operators. Such questions can relate to the difficulties encountered during the planning and implementation stages. However, the number of evaluation questions is limited and additional questions should be the subject of a selection process. |
|||||||||||||
HOW ARE THE FINDINGS PRESENTED?Notes and reports of the organisation stageProblem diagrams are established during the organisation stage, where the reports and notes should be provided. At this stage, the construction of the diagram must be precisely described. The precise sources identifying the problems (exact quotations, references to original documents) must be provided. References to specific documents, interviews, and expertise must support the location of the problems in the diagram, and the assumptions developed during the diagram's construction must be explained. Final reportThe problem diagram should be incorporated into the final report because it is not only a stage in the development of the evaluation, but also an illustration of a possible analysis of the evaluation's strategy. The report itself may include a short presentation of the diagrams, in addition to its development in the annexes. Verbal presentationThe evaluation team will need to present its work (methodology and findings) to different groups of people (the evaluation reference group, participants in the debriefing session, etc.). The problem diagram is a very efficient tool for these situations, providing that it is readable without being over-simplistic. To do so, a main diagram and several sub-diagrams developing fundamental sections of the latter should be presented, each of them not exceeding more than 20 items. |
|||||||||||||
WHAT ARE THE PRECONDITIONS FOR THE DEVELOPMENT OF THE PROBLEM DIAGRAM?Human resources and working arrangementsThe need for human resources varies with the tasks to be achieved and the situation encountered: reconstruction of the problem diagram with the analysis of the situation displayed in the documentation; retrospective reconstruction of the problem diagram as it should have been established to support the relevance of the strategy or programme under assessment. The right column of the table below illustrates a wide range of working days. Type of work required for the design of a problem diagram (usual situation)
Travelling expensesStrategy papers are drafted under the responsibility of the head office (DG RELEX or DG DEV), which is also responsible for the planning stage. EuropeAid is in charge of project drafting, which explains why the majority of the useful documentation (including the complete baseline documentation) is in Brussels. Some tasks may be devolved to Delegations in the future, whose responsibilities in this area will grow. In a country evaluation context, the retrospective construction of the diagram requires at least 2 or 3 experts working on a mission of 5 to 10 days in the country under consideration. Computer devicesMost of the graphic problems can be solved with software like MS PowerPoint. The evaluator can also benefit from specific software designed to assist decision-making, such as MS Visio, TeamUp-PCM for Windows, and Palisade Precision Tree. . |
|||||||||||||
IN COUNTRY EVALUATION |
|||||||||||||
. WHEN SHOULD IT BE USED?Donors must explicitly or implicitly explain the decision to provide assistance to a country or a region in a given timeframe. For this purpose, they must determine the objectives of the assistance designed to solve a range of problems (the objectives can be the donor's or the beneficiary's). The purpose of an evaluation is not to analyse the problems and establish the related diagram, but rather to reconstitute the analysis which has already been undertaken and convert it into a diagram. Consequently, in country evaluations, the problem diagram is particularly appropriate for the following two situations. To examine the explicit problems included in the documentationThe conversion of the analysis into a diagram (Stage 2: how to construct the problem diagram) enables the evaluator to check the coherence of the analysis and its relevance to the objectives. For this purpose, a problem and an objectives diagram should be reconstructed (how should problem diagrams be used?) and be compared. To analyse the non-explicit problems in the documentationA study of the analysis of the situation facilitates the establishment of a problem diagram, whose identified problems should be solved by the assistance. This stage should ease the reconstitution of an objective diagram, or as a minimum, validate the draft of the diagram established in accordance with the documentation. |
|||||||||||||
WHICH PROBLEMS MUST BE TAKEN INTO ACCOUNT?In a bilateral assistance strategy, the European Commission is the donor, and the States and national public institutions are usually its partners. The European Commission and the States have their own overall objectives, corresponding to their own issues. Yet, to a certain extent and degree of detail, the problems become common. The donor's range of problemsThis type of problem corresponds to strategic issues meant to be resolved with financial assistance (for example, regional instability, weak absorption capacity of the local market, uncontrolled immigration). The country beneficiary's range of problemsThis range of problems is related to major political issues to which the States are confronted (macro-economic unbalances, social instability, weak growth, social or geographical disparities, etc). The range of problems common to the donor and the country beneficiaryThis type of problem has common origins which are relevant for both the donor and the country beneficiary, and are relative to the fields of the assistance scope (for example, health, SME, irrigation and environment). Except for participatory evaluations, the evaluation team will prioritise the first and third type of problems, because they are directly relevant to the donor. |
|||||||||||||
WHICH DOCUMENTATION SHOULD BE REFERRED TO?Where can documents analysing problems be found?Documents related to the European Commission strategy Documents related to development assistance programmes Thus, the construction of a problem diagram requires the evaluator to distinguish context problems from main problems. Which documents should be selected?Successive versions of the analysis of the situation can seldom be found (in strategy or programme papers). After ensuring that previous drafts do not yield more in-depth and valid analyses, the evaluation team mainly works on the final version. The baseline documentation comprises the (CSP) and the corresponding National Indicative Paper (NIP), covering the timeframe of the evaluation scope. They are usually provided to the evaluation team at the start of the mission. Often, the evaluation period (defined in the Terms of Reference) is covered by several strategy and programming papers, each of them including analyses and information about the situation of the country, specific sectors or regions. Usually, the situation and problems are analysed at the start of the period covered by the strategy and/or the programme under assessment. Yet, the analysis may be affected by difficulties concerning the information, such as:
These difficulties, which may even weaken the relevance of the objectives, should not be underestimated by the evaluation team. |
|||||||||||||
WHICH TYPES OF PROBLEM DIAGRAMS CAN BE USED?Complete problem diagramUsed in participative planning, the complete problem diagram sketches on each side of the core problem the problems identified by the stakeholders. Simplified problem diagramThe evaluation only focuses on the lower part of the diagram, because the projects usually deal with the causes of the problems. Complex problem diagramUsually, the evaluation team limits its task to the reconstruction of a simple tree shape diagram (simplified or complex, depending on the case) which is derived from the identification of the core problem. Network diagramThis type of diagram is faithful to the complexity of an actual situation. Yet, its reconstitution requires a wealth of information which is seldom available. Furthermore, the conversion of such information into a diagram represents a challenging task. |
|||||||||||||
EXAMPLES |
|||||||||||||
|
|||||||||||||
BIBLIOGRAPHY |
|||||||||||||
The logical framework and the objectives tree
Use of the logical framework in evaluations:
The other approaches:
|
Check lists
CHECK-LIST FOR EVALUATORS |
||||||||||||||||||||||||||
.
. |
||||||||||||||||||||||||||
CHECK-LIST FOR MANAGERS |
||||||||||||||||||||||||||
.
|
Objectives diagram and the effect diagram
SUMMARY |
||||||||||||||||||||
. Why are these tools used in evaluation?These diagrams are usually used as organisation tools. They provide a framework for the information collection and the undertaking of in-depth interviews during the assessment of the programme or strategy. They relate direct outcomes and field effects with the expected objectives and effects of the programme. The construction of the diagrams should therefore constitute one of the first tasks of the evaluation. The objectives diagram illustrates the objectives classification, from the global objective to detailed operational objectives. The effect diagram displays the classification of the results, outcomes and effects of what is intended from the implementation of the objectives system.
What are the possible uses of these diagrams?In the evaluation context, diagrams are used to reconstruct the intervention logic of the European Commission's support to the country. This reconstructed logic will be shaped into one or more logical diagrams of effect. Prior to the preparation of effect diagram(s), the team will have to prioritise the stated cooperation objectives and translate these into intended effects. These intended effects will form the "boxes" of the diagram(s). Possible "holes" in the intervention logic will be indicated and filled on the basis of assumptions to be validated by the reference group. The effect diagram(s) will help to identify the main evaluation questions.
The faithfully reconstructed objectives diagram displays the objectives system and provides the evaluator with a first approach to the strategy and policies inner quality. Indeed, an unclear, incomplete or incoherent diagram means a lack of relevance in the resulting planning or a lack of faithfulness to the initial objectives system. During the implementation of the strategy or programme, external events, such as the evolution of the world commodities market price, elections, political changes, or the conclusion of international agreements can influence the achievement of the objectives and expected outcomes. Comments dealing with the discrepancies between the expected and the observed outcomes should take these events into account. A would-be coherence A strategy or a programme seldom address the full scope of the overall objective, which is limited to choices for each row of objectives or effects. Unless the evaluation can find an explicit explanation of the choices made in strategic documentation, it must provide an answer to series of questions:
How are they constructed?After determining the evaluation scope, the evaluators should construct a diagram illustrating the objectives presented in the strategy and planning documents. The objectives diagram is drawn from this diagram. When the evaluation scope covers one or more strategy papers (geographical) or strategic policies (sector-based, thematic), it is recommended that one diagram per document is created (unless there is a logical continuity in the strategy or the policy). When a logical framework has supported the drafting of a programme, it clearly states the various levels of objectives targeted by the programme. The logical framework is consequently a reference point for the evaluation, as a consequence of the presentation of the objectives diagram's basic constituents. Although the establishment of result-based logical frameworks has not yet been generalised, the effect diagram can usually be deduced from the logical framework of the intervention's objectives. The objectives diagram and effect diagram can also be used in evaluations for projects and programmes whose rationale is not explicit in the logical framework. In practice, the objectives and intended effects of complex policies and strategies often lack explicit presentation and logical structure, whereas the justification for an evaluation is to be able to answer the following question:
What are the preparation stages for the construction of the diagrams?Stage 1: delimit the evaluation scopeThe terms of reference include information about the timeframe and the financial tools to be assessed. Note that projects and programmes originating from previous documentation can also be implemented during this timeframe. Stage 2: identify the objectives and effectsCollect the documentation required for the establishment of the diagrams. It should comprise:
An overall analysis of the break points in the global strategy should be undertaken, without including the objectives of the projects and programmes in progress and scheduled for a different timeframe from the one assessed. Establish a list of the objectives recorded in this documentation. Stage 3: construct a faithfully reconstructed objectives diagramA provisional classification could be carried out on the basis of the distinction between three levels of objectives:
La définition des relations logiques occupe une place centrale dans le classement. Elle est avant tout affaire d'expérience. Stage 4: convert the faithfully reconstructed objectives diagram into a logically reconstructed objectives diagramSome faithfully reconstructed objectives diagrams may reveal logical defects in strategy or political papers, such as:
In order to prepare a comprehensive and coherent objectives diagram, the evaluation team will need all available documentation, its own expertise and, if required, that of experts. Each of these rationalisation operations should be explained in a technical note. Stage 5: construct the effect diagramObjectives diagrams and intended effect diagrams share the same rules of construction. The effect diagram is constructed from the conversion of each of the objectives into the corresponding intended effect:
Computer devicesMost of the graphics can be addressed with software such as MS PowerPoint. How are the findings presented?Objectives diagrams are established during the organisation stage, where reports and notes should be provided. At this stage, the diagram's construction must be precisely described. For the faithfully reconstructed objectives diagram, the sources of the objectives/effects (quotations, references to the original documentation) must be provided. References to documentation, interviews and expertise must support the objectives' location in the diagram, and the assumptions developed during the construction of the diagram must be explained.
The process through which the logically reconstructed objectives diagram has been extracted from the faithfully reconstructed objectives diagram must also be clearly explained. Verbal presentationsThe evaluation team will need to present its work (methodology and findings) to different types of people (the evaluation reference group, participants in debriefing sessions). The objectives diagram and/or the effect diagram are very efficient tools for this purpose, providing that they are readable without being over-simplistic. To do so, a main diagram, and several sub-diagrams developing fundamental sections of the main diagram, should be presented, each of them not exceeding 20 items (boxes). What are the preconditions for its use?
What are the advantages and limitations of the tool?
|
Detailed presentation
.
This section is structured as follows:
- What is an objectives diagram and what is an effect diagram?
- Why and when?
- How should the objectives and effect diagram be constructed?
- How are the diagrams used?
- How are the findings presented?
- What are the preconditions for establishment of an objective diagram and an effect diagram?
- In country evaluations
- Examples
- Bibliography
.
WHAT IS AN OBJECTIVES DIAGRAM AND WHAT IS AN EFFECT DIAGRAM? |
|||||||||
. WHAT DO THEY REPRESENT?Objectives and effects: definitionsDevelopment assistance (in terms of projects, programmes and strategies) usually focuses on an objective to be achieved, or an intended effect. The objective is expressed in terms of an intervention, whereas the impact is expressed in terms of consequence of the intervention. ![]() Various glossaries of evaluations provide definitions for these two terms, such as in the OECD's glossary (Glossary of Key Terms in Evaluation and Result-Based Management). ![]() The objective of the external assistance covers a variety of aspects, from factual interventions (roads, training, etc.) to macro-economic and social changes. In evaluation, the term impact conveys the idea of wide-scope and long-term effects. By convention, medium term effects are called outcomes, and short-term effects results. The tool will refer to the term effect, so as to include the development assistance's range of outputs (results, outcomes and effects). When the diagram is used as a structuring tool, the effects presented in the diagram are the effects intended from the assistance. The diagram can also be used as an evaluation tool for the analysis of the programme's outcomes. The effects presented in the diagram are the observed effects. At the planning stage, the assistance goals are usually expressed in terms of objectives. At the evaluation stage, the concepts of objectives and effects can be used either way. The role of a diagram in the planning processThe objectives systemUsually, a long term and global strategic objective, assumed to be a first level objective, is fulfilled through the completion of a range of second level objectives. This is true even where the first level objective is straightforward. Each second level objective depends on the completion of several third level objectives, and so on, down to operational objectives (the intervention projects). Therefore, an objective is usually understood as a means to achieve a superior level aim, while depending on the completion of subordinate means or objectives. The objectives system is the presentation of all the objectives of all levels with their respective links. The effect systemThe outputs of the implementation of a project, a programme or an assistance strategy include direct results, the short-term outcomes which are linked to these results, and the longer term impacts (direct and indirect). Such outputs can be called 'effects' and are linked together in a range of causal relations or synergies at the basis of the effect system. The objectives treeThe need for a logical classification of the range of objectives in objective-based planning is at the origin of the drafting process of programmes and projects. The objectives system is usually presented as a system of roots and a trunk, hence its common name, the objectives tree. Such tree-like illustrations, however, are subject to some basic rules:
The level expresses the place of the objectives in the cause-and-effect system. In the diagram, the level is represented by the row. ![]() When these rules are applied, objectives trees are a simplified illustration of the objectives system defined during the drafting of the strategy (or the programme). In the context of the European Commission development assistance, objectives trees show the classification of objectives to be achieved in a geographical (region or country), sector-based or thematic strategy. They range from the European Union's long-term global objective, down to the activities carried out in operational programmes. |
|||||||||
The role of the objectives diagram and effect diagram in an evaluationWhat actually happens in thematic, sector-based or geographical evaluations is more complex than that illustrated in objectives and effect systems. Indeed, strategies and programmes under evaluation do not systematically result from objective-based planning. Even if they do, the logical classification of objectives may result from decisions based on circumstances rather than from a rational selection derived from the fundamental issues. This situation may result in:
The use of standard objectives trees in evaluations assumes that the definition of the objectives is rational and the strategic scope can be reduced to a simplified illustration. These limitations explain why the use of objectives diagrams, whose shape can fluctuate more, is favoured to strictly codified objectives trees. The effect diagramInternational development assistance has tended to evolve progressively in the fields of planning and management from an objective-based approach to a result-based approach. Thus, the evaluation must take into account the effect system linked to the programme or the strategy to be assessed, and organise it. The term effect diagram is used to describe the theoretical organisation of the effect system (outputs, short-term results and intermediate impacts) which leads to the overall intended impact. The effect diagram displays the classification of the results, outcomes and impacts of what is intended from the implementation of the objectives system. Its tree-shape connects the actual activities which have been planned, and the outputs, which should produce direct results, to the medium term intermediate impacts and the long-term global impacts. The tree can also be read from the bottom to the top, the long-term overall impact being reached after the implementation of intermediate impacts, results, outputs and interventions. |
|||||||||
WHERE DOES THE OBJECTIVES DIAGRAM COME FROM?The objectives diagram comes from objective-based management whose use expanded during the 1960s in the economic sphere of the United States and in Western economies. It was first adapted to the requirements of the United States Department of Defense and, in the late sixties, to the USAID country assistance management as a component of the logical framework. Thus, the first well-known use of objectives diagram (including USAID), relates to the drafting of logical frameworks. ![]() The objectives diagram belongs to a series of country assistance decision tools, and can be combined with the problem diagram in the context of Goal Oriented Project Planning (GOPP, originally called ZOPP in German for Zielorientierte Projektplanung). At the start of the 1980s, GTZ (German Technical Cooperation) conducted a methodological revision of ZOPP, to be used as an educational tool to determine the stages of the participatory process which produces the logical framework. The typical sequence can be illustrated as followed: Analysis of the problems of a simplified sequence in the project cycle
![]() In this context, the objectives diagram is:
Following GTZ's initiative, various public donors and operators (agencies and NGOs) have used the problem tree as a component of project planning and management. Its use as a participatory and appropriation tool has not always played the paramount role defined by the ZOPP designers, indeed, in many cases, the stakeholder analysis was insufficiently comprehensive and the participation procedures were extremely formal, resulting in the problem tree becoming a formal exercise, even a manipulation tool. The logical framework was therefore limited to rigid project management tool, ignoring the fact that it is first and foremost a learning and negotiation tool. As a consequence, emphasis was put on over-simplistic wording, forgetting that these corresponded to the educational objective of the tool. The biggest criticism of the objectives diagram is the assumption that a consensus can be reached for development objectives and strategies among stakeholders (donors, operators and beneficiaries). Inadequacies and modifications at the programme drafting stage should be examined during the evaluation, which should focus particularly on the drafting process of the objectives system and avoid tools requiring or yielding a simplistic representation of the system. |
|||||||||
WHAT SHAPES CAN AN OBJECTIVES DIAGRAM AND AN EFFECT DIAGRAM TAKE?Tree shape or complex diagramThe standard objectives tree is subject to the following basic rules:
When trees are drafted in the context of objectives-oriented planning, the first row objective is usually called the "global objective". The lower row objectives have various names, but preferably "intermediate objectives". Several rows can be dedicated to intermediate objectives. The objectives located at the bottom of the tree are called "operational objectives". The objectives diagram set out below is a complex objectives system:
Complex objectives diagram ![]()
Examples of more complex objectives systems can be found, having two global objectives and subdivisions of intermediate objectives, however, such objectives diagrams cannot present feedback links. A structured presentation of the effect system (the most recent one) has taken into account the limitations of the tree shaped representation and adopted the flexible shape of the diagram. As a consequence, the systems of results, outcomes and effects can have the same degree of complexity as the objectives system, and can be illustrated by similar diagrams. Effect diagram ![]() Vertical or horizontal diagramDiagrams can either be vertically or horizontally oriented. |
|||||||||
WHAT IS A LOGICAL LINK?DefinitionThe objectives and effects illustrated in the diagrams are related to each other with horizontal or vertical links. These links are called "logical" when expressing an inference relation (induction or deduction) which has been validated by experience. They highlight the fact that:
The logical nature of the links between objectives/effects are essentially based on experience and development theories (derived from this experience). Experience and theory can determine which subordinated objectives/effects should result from the implementation of objectives/effects of a certain level, in order to be properly fulfilled. However, in terms of development, experiments do not always result in the same conclusions and, as a consequence, should not be considered universally applicable. This can be accounted for by the dependency of the objectives/effect on various factors, often not well-understood. For example, even when experts agree on the relevance of a global objective, they may disagree on the definition of intermediate objectives, and in particular, operational objectives. The content of overall objectives such as poverty alleviation is continuously debated and evolving. Verification of the logical nature of the diagramThe logical links between objectives/effects (i.e. the logical nature of the objectives diagram and effect diagram) can simultaneously be checked with:
These two groups of actors may validate the logical nature of the diagram unanimously or individually. The evaluation should clearly present the process of the validation and its outcome. . |
|||||||||
WHY AND WHEN ? |
|||||||||
. WHEN SHOULD AN OBJECTIVES DIAGRAM BE USED?Three fundamentals in development assistance programmes ![]()
An assistance programme can be viewed from different perspectives: its results, internal procedures, partnerships, etc. In terms of efficiency, its immediate results and/or impacts should be understood in relation to the various objectives and/or effects defined in the assistance programme. This implies that the objectives must be identified, or easily recognisable, and classified, or easily put into a hierarchy. In programmes established on the basis of the logical frameworkWhen a logical diagram has supported the drafting of a programme, it clearly states the various levels of objectives which are targeted by the programme. The logical framework is consequently a reference point for the evaluation, as a consequence of the presentation of the objectives diagram's basic constituents. Although the establishment of result-based logical frameworks has not yet been generalised, the effect diagram can usually be deduced from the logical framework of the intervention's objectives. The evaluator should reproduce the objectives diagram included in the programme, or, if it is not available, reconstruct it with the help of the logical framework's first column, and, if necessary, convert it into an effect diagram. In more complex policies and strategiesThe objectives diagram can also be used in evaluations for projects and programmes whose rationale is not explicit in the logical framework. In practice, the objectives and intended effects of complex policies and strategies often lack explicit presentation and logical structure, whereas the justification for an evaluation is to be able to answer the following questions:
In order to carry out the evaluation, the evaluator must determine and rank the objectives/effects of the strategy or the policy. In this context, an objectives diagram combined with an effect diagram is an effective tool for the reconstruction and representation of the objectives system and/or effect system. For programmes established on the basis of the logical framework, as well as in more complex policies and strategies, the construction of the objectives diagram and effect diagram should be one of the evaluation's first tasks. The diagram orients the first stage of the information collection and the undertaking of interviews, whose goal is to develop an in-depth knowledge of the policy or programme to be assessed. |
|||||||||
WHAT ARE ITS ADVANTAGES AND LIMITATIONS?Its advantagesLogical classification of the objectives and the effectsThe diagram explicitly illustrates the objectives/effects and their various rows, from the global objective (more or less long term), down to the range of activities that have already been implemented or are to be undertaken. It reveals the logical links between objectives/effects, or the lack of a linkage. Definition of the evaluation questionsThrough the diagram, the evaluator examines a range of questions whose aim is to help answer the crucial question: To what extent have the stated objectives and the intended effects been achieved? Presentation of the strategyWhen the objectives diagram is well structured and clearly presented, it is a valuable educational tool which facilitates the understanding of the strategy. |
|||||||||
Its limitationsReplacing the tree illustration for the diagram avoids most of the construction constraints and their inherent limitations (with one exception: the representation of retroactive links). A simplified representation of realityAs a chart, the diagram is a simplified representation of reality, and its educational value depends on the selection of a sensible degree of simplification. In order to avoid an over-simplification of the facts, the evaluator can develop sub-diagrams focusing on specific parts of the main diagram. A would-be coherenceThe diagram establishes a logical link between an objective and another belonging to the row above, or between an effect and another belonging to the row above. Each subordinate objective is presented as logically dependent on a higher row objective, and each effect is presented as logically dependent on a higher row effect. The outputs of the activities implemented or scheduled by the programme appear as a contribution to the global objective/impact and support the coherence of the objectives and effect systems. Yet, the objectives diagram and the effect diagram (whether they are directly drawn from the programme's logical framework or reconstructed from explanatory documentation) conceal a series of choices. Each objective or effect of a given row is achieved through the implementation of subordinate objectives or effects whose selection must be explained. Indeed, a strategy or a programme seldom address the full scope of the overall objective, which is limited to choices for each row of objectives or effects. Unless the evaluation can find an explicit explanation of the choices made in strategic documentation, it must provide an answer to series of questions:
Selected and rejected objectives ![]()
Selected and rejected effects ![]()
If the evaluator does not investigate such questions, the evaluation of relevance and coherence of a programme or strategy may be superficial. |
|||||||||
WITH WHICH TOOLS CAN IT BE COMBINED?Problem diagramThe objectives diagram often comes (but not systematically) from the problem tree. Problem diagrams illustrate the overall aspects of a situation which should at least be partially improved by the strategy, whose ranked objectives should be deduced from the diagram. In the planning process, such a deduction is not systematic. Indeed, for various reasons, all the problems listed in the diagram cannot be analysed with the same priority. The relationship between the problem and the objectives diagrams
![]()
In evaluations, the reconstruction of the problem diagram includes a stage during which the problems caused by the context are distinguished from those following the implementation of activities. The resulting diagram should be easily translated into an objectives diagram: the focal problem should correspond to the global objective, and each level of problems should be matched by an equivalent row of objectives. Logical frameworkIn strategy or planning papers, the objectives diagram is often linked to the development of the logical framework, which provides to the diagram its left column and general structure. Decision diagramThe analysis of the programme's development (decision diagram) can be used to explain the reasons for the selection of objectives for various rows (and particularly the highest row objectives). In mid-term programme evaluations, the evaluator may be asked to re-establish some of the tools which have been previously mentioned in this methodology, in addition to the objectives diagram. |
|||||||||
WHAT ARE THE PRECONDITIONS FOR ITS USE?Study of the initial documentationWhen the objectives diagram and the effect diagram are found in the initial documentation of the programme, or when they can be drawn from the logical framework, they should be presented, with the study of the choices made during their construction. This implies a study of the documentation which was used in the construction of the diagrams and interviews with stakeholders directly involved in their drafting. In sector-based evaluations, baseline documents can be found in all the sector-based strategy documentation originating from the European Commission (such as communication papers, guidelines, handbooks, etc.), from the partner States and other donors. Baseline documents also include other strategic documents (such as CSPs) and their corresponding national indicative programmes (NIPs) which are relevant to the evaluation scope under study, as well as bilateral agreements such as Association Agreement, Agreement Cooperation and their monitoring documents. When key actors and documentation are not availableWhen neither the documentation nor stakeholders are available, the evaluator should discuss the relevance of the objectives system presented in the diagram with specialists, and formulate a judgement. When the evaluator has reconstructed the diagram on the basis of the available documentation, the evaluation team must:
If archives and stakeholders are not accessible, the evaluation team must explain and describe their own assumptions underpinning the construction of the diagram. The team must therefore provide the reader of the diagram with the sources (quotations and documentary references) on which the diagram is based. |
|||||||||
HOW CAN THE OBJECTIVES DIAGRAM BE USED IN PROJECT EVALUATIONS?A means to concentrate on the intervention rationale of the projectUsually, projects are provided with a logical framework. The analysis should focus on this logical framework, and study the vertical and horizontal coherence of the matrix with the context. In the absence of a logical framework, the intervention rationale should be reconstructed on the basis of the project's documentation (Stage 2: identify the objectives and the impacts). There may be a delay between the design of the project and its start. In this case, the relevance of the logical framework should be checked when the project is launched. Any other interventions in the field of application of the project should be taken into account, in order to assess the coherence of the project's interventions with the general context of the assistance. Which objectives should be taken into account?The evaluation studies the objectives presented in the logical framework, or in the reconstituted logical framework. The ownership of the project's objectives should be measured for the various types of stakeholders, particularly in the context of a participative evaluation. Once completed, the logically reconstructed objectives diagram results in two concurrent evaluative approaches:
. |
|||||||||
HOW SHOULD THE OBJECTIVES AND EFFECT DIAGRAM BE CONSTRUCTED? |
|||||||||
. The evaluation should first identify the outcomes of the assistance and compare them with the objectives to be achieved. The structuring stage requires the analysis of the objectives, from which the intended effects are deduced. After determining the evaluation scope, the evaluators should construct a faithfully reconstructed objectives diagram from the strategy and planning documents. A logically reconstructed objectives diagram and an effect diagram (intended effects) will be derived from the faithfully reconstructed objectives diagram. STAGE 1: HOW IS THE EVALUATION SCOPE DELIMITEDUsually, the evaluation scope is strictly defined in the terms of reference which includes information about the timeframe and the financial tools to be assessed. In any case (concerning the evaluation timeframe or the financial tools), the managing authority must provide the evaluation team with a precise definition of the scope, although the evaluation team must feel free to warn the latter about the consequences of a too restrictive scope. Period of the evaluationThe period to be assessed fully or partially corresponds to the period covered by the strategy and/or planning papers. However, during this timeframe, projects and programmes which are being implemented may have already been described and authorised in previous documentation. What importance should be given to the objectives or the effects of such projects and programmes? Two situations should be considered, depending on the nature of the projects and programmes implemented during the period under study (i.e. their objectives are in continuity with or differing from the objectives planned in previous projects and programmes). Projects and programmes whose objectives differThe evaluator should undertake a general analysis of the differing topics in the overall strategy and type of activities, and avoid the study of specific objectives (or effects) included in these projects and programmes. Projects and programmes whose objectives are in continuityIn this case, the study of their specific objectives (or effects) should be included in the analysis of the objectives. Financial toolsExplicit terms of referenceTerms of reference can focus the evaluation on the activities carried out by a specific institution (a General Directorate of the European Commission) with a specific financial tool. The scope is strictly limited. The evaluation team should however examine other institutions, include them in the category "other donors", and seek the elements of coherence in their activities which could be missing from the objectives system or the effect system of the institution under evaluation. The evaluation team should also examine the coordination between activities under assessment and others, and formulate its conclusion on the possible effects of insufficient coordination. Inexplicit terms of referenceWhen terms of reference do not determine the evaluation scope, the analysis of the objectives or the effects should cover the range of activities, whatever the implementing institution and the financial tool implemented. |
|||||||||
STAGE 2: HOW TO IDENTIFY THE OBJECTIVES AND EFFECTSCollect the documentationAs a first step, the documentation needed for the construction of the diagram is collected in the relevant Unit of DG RELEX or DG DEV. It comprises:
The documentation varies with the type of evaluation (geographical, sector-based and thematic). Record the references to the objectives in the baseline documentationThe task should begin with a careful analysis of the baseline documentation, to assemble a record of all references (quotations, sources) to the objective (understood as the target of the European Commission's assistance) and any explicit mention of the links between objectives. The documentation seldom uses the term "objectives" and only introduces the term in expressions such as "with a view to", "in order to", "so as to", etc. Objectives can be called "priorities", "fields of intervention", or sometimes "focal sectors". A precise table of such terminology must be established. Distinguish strategic from background objectivesLes citations relevées dans les documents de référence sont traduites en objectifs. Chacun des objectifs est assorti d'un numéro d'identification renvoyant aux citations. The documentation identifies the objectives which should be targeted by the assistance. The remaining objectives are treated as background items. In an annex to the reports (how are the findings presented?), the evaluation team should present the list of all the objectives, marshalled into two categories: strategic and background objectives. |
|||||||||
STAGE 3: HOW TO CONSTRUCT A FAITHFULLY RECONSTRUCTED OBJECTIVES DIAGRAMClassify the objectives by rowsSeveral situations can be encountered:
Each of the classifications by row should be explained by means of an interpretation expressed as an assumption. A provisional classification could be carried out on the basis of the distinction between three levels of objectives:
The definition of the logical relationship is paramount in the classification. This task is a question of experience, because the decision of the evaluation team that two events are logically connected depends on the judgement of experts in the field under study, and managers responsible for the implementation of the strategy and policies. Establish a temporary diagramThe first faithfully reconstructed objectives diagram can be drawn from this first classification. It usually reveals incoherencies that will have to be corrected in accordance with the evaluator's interpretation (these corrections must be indicated in the diagram). Should the construction of the diagram start with the overall objective Usage has shown that these two types of objectives are more easily identified than intermediate objectives, which are also more difficult to rank. Thus, it is recommended that the roots of the diagram and the extremities of the branches are developed concurrently. In the diagram, each objective is presented as a box with a heading. The evaluator should also indicate the identification number which refers to the quotation list. Providing the boxes with colours corresponding to the level to which the objective belongs may also be helpful. Each link should refer to one or more quotations (in accordance with the identification number). The diagram should also mention the specific means (financial and non financial) used in the implementation of each operational objective. Identify the authors of the objectives' wordingBefore testing the temporary diagram, the evaluator should identify the main and secondary authors of the objectives' wording (writing of a chapter, sector-based contributions, participation in working meetings, etc.). It may be useful to check the consequence of the wording on the definition of the operational objectives of the programme planning. Test the temporary diagramThe authors of baseline documentation and additional documents should test the diagram in order to validate the classification of the objectives by rows and links. The aim is to check that the evaluator's interpretations accurately reflect the authors' intentions. If the authors are not available or, with a view to complement their contribution, the evaluator can seek the participation of other actors responsible for the drafting process (both from the writing and the discussion process). Contacting the authors is usually possible when the documentation is recent and the authors are still in position or contactable in one of the services. This task is challenging when the documentation is old and its authors are not easily identifiable, nor reachable. At the end of the interviews, the evaluator may also encounter difficulties evaluating what was actually done during the drafting process and the justifications of respondents for the decisions taken. Establish the final version of the faithfully reconstructed objectives diagramThis version takes into account the opinions collected during the test of the temporary diagram. It should be considered an accurate report of the initial intentions of the European Commission, taken from the official documentation. |
|||||||||
STAGE 4: HOW TO CONVERT THE FAITHFULLY RECONSTRUCTED OBJECTIVES DIAGRAM INTO A LOGICALLY RECONSTRUCTED OBJECTIVES DIAGRAMAdvantages and limitations of a faithfully reconstructed objectives diagramAdvantages of the faithfully reconstructed objectives diagramSome faithfully reconstructed objectives diagrams display a strong internal coherence: apart from the overall objective or the overall expect impact, each objective 'x' leads to an higher (immediate or not) objective and is supported by one or more subordinate objectives. This is particularly the case when faithfully reconstructed objectives diagrams are extracted from a carefully established logical framework. Faithfully reconstructed objectives diagrams can be deemed as completely logical and stand as logically reconstructed objectives diagrams. Logical links of a given objective 'x' (in objectives diagrams) ![]() The establishment of a logically reconstructed objectives diagram is not required when the objectives system of the strategy and planning documents is thoroughly and consistently presented. Limitations of the faithfully reconstructed objectives diagramSome faithfully reconstructed objectives diagrams may reveal logical defects in strategy or political papers, such as:
In these cases, the definition of the evaluation thematic scope and the evaluation questions is difficult. The team should therefore draft a logically reconstructed objectives diagram. Preparation of the logically reconstructed objectives diagramIn order to prepare a comprehensive and coherent objectives diagram, the evaluation team will need all available documentation, its own expertise and, if required, that of experts. This task should be completely transparent (how are the findings presented?): whenever possible, the diagram should display the reclassification of objectives, the changes affecting links between objectives, and the introduction of missing intermediate objectives. Each of these rationalisation operations should be explained in a technical note. |
|||||||||
Stage 5: how to construct an effect diagramObjectives diagrams and intended effect diagrams share the same rules of construction. The effect diagram is constructed from the conversion of each of the objectives into the corresponding intended effect. ![]()
. |
|||||||||
HOW ARE THE DIAGRAMS USED? |
|||||||||
Objectives diagrams and effect diagrams are a crucial tool in the organisation stage of the evaluation. They play many roles:
Description and analysis of the strategyObjectives diagrams demonstrate a synthesis of the geographical strategy displayed in the official documentation. When the evaluation scope covers one or more strategy papers (geographical) or strategic policies (sector-based, thematic), it is recommended that one diagram per document is created (unless there is a logical continuity in the strategy or the policy). They illustrate the objectives/effects classification, from the overall objective/impact to projects and programmes scheduled to implement the objective/effects. The logically reconstructed objectives diagramand the effect diagram are informative means to reflect upon:
The faithfully reconstructed objectives diagram displays the objectives system and provides the evaluator with a first approach to the strategy and policies inner quality. Indeed, an unclear, incomplete and incoherent diagram means a lack of relevance in the forthcoming planning or a lack of faithfulness to the initial objectives system. These analyses must be presented in the notes and reports of the evaluation organisation stage. They must also appear as an abstract in the final report. Definition of the themes for evaluation questions The previous analyses reveal questions about:
These questions result in the determination of a range of themes which could be investigated during the following stages of the evaluation, particularly through evaluation questions. However, evaluation questions cannot be automatically deduced from these analyses. Other questions can emerge during the organisation stage, formulated by the main strategy implementation operators (since they are the most well-informed actors about the problems encountered during the planning and the implementation stages). However, the number of evaluation questions is limited and additional questions would have to be the subject of a selection process. |
|||||||||
HOW ARE THE FINDINGS PRESENTED? |
|||||||||
Notes and reports of the organisation stageObjectives diagrams are established during the organisation stage, where reports and notes should be provided. At this stage, the diagram's construction must be precisely described. For the faithfully reconstructed objectives diagram, the sources of the objectives/effects (quotations, references to the original documentation) must be provided. References to documentation, interviews and expertise must support the objectives' location in the diagram, and the assumptions developed during the construction of the diagram must be explained. The process through which the logically reconstructed objectives diagram has been extracted from the faithfully reconstructed objectives diagram must also be clearly explained. Final reportDiagrams should be incorporated into the final report because they are not only stages in the development of the wording of evaluation questions, but also an illustration of a possible analysis of the evaluation strategy. The report may include a short presentation of the diagrams, in addition to a full explanation included in the annexes. Verbal presentationsThe evaluation team will need to present its work (methodology and findings) to different types of people (the evaluation reference group, participants in the debriefing session). The objectives diagram or the effect diagram are very efficient tools for this purpose, providing that they are readable without being over-simplistic. To do so, a main diagram, and several sub-diagrams developing fundamental sections of the main diagram, should be presented, each of them not exceeding 20 items. |
|||||||||
WHAT ARE THE PRECONDITIONS FOR THE ESTABLISHMENT OF AN OBJECTIVES DIAGRAM AND AN EFFECT DIAGRAM? |
|||||||||
Human resources and working arrangementType of work required for the design of an objectives diagram ![]() Observation :
Travelling expensesStrategy papers have been drafted under the responsibility of the head office (DG RELEX or DG DEV). The head office is also responsible for the planning stage. EuropeAid is in charge of the project drafting. This implies that the majority of the useful documentation (the whole baseline documentation) can be found in Brussels. This situation may devolve to Delegations in the future whose responsibilities in this area will grow. The current needs (2004) in travelling expenses to Brussels
These elements (and the corresponding budgetary components) will have to be reviewed when the Delegations take broader responsibilities for strategy development (2007). Computer devicesMost of the graphic problems can be addressed with software such as MS PowerPoint |
|||||||||
IN COUNTRY EVALUATIONS |
|||||||||
. IS THE USE OF AN OBJECTIVES DIAGRAM AND AN EFFECT DIAGRAM COMPULSORY?Every donor must implicitly or explicitly explain the decision to provide assistance to a country or a region in a given timeframe. For this purpose, the objectives to be achieved or the intended effects which form the basis for the assistance are described in documents which are regularly published and present the overall objectives and priorities. Such documentation can either display orientations, an individual programme, or both. Such documents may present programmes which are jointly drafted by the donor and beneficiary. In this case, the interventions under study should be that of the European Commission and of the country partner. The outcomes and intended impacts are the outputs of the implementation of these two types of interventions. In national programmesNational Indicative Programmes (NIP) can be inserted in the logical framework which can be a single document or correspond to each main sector of intervention. In this case, the evaluation uses the logical frameworks (i.e. objectives and implementation indicators) as the reference for efficiency criteria. Its goal will be to check in what extent the objectives and effect systems have been completed during the implementation of the programme. In country strategiesCountry Strategy Papers (CSP) do not usually rely on an overall logical framework because they cover a long timeframe and are implemented through various successive programmes, each of them supported by a logical framework. Consequently, the evaluators should assemble the objectives and effects system, and draft a diagram for each programme. |
|||||||||
WHICH OBJECTIVES AND EFFECTS MUST BE TAKEN INTO ACCOUNT?In strategies and complex bilateral assistance programmes, the European Commission is the donor and the States and national public institutions are usually its partners. The European Commission and the States have their own overall objectives, corresponding to their own issues. Yet, to a certain extent and degree of detail, the objectives overlap. In a bilateral assistance contract, three objectives or effect systems can be developed:
Except for participatory evaluations, the evaluation team will prioritise the first and third objectives system. |
|||||||||
WHAT ROLE DO OBJECTIVES DIAGRAM AND EFFECT DIAGRAMS PLAY IN COUNTRY EVALUATIONS?Once completed, the logically reconstructed objectives diagram and the effect diagram provide two parallel approaches to the country strategy evaluation:
These approaches are illustrated in the following table. The main branch of the logically reconstructed objectives diagram splits into two: the right branch deals with evaluation questions, the left branch with the overall analysis of the strategy. Role of the objectives diagram in the conduct of an evaluation structured by evaluation questions ![]()
Footnotes Two main aspects of relevance are examined:
The evaluation questions connect the intended effects (results, outcomes, secondary impacts) with the specific and general means which are used to achieve the objectives. The overall analysis studies the strategy building, i.e.:
|
|||||||||
WHICH TYPES OF OBJECTIVES DIAGRAMS OR EFFECT DIAGRAMS CAN BE USED FOR COUNTRY EVALUATIONS?Two types of diagram, setting the framework for the evaluation, illustrate the objectives of the programme or the strategy to be assessed. The faithfully reconstructed objectives diagramThe faithfully reconstructed objectives diagram is useful for two reasons:
The logically reconstructed objectives diagramThe logically reconstructed objectives diagram determines the thematic scope of the evaluation. It establishes the reconstructed classification of the objectives system as they are presented in logical frameworks or reconstituted in the faithfully reconstructed objectives diagram. It is also the logical reconstitution of the objectives' classification implied in strategy papers (CSP, RSP) and planning papers (NIP, RIP). It contributes to the assessment of strengths and weaknesses on which the evaluation questions should focus:
The logically reconstructed objectives diagram is particularly needed because the strategy papers and documentation concerning the planning of geographic assistance do not provide the evaluator with a systematic presentation (with or without a diagram) of the objectives system. It should be summarised within the final report of the evaluation, and fully presented in the annexes. Additional observationDiagrams are not the only tools available to analyse objectives and effects. For example, evaluators frequently use a matrix. |
|||||||||
HOW MANY EFFECT DIAGRAMS MUST BE ESTABLISHED?A single diagramIf the analysis of the situation does not radically evolve during the evaluation and if the activities focus on a single intended overall impact, a single diagram is enough. Many diagrams, depending on the contextWhen the strategy evolves, effect diagrams should be constructed for each change in strategy of the European Commission. When the evaluator cannot determine a single overall impact and when each of the global impact depends on an independent effect system, the evaluator should establish as many diagrams as overall impacts. Intermediate diagramsIntermediate diagrams, corresponding to each of the assistance priorities (or sectors) for the period(s) covered by the evaluation, can be useful for the construction of an objectives diagram or effect diagram. They are particularly adapted for programmes with a logical framework provided for each of the priorities. Intermediate diagrams can be useful for the determination and justification of evaluation questions. Magnifier diagramsWhen overall diagrams are very complex, their readability and information quality may suffer. A simplified overall diagram should be prepared, and detail diagrams should be drawn from the overall diagrams, in order to emphasize some of their specific parts. |
|||||||||
WHICH DOCUMENTATION SHOULD BE REFERRED TO IN COUNTRY EVALUATIONS?Baseline documentationThe baseline documentation includes strategy papers (CSP) and the related indicative programmes (NIP) relevant to the period covered in the evaluation scope. Bilateral agreements, such as the Association Agreement, Cooperation Agreement and their monitoring documents should also be included, as well as multilateral agreements (the Lome and Cotonou agreements) and their related documents (for example, the Barcelona Declaration). In most cases, they should be provided to the evaluator at the start of the evaluation. However, the evaluator should check that:
Additional documentationThree types of documents must be available :
Other informative documentationOther documentation may be useful, such as successive versions of the baseline documentation (the date of the final version must be checked), including the various contributions and reactions of the Commission's services, their national partners and other donors (particularly Member States). It is valuable, but often difficult to identify and seldom available or accessible. . |
|||||||||
EXAMPLES |
|||||||||
. |
|||||||||
BIBLIOGRAPHY |
|||||||||
. The logical framework and the objectives tree
Use of the logical framework in evaluations
The other approaches
|
Check lists
CHECK-LIST FOR EVALUATORS |
||||||||||||||||||||||||||||
.
. |
||||||||||||||||||||||||||||
CHECK-LIST FOR MANAGERS |
||||||||||||||||||||||||||||
.
|
Decision diagram
SUMMARY |
||||||||||||||||||||
. Why is this tool used in evaluation?Strategies result from a procedure, taking into account global objectives, contextual elements as well as the position of the stakeholders. This procedure is usually empirical. The decision diagram sketches the strategy drafting process, particularly for the selection of key information, the participation of stakeholders in the process, and the management of the implementation arrangements. It highlights the choices that have been made when the strategy was elaborated as well as the consequences of the selected objectives and their forecast impact. The decision diagram is a useful complement to the objectives diagram The documents which present co-operation strategies usually display a range of objectives among which one or occasionally more global objectives, operational objectives relative to development assistance projects, and a range of intermediary objectives at various levels can be identified. The selection carried out by the authors of the strategic and programming documents depends on various sources:
The aim of the decision diagram is to describe the impacts of such orientations, contextual data and analyses. Indeed, each box of the diagram's central column represents a choice (selected and rejected objectives), while the boxes on each side illustrate the flow of inputs which represents the external justification for these choices. What are the possible uses of these diagrams?The decision diagram highlights:
The decision diagram facilitates the analysis of the strategy in terms of internal coherence (logical succession of the choices) and external relevance (contextual elements and position of the stakeholders). When the terms of reference of an evaluation require an analysis of the partnership, the diagram is used to highlight the intervention of the main partners (governments, Member States and other donors) in the strategy design, the establishment of the programmes and the selection of the projects. The diagram can perform the same role for the analysis of the 3 Cs (Coherence, Co-ordination, Complementarity). How is the decision diagram constructed?
The drafting process of the decision diagram continues in two steps:
What are the preparation stages for the construction of the diagrams?Stage 1: determination of the points when the decisions were made
Usually, the establishment of the decision diagram follows the construction of the objectives diagram. The objectives diagram is the basic tool for identifying the points at which the decision-making takes place. Apart from the definition of the overall objective, each intersection of the diagram represents a decision-making point. It stimulates the following questions: Why have these objectives been selected? Why have others been rejected? Stage 2: drafting of questionsThe questions focus on the justification of the selection of objectives. The identification of the rejected objectives helps the accuracy of the wording of questions. How are these objectives identified? At each decision-making point, the evaluator may encounter four situations:
In the first three situations, the evaluators should investigate which of the objectives were planned but eventually rejected, and the reasons for rejection. In the last situation, would-be objectives should be identified. Stage 3: collection of useful informationBefore the construction of the diagram, the quality of the sources of information should be checked. If this verification reveals insufficient sources, the construction of the diagram should be abandoned. Usually, written information can be found in four types of documents:
The evaluation team should formulate its conclusions about the quality of the sources in terms of quantity, relevance, reliability and accessibility. This judgement is presented to the managers, who take the final decision. Stage 4: constructing the temporary diagramPreliminary selection of the relevant information is carried out on the basis of the objectives diagram. It requires:
Provisional answers to the evaluation questions are formulated on the basis of the information collected. Some of the questions may not be answered at this stage. The objectives diagram supports the decisions chain. It identifies four (sometimes five) decision-making points dealing with:
The drafting of strategies and programmes is not strictly and exclusively driven by such a rationale. Implicitly or explicitly, the designers of the strategy and programme begin with an overall objective. Thereafter, they examine which means at the disposal of the European Commission are able to achieve this objective. The successive choices can be synthesised in two points:
At the left of the sketch, a magnifier explains the decision's outcome:
A global fishbone shaped diagram is thus completed through detailed diagrams corresponding to each decision-making point.
Stage 5: testing of the temporary diagramThe decisions and their explanation must be confirmed by the main actors responsible for the drafting of the strategy and the programming, including the European Commission's services (head office and delegations), the representatives of the other stakeholders (Member States, NGOs, etc.), the usual interlocutors in beneficiary countries and/or their government. Stage 6: constructing the final diagramThe process of testing of the temporary diagram may question some of its parts when the justifications do not illustrate the real strategic and programming drafting process. In this case, the information should be reviewed and augmented by another consultation round. The final and temporary diagrams have the same shape (a main diagram and the sketch of the point of decision-making). The final diagram includes an explanatory table about the analysis of the information collected. What are the preconditions for the diagrams use?
What are the advantages and limitations of the tool?
|
Detailed presentation
.
This section is structured as follows:
- What is a decision diagram?
- Why and when should the decision diagram be used?
- What are the decision diagram's construction stages?
- How should the decision diagram be used?
- Examples
- Bibliography
.
WHAT IS A DECISION DIAGRAM? |
||||||||||||||||||||||||||||||||||||||||
. WHAT DOES THE DECISION DIAGRAM REPRESENT?General definitionThe decision diagram shows the process during which the strategic objectives and the overall policies of cooperation with developing countries, which are defined by the European Union's assistance agreements, are converted into short-term and medium-term bilateral co-operation decisions. It illustrates the successive official and informal decisions, resulting in the programming of development assistance. Decision factorsTwo types of decision factors can be found:
External factors appear during the decision-making process as a flow of information originating from the European Union's institutions, or other sources (strategic trends, context of the country under consideration, governmental policies, position of the Member States and other donors, etc.). They are called inputs. |
||||||||||||||||||||||||||||||||||||||||
WHERE DOES THE DECISION DIAGRAM COME FROM?Various fields of applicationDecision-making is an important theme in management theory. Various tools are proposed for the determination of public policies (such as in the field of defence), medical and managerial decisions. They usually lead to a series of choices for alternative solutions which have to be optimised by rationalisation tools, such as cost-effectiveness analysis. In the analysis of the strategic decision-making process, the retrospective use of such a tool was suggested for the first time (in a simplified version) for the evaluation of the European Commission's strategic co-operation with Egypt. The tool, however, was not exploited fully during this evaluation. Reasons for its useIn the management field, the role of the decision diagram is to rationalise the decision-making process and to make it more efficient for the resolution of problems and the achievement of objectives. It provides an answer to the following question: Which elements must be taken into account to make the most appropriate operational Decision diagrams can be supported by a variety of software (computer devices), providing the user with guidance throughout the decision-making process. Thus, decisions diagrams are, by nature, a normative and simplifying tool. The evaluator needs a tool which describes a complex approach with undefined rationale and several stakeholders with different aims. Such approaches are more focused on consensus than efficiency, and do not necessarily rely on any rationalisation of decision-making. Can the diagram be combined with other tools?The decision diagram can be directly combined with the objectives diagram and effect diagram (and indirectly with the problem diagram). It was originally designed to overcome the deficiencies of the objectives diagram, but it is also useful in itself. Its construction can be supported by a SWOT analysis and/or a socioanthropological analysis, in which information about the context of the societies affected by the strategy is provided. |
||||||||||||||||||||||||||||||||||||||||
WHAT SHAPE CAN IT TAKE?The fishbone shape can be used to illustrate the three main components included in decision-making. Thus, a theoretical decision diagram for the European Commission's country development assistance will present:
Standard decision diagram (fishbone shape) ![]()
The diagram illustrates the decision-making stages and the main external flow of information supporting the drafting of the strategic or political decision. . |
||||||||||||||||||||||||||||||||||||||||
WHY AND WHEN SHOULD THE DECISION DIAGRAM BE USED? |
||||||||||||||||||||||||||||||||||||||||
. WHEN SHOULD THE DECISION DIAGRAM BE USED?During the design of the strategiesThe European Commission's external intervention strategies result from a complex procedure, taking into account the contextual elements as well as the position of the stakeholders. This procedure is usually empirical and lacks transparency, although more rationalisation and explanation may lead to more appropriate decisions (selection of priorities, adoption of programmes). These decisions could have a positive influence on the country partners' ownership of the strategies. Using a decision-making tool such as the decision diagram could facilitate the organisation of the design process, particularly for the selection of key information, the participation of stakeholders in the process, and the management of the implementation arrangements. This methodology, however, focuses on the tool's usage in the particular context of the evaluation (during the organisation stage). Its use in the strategy design stage requires an adaptation of the methodology. During the evaluation of the strategiesIllustration of justified choicesCo-operation strategies and intervention policies designed by national and multinational donors have explicit and implicit objectives. The documents which present them usually display a range of objectives more or less in order (see the complex objectives diagram), and among which one or occasionally more overall objectives, operational objectives relative to development assistance projects, and a range of intermediary objectives at various levels can be identified. Thus, the operational items of a programme (planned projects and/or in process) which depend on the strategic or political papers, are supported by the objectives system given in these papers. In an objectives diagram, the selection of a higher-order objective and its division into several intermediary objectives, down to the operational objectives often results from non-explicit choices. Unless decisions are deemed arbitrary, the selection carried out by the authors of the strategic and programming documents depends on various sources:
The decision diagram illustrates the impact of these information flows on the successive choices (i.e. the progressive elimination of alternative options) made by the decision-makers. Indeed, each box of the diagram's central column represents a choice (selected and rejected objectives), while the boxes on each side illustrate the flow of inputs which represents the external justification for these choices. |
||||||||||||||||||||||||||||||||||||||||
A strategic sequence and an operational one in the decisionStrategy papers are established for each 7 year budgetary period. They deal with the events, orientations and changes having a significant and long-lasting impact, which can be worldwide (such as the attack on the World Trade Centre in 9/11/2001) or regional (the second Intifada in 2000, or the military invasion of Iraq in 2003). If such events happen during the strategic period, strategy papers can be revised (for example, in the mid-term revision). However, these documents usually remain unchanged until the end of their term. Programme durations are sometimes shorter than strategy papers (usually 2 to 3 years). They can be affected by substantial changes in the political orientations of the partner government, natural disasters (typhoons, droughts), the signature of association or co-operation agreements, political shifts in the main donors' activities, etc. The revision of existing programmes or the drafting of new programmes is common in the programming activity during the budgetary period. In addition to a short presentation of the whole decision-making process (in the form of a summary diagram), the evaluator should divide the theoretical diagram (standard decision diagram) in two. This should result in a strategic decision diagram on one hand, and an operational decision diagram on the other hand. Standard strategic decision diagram ![]()
Standard operational decision diagram ![]()
Such a separation is interesting because it limits the number of boxes in each diagram and improves their readability. |
||||||||||||||||||||||||||||||||||||||||
Evaluation of the decision processThe design of the strategy and programming papers is theoretically collaborative. These documents are the subject of negotiations between the European Commission and the partner governments of the ACP regions. In other regions, programmes are the subject of an agreement between partners, whereas the communication of strategies is for information only. Member States must be consulted on all strategy and programming documents. Practice shows, however, that this collaboration is sometimes only formal. As a consequence, the European Commission insists on an in-depth examination of the decision procedure which could be part of a country evaluation, or stand as a thematic evaluation on its own. In both cases, the decision diagram is a very relevant tool. |
||||||||||||||||||||||||||||||||||||||||
WHAT ARE ITS ADVANTAGES AND LIMITATIONS?AdvantagesClarification of the strategyThe diagram highlights:
Complementing the objectives diagram and effect diagram, the decision diagram facilitates the analysis of the strategy in terms of internal coherence (logical succession of the choices) and external relevance (contextual elements and position of the stakeholders). Illustration of the main partners' interventions and the analysis of the 3 CsWhen the terms of reference of an evaluation require an analysis of the partnership, the diagram is used to highlight the intervention of the main partners (governments, Member States and other donors) in the strategy design, the establishment of the programmes and the selection of the projects. The diagram can perform the same role for the analysis of the 3 Cs (Coherence, Co-ordination, Complementarity). LimitationsThe major limitations in the use of the diagram are:
Baseline informationDirect information about the factors influencing the strategic and programming drafting process is sometimes scarce in the official papers which present the strategies and programmes. Other public documents, such as evaluations, can provide indications but the information they provide is limited. The restricted documentation (studies, notes, correspondences) is in principle more informative, but often cannot be consulted, or its access is difficult. Verbal information has the same limitations concerning availability and accessibility. Causal interpretationEven when potentially informative elements are provided (for example, the intervention of other donors, or the priorities of the government), causal relations between decisions are not always explicit. An interpretation is sometimes required, which can be the source of a risk of error, particularly in cases where several causes support the decision-making. |
||||||||||||||||||||||||||||||||||||||||
WHAT ARE THE PRECONDITIONS FOR ITS USE?Existence of an objectives diagramThe presence of an objectives diagram in strategy or programming papers is a precondition for the construction of a decision diagram. If such is not the case, the evaluator will need to construct an objectives diagram. Existence and relevance of the informationThe availability (effective or to be assured) of reliable written or verbal information (stage 3) is also an important precondition. This information needs to be sufficient and allow verification. Tight collaboration with the European Commission's servicesA trustworthy collaboration between the evaluators and the European Commission's services (competent services of the DG Development, DG External Relations, and Delegations) would guarantee the access to the information and its interpretation. . |
||||||||||||||||||||||||||||||||||||||||
WHAT ARE THE DECISION DIAGRAM'S CONSTRUCTION STAGES? |
||||||||||||||||||||||||||||||||||||||||
. Once the preconditions are fulfilled, the drafting process of the decision diagram continues in two steps:
The process cannot start with the collection of documentation because useful information cannot be found in one or more identified documents. It has to begin with the definition of the scope of such a collection. The determination of the objectives at various levels (overall, intermediate and operational) implies a series of choices which can be translated into a question: why such an objective and not another one? The question's wording therefore orients the collection of the information judged to bring answers. The relevance of these answers is checked, and may result in seeking new information (written or verbal), and even rewording some of the questions, particularly those concerning the objectives which are rejected. |
||||||||||||||||||||||||||||||||||||||||
STAGE 1: DETERMINATION OF THE POINTS WHEN THE DECISIONS WERE MADEGeneral principleThe decision diagram (existing or to be constructed) identifies the points at which the implicit or explicit selection of the objectives were made. Whilst the objectives diagram assesses the internal coherence of the objectives system, the decision diagram illustrates the external factors corresponding to each choice in the decision-making and resulting in the selection of one objective over another. Determination of the period allocated for the selection of objectivesUsually, the establishment of the decision diagram follows the construction of the objectives diagram. This sequence should be respected. The objectives diagram is the basic tool for identifying the points at which the decision-making takes place. Apart from the definition of the overall objective, each intersection of the diagram represents a decision-making point. For example, in the objectives diagram, 3 intersections are illustrated, which stimulates the following questions: Why have these objectives been selected? Why have others been rejected? The decision diagram questions the choices concerning:
The planning process should be both strategic (implementation of the best means to achieve objectives of different levels) and empirical (selection of the means likely to achieve specific objectives). Whatever the planning process, the choices will be made in accordance with the objective diagram. Each of the decision-making points should be given an identification code (which may include a date), to facilitate its transfer to analytical tables (see stage 3). |
||||||||||||||||||||||||||||||||||||||||
STAGE 2: DRAFTING OF QUESTIONSIdentification of the rejected objectivesThe questions focus on the justification of the selection of objectives. The identification of the rejected objectives helps the accuracy of the wording of questions. How are these objectives identified? At each decision-making point, the evaluator may encounter four situations:
In the first three situations, the evaluators should investigate which of the objectives were planned but eventually rejected, and the reasons for rejection. In the last situation, the identification of the objectives will have to be made by the evaluation team, or with the help of external expertise. The propositions developed this way can be supported by:
Wording of questionsWhatever the situation, the outcome is the answer to the following question: Why such an objective has been rejected? Several questions of this type may be needed at each decision-making point. |
||||||||||||||||||||||||||||||||||||||||
STAGE 3: COLLECTION OF USEFUL INFORMATIONBefore the construction of the diagram, the evaluation manager or, failing that, the evaluation team should check the quality of the sources of information. If this verification reveals poor or unreliable sources, the construction of the diagram should be abandoned. What is useful information?Useful information, written or verbal, should not be identified in a too limited way, nor in a comprehensive way. Indeed, the identification of appropriate information for any type of evaluation can only be stated in general. Written informationUsually, written information can be found in four types of documents.
Verbal informationTo complement the written information, the evaluator can ask the authors and contributors of the main documents (strategy papers and programmes) to explain the reasons for their choices. What specific information is required for country evaluations?Written informationThe following list, relating to the kind of available documents and the nature of their information, is not intended to be comprehensive.
Most of these documents can be found for the strategies established after 2001. Prior to this date, the situation is more challenging because the documents are not routinely made available, even when the relationship between the evaluators and the Commission's services is productive. Verbal informationIn addition to - or in substitution for - written information, the authors and contributors to the main documents (strategy and programming papers), as well as the decision-makers responsible for the drafts in process, may be valued informants, providing that they are available and in post for a significant period in the head office (DG Dev and DG Relex, EuropeAid), or in one of the delegations. How is useful information collected?Collection of written informationWritten information should be collected during the preliminary stage of the country evaluation from the relevant services of the Commission, such as the DG Relex and DG Dev (country and regions), and in the European Commission's Delegation in the country under consideration. Collection of verbal informationThe main interlocutors are the managers of the services previously described. If they are unavailable, former managers can also be interviewed if they can be easily contacted. Is the information collected sufficient?If the evaluation managers have not identified a list of information sources, the evaluation team should formulate its conclusions about the quality of the sources in terms of quantity, relevance, reliability and accessibility. This judgement is presented to the managers, who take the final decision. |
||||||||||||||||||||||||||||||||||||||||
STAGE 4: CONSTRUCTING THE TEMPORARY DIAGRAMPreliminary analysis of the informationPreliminary selection of the relevant information is carried out on the basis of the objectives diagram. It requires:
Following the analysis of the texts, additional information resulting from the documentation collected may be required, or new documents could be requested and investigated (this method yields uncertain findings). Drafts of the explanationsProvisional answers to the evaluation questions are formulated on the basis of the information collected. Some of the questions may not be answered at this stage. A table of questions can illustrate the results of the information analysis and take the following shape: Table A: results of the collected information analysis
In Table A, the columns gather data concerning:
Construction of a temporary diagramThe objectives diagram supports the decisions chain. It identifies four (sometimes five) decision-making points dealing with:
Most of the time the drafting of strategies and programmes is not strictly and exclusively driven by such a rationale and is, in fact, often empirical. Implicitly or explicitly, the designers of the strategy and programme begin with an overall objective. Thereafter, they examine which means at the disposal of the European Commission are able to achieve this objective. The successive choices can be synthesised in two points:
Preceding each point, and on both sides of the decisions chain, flows of inputs identified from the analysis of the written and verbal information collected, can be found. At the left of the sketch, a magnifier explains the decision's outcome:
A global fishbone shaped diagram is thus completed through detailed diagrams corresponding to each decision-making point. Sketch of the moment of the decision-making: example of first row intermediary objectives ![]()
|
||||||||||||||||||||||||||||||||||||||||
STAGE 5: TESTING OF THE TEMPORARY DIAGRAMThe decisions and their explanation must be confirmed by the main actors responsible for the drafting of the strategy and the programming, including the European Commission's services (head office and delegations), the representatives of the other stakeholders (Member States, NGOs, etc.), the usual interlocutors in beneficiary countries and/or their government. The observations of the respondents can be recorded in the following table. Table B: Results of the testing of the temporary diagram
In Table B, the columns gather data concerning:
|
||||||||||||||||||||||||||||||||||||||||
STAGE 6: CONSTRUCTING THE FINAL DIAGRAMIf the observations of the decision-makers consulted are not critical, the temporary diagram is corrected incorporating the observations. This operation produces the final diagram. The process of testing of the temporary diagram may question some of its parts when the justifications do not illustrate the real strategic and programming drafting process. In this case, the information should be reviewed and augmented by another consultation round. The evaluators and the evaluation managers should decide whether they need to construct a new temporary diagram and test it. Indeed, the observations collected during the test may be sufficient to avoid another consultation and to establish the final diagram directly. The final and temporary diagrams have the same shape (a main diagram and the sketch of the point of decision-making). The final diagram includes an explanatory table about the analysis of the information collected. . |
||||||||||||||||||||||||||||||||||||||||
HOW SHOULD THE DECISION DIAGRAM BE USED? |
||||||||||||||||||||||||||||||||||||||||
. HOW MANY DIAGRAMS SHOULD BE ESTABLISHED?
Diagram of strategic decisionsUsually, the number of the diagrams to be constructed should correspond to the number of the strategy papers drafted during the evaluation period. Evaluators and evaluation managers may decide to work on the most recent document only, when required. Indeed, some of the early documentation needed for the establishment of the decision diagram may be too old to guarantee its availability and that of its authors. Whatever the case, it is recommended that a decision diagram for each possible medium-term strategic revision is prepared. Diagram of operational decisionsSuch a diagram is usually prepared for each development assistance programme. As strategy papers may cover several programmes, the evaluators may have to study 4 or 5 programmes for an evaluation in a twelve-years timeframe. The construction of such a large number of diagrams represents a significant amount of work and cost, and yields uncertain results, being dependant on the availability of the information (documents and informants). Thus, the evaluators and evaluation managers may agree to focus on programmes relating to the most recent strategy paper. |
||||||||||||||||||||||||||||||||||||||||
HOW SHOULD THE DECISION DIAGRAM BE USED?Recommendations for country evaluationsThe general recommendations about the use of decision diagrams are derived from the experience drawn from their use in the particular context of country evaluations. Decision diagrams should be established whenever the evaluation studies the relevance of the objectives with regard to the European Union's goals, the country specificities and the overall context. Previous evaluations show that, excluding extreme situations, objectives and interventions are usually relatively coherent, although the appropriateness of the choices made is not always indisputable. The decision diagram is likely to alleviate this particular limitation. The use of decision diagrams in thematic and sector-based evaluations appears possible, but should be tested, and adapted where necessary. Evaluation of the relevance of the strategy and planningThe explanations and justifications for the choices stated in the documentation and from informants should be evaluated. Thanks to their experience or with the support of external experts, evaluators should formulate their own views about the relevance of the explanations which are provided for each decision-making choice. They should be able to judge to what extent the rejection of an objective seems justified. The selection of the overall objective (following the rejection of other possible overall objectives) should be consistent with the strategic goals formulated by the European Union's institutions (Council, Parliament, Commission). The justification of the subordinate objectives is based on coherence with the overall objective or the ones resulting from it, and with the decision chain's inputs. Evaluation of the quality of the partnershipThe partnership between the country's authorities and the representatives of the Member States can be evaluated through the study of specific input flows. Thus, a judgement can be formulated about the importance given to the following topics during the strategy and planning decision process:
Evaluation of the 3CsThe following can also be analysed:
Additional toolsThese tools can be used when the decision-making factors are not explicit in the documentation collected, or when their interpretation is uncertain. They are especially useful for evaluations where the terms of reference explicitly include an analysis of the decision-making process (thematic evaluations about the decision-making process). The evaluators can present their successive judgements in a series of tables completing the diagram. Validation of the diagram by the decision-makers consulted during the testTable C: Evaluation proposed for the selection of objectives
In Table C, the columns gather data concerning:
Validation of the diagram by an expert panel including the Commission's decision-makers and independent expertsThe evaluation may be subjected to a panel composed of representatives of the Commission's services (geographic departments, delegations), and independent and recognised experts. The findings could be reported in the following table: Table D: Final evaluation of the selection of objectives
In Table D, the columns gather data concerning:
|
||||||||||||||||||||||||||||||||||||||||
HOW ARE THE FINDINGS PRESENTED?Intermediary documentationThe intermediary documentation is addressed to the managers of the evaluation and the steering committee. It must include all the graphs and tables in detail. Final reportThe final report is addressed to a wider public, more interested in the findings than the evaluation's methodology. It is recommended that analytical graphs and tables should be placed in an annex and a synthesis of the findings should be shown in the main report under three headings:
A simplified diagram of the strategic and operational decisions may usefully be included in the main report. Presentation of the final studyIt may be useful for the evaluators to use the standard decision diagram during the presentation of the evaluation's findings. |
||||||||||||||||||||||||||||||||||||||||
WHAT ARE THE PRECONDITIONS FOR THE ESTABLISHMENT OF A DECISION DIAGRAM?Human resources and working arrangementsType of work required for the design of a decision diagram ![]()
Travelling expensesStrategy papers used to be prepared under the responsibility of the Commission's Directorates-General (Relex or Dev) which are also responsible for the planning stage. EuropeAid is in charge of the project design. Thus, the majority of the useful documentation (the whole baseline documentation) can be found at the Commission's Headquarters in Brussels. The delegations have taken part in the drafting process of the most recent documents and their responsibilities in this area will grow. In addition, decision-makers from ACP countries benefiting from European assistance may in principle take part in the drafting process of the strategies and programmes which have been co-signed. Computer devicesSpecialist software can support decision-making. Yet, these devices do not seem relevant to the evaluation context, where spreadsheet programs or PowerPoint should be sufficient. . |
||||||||||||||||||||||||||||||||||||||||
EXAMPLES |
||||||||||||||||||||||||||||||||||||||||
. |
||||||||||||||||||||||||||||||||||||||||
BIBLIOGRAPHY |
||||||||||||||||||||||||||||||||||||||||
.
|
Check lists
.
CHECK-LIST FOR EVALUATORS |
||||||||||||||||||||||||||||||||
.
. |
||||||||||||||||||||||||||||||||
CHECK-LIST FOR MANAGERS |
||||||||||||||||||||||||||||||||
.
. |
Interview
SUMMARY |
||||||||||||||||
. Why is this tool used in evaluation?The interview is an information collection tool which usually takes the shape of a face-to-face discussion between the evaluator and the interviewee. In evaluation, the use of interviews is simple, quick, and affordable, which makes its use inevitable.
What use can be made of the interview?In evaluation, the interview collects different kind of information:
The interview may be used as a quantitative collection tool; however, it is mostly a qualitative device. Information, including facts that can be checked, points of view, analyses and opinions should be clearly distinguished. Three types of interviews can be carried out: Unstructured interviews This type of interview is particularly interesting at the start of an evaluation, in order to get a global view of the subject, and identify the major topics and issues. Semi-structured interviews This type of interview is the most frequently used, particularly when the evaluator knows sufficient about the aims and the main questions to pose during the evaluation. Structured interviews This type of interview is useful when a large number of interviews must be carried out and when the evaluator wants to minimise the risk of bias from the interviewer. Establish homogeneous interview grids when several teams are responsible for conducting the interviews. Semi-structured interviews are the most commonly used tool in evaluation and are the subject of further guidance. How is the interview carried out?How is the interview prepared?The evaluator should first prepare the list of questions to be asked during the interviews. The schedule of questions indicates the categories of respondent to be interviewed, within which the evaluator chooses those most capable of providing the information. The evaluator must determine:
Once the categories of respondent are defined, the evaluator can schedule the interviews and try to find a balance between the rational and optimal use of his/her own time and a flexible and "human" vision of the other's time. Questionnaire grids (the evaluation's strategic questions), and interview guidelines derived from them (questions asked during the interview), vary with the categories of respondent, the latter's links with the evaluated issue and the type of interview (unstructured, semi-structured or structured interviews). Grids should include all themes and questions which the evaluator wants to discuss with the respondents. The questionnaire grid is an intermediary between the evaluation study's design and its implementation Interview guidelines provide the interview with a framework which is not binding on the evaluator. How is the interview conducted?
|
Figure 2: the preconditions for its use | |
The time span | The preparation for the interview does not take long.
The number of interviews which can be carried out during the day is limited. In practice, at the interviewee's request, the expert may conduct an interview with several respondents at the same time. This particular use of the interview increases the opportunity of collecting the information required in a relatively short time. |
Human resources | Interviews must be conducted by a trained professional. The necessary skills are:
|
Financial resources | Possible transportation costs
Costs depend on the number of interviews; however, the interview itself does not lead to substantial costs |
What are the advantages and limitations of the tool?
Figure 3: The advantages and limitations of the tool | |
Advantages | Quick and easy to use.
Short delays and low costs. Appropriate tool to meet a limited number of key respondents. Essential tool to develop analyses and understand the stakeholders' perceptions of the programme. |
Limitations | At a reasonable cost, only few people can be interviewed.
Problems relating to the respondent's 'representativeness' particularly for social groups and beneficiaries. The information must be checked and interviews are generally combined with other analytical tools. |
Detailed presentation
This section is structured as follows:
WHY AND WHEN?
WHAT ARE THE REASONS FOR USING INTERVIEWS?
What is the purpose of an interview?
The use of interviews in evaluation is inevitable. This tool collects information and points of view, and analyses them at each stage of the evaluation.
The interview usually takes the form of a face-to-face discussion between the evaluator and the interviewee.
Where does this tool come from?
Interviews are used in many fields, such as psychology, ethnology, sociology, anthropology, and philosophy.
In psychology, this tool focuses on motivation, or the reasons for a person's behaviour. The interview is thus used to study human actions and to help with the understanding of the human psyche.
In ethnology, the interview aims mainly at collecting direct observations. In sociology, the interview is used in all activities.
Three types of interviews and their contribution to the evaluation
Unstructured interviews
The interviewee expresses himself/herself freely and can discuss unplanned topics, because there is no predetermined set of questions. The evaluator intervenes only to generate and develop questions relating to the interviewee's comments.
This type of interview is particularly interesting at the start of an evaluation, in order to get a global view of the subject, and identify the major topics and issues.
Semi-structured interviews
The evaluator modifies the interview guide's instructions with additional questions, in order to develop useful areas of inquiry during the interview.
This type of interview is the most frequently used, particularly when the evaluator knows sufficient about the aims and the main questions to pose during the evaluation.
Structured interviews
The evaluator follows strictly the interview guide's instructions. He asks different interviewees the same set of questions, in the same order, and using the same words. The evaluator avoids generating and developing additional questions, and the interviewee is not given the opportunity to express himself/herself freely. Answers to each question tend to be short. Structured interviews are seldom used in evaluation, where the evaluator needs to adapt to the situation. However, they can be used to classify points of view and information about the impact of a project/programme by categories. Thereafter, the evaluator can use the results of these interviews to design a questionnaire, with a view to analysing the impact of the project/programme.
Semi-structured interviews are the most commonly used tool in evaluation and are the subject of further guidance.
WHEN SHOULD THE INTERVIEW BE USED?
What kind of information does the interview collect?
The interview may be used as a quantitative collection tool; however, it is mostly a qualitative device.
In evaluation, the use of interviews is simple, quick, and affordable.
The interview collects:
- Facts and information for the verification of facts
- Opinions and perspectives
- Analyses
- Suggestions
- Reactions to the evaluator's hypotheses and conclusions
However, a series of interviews is not sufficient to quantify an event, because the answers are not standardised. This is the main difference between the interview and the questionnaire.
The added value of the interview
Among other advantages, this tool is essential for the development of analyses because it collects information taken directly from the context. Thus, it provides a good indication of what motivates stakeholders to act, their various perceptions of the programme's aims, problems encountered and effective outcomes.
Can the interview be combined with other collection tools?
The interview is a useful device for developing hypotheses and analyses. It can highlight the programme's aims and dynamics, the stakeholders' rationale, and the organisation of the various opinions and perceptions of the programme.
Depending on the type of observation tool used and on the stage of the evaluation, the interview can be used in combination with other tools:
- To test the main questions to address at the beginning of the evaluation with a view to preparing a series of focus groups.
- To be the main observation tool, and be supported by a questionnaire (if the country's general context allows it) or, for example, a focus group of beneficiaries.
How should the interview be conducted?
Usually, the interview takes the form of a face-to-face discussion. This arrangement is particularly effective, as the relationship builds on trust and interviews supplement written information.
However, beyond verbal information, other elements should be taken into account, such as the context, the general mood of the session, the people likely to influence the interviewee, etc, as well as the interviewee's reactions: hesitation, silence, eye contact, etc.
From whom does the interview collect information?
The interview is a suitable tool for collecting information, analysing and forming conclusions from a limited (but essential) number of respondents, such as:
- Partners and people in charge of the evaluated policy or programme
- Strategic institutional stakeholders
- Main operators and people in charge of the programme implementation
- Representatives of beneficiaries
AT WHAT STAGE OF THE EVALUATION SHOULD THE INTERVIEW BE CONDUCTED?
Types of interview appropriate to various stages of the evaluation
Stages of the evaluation | Type of interview | The interview's contribution to the evaluation |
Desk phase: implementation of the methodology and preparation of the mission to the country |
Preparatory interview
used for the design of questionnaire grids and the selection of respondents. At this stage, interview guides should be flexible and aimed at highlighting the topics on which the interview is based. |
It strengthens the basis for the choice of major topics and issues, and completes the questionnaire grid.
Its structure should be flexible. |
Collection of information in the country and from the European Commission | The interview is
designed to collect information and perspectives. Several interview guides should be developed to correspond to the different categories of respondents, and to the major topics and issues. |
It collects information on the programme's objectives and outcome from the people in charge of the programme, operators, stakeholders and beneficiaries. |
Analysis and preparation of judgements | In-depth interviews for presenting and investigating the issues, used to collect reactions to the evaluator's findings and analyses. At this stage, the evaluator can test the relevance and the feasibility of his/her conclusions. | It collects feedback from respondents. It gives in-depth information and reformulates old questions. |
WHAT ARE THE ADVANTAGES AND THE LIMITATIONS OF THE INTERVIEW?
The advantages

The interview is appropriate for project evaluation, as well as for more complex evaluations, such as sector evaluations and country/region evaluations, where it should be combined with other tools.
The limitations

WHAT ARE THE PRE-CONDITIONS FOR THE USE OF THE INTERVIEW IN EVALUATION?
The time span
The preparation for the interview does not take long.
One expert will not be able to conduct many interviews per day and, therefore, the number of interviews which can be carried out during the mission is limited. In practice, at the interviewee's request, the expert may conduct an interview with several respondents at the same time. Thus, this particular usage of the interview increases the opportunity for collecting the information required in a relatively short time.
Human resources
Interviews must be conducted by a trained professional. The necessary skills are:
- Thorough knowledge of the major topics and issues addressed in the evaluation
- Excellent interviewing skills:

- The ability to quickly understand the respondent's perspective (his/her interest in the interview, whether he/she has expressed himself freely, whether he/she has committed himself) in order to be interactive and, where appropriate, modify the questions
Financial resources
Costs depend on the number of interviews and their physical location in the country. However, apart from professional fees and transportation costs, the interview itself does not lead to substantial costs.
HOW THE INTERVIEW IS CARRIED OUT?
HOW IS THE INTERVIEW PREPARED?
Stage 1: list the categories of stakeholders and interviewees
The evaluator should define the categories of stakeholders he needs to meet, and then select a sample group for each category.
Examples of stakeholders' groups
- Agents in charge of the operational strategy / the implementation
- Actors specialised in a sector of intervention
- Operators working at a local level / at a national level
- Technical operators: representatives / on-site operators
- Public sector agents (working for the state, the administration, etc.) / private sector agents (working for the civil society, NGOs, etc.),
- Beneficiaries
The list of the stakeholder categories will evolve as the evaluation progresses. Thus, additional interviews should be anticipated.
Some choices are obvious (for example, the agent in charge of the implementation of a nationwide programme), but others are more complex.
Moreover, the evaluator must consider how he will meet the interviewee, which depends on:
- The respondent's accessibility (will the evaluator meet him in person or only a representative)
- The respondent's availability
What are useful questions to bear in mind for the selection of interviewees?
- Within the relevant institution, who knows the programme/the politics well?
- Who is in charge?
- Who is in charge of implementation, of monitoring?
- Of which aspects of the programme is the respondent in charge?
- Who will be able to trace the programme's evolution, in the context of a rapid turnover of the implementation team?
During the interviewees' selection, the evaluator should be careful to distinguish between direct witness' testimonies and second-hand testimonies.
Stage 2: design the questionnaire grid
Questionnaire grids, and interview guides derived from them, vary with the categories of respondent and the latter's links with the evaluated issue.
Grids should include all instructions, themes and questions which the evaluator wants to discuss with the respondents.
The stages required to organise the grid

The questionnaire grid is an intermediary between the evaluation study's design and its implementation.
Design of the interview guide

The evaluator should allow the interview to remain as close to an open discussion as possible and accept discursive answers, inconsistencies, pauses, hesitations, etc.
The questions can be written in the questionnaire grid, or be asked spontaneously, in order to complete, to probe, to give new insights and to challenge the respondent's answers.
Stage 3: schedule the interviews
This organisational stage depends on the evaluation stage within which the interview is implemented, and on the respondent's availability.
The selection criteria must clearly indicate the priority of the people to be met. Furthermore, the evaluator must be careful about:
- The risk of key officials being unavailable
- His/Her short mission span
- Travelling time
- Scheduling too many interviews, leading to the cancellations of appointments due to poor planning or unexpected events
- Respecting his/her appointment, such as the time, although conventions vary from one country to another
- Conducting the interview's introductory stage slowly, in order to establish rapport and respect traditions and customs
- Unplanned and additional interviews which have to be organised because a respondent has suggested an "interesting" informant who may be interviewed, or because a key interviewee is now available
- Tiredness: too many interviews conducted in a day may affect the capacity of the interviewer to listen
The evaluator must find a balance between:

Scheduling 4 to 6 interviews a day seems to be a good compromise between the two visions.
Make the appointments
This stage constitutes the first contact between the interviewee and the evaluator. Therefore, the choice of the person in charge of this task is very important.
It is often better to schedule appointments in advance. However, time should also be allocated for unplanned interviews, particularly in the context of country/region evaluation.
The evaluator must know how to resist pressure to meet as many respondents as possible. Organising focus groups can solve this problem.
Send interview guides prior to the appointment
This stage is optional, however, it often eases the interview's course by giving the respondent the opportunity to:
- Understand the topics on which he/she has to talk
- Collect all the necessary documentation
- Ask the evaluator whether the participation of other respondents during the interview is possible
It will save time for both the respondent and the evaluator, and help them feel at ease. Most of the time, the quality of the interview depends on this.
Conduct additional interviews
They are often necessary after the first elements are collected and during the progress of the evaluation. Thus, they must be programmed into the planned timetable and budget.
HOW IS THE INTERVIEW CONDUCTED?
Stage 1: establish a rapport
The interview is easy to conduct but its findings and value vary greatly with the way it is conducted.
The evaluator's duty
- To be aware of and respect local habits and customs (such as rules of politeness, silences, wording of questions, time devoted to conversation prior to core questions, etc.)
- To anticipate any language difficulties (codes, interpreter's competence, meaning of terms such as "development", etc.)
- To adjust himself/herself to the interviewee, his/her role and his/her hierarchical rank in the institution (for example, the evaluator may be asked not to reveal some information)
- To take into account the interviewee's material, social and hierarchical environment which may greatly influence his/her attitude
- To explain the purpose of the interview, how the respondents have been chosen, and the intended use of the information
- To establish (and sometimes to negotiate) the rules, such as the interview's length, the recording of the interview, etc.
- To inform the interviewee at the start whether his answers will be used anonymously
- To ensure that the interviewee has understood the aims of the interview and is willing to respond
To find out more:
- What are the tasks to be completed during the preparation for the interview in country/region evaluation?
Stage 2: Adjust the respondent's answers to the interview subject
The evaluator must adjust to his interlocutor's role and hierarchical rank in the institution. As a consequence, he/she must be aware of the specificities of the respondent's answers, such as:
- The way he/she understands the questions and deals with them
- The possible difficulty in expressing points of view, or criticisms
- Ideas he/she might want to point out with the evaluator's mediation
- Considering the evaluator as an auditor
The evaluator must also adjust to the interlocutor's attitude: his/her personal perspective, points of interest not planned in the interview guide, etc.
As a consequence, the evaluator's flexibility is the key to successful data collection in an interview. During the interview, however, he/she must control its progress by staying within the bounds of the subject and avoiding dwelling too much on one topic.
Stage 3: follow the interview guide and deepen the questioning
Flexibility and control are two elementary rules in an interview. The respondent's information is being analysed simultaneously through the evaluator's capacity to listen. Thus, the evaluator should not express himself/herself in a way which is detrimental to the interview, nor let the interviewee talk without limit. In practice, the evaluator's ability to react should provide him/her a balance between flexibility and control.
Three types of reactions
- Contradiction (the evaluator highlights the respondent's contradictions or expresses the contrary views of other respondents, whose identity may have to be concealed)
- Notification (the evaluator informs the respondent that he/she is about to address a new theme or a new question)
- Clarification (the evaluator asks the respondent to develop a specific point)
Types of clarification
Questions | Repeat verbatim the respondent' s comments in the interrogative form |
Repetition | Return to items which have already been discussed |
Interpretation | Summarise the content of what has been said, check its interpretation with the respondent (and correct it, if necessary), and move on to the next point |
Intervene directly in the interview
Although the interview guide provides a useful structure for the interview, the evaluator must be free to generate and develop questions beyond it.
Making direct observations during the interview enables the evaluator to collect specific information about the respondent's attitude, behaviour and customs, and / or about the people around them. With these observations, the evaluator can detect discrepancies between the respondents' attitude and their words.
To find out more:
- Three types of interviews and their use in evaluation
The nature of the collected information
The evaluator must distinguish reliable facts from points of view, personal analyses and opinions. To do so, he/she should press the respondent to support his/her allegations with facts or actual examples which the evaluator can check, and which reinforce the respondent's comments.
Avoid asking difficult questions
During an interview, questions should be answerable and should avoid discouraging respondents. Thus, interviewees should be spared from having to provide information stemming from archives or voluminous report.
Control the information
In an interview, the first level of control occurs with the triangulation of questions. This methodology is based on the principle that three different sources are needed to validate the reliability of the information. Triangulation is used in the interview in two ways:
- In interviews, the evaluator asks systematic and similar questions to at least three different respondents. Prior to this triangulation, the information is not considered to be reliable.
- In a single interview with one respondent, the evaluator asks a question in three different ways, in order to check the information and observe the possible fluctuation in the respondent's analyses and interpretations of an event.
Stage 4: Conclude the interview
Conclude the interview
The evaluator should close the interview with positive reflections or an open conversation, and maintain a polite approach. He/She should also decide whether another appointment with the respondent is necessary.
Keep track of all the information
The evaluator should read his notes shortly after the interview, structure them and add, if necessary, non-verbal details such as the respondent's behaviour, trouble, silences, interruptions, his relationship with him, the atmosphere, and especially the respondent's suggestion of other people to be interviewed (who will have to be contacted) and read the specified documentation, etc.
These elements are invaluable; they are not detectable in audio tapes, nor short notes. Thus, the evaluator's impressions at the end of the interview must be considered a valid source of information.
Protect the confidentiality of the interview
If the respondent's willingness to respond depends on the evaluator's assurance of confidentiality, this principle must be respected, along with respect for the respondent's private life.
If necessary, validate the content of the interview report with the respondent
Depending on the intended use of the respondent's answers, this validation can be important if, for example:
- the respondent's verbatim comments are cited in the report
- They stand as evidence on their own
- The evaluator finds them ambiguous and is afraid of making a poor interpretation, or worse, a misinterpretation
- The respondent has an official or sensitive duty
- The respondent's willingness to respond depends on this validation
Anticipate what will have to be done after the interview
- The intended use of the information collected
- Its analyses
- The types of debriefing, which should meet the client's and respondents' expectations (sometimes, interviewees request feedback about the use of their answer)
IN COUNTRY/REGION EVALUATION
WHAT ARE THE TASKS TO BE COMPLETED DURING THE PREPARATION FOR THE INTERVIEW?
The hypotheses and questionnaire grid development
In country/region evaluation, the evaluator needs to define the strategic questions quickly, in order to be efficient during the information collection.

The evaluator's tasks during this first stage
Select the types of information to collect
The evaluator should:
- Organise the information to be collected on strategic questions between key and secondary interest
- Devise a methodology to limit the collection of documentation. As country/region evaluation covers so many diversified fields, the evaluator risks dissipating his efforts if this stage is neglected
Develop and organise the questions
The evaluator should classify key questions by their aims and categories of interviewees.
Determine the list of the respondents
Settle the order in which respondents will be met
THE TYPES OF INTERVIEWS?
Different types of interviews can be distinguished during the two main stages of country/region evaluation
During the first stage
Preliminary interviews setting the programme's purpose and boundaries
These consist of meeting people who have a global vision of the policies implemented, their objectives and impact, and who belong to relevant institutions. These interviews can avoid bias in the process of an evaluation arising from reliance on officials in charge of policy definition and monitoring.
Interviews with different categories of stakeholders
During these interviews, the evaluator meets the stakeholders who have been participating in the process of drafting, definition, negotiation and the operational choices for the implementation of policies.
Thus, the relevance and impact of geographical strategies rely on what different stakeholders have done in the process of drafting and negotiation.
During the second stage
Interviews with operators and beneficiaries
The use of interviews to analyse the impact of a programme on beneficiaries in country/region evaluation is not usually advisable because of the large number of people involved, representative problems, etc. However, interviews can be conducted prior to the implementation of other data collection tools, such as focus groups and questionnaires.
Thus, interviews should focus on particular categories of beneficiaries and/or precise questions dealing with various hypotheses.
HOW ARE RESPONDENTS CHOSEN?
In country/region evaluation, this issue is crucial. It is less a question of knowing how to collect information, than of determining what the information's content should be, and from whom it can be obtained, in the context of short missions and wide fields of inquiry.
Choosing the categories of respondents is particularly important:
- To identify people who have benefited from the implemented policies
- To identify those who have played a strategic role in the programme
- To identify those who have played an intermediary role
- To identify those who could have been behind obstructions or unplanned effects (such as groups with diverging interests, intermediary groups intervening in the middle of the process, target groups of the policies / the programme, etc.)
- To analyse current strategies and the various rationales considered
Distinguish between the different roles of officials from institutions
The list of officials who should be met can be long because of the nature of the country/region evaluation, which can cover numerous programmes and projects.
Thus, a selection must be made consistent with the budget allocated to the evaluation, and especially when the mission is very short. The evaluator should make a distinction between key interviews and "etiquette" interviews; he/she should also think of the quickest and cost-effective way to conduct them.
Once the roles of officials are identified, the evaluator must locate the people who have agreed to be interviewed, corresponding to this typology.
This selection belongs to the evaluation process and becomes clearer as the first hypotheses are being developed.
Go beyond formal statements
In country/region evaluation, the evaluation's political dimension and economic implications tend to complicate some interviews with institutional authorities and respondents concerned with their hierarchy. The evaluator should be able to "decode" the information he/she is given, by taking into account the context in which the interview is conducted.
A way to avoid formal statements implied by this situation is to carry out an informal interview (for example, away from the institution to which the respondent belongs).
Balance the interviews between officials and beneficiaries
The evaluator must be careful to avoid devoting all the time in his/her mission to interviews with officials, to the detriment of beneficiaries. He/She must check the interviewee's "representativeness", which is always relative and sometimes self-proclaimed. Several questions may help find the respondent's interest in the evaluation and his/her "representativeness":
- To which group does the respondent belong?
- What is his/her professional background?
- What is his/her influence on the programme's strategy?
- Does he/she have direct or indirect interests in the programme? Are they explicit or implicit?
- What is his/her opinion about the policies / the programme?
- How did he/she become a representative (has he/she been elected, designated)?
- Is he/she representative of the entire group he/she stands for?
- How many respondents should be interviewed to collect representative information and perspectives?
HOW DOES THE EVALUATOR SCHEDULE THE APPOINTMENTS?
The pre-conditions for appointment making
During the preparation for his/her interviews, the evaluator should be practical and:
- Fix an order of importance for the collection of data and, after developing the first set of questions, select who should be interviewed
- See: What specific tasks have to be completed during the preparation of the interview in country/region evaluation?
- Identify quickly on-site informants and stakeholders. They often hold a position at the centre of many sources of information. Thus, they can advise the evaluator, and suggest "good" respondents, who will be able to give accurate answers.
- Identify respondents who would be better interviewed in a group
- Consider the possibility of interviewing people on the telephone, which could save time
- Avoid overestimating the number of interviewees that can be met during a short mission
How to schedule the appointments?
The particular context of country/region evaluation implies taking additional precautions, in addition to the usual precautions, in the making of appointments.
The evaluator should:
- Schedule in advance at least part of the interviews that he/she needs to conduct on-site
- Ensure that the evaluation's clients (for example, the European Commission Delegation or the national authorities), or the local consultants organise the first set of interviews on his/her arrival
- Make time available for unscheduled appointments made necessary because of new perspectives identified during his mission. As a consequence, the evaluator's hypotheses can be extended or modified.
With a view to being more efficient, the evaluator could:
- Use the appointment making process as an opportunity to communicate to the respondents all the information they need to know: the purpose/ of the evaluation, what questions they should prepare for, and the interview/ duration. This is also the appropriate moment to check the respondent's profile against the planned content of the data collection.
- Distribute an interview guide prior to the interview. This is appreciated by all parties because it is time-saving and directs the respondent's information gathering to specific documentations.
Check lists
CHECK-LIST FOR EVALUATORS |
||||||||||||||||||||||||
.
. |
||||||||||||||||||||||||
CHECK-LIST FOR MANAGERS |
||||||||||||||||||||||||
.
|
Focus group
SUMMARY |
||||||||||||||||||
. Why is this tool used in evaluation?A focus group is a form of group interviewing which comprises individuals involved in a development policy or intervention. It is set up to get information concerning the people's opinions, behaviours, or to explain their expectations from the said policy or intervention. In that sense, a focus group is a fast result-driven qualitative survey. The focus group is useful in evaluations of projects or programmes, and particularly for field studies with beneficiaries and intermediary stakeholders. When a focus group is organised after the implementation of a programme with a view to assess its impact, it helps understanding, analysing and identifying the reasons beneath the opinions expressed by the participants. What use can be made of the focus group?
The focus group is a mean to collect information and points of view quickly. When it involves stakeholders with different points of views, it eases the expression and explanation of the discrepancies within those points of view, as well as enabling an in-depth study of the stakeholders' opinions. Such is the case for a health sector focus group involving doctors from the private and public sectors. Less frequently, the focus group can stand as a restitution tool at a local scale. In this case, the tool focuses on the observations and the field analyses' first conclusions. Through the presentation of the survey's first outcomes, this type of focus group collects the reactions of the stakeholders targeted by the intervention. In an impact analysis, the focus group can identify the various groups of stakeholders involved in the intervention, and check their reactions towards a given problem. The objective is to detect diverging opinions within a group composed of allegedly homogeneous opinions. Using a focus group in the impact analysis of the construction of a dam, regrouping people in favour of the project for economic reasons may reveal more precise diverging opinions within the group. The focus group is the only tool with which the evaluator can analyse and test the information given. It helps grasping the participants' behaviours, their understanding and perception of an intervention, which would not be possible with an interview. Group interviewing can collect a variety of points of view and perceptions stimulated by the interaction between participants. Each of the focus group's members is permanently driven to prove one's statement. How is a focus group investigation carried out?What are the conditions for use of the tool?Before organising a focus group, the evaluator should first define the stakes and goals of the evaluation, and determine a theme to which the tool will provide answers. The resources allocated to this task indicate the number of focus groups that the evaluator can forecast. The categories of stakeholders targeted by the evaluation are another component to have in mind while choosing the type of focus groups, knowing that its composition depends on the objectives : an in-depth objective requires a socially homogeneous group, whereas the testing of a theme can only be realised with a group of diverging points of view.
Two types of focus groups can be organised to assess the impact of a policy on a new school course: a first one involving the course's teachers, and a second one gathering inspectors, and the school's directors and teachers. What are the stages for the setting up of the focus group?
How is the focus group implemented?Who implements focus groups?The moderator implements the focus g roup. He/she should be well informed of the evaluation's topics and goals, be familiar with all the techniques relating to group interaction and speak the language of the participants. If the evaluator does not qualify for one of these skills, he/she must be assisted by a local moderator. The latter should be introduced to the tool's goals within the context of the evaluation, and trained to foster group interaction. Key informants are often helpful to the evaluator during the selection of participants and the identification of active participants who can foster the debate. An observer can be invaluable to the evaluation by keeping track of the opinions expressed during the session. How is the group interaction fostered?Prior to the session, the moderator should meet the participants in order to motivate them in becoming actively involved in the focus group. The participants should understand the principles underlying the session's process, and think about the topic before the session. This is particularly recommended for focus groups with users and beneficiaries. A focus group investigation is organised with countrywomen and the chief of the village has the responsibility to recruit them. In this context, the moderator should benefit from a quick meeting with the women to make himself/herself known, establish a relaxed atmosphere and suggest the first topics to reflect upon. Focus groups are not just the sum of individual interviews. Thus, the moderator must always initiate and maintain a dynamic interaction between the participants. A session can be organised in a reactive way, where participants react to the evaluator's analysis, information, etc. as well as in a pro-active way (the information and testimonies of the participants support the development of collective analyses and suggestions). The moderator should organise the session into stages, including a mid-term debriefing to the group. How to keep track of the information?This stage should not be under-estimated for focus groups conducted in a local language. It consists of the transcription of a session verbatim from the notes which have been taken during the session and its recording (if it has been scheduled). In the absence of recording, it can be interesting to organise a debriefing session, in order to validate the content of the focus group's transcription. Example of the focus group investigation in a country evaluation: the Benin mission Four focus groups were set up to bring elements of explanation to one of the evaluation questions (relative to the decrease of the number of patients going to the health centres which have benefited from the EC's assistance). The evaluation team decided to compose and conduct them differently, in order to check the conditions in which focus group investigations should be prepared and carried out in country evaluations. Two focus groups with beneficiaries were conducted by local moderators ; 2 members of the evaluation team were responsible for the moderation of 2 focus groups: one involving doctors from the public and private sector of the Cotonou district, and another involving midwives and nurses from Cotonou. Conclusions from the course of the focus groups and their outcomes can be drawn: the focus group is an efficient collection tool when it gathers socially homogeneous groups within the same socio-professional category. Yet, caution should be taken when participants share too close experiences about the questions under consideration. Their testimonies may turn out to be too identical (for that reason, the focus group with the doctors worked better than the one involving the midwives and the nurses) . The different course taken by the two focus groups with beneficiaries points out that a particular care should be taken for the selection (when recruiting the) of participants. For example, the moderator should ensure the presence of active participants within the group, in order to foster participation (the moderator should also check that leaders do not impose their points of view on the rest of the group). The moderator should also motivate the participants by meeting them a day before the meeting What are the preconditions for its use?
What are the advantages and limitations of the tool?
|
Detailed presentation
.
This section is structured as follows:
- Why and When is a focus group used?
- How is a focus group investigation carried out?
- In country / region evaluations
- Examples
.
WHY AND WHEN IS A FOCUS GROUP USED? |
||||||||||||||||||||||||||
. WHAT ARE THE REASONS FOR USING A FOCUS GROUP?The evaluators use focus groups for two different reasons:
In the evaluation field, focus groups are used to achieve different objectives, at different stage of the evaluation and with various participants. For what purpose should a focus group be used?A focus group can be used as:
Les fonctions de l'entretien collectif ![]() For what outcomes?Focus groups are used in order to:
|
||||||||||||||||||||||||||
WHERE DOES THIS TOOL COME FROM?The focus group comes from sociology, anthropology and the marketing field. A controversial issueSome schools of thoughts refuse to include focus groups in the "family" of group interviewing. As there are several techniques and methodologies available to design a group and conduct sessions, each member of the "family" has its own scientific basis. The choice between these various methodologies depends on the focus group's purpose. Specific examples
|
||||||||||||||||||||||||||
WHAT IS THE ROLE OF A FOCUS GROUP IN EVALUATION?In what kind of evaluation?Focus groups are frequently used in project or programme evaluations, and especially in field studies focusing on the beneficiaries' situation. What type of focus group for which stage in the evaluation?Preparatory focus groupsThey provide grounds for the selection of the major topics and issues, and complete the questionnaire grid. Focus groups for information collectionAt the heart of the evaluation, these focus groups collect information and analyses:
In-depth focus groupsDuring the evaluation, in-depth focus groups gather operators and beneficiaries together to analyse, collect and examine their reactions to the first draft of conclusions and recommendations. With which tools can the focus group be combined?Focus groups can complement interviews, and be a useful alternative to surveys (which often appear to be difficult to plan and manage, and expensive). With case studiesFocus groups can be used to conclude case studies in sector and country/region evaluations. For example, the evaluator can set up focus groups to compare opinions, and focus groups which encourage reactions to the evaluator's suggestions. With interviewsIn this case, focus groups aim at examining and challenging elements of individual analyses and opinions which have been expressed in earlier interviews. |
||||||||||||||||||||||||||
WHAT ARE ITS ADVANTAGES AND LIMITATIONS?The advantagesA means to interview numerous respondentsBy definition, the number of respondents from whom information, analyses and opinions are to be collected, is greater in focus groups than in individual interviews. Consequently, the focus group's reference sample is bigger.
However, participants in a focus group do not constitute a representative sample of the programme's reach, and findings cannot be interpreted statistically because the information collected is qualitative. A tool which facilitates the study of the programme's effectsThis tool is cost effective and can also provide the evaluator with valuable information about the effects of policies / programmes on specific groups of beneficiaries or stakeholders. Considering effect analyses, the evaluator can use this tool to collect the opinions of strategic interest groups about the aims of the policies / programmes under evaluation. A source of creativity, if the group dynamic is controlledFocus groups facilitating the expression of divergent points of view provide opportunities:
Whilst not designed for this purpose, these sessions, can overcome obstacles and contradictions which are often linked to misunderstandings rather than contradictory interests. The limitationsComplex organisation and preparationPrior to the organisation of focus groups, the evaluator usually needs to conduct preliminary analyses on strategic groups which have been formed within the various stakeholders.
Restricted participant expressionSometimes, the evaluator fails to elicit the participant's views due to:
Time allocation and costs (which can easily increase)Conducting several focus groups can be problematic in terms of time allocation and costs, especially due to:
. |
||||||||||||||||||||||||||
HOW IS A FOCUS GROUP INVESTIGATION CARRIED OUT? |
||||||||||||||||||||||||||
. HOW IS A FOCUS GROUP SET UP?Choose the type(s) of focus groupFor each type of focus group, the evaluator must specify:
Schedule the implementation of the focus groupTwo elements must be taken into account:
Within the allocated time, the evaluator may have difficulty in dealing concurrently with the methodological quality of the focus group and implementation delays imposed on him/her. Identify the strategic interests groupsPreliminary documentation analyses and interviews are necessary before the focus groups can be set up. This work helps the evaluator to determine the various strategic interests groups among categories of stakeholders and beneficiaries. These groups usually identify themselves by their expectations, their points of view, their social perceptions and their specific strategies relating to the policy or the programme under evaluation. Select the participantsOnce the strategic groups are identified, the evaluator must select from them the people who will be invited to take part in the focus group. To do so, the evaluator needs a mediator or a local key informant, particularly when representatives of populations living in remote areas, or ones who only speak in dialects are invited. The selection of the participants depends on their availability, the equipment and human resources (an interpreter, for example) at the disposal of the evaluator. In addition to the moderator's skills, which play a crucial role in the overall conduct of the session, the group interaction must also be considered. It can be fostered by one or more leaders who often intervene in the debate and stimulate the other participants. The moderator and the key informant should thus ensure that such leaders participate in the session. The participants' motivation to become actively involved in the focus group should have been stimulated by an in-depth presentation of the topic of the session. They should understand the principles underlying the session's process, and think about the topic before the session. This is particularly recommended for focus groups with users and beneficiaries, who are not familiar with this type of approach. Construct the moderator's guideThe moderator's guide should be designed sufficient in advance to allow the moderator time to study it. The evaluator should specify the topics under study and the themes to be developed during the session, although participants are free to develop other questions linked to the suggested themes. Suggestions should therefore be highlighted in the text (including the information which is to be collected by the moderator). Advice:
Plan the focus groups meetingsThis stage mainly depends on the availability of:
Planning the focus groups also requires sufficient progress in the preparatory work relating to the focus group implementation. |
||||||||||||||||||||||||||
WHO IMPLEMENTS FOCUS GROUPS?Usually, the moderator (or the facilitator) implements the focus groups. He/she should be selected for his/her capacity to create and maintain the group interaction. The quality of the focus group depends on this capacity, as well as a focused and relevant exchange between participants. He/she should therefore be well informed of the evaluation's topics and goals. What are the necessary skills?Focus groups must be conducted by a skilled professional who:
How should the moderator be trained?Although moderators are selected for their abilities to moderate a group, few of them will really know how to moderate a focus group. They should be introduced to the topics and goals of the evaluation, and be trained in the type of techniques which will ensure that the focus group process is successful. A training day should be scheduled. This training day includes the presentation of the moderator's guide, indications about the goals of the focus group and how the evaluator wants it to be conducted. Before the session, the evaluator should check with the moderator that all the terms and expressions used in the moderator's guide are understood. Advice to the moderators:
What are the moderator's tasks?Guide the sessions
Supervise session progress
Encourage the discussion
What are the tasks of the local key informant or intermediary?The participation of key informants during the selection of participants is often helpful to the evaluator. The interaction within the group depends on its composition, and of the presence of one or two very active participants, who can foster the debate. The key informant can ensure that such individuals are included in the group. The key informant can also prepare the participants by explaining the organisation of the session, the topics and the objectives of the focus group. What are the observer's tasks?In addition to the moderator's activities, an observer can be invaluable to the evaluation by keeping track of the opinions expressed (shared and contradictory) during the session. |
||||||||||||||||||||||||||
HOW ARE THE PARAMETERS OF THE GROUP DETERMINED?How are the groups designed to match the evaluation's purpose?Determining the group's composition in support of the evaluation's purpose ![]()
When this tool is used to help a group articulate its of points of view, two possibilities can be considered for the composition of the group: Groups constituted by people sharing similar opinions and professional activitiesWhen the group comprises socially homogeneous members, participants can speak freely. The debate, however, can be stimulated when participants do not have identical views on the questions. Differing analyses, viewpoints and experiences can be challenged on the basis of a common understanding.
A focus group with a homogeneous composition is used to probe information and points of view expressed by stakeholders sharing the same strategic interests (in terms of social and professional status). Thus, the focus group provides the evaluator with an overall view of the experiences, the opinions, the behaviour and the needs of participants. Groups constituted by people having different opinions and professional activitiesIn this case, the evaluator expects that the participants' analyses and points of view about the issue will be divergent. Consequently, the focus group is used as a means to express these divergences publicly. Each category of participants is given the opportunity to explain and defend his/her opinions. The focus group is a means to collect statements about perspectives on an issue. The evaluator can draw the participants' attention to the similarities, the differences and the complementarities of their various points of view. The momentum of this divergent discussion provides the evaluator with elements for new analyses and insights relating to the causes of these divergences. Specific casesThe composition of the groups may sometimes alter the content of the information and the analyses under discussion. Example
The evaluators can also set up groups of agents and officials working for the same institution. In this case, they should be careful about the impact of a point of view expressed in public. Example
The participants can intentionally redirect the focus of the discussion for different reasons. They can speak solely amongst themselves, or they can defend the strategic interests of their group and attempt to manipulate the evaluator, or even accuse him/her of bias. How is the size of the group determined?The size of a focus group varies greatly. However, it should not exceed 10 or 15 people, in order to let everybody talk effectively. |
||||||||||||||||||||||||||
HOW IS GROUP INTERACTION FOSTERED?Presenting the team in charge of the focus group to the participantsPrior to the focus group session, the context of the evaluation and its objectives should be reviewed and the team should be introduced. This will help avoid misunderstandings which otherwise may change the content of the participants' statements.
The evaluator must therefore remember that when he/she organises a focus group with beneficiaries, he/she is perceived as the representative of international donors, which makes him/her an agent for dysfunctions highlighted by participants. The evaluator should be careful about participants' natural tendency to consider the evaluator as a means of communicating with other groups (of civil society or the political arena). This is not a handicap for the group interaction, however, because the exchange between participants is livelier than a collection of personal experience statements. Which methodologies should be selected to conduct a session?Several techniques for conducting focus group sessions are at the disposal of the moderator, to avoid boredom and retain the attention of the participants, and to develop the progress of group interaction. Focus groups are not just the sum of individual interviews. Thus, the moderator must always initiate and maintain a dynamic interaction between the participants. The quality of the interaction depends on the moderator's skills to conduct a focus group.
Recommendations about the moderator's behaviourWhat the moderator should encourage and avoid during a focus group
Focus groups do not yield quantitative data and are not representative of the whole target group (beneficiaries, for example). Thus, the evaluator should remember that the expression of personal opinion is susceptible to a group bias (need for belonging and recognition, consequences of leadership, exaggerated assertions, artificial conflict originating from the debate on various opinions, etc.). How to manage the tempo of the focus group?Several elements must be taken into account:
The same session can be sometimes reactive to the outline of an analysis, information, etc. and sometimes pro-active, i.e. the group is able to yield rational information and analyses by itself. Which type of interaction should be encouraged?Proactive participation needs to be very structured. To this end, the moderator can use visual or audio displays, and introduce practical exercises, simulations, and time dedicated to syntheses and summing up, etc. |
||||||||||||||||||||||||||
HOW TO KEEP TRACK OF INFORMATION?Taking notes on a board, or using other collection devices during a session can help the group to keep track of the information debated and the subsequent synthesis. The evaluator must conclude the session with a debriefing of what has been said, in order to check for any disagreements or misunderstandings. This stage should be kept informal, in order to avoid time-consuming formal validation. The moderator or the observer should categorise participants: people who often express themselves and people who seldom do; sub-groups conversing with other sub-groups and sub-groups who seldom do, etc. These notes are important to put the verbatim comments in its proper context. After the session, the evaluation team compares notes, determines and fills any missing points. The evaluators should also study the group interaction: if it has been impeded, they will have to restructure future focus groups. A good way to keep track of the comments concerning the organisation of the session and the content of the discussion itself is to record the session. The observer can collect all the verbal expressions, while focusing on the most important aspects (which will be developed in the report of the session) and intervening in the session when the moderator omits questions or when the questions are not properly answered. This is particularly recommended for focus group conducted in a local language, and where notes are generally taken in another language (French, English, Spanish, etc.). Indeed, the session's verbatim record is often lost during this instant translation, and the recording fills the missing parts of the notes. The evaluator should plan an additional working day in the focus group schedule, dedicated to the transcription of the notes. |
||||||||||||||||||||||||||
HOW ARE THE FINDINGS ANALYSED?This stage is highly dependent on the quality of the focus group's organisation. In structured focus groupsThe best way to analyse findings from structured focus groups is to gather all the answers and identify the people in the group who have defended certain opinions on a particular question, and the people who have opposed to these points of view. Thereafter, the findings can be easily organised into sub-groups and the evaluator can use them directly. In less structured focus groupsThese focus groups tend to be a conversation between several people. The participants express many views but can agree on only a few conclusions. Thus, the analysis of the findings is similar to the one carried out for interviews. |
||||||||||||||||||||||||||
WHAT ARE THE FINANCIAL REQUIREMENTS FOR ITS USE?The full cost of a focus group can be fairly small. However, cost varies with the number of sessions and the difficulties of organisation (distances, etc.). Elements of cost to be taken into accountIn addition to the remuneration of the evaluators' working days and their travel expenses, the budget must also include:
Should the evaluator pay for participant's time?Some evaluators provide the participants with a per diem, in addition to the transportation and catering expenses. The award of a per diem may lead to bias in the responses of participants, which can undermine the validity of the findings. However, depending on the participant's occupation (for example, a shopkeeper or a farmer), the corresponding loss of income of a session lasting half-a-day or a full day can be financially damaging. Thus, the evaluator should be pragmatic and provide only a limited number of participants with financial compensation. The compensation should be exceptional and the amount low, because the respondents should not participate in focus groups for money, or get the impression that their opinion is being "bought". . |
||||||||||||||||||||||||||
IN COUNTRY/REGION EVALUATION |
||||||||||||||||||||||||||
. WHY USE A FOCUS GROUP IN THIS CONTEXT?In country/region evaluation, the main challenge is not how to collect information, but to determine what information to look for, from whom, about a wide range of questions, and during a limited period of time. What are the specific difficulties in country/region evaluation?The selection of beneficiaries and the understanding of their strategic goalsOne of the country/region evaluation's purposes is to determine precisely the categories of stakeholders who have benefited from the implemented policies in some way. Thereafter, the evaluation must provide an analysis of the various stakeholders' strategies, and rationales considered, within the context of the evaluated policies. Numerous stakeholdersIn country/region evaluation, setting up meetings with numerous and varied stakeholders (such as institutional officials, implementation operators and agents, beneficiaries, etc.) can be challenging. Sample preparationThe sampling of individuals and institutions from whom the evaluator collects information and points of view can yield insufficient results. Indeed, for reasons of timetable management, this selection is often carried out at the beginning of the evaluation process, and may not be adapted to the evaluation's purpose and strategic problems. Lack of timeUsually, the evaluators spend most of their time interviewing officials, to the detriment of impact studies with beneficiaries which are particularly difficult to carry out in country/region evaluation. To what extent is the focus group adapted to these difficulties?Historically, focus groups have seldom been used in country/region evaluation.
Dependant on the nature of the evaluation (ex-ante, mid-term or ex-post), the evaluators deal with these three tasks in a different way. For example, in an ex-ante evaluation, a focus group is conducted to highlight the needs and priorities of the stakeholders who may be affected by a programme or a policy. In this case, impact studies are not a priority. They may be useful, though, to assess past policies or the impact of funding institution's activities, with a view to improving the drafting of new policies. |
||||||||||||||||||||||||||
HOW TO TAKE ADVANTAGE OF THE MAIN FEATURES OF THE FOCUS GROUP?In response to the particular difficulties of country/region evaluation, focus groups concentrate on three objectives. Atool for defining and analysing major topics and issuesIn this context, focus groups can be combined with interviews. First stage: interviewsThe evaluator carries out highly focused interviews, in order to better comprehend the purpose of the evaluation. They are organised with people:
Second stage: focus groupsOn the basis of the findings of these preliminary interviews, the evaluator conducts one or more focus groups with:
As a consequence, the categories of people who have been interviewed individually at the first stage are increased. A broad and rapid information collection toolThis tool obtains information:
A tool for the analysis of impactIts major objective is to determine which categories of stakeholders have been affected by the policy (positively or negatively). Beneficiaries not identified at the policy definition stage may also be included in the evaluation process. Consequently, in country/region evaluation, impact analysis also deals with a series of intermediary stakeholders, from decision-makers to final beneficiaries.
Three types of focus groups
. |
||||||||||||||||||||||||||
EXAMPLES |
||||||||||||||||||||||||||
. |
Check lists
CHECK-LIST FOR EVALUATORS |
||||||||||||||||||||||||||||
.
. |
||||||||||||||||||||||||||||
CHECK-LIST FOR MANAGERS |
||||||||||||||||||||||||||||
.
|
Survey
SUMMARY |
||||||||||||||||||||||||
. Why is this tool used in evaluation?A survey is an observation tool which quantifies and compares information. Data are collected from a sample of the population targeted by the evaluation. A survey questionnaire is a schedule of questions collecting information needed for the evaluation. Respondents are not supposed to answer the essential issue under investigation: in a good questionnaire, questions derive from the evaluation questions and are sufficiently basic for the respondent to answer them. Questionnaires often combine both types, with a preference for structured items and a few open-ended questions (yielding information which is more diverse and/or precise, but less amenable to statistical analysis). Structured questionnairesStructured items are questions which respondents must answer in a specific way by choosing from a limited and predetermined set of answers. The questionnaire format is designed to obtain information about facts, to find out whether respondents agree to a suggestion, to record their opinions on a set of assertions, etc. Open-ended questionnairesIn open-ended questionnaires, respondents answer a precise question and interviewers take notes. Thus, open-ended questionnaires are similar to structured interviews, as open-ended items allow a variety of approaches and depth in response. What use can be made of the survey in evaluation country?A survey is the best tool for collecting information from the population, and to compare and quantify the various opinions expressed. Its use is particularly relevant to know the final beneficiaries' degree of satisfaction concerning a policy. Structured questionnaires have the advantage of allowing a cost-effective statistical analysis. Yet, they can be unfitted during the survey's implementation course when the evaluator needs to refine some questions. Through a daily check with the interviewers, the evaluator can decide to develop or add questions during the interviews, with a view to undertake a more precise analysis. How is a survey carried out?
An efficient time management is a prerequisite for the survey. During the tool's testing mission in Benin, the local partner was identified a month and a half prior to the field mission; the sample and questionnaire design, and interviewers training were organised on-site 2 weeks before the arrival of the testing team. How is the questionnaire developed?Relevant questions for the evaluators
Structuration of the questionnaire
How is the survey carried out?Design the samplesThe methodology selected depends on the determination of the population that constitutes the target group of the survey. This determination is linked to:
Various types of sampling can be developed: simple (random sampling), stratified, cluster sampling, probability proportional sampling, progressive, etc. Questionnaires among households during the testing mission in Benin: " The method to collect data is the itinerary method with which the interviewer can identify the survey's targets by counting the households while covering each street of the scoped area, register them in a household counting slip, proceed to the selection of the household samples and interview the head of family an/or his wife. The counting should start from the chief of the village's house and progress clockwise, segment after segment, so as to cover the whole village and reach the number of households to be surveyed. The number determining the first household to be surveyed is random and given by the survey's centre. The numbers of the other households to survey will be determined by the drawing number established by the survey's centre. For example: if the first number is 3 and the drawing number is 5, the first household to be interviewed will be the third of the survey's list. The other households will therefore have the number 8, 13, 18, 23, etc. " Conduct a pilot questionnaireA good quality survey relies on the clarity of the question wording, the ease of response, the questionnaire's length and flow, problems encountered by the interviewers, etc. How is the questionnaire conducted?The choice of a specific type of survey depends mainly on the context:
The evaluator should keep in mind that, because of linguistic reasons (mostly with final beneficiaries) and time spent in the preparation of the survey, a local partner is strongly advised. The latter should be able to provide the evaluator with human resources - interviewers, statisticians, demographers, etc. - and material resources - transportation, IT, demographic data, etc. - all of which the evaluator may not have available on-site. How are the findings analysed and treated?
|
Figure 3 : the preconditions for its use | |
The time span | Carrying out a survey requires great care at the preparation stage, and an allocation of time in proportion to the importance of the survey, the extent of sampling and field difficulties.
The elaboration of the questionnaire dedicated to specific groups requires sufficient data and hypotheses, which means that the survey cannot take place at the start of the evaluation. |
Human resources | Where cultural and linguistic specificities are important, it is better to have locally recruited interviewers.
Specialist organisations may sometimes be able to support the evaluator's recruitment process. The evaluator should organise one or more training/debriefing days for the interviewers. |
Financial resources | Remuneration of the interviewers.
Transportation expenses. |
.
What are the advantages and limitations of the tool?
Figure 4 : The advantages and limitations of the tool | |
Advantages | Yields quantified and reliable data
Useful to identify changes and make comparisons between opinions Enables the evaluator to survey a large number of final beneficiaries Enables the evaluator to work on a target population and on a limited scale Identifies the outcomes of programmes and policies |
Limitations | Requires implementation delays exceeding the average time scale of an evaluation mission
Requires important resources and logistics provided by a reliable local partner Requires pre-existing data on the initial situation Requires a large number of staff to conduct the survey and analyse the findings May present difficulties during the development of representative sampling |
Detailed presentation
.
This section is structured as follows:
.
DEFINITION |
|||||||||||||||||||||||||
. WHAT IS A SURVEY?What are the reasons for using a survey?A survey collects data over broad populations, by means of a standardised set of questions. Are there different types of questionnaire?There are two main types of questionnaire:
Usually, questionnaires combine both types, with a preference for structured items and few open-ended questions (yielding information which is more diverse and/or precise, but less amenable to statistical analysis). Structured questionnairesStructured items are questions which respondents must answer in a specific way by choosing from a limited and predetermined set of responses. The questionnaire format is designed to obtain information about facts, to find out whether respondents agree to a suggestion, to record their opinions on a set of assertions, etc. Structured items do not collect detailed information, but as the categories of response are predetermined, they provide the basis for efficient statistical analyses.
Types of questions in structured questionnaires
In structured questionnaires, the respondent may be inclined to respond in a way which does not accurately reflect his opinion, because:
Open-ended questionnairesIn open-ended questionnaires, respondents respond to a precise question and interviewers take notes. Thus, open-ended questionnaires are similar to structured interviews, as open-ended items allow a variety of approaches and depth in response.
Types of responses in open-ended questionnaires
|
|||||||||||||||||||||||||
Limitations of its useThe survey is an outstanding observation tool for the collection of information from a large number of final beneficiaries. Although structured questionnaires can measure the relative weight of the opinions expressed, they must be conducted with a carefully constructed sample so as to be statistically valid. Thus, facilities and means are required, such as:
The use of open-ended questionnaires can be an alternative to structured questionnaires when the evaluation does not need a statistical analysis of the data collected. Past experiences have shown that the information collected by this type of questionnaire eases the understanding and analysis of the population's reactions, and provides answers to the evaluation's requirements. |
|||||||||||||||||||||||||
WHERE DOES THIS TOOL COME FROM?Its originsThe use of surveys and sampling techniques emerged after the First World War, during electoral forecasting. Polls institutes were the first to use surveys. Nowadays, surveys and sampling techniques represent the social sciences' reference tools. The demands from quantitative sociologists have led to the development of surveys and sampling techniques to be able to benefit from the statistical data. Its use in other fieldsIn addition to opinion polls institutes, the marketing field is a significant user of surveys. In anthropology and ethnology, observation surveys are designed as semi-structured interviews. This type of survey can be used in some case studies, but is seldom used in evaluations due to constraints on time and budget. Indeed, these surveys are more useful for researchers, although their findings may provide the evaluator with insights into the general context and the socio-cultural dynamic of the region under evaluation. . |
|||||||||||||||||||||||||
WHY AND WHEN IS THE SURVEY USED? |
|||||||||||||||||||||||||
. WHEN AND HOW SHOULD A SURVEY BE USED?In which situations is this tool appropriate?An observation tool collecting quantitative and qualitative informationThe survey is an observation tool which collects:
In developed countries, opinion pollsters and market researchers often conduct surveys because they yield quantitative data, whereas in developing countries, they are less common because of constraints of costs and difficulties in implementation. As a general rule however, surveys remain an appropriate tool to collect information in studies and evaluations. Surveys require good definitions, and careful management of the information under investigation and of the people from whom the information is to be collected. Surveys should be implemented after a piloting stage. Theoretically, surveys can provide good indications of change, if carried out with the same sample and within a period of time sufficient to yield information about the impact of a policy or a programme. In practice, however, these conditions seldom prevail in developing countries. Warning about its use in evaluationThe evaluator will experience fewer difficulties designing a precise set of questions in project and programme evaluation than in country/region evaluation, where the process of identifying real beneficiaries and stakeholders can be difficult. |
|||||||||||||||||||||||||
What type of information does a survey yield?In evaluation, surveys are mainly used to assess the impact of policies and programmes on various categories of people.
Surveys are carried out:
|
|||||||||||||||||||||||||
What are the advantages and limitations of a survey?
Despite the tool's advantages (especially comparing with other observation tools), limitations in the time span, the financial resources and the technical means can limit its use in evaluations. |
|||||||||||||||||||||||||
Can a survey be combined with other tools?Structured questionnaires bring complementary quantitative information to other information collection tools (which only yield qualitative data), such as:
These tools can be used during the exploratory stage of a survey (to identify or detail working hypotheses) to ease the drafting of the questionnaire with:
A series of interviews or focus groups can be organised prior to a structured questionnaire, in order to highlight the context of the outcomes, and to support the understanding of the data and its interpretation (for example, to describe a custom, or explain an individual perception). Surveys' findings also support other tools which are more complex, such as:
|
|||||||||||||||||||||||||
WHAT ARE THE PRE-CONDITIONS FOR ITS USE IN EVALUATION?The time spanCarrying out a survey requires great care at the preparation stage, and an allocation of time in proportion to the importance of the survey, the extent of sampling and field difficulties. Human resourcesSetting up a survey necessitates the appointment of interviewers. Required skills for interviewers
The recruitment
Financial resourcesMost of the costs of surveys are remuneration of the interviewers and transportation expenses, if the survey is carried out using widely dispersed interviewees. It is not possible to specify exact amounts because expenses vary greatly with the type of country in which the survey is implemented. However, the transportation budget is critical in countries where transport is scarce. Generally, in assessing the financial requirements, the following should be taken into account:
Surveys belong to the category of tools which are the easiest to use, allowing the evaluator to obtain information from a large number of people. Surveys by telephone are cheaper than face-to-face surveys. Surveys by mail are the least expensive, although the response rate is usually low and limits information about the respondents profile, especially in countries where computer networks are poorly developed. |
|||||||||||||||||||||||||
WHAT IS THE ROLE OF A SURVEY IN EVALUATION?Why should it be used in evaluations?Although open-ended questionnaires require skilled resources for drafting, the organisation of implementation and their analysis, the time spent in data collection on site is limited in most evaluations for financial reasons. Yet, in certain cases, the evaluation questions require the collection of information from final beneficiaries. In this context, the questionnaire is crucial because it collects information about the viewpoints and practices from a large range of people. When the completed structured questionnaire is analysed statistically, it quantifies the opinions surveyed, which can be supportive of an impact analysis, or the drafting of evaluation or programme indicators. This type of analysis is however difficult to undertake and requires more resources than in any other analyses. What are the requirements for the use of a survey?
|
|||||||||||||||||||||||||
WHY IS THE SURVEY USED IN COUNTRY EVALUATIONS?As surveys need specific conditions for their implementation in country/region evaluations, they have proved to be more appropriate for countries at a high level of development. In developing countries, however, the evaluator may survey target groups to obtain information (such as quantitative data) which cannot be collected by other means. In country/region evaluations, surveys have a selective purpose and focus on geographically defined categories of stakeholders. Consequently, the evaluator needs to closely supervise the interviewers, or delegate the supervision to a reliable local partner. The evaluator can be confronted with continuing problems relating to the collection and speedy selection of the required information from abundant sources. In addition, it may be difficult to reach final beneficiaries. But, by interviewing a large number of beneficiaries, surveys can provide the evaluator with part of the information on the policy under evaluation, such as:
Currently, surveys and questionnaires are rarely used in country/region evaluations but as surveys can provide useful analyses of impact relating to various groups (and particularly beneficiary groups), the number carried out in the future should increase. . |
|||||||||||||||||||||||||
HOW IS A SURVEY CARRIED OUT? |
|||||||||||||||||||||||||
. HOW IS A QUESTIONNAIRE DEVELOPED?What are the relevant questions for the evaluator?What is required?As the design of the questions depends on the target population, and on the information, the analyses and opinions which are sought by the evaluator, the questionnaire should start with a list of data about the study's topic, including:
This list organises the stages of the survey and the drafting of the implementation timetable. Is this question strictly necessary?Asking this implies that the purpose of a study can only be fulfilled if the question is posed. Therefore, if the answer is 'no', the question should not be included in the questionnaire. Will a single question be sufficient to obtain this information?Depending on the complexity of the information required, several questions may be necessary, which is particularly important for elements that seek to examine causality. Is the respondent in a position to answer the question?Three main reasons may account for a lack of response:
Will the respondent provide the evaluator with accurate information?This issue mainly deals with information whose nature may be sensitive, such as:
In these cases, respondents may be tempted to alter or avoid a topic. The evaluator may anticipate or avoid this reaction by the skilful choice of questions or by structuring the questionnaire. |
|||||||||||||||||||||||||
How should a questionnaire be structured?The questionnaire's basics
Introductory and transitional documentsThese documents are necessary for the clarity and the natural flow of the questionnaire. They must be adapted to the survey's type, its topic, and the socio-cultural characteristics of the sample.
Sequence of questions
|
|||||||||||||||||||||||||
What is a typical format for a questionnaire?Presentation of the interviewer and the surveyThis stage, sometimes overlooked, is useful to introduce the questionnaire to respondents. Introductory questionsIntroductory questions aim at arousing the interest of respondents and establishing a rapport in the case of a face-to-face survey. They convey the general scope of the topic and are formulated to obtain a positive response, even though this response may not be directly useful for the survey. Background questionsBackground questions are used to check whether a respondent has the required information or, if necessary, give the interviewer contact details of an informed respondent. Background questions often take the format of dichotomies. Warming-up questionsWarming-up questions contribute to progressively focusing the questionnaire on the topic of the study. Usually, they take the form of straightforward questions (for example, behaviour questions). Specific questionsSpecific questions represent the core of the questionnaire: responses to these questions must provide the study with critical and focused information. At this stage, the respondent should be completely at ease with the interviewer and concentrate on the topic. The interviewer may ask him/her personal questions, or press him/her to reflect upon his/her motivation, opinions and judgements. Attribute questionsWith the help of responses to attribute questions, the evaluator describes respondents and incorporates them into sub-groups. These questions yield two types of information:
|
|||||||||||||||||||||||||
How should a questionnaire be written and presented?The questionnaire's basics
The vocabulary
The style
Ways of introducing the questionnaire to respondents
The cover letter should be carefully prepared, especially its general appearance. |
HOW ARE DESK-BASED SURVEY TECHNIQUES CARRIED OUT?Why are desk-based survey techniques carried out?The evaluator develops a line of questioning supported by his working hypotheses, with a view to designing the questionnaire. In many cases, pre-field techniques are used to validate these hypotheses and the resulting questions. More generally, they are carried out to improve the questionnaire, prior to the start of the survey. Techniques for the validation of the questionnaireComplementary techniques may be used to validate the questionnaire and the whole drafting process. These techniques include:
These techniques enable the evaluation team to agree on the same definition of the study's key concepts. This stage is necessary for the questionnaire wording process. |
||||||||||||
HOW ARE SAMPLES DESIGNED?The constraints of the population selectionThe selected methodology depends on the determination of the population that constitutes the target group of the survey.
Basics for the sampling techniqueSample quality is key to the reliability of the findings. The sample should reflect as closely as possible the whole population. It should be a scale model of the population. The evaluator's dilemmaIn programme or co-operation policy evaluation, samples are often hard to design: registry documents, professional index, etc., relating to the population under study are seldom available, and information about the population's characteristics are often non-existent, unavailable or unreliable. The situations in which the survey is to be conducted are often very difficult. The various types of sampling
How is the size of the samples determined?Basics: the findings' accuracy increases with the number of respondents. ![]() Therefore, more extensive sampling than the survey's requirements must be scheduled, so as to anticipate possible refusals, non-attendance of a session, disruptions, etc., which often occur during the implementation of a survey. The respondent listThis list should include the respondent's name, profession, contacts, and, if required, his selection characteristics. |
||||||||||||
HOW IS THE PILOT QUESTIONNAIRE CONDUCTED?Why is the questionnaire piloted?A good quality survey relies on the piloting of the questionnaire and of the survey's methodology, in order to identify problems relating to the clarity of the question wording, the ease of response, the questionnaire's length, problems encountered by the interviewers, etc. The content of the pilot questionnaire
Because surveys are expensive and difficult to administer (constraints of time), they are not usually systematically piloted during an evaluation, although this could add a qualitative dimension. |
||||||||||||
HOW IS THE SURVEY CARRIED OUT?The different types of surveyThe choice of a specific type of survey depends mainly on the context:
The various types of questionnaire and their particular characteristics
AdviceThe carrying out of surveys with beneficiaries may pose logistical problems, especially in far-reaching areas. Among other distortions, the evaluator must be careful to avoid an over-representation of the most accessible areas (for example, nearby main roads), to the detriment of remote areas which may never be selected in field inquiries. If these elements are overlooked, the evaluator may miss specific elements of social organisation, such as access to equipment and utilities. |
||||||||||||
HOW ARE THE FINDINGS ANALYSED AND TREATED?Coding questionnairesAfter the receipt of the completed questionnaires, the code development stage can begin. This stage is necessary for open-ended questions which call for statistical analysis. Codes should reflect the type of analysis which has been selected. Counting findings and correlationsAfter coding the questionnaires and entering the codes in a data processor, the evaluator can process the responses which should yield findings corresponding to the working hypotheses.
The treatment of responses
Analysing findingsThe analysis of the findings relates to the working hypotheses. Pre-field techniques and documentation collected are also used during this analysis stage. Going back to the original responses may sometimes be useful to investigate certain elements (for example, to make additional "overlapping" of responses).
The different types of analysis
Numerous statistical methodologies can be developed, whose usage is more or less complex. In the evaluation field, simple treatments are often sufficient for the analysis stage, even though establishing correlations between items enriches the analysis. Margin of errorsErrors almost always occur in the treatment of findings (the survey's conditions, sampling errors, constraints of time, costs, etc.) and a margin of errors should not be overlooked during the analysis of findings. Developing the survey reportThe survey report must include an analysis of basic findings. It may contain, for example:
The evaluator must ensure as a minimum that the survey report has:
. |
||||||||||||
GLOSSARY |
||||||||||||
. Sample/samplingSamples are designed to gather people who are to be effectively surveyed. People are selected from the evaluation's target population, and this selection depends on the methodology chosen for the design of the samples (see How are samples designed?). Working hypothesesApart from assuming an empirical approach for the evaluation (which is almost impossible), scientific researches usually: (1) develop hypotheses, and (2) check them. This principle supports the survey's design. The survey's findings enable the evaluator to check the hypotheses at the core of the survey's development, to reinforce them and/or to correct them. Target groupThe target group represents the part of the population on which the survey focuses. A typology or descriptive variables determine it. TypologyThe typology is a methodology dividing a population into types or homogeneous sub-groups with specific criteria (economical, social, demographic, etc.). Within each sub-group of the typology, respondents share similar characteristics (corresponding to the typology's criteria), or even the same values for certain predetermined variables.
|
||||||||||||
EXAMPLES |
||||||||||||
|
Check lists
CHECK-LIST FOR EVALUATORS |
||||||||||||||||||||||||||||||||||
.
. |
||||||||||||||||||||||||||||||||||
CHECK-LIST FOR MANAGERS |
||||||||||||||||||||||||||||||||||
.
|
Expert panel
SUMMARY |
||||||||||||||||||
Why is this tool used in evaluation?An expert panel usually comprises independent specialists, recognised in at least one of the fields addressed by the programme under evaluation. The panel specialists arrive at conclusions and recommendations through consensus. Depending on the project proposal, these recommendations deal with the implementation or the impact of a programme, or part of it. The expert panel is specifically appointed for the evaluation, and in conformity with standard procedures. The panel holds meetings and provides conclusions and recommendations in accordance with a precise and replicable working plan. The use of an expert panel in country/region evaluations can be helpful in several situations, such as:
How is a panel expert carried out?What criteria should be used to appoint the panel?The pre-requisite for the expert's selection is his/her professional experience. He/She should have specialised in the field under evaluation, and be recognised and respected by his/her peers. Experts must be independent of the programme under evaluation, because they should not be judge and judged. Independence regarding the programme under evaluation is very important since the expert cannot be the judge and the judged.
The ability to work in a group, listen to other experts and be open-minded is an essential criterion. Otherwise, working conditions may quickly turn out to be unmanageable, which would impede the panel process. How are experts selected?In a straightforward selection, the evaluation managers have access to a list of acknowledged experts in specific fields, and limit their selection process to ensuring the expert's independence regarding the programme under evaluation. In gradual selections, preferred profiles of experts are developed with respect to the topics under scrutiny in the evaluation. Elements to be taken into account in developing the panel profile are as follows:
What are the procedures for the management of the panel?There is no unique working process, and the expert panel should be encouraged to plan and implement its own workplan. Experts can focus their work on documentation and sessions, or broaden it to include meetings with project managers, field visits, implementation of surveys, etc. The first panel session must result in the experts having a full understanding of their role in the evaluation. During this session, the applicable methodology for the management of the panel's work must be discussed and validated. The discussion usually focuses on:
The next sessions (ranging from 3 to 5) will be directly linked to the panel's work. They will systematically deal with:
What is the role of the panel chairman?The panel chairman plays a crucial role. He/She guides the study panel, proposes the working arrangements, records findings, encourages contributions, facilitates debates and is the chief spokesperson for the panel. The quality of the working arrangements often depends on the chairman's leadership.
The chairman as Panel facilitator schedules the work of the panel and its production, and steers the panel's progress toward consensus. The chairman as Report Architect and Integrator ensures a critical overview to the panellists' outputs, so as to improve the debate. The chairman as Project Manager ensures that the available resources are sufficient and properly employed throughout the study. He/She ensures that the panel's sessions have been properly prepared by the technical writer. The chairman as Spokesperson represents the panel in various bodies (such as monitoring committee and meetings with the commissioning agency and the press). How does the expert panel report on its work?The report, which supports the experts' contribution to the evaluation, is the only output from the panel which is made available to the commissioning agency. The report's structure should include: an executive summary, the mission's terms of reference, the composition of the panel, the evidence gathered and reviewed, the analyses carried out, The conclusion of the experts in the context of the report's consensus findings. What are the preconditions for its use?
What are the advantages and limitations of the tool?
|
Detailed presentation
.
This section is structured as follows:
- Why and when is an expert panel established?
- How is the panel process managed?
- Examples
- Bibliography
.
WHY AND WHEN IS AN EXPERT PANEL ESTABLISHED? |
. What is the role of an expert panel in an evaluation? |
What is an expert panel?DefinitionAn expert panel usually comprises independent specialists, recognised in at least one of the fields addressed by the programme under evaluation. The panel specialists arrive at conclusions and recommendations through consensus. Depending on the project proposal, these recommendations deal with the implementation or the impact of a programme, or part of it. When consensus is not reached for particular questions, the panel must report on the various perspectives of the experts. The expert panel is specifically appointed for the evaluation, and in conformity with standard procedures. The panel holds meetings and provides conclusions and recommendations in accordance with a precise and replicable working plan, which accounts for its reliability. This tool is therefore designed to take advantage of the experts' knowledge in assessing policies, programmes and projects implemented in the field of their expertise. Various types of expert panels in evaluationInternational funding institutions such as the World Bank, the European Commission, Scandinavian countries, Canada and the United States regularly appoint expert panels to evaluate programmes. In environmental evaluations, funding institutions such as the World Bank use expert panels to assess the quality of the service providers' work and suggest improvements. These panels examine the evolution of the project studies, from their preliminary stages to completion, including the first years of implementation. The Delphi Method is another type of expert panel used for evaluation, based on an anonymous and repeated postal survey with experts (see Means Documents). Where does the expert panel come from and how has it evolved?The panel's originThe concept of the expert panel originates in the research field. It derives from the peer review of scientific work processes or programmes, where expert panels have developed as an evaluation tool. Its evolutionTwo main developments over the past few years are noteworthy.
|
When is an expert panel appropriate for an evaluation? |
What are the pre-conditions for its use?Expert panels are appropriate for many evaluation situations, and particularly for:
Various tasks of an expert panel ![]()
With which tools can the expert panel be combined?The expert panel can be combined with almost all the usual evaluation tools. As well as combined with other tools, the expert panel's work is sometimes complemented by external studies. This usage has become widespread in fields such as research evaluations because it eases the task of the experts and provides them with information about the programme under evaluation. External studies include preliminary studies, surveys, database analyses. . |
What are the advantages and limitations of an expert panel? |
. The advantagesThe experts' knowledge of the subjects under evaluation is the principal advantage of this tool. It fosters:
The limitationsThe tool's limitations which should be minimised essentially derive from a series of risks:
Even when these risks are controlled, social science detractors remain sceptical about the reliability of expert panel's conclusions. . |
What are the pre-conditions for an useful panel contribution to an evaluation? |
. The time spanOne of the advantages of an expert panel is in its speedy assembly process. For an evaluation, only 3 to 6 months work needs to be scheduled, and even less time for panel advice on a technical field within an evaluation. Human resourcesThe core aspect of an expert panel is the issue of human resources. Experts must have recognised expertise in the field under evaluation, be independent of the programme being assessed, be able to work in a group and be available for a continuous work throughout the evaluation. Financial resourcesThe expert panel is known for its cost-effectiveness, but in the case of travel into remote countries and field visits, the expenditure allocated for experts' salaries and expenses will need to be increased. Budget line items normally taken into account while preparing estimates are as follows:
. |
Why and how is an expert panel established in country/region evaluations? |
. Reasons for establishing an expert panelThe use of an expert panel in country/region evaluations can be helpful in several situations, such as:
Conditions for its use in country/region evaluationsNo specific conditions are required for the use of expert panels in country/region evaluations. However, experts must be familiar with the context of the country assistance under study. Examples of its contribution to country/region evaluationsNo example of its contribution to country assistance has been found yet, but this tool is frequently used in programme evaluations in Western countries. A series of examples highlight the use of expert panels in the development of terms of reference documentation, recruitment profile processes, samples of reports. . |
HOW IS THE PANEL PROCESS MANAGED? |
. What criteria should be used to appoint the panel? |
Core criteria of the expert panel's composition
![]()
Professional experienceThe pre-requisite for the expert's selection is his/her professional experience. He/She should have specialised in the field under evaluation, and be recognised and respected by his/her peers. The credibility of his/her conclusions is highly dependent on these elements. IndependenceExperts must be independent of the programme under evaluation, because they should not be judge and judged. Thus, experts having a direct conflict of interest (such as experts participating in the programme or belonging to a body which benefits from it) should not be appointed. However, securing sufficient independence is difficult in fields where there are few acknowledged experts. In this case, the panel, and not the experts, must be independent, and differing points of view must be dealt with in an even-handed way. In addition, the experts' appointment to the panel is personal, which means that they do not represent their home institution and cannot be substituted on the panel. Ability to work in a groupThe ability to work in a group, listen to other experts and be open-minded is an essential criterion. Otherwise, working conditions may quickly turn out to be unmanageable, which would impede the panel process. Other criteriaIn addition to these core criteria and specific requirements relating to the expert's profile, some commissioning agencies may have their own views on the panel's composition (such as the expert's nationality, a balanced representation of differing points of view, the participation of specific categories of panellists such as beneficiaries, consumers). Ideally, experts should also speak the language used in the area subjected to evaluation. |
How are experts selected ? |
Selection process for the panel members ![]()
Procedures for the recruitment of expertsVarying with the importance of the task and the complexity of the themes, the recruitment of panel members can be relatively straightforward and speedy, or require a time-consuming selection process. Straightforward selectionIn a straightforward selection, the evaluation managers have access to a list of acknowledged experts in specific fields, and limit their selection process to ensuring the expert's independence regarding the programme under evaluation. Gradual selectionGradual selections have become a common procedure. Preferred profiles of experts are developed with respect to the topics under scrutiny in the evaluation. These profiles are critical to satisfactory recruitment to meet the evaluation's needs. If the panel's intervention only focuses on technical issues, the expert's lack of independence within the fields under consideration will usually have minor impact on the evaluation, because the evaluator is responsible for the main tasks. However, if the panel has to carry out a significant part of the evaluation, it should include an expert experienced in evaluation processes and a socio-economist, in addition to experts specialised in the topics under review. Depending on the nature of the mission assigned to the experts, additional fields of competence may need to be proposed. Development of the panel profileThe Royal Society of Canada details elements to be taken into account in developing the panel profile, as follows:
The selection processOnce these profiles have been developed, the institution managing the evaluation should establish a "long list" of experts and remove experts with possible conflicts of interest. Thereafter, the institution contacts the selected experts. Depending on the number of experts required (5 to 10 in most cases), the institution should pre-select a larger number of potential experts (2 to 3 times more than the final list), in order to ensure the availability of experts. Within this "long list", some funding institutions may determine nominees and alternates. This procedure is also recommended for the nomination of the chairman and key experts. Once the selection process is over and the experts are recruited, the commissioning agency (or the experts) proceeds with the appointment of the chairman. As the role of the chairman is essential for the efficient working environment of the panel, the importance of this appointment is paramount. He/She will be responsible for setting the tone and rhythm of the panel's work. Secretarial work should be the responsibility of a technical writer, whose ongoing availability should compensate the pressurised timetables of the experts. The tasks of the technical writer include the production of working reports, the incorporation of suggestions, and, if required, the monitoring of external studies. |
What are the pre-conditions for the expert panel's work? |
A precise definition of the panel's field of workEvaluators must take time to explain to the experts the context of their work and provide them with information about the programme, the procedures, and evaluation methodologies. In doing so, evaluators help the experts formulate their conclusions with a full knowledge of the environment of area assessed. Usually, the content of the panel's working sessions is confidential. Availability of documentation for the expertsTerms of referenceTerms of reference are often provided to ease the panel's working arrangements. They should specify:
Reference interview guidesIn evaluations where calls for expert panels are systematic, developing procedures for the management of the panel process, as well as interview guides may ease the panel's work (for example, during interviews with officials, field study, interviews with beneficiaries). These procedures can also benefit repetitive evaluations, through the standardisation of reports, which allow for comparisons. Interview guides may be provided to evaluation teams unfamiliar with evaluation process, with a view to assisting them in their tasks. |
What are the procedures for the management of the panel? |
Fundamentals for the expert panel's workThere is no unique working process, and the expert panel should be encouraged to plan and implement its own workplan. Experts can focus their work on documentation and sessions, or broaden it to include meetings with project managers, field visits, implementation of surveys, etc. The choice of working process is highly dependent on the area studied, the experts mission, and the information and resources at the disposal of the experts. Experts are expected to investigate and analyse the assigned topics and present their conclusions in a written report. The quality of the drafting and writing of the report is crucial and must be given careful attention. Expert panels are usually expected:
Some commissioning agencies believe that the search for consensus usually results in the experts producing an anodyne and unrealistic report. They require a report highlighting the experts' different points of view and the reasons for these differences. Confidentiality of the panel's debates and intermediate findings is another rule impacting on the panel's working arrangements. Guidelines for the panel sessionsThe first panel sessionThe content of the first panel session derives from the terms of reference. This session must result in the experts having a full understanding of their role in the evaluation. During this session, the applicable methodology for the management of the panel's work must be discussed and validated. The discussion usually focuses on:
During this session, panellists should be reminded of the general rules (such as independence and consensus) because although experts are generally familiar with the topics under study, they are often less well informed about evaluation principles. The possibilities of conflict of interest with the programme to be studied should be closely examined, discussed and resolved during this session. The next sessionsThe following sessions (ranging from 3 to 5) will be directly linked to the panel's work. They will systematically deal with:
With a view to ensuring the confidentiality of the panel's work, certain commissioning agencies recommend that records, summaries and intermediary reports of the sessions are destroyed. Only the final report is kept as the formal output required from the panel. The organisation of the missionThe organisation of the mission depends on the panel's tasks. Most of the experts will be unfamiliar with evaluation techniques and may live far from the session's location. Thus, the production of the expert panel work programme should be scheduled well in advance, preferably as soon as the mission starts. This work programme should be adhered to whenever possible. Experts responsible for tasks between two sessions, such as field visits, will be expected to work at least in pairs, in order to avoid bias of interpretation or empathy (the limitations). The formation of small groups should reflect the various points of view represented on the panel. |
What is the role of the panel chairman? |
The panel chairman plays a crucial role. He/She guides the study panel, proposes the working arrangements, records findings, encourages contributions, facilitates debates and is the chief spokesperson for the panel. The quality of the working arrangements often depends on the chairman's leadership. The various roles of the panel chairman ![]()
The chairman as Panel Facilitator
The chairman as Report Architect and IntegratorThe chairman guides the study, defines methodologies, reviews outputs, ensures that timetables are respected, and records the findings of the panellists, which includes:
The chairman as Project ManagerThe chairman ensures that the available resources are sufficient and properly employed throughout the study. He/She is in permanent contact with the commissioning agency on financial and technical issues. If sub-contractors work for the panel, the chairman is responsible for the management of their studies, the supervision of their progress and their successful completion of their work. He/She ensures that the panel's sessions have been properly prepared by the technical writer, and that all documentation and means required for their participation in the sessions are provided to the experts in a timely fashion. If the mission requires a readjustment of its budget, its time allocation or its objective, the chairman presents this to the commissioning agency in order to reach an agreement. The chairman as SpokespersonThe panel will need to be represented in various bodies (such as monitoring committees) and possibly in meetings with the commissioning agency and the press. As it is impracticable to gather all the experts for these meetings, the chairman serves as the spokesperson for the panel. He/She may delegate certain tasks to other panel members, but he/she should conduct the most important meetings. |
How does the expert panel report on its work? |
Synthesis of the panel's studyAt the end of their mission, the experts report on their work. The report, which supports the experts' contribution to the evaluation, is the only output from the panel which is made available to the commissioning agency. Consequently, the report should be carefully prepared. Guidelines for the final reportThe report's structure depends on the nature of the mission. In technical or scientific missions, the report should at least include:
Reaching a consensus is the most challenging task for managers of expert panels, because a consensus strengthens the value of the panel's conclusions. In this context, the role of the chairman is crucial in seeking consensus, or formulating the final position of the panel, even if it include some dissenting views. . |
EXAMPLES |
|
BIBLIOGRAPHY |
General references
Specific references for the use of expert panels in country/region evaluations
|
Check lists
CHECK-LIST FOR EVALUATORS |
||||||||||||||||||||||||||||||||||
.
. |
||||||||||||||||||||||||||||||||||
CHECK-LIST FOR MANAGERS |
||||||||||||||||||||||||||||||||||
.
|
Detailed presentation of the tool
WHAT IS THE ROLE OF AN EXPERT PANEL IN AN EVALUATION? What is an expert panel? Definition The expert panel is specifically appointed for the evaluation, and in conformity with standard procedures. The panel holds meetings and provides conclusions and recommendations in accordance with a precise and replicable working plan, which accounts for its reliability. This tool is therefore designed to take advantage of the experts' knowledge in assessing policies, programmes and projects implemented in the field of their expertise. Various types of expert panels in evaluation In environmental evaluations, funding institutions such as the World Bank use expert panels to assess the quality of the service providers' work and suggest improvements. These panels examine the evolution of the project studies, from their preliminary stages to completion, including the first years of implementation. The Delphi Method is another type of expert panel used for evaluation, based on an anonymous and repeated postal survey with experts (see Means Documents). Where does the expert panel come from and how has it evolved? The panel's origin Its evolution
To find out more:
|
Case study
SUMMARY |
||||||||||||||||||||
. Why is this tool used in evaluation?The implementation of case study reviews of one or more actual examples, in order to gain an in-depth knowledge of the topic and, if possible, to learn about the entire evaluation programme. In complex situations, case studies are the preferred evaluation tool when "how" and "why" questions are being posed, They allow a detailed examination of the actual elements in line with the evaluation goals. The purpose of the case study is to provide a picture, which is often more reliable than the outputs from other tools in context of the scarcity of basic data (which is often the case in country evaluations).
If case studies include the analysis of documents, statistical and implementing data, they are mostly known as a field observation tool and a means to interview people directly involved in the programme, such as the officials and stakeholders. How is a case study carried out?What are the conditions for the use of this tool?To ensure that a case study is credible and yields satisfactory results the geographic evaluations specific context needs to:
How is the instance selection undertaken?This selection is crucial because an incorrect basis for selecting an instance can lead to a flawed evaluation outcome and can jeopardise its generalisation. The United States General Accounting Office suggests 3 possible keys for instance selection:
Example of the country's selection criteria for the European assistance evaluation in the water sector: How is the case study organised and planned?A modus operandi defining how to carry out one or more case studies is always useful.
Among other advantages the carrying out a pilot case study:
How are the data collected?
How are the results analysed and interpreted? The analysis overlaps with the data collection stage, and this is particularly true for case studies in which:
What are the preconditions for its use?
What are the advantages and limitations of the tool?
|
Detailed presentation
This section is structured as follows:
- What is a case study?
- What are the various types of case study?
- Why and when?
- In country/Regional evaluation
- How is a case study carried out?
- Examples
- Bibliography
WHAT IS A CASE STUDY? |
. What is meant by a "case study"?The manager familiarises him/herself with the strategy, and particularly with the sectors and cross-cutting issues concerned. He/she identifies the key actors and composes the reference group in which the delegation is involved. The head of unit sends a note to the services concerned (example). The case study is probably one of the most diversified evaluation tools, whose goals and content can vary greatly. So much so, that it is sometimes difficult for experts to be sure whether or not they are dealing with case studies. To illustrate how case studies can benefit evaluations, two complementary definitions from the various published works can be cited: According to the United States General Accounting Office (GAO 1970), one of the main evaluation institutions which has used and rationalised case studies in evaluation tasks, “a case study is a method for learning about a complex instance, based on a comprehensive understanding of that instance obtained by extensive description and analysis of that instance taken as a whole and in its context”. According to R. Yin (YIN 2003), whose interest focuses on applied research in social sciences, “a case study is an empirical inquiry that investigates a contemporary phenomenon within its real-life context, especially when the boundaries between phenomenon and context are not clearly evident”. Thus, the implementation of case studies should start with a review of one or more actual examples, in order to gain an in-depth knowledge of the topic and, if possible, to obtain lessons for the entire evaluation programme. The case study may include the examination of documents, of statistical data, or of programme implementation data, but, in particular, it includes the direct observation of an intervention and interviews with people directly concerned in its implementation and effects. This necessitates fieldwork by the evaluators, and this aspect is one of the strengths of the case study. With its wide variety of use, the case study is a tool that can be used concurrently with almost all the stages in the evaluation process: data collection, observation, analysis and even in judgement. |
Where does the tool come from? |
This tool comes from the research field and has been incorporated subsequently into evaluation practice. Its use has expanded since the 1950s, and among research methods in social sciences (survey, experimentation, archives analysis…), the case study has established itself as a tool that can be used in specific situations where other tools have proved to be less effective. Nonetheless, its general use has not developed easily, for the use of tools producing quantitative data (survey, mathematic tools, etc…) often compete with case studies. In this context, its advocates and users have established rigorous methods that allow them to obtain reliable results. Fields in which this tool has been successfully used include social and political sciences, psychology, medicine, management, international relations, etc… In the evaluation field, the use of the case study has covered a wide range of subjects. The first case studies undertaken were similar to research tools, and often carried out by a sole evaluator and essentially descriptive. Their main objective was to illustrate the implementation of a programme and its results through actual examples. They were often based on empirical methods. Since then, the methodology has been progressively formalised and the approach more structured. However, the case study still remains a tool that allows the evaluator a degree of "freedom". The evaluator can place more emphasis on one aspect of the study, even though it was not designed that way (for example, the evaluator can decide to investigate the reasons why two groups of people have opposite views on a subject). This progressive building of methodology and experience has led, inter alia, to the identification of various types of case study. . |
WHAT ARE THE VARIOUS TYPES OF CASE STUDY? |
. There are six main types of case study. It is possible to use them in combination, if the context allows it. Actual implementation varies a little from type to type, however, it follows common conventions. These conventions are described in the section "How is a case study carried out?" The evaluator can choose to implement more than one study within a single evaluation. These are called multiple case studies, whereas when case studies are conducted in several sites, they are called multiple sites case studies. |
The illustrative (or descriptive) case study |
The illustrative case study primarily describes one or a limited number of instances. It can be depicted as a tool which allows the evaluator to start from a general perspective and then highlight a specific element. Because it is a very descriptive tool, it can go deeper into an evaluation task and record field elements, increasing the credibility of the argumentation. This field approach, based on the observation of actual events enables the reader of an evaluation report to fully understand the context of implementation, and any gap between what should have happened and what actually happened. Its principal role is to highlight this gap and explain why it has appeared. Based on an analysis of available data and a series of interviews, its goal is to collect:
From this information, the illustrative case study helps with the understanding of events, and facilitates and supports the analysis made in the evaluation report. As with other case studies, the illustrative case study focuses on the "How" and "Why" questions. However, the limited number of case studies conducted may not be sufficient to draw generalisations from what has been illustrated. |
The exploring case study |
This type of case study will probably be the most commonly used in country/region evaluation in the future. The exploratory case study mainly contributes to clarifying a situation in which information is scarce, and is undertaken before starting the field stage of an evaluation. This situation is often encountered in country/region evaluation. The exploratory case study helps the evaluators assess the facts relating to programme implementation, the local context and, if possible, the findings, or the difficulties in identifying and measuring them. By giving this key information, the exploratory case study provides the basis for the establishment of the evaluation questions and the investigation methods that will have to be used. It should offer practical advice as to what questions should be asked, how and to whom. This type of exploratory study can include one or several studies, depending on the variety of contexts to be taken into account. As exploratory case studies are not designed to provide conclusions, but only to prepare the ground for them, the level of investigation is less rigorous than in a descriptive case study, for example, and the chain of evidence is less well developed. The reporting of such studies can take many forms but their execution needs less work than in other case study types, because their purpose is principally to ease the work that follows. Thus, they contribute more to the evaluation methodology than the accumulation of evidence. This aspect is a potential pitfall for this type of study: it can lead to premature findings released as conclusions, in spite of a lack of depth in the study. Evaluators must proceed carefully and thoroughly evaluate their subsequent case studies once they are completed, in order to avoid confirming initial findings instead of testing them. |
The critical instance case study |
The critical instance case study examines a limited number of sites for one or two specific purposes (i.e. to investigate problematic projects). It analyses in detail cause-and-effect questions relating to the issues of concern, and confirms or invalidates hypotheses that have been formulated before the beginning of the case study. This aspect makes the critical instance case study methodology inappropriate for a country/region evaluation, because it is too focused on situations of extreme specificity: it is only mentioned here for completeness. Some critical instance case studies may reveal broader issues. In this case, they can be used in country/region evaluation, but illustrative case studies should be used if the objective is to include in the evaluation a fair and balanced picture of the situation. |
The programme implementation case study |
Along with programme effects case studies, programme implementation case studies are probably be the most appropriate tool in country/region evaluation. Their goal is to examine whether programme implementation:
Programme implementation case studies are designed to answer evaluation questions relating to implementation (and particularly to effectiveness and efficiency). They should be designed to make generalisations from their findings possible. In country/region assistance evaluation, multiple sites studies are therefore almost a necessity. The instance selection is consequently very important because it will be the basis for future investigations. |
The programme effects case study |
This is the most appropriate case study for country/region evaluation because it analyses the effects of programmes and strategies. It is designed to study the observed outcomes (positive or negative, expected or not) and to check whether or not they are the result of the implementation of programmes or strategies. The programme effects case study often includes an initial stage of implementation analysis, which corresponds to the coverage of programme implementation case studies, and is essential for a full understanding of the context. Thereafter, it focuses on the results of implementation and attempts to demonstrate how and why the programme's outcomes are positively linked to the observed changes. As with programme implementation case studies, programme effects case studies also frequently aim at obtaining information that can easily be generalised. Therefore, multiple sites studies are almost inevitable. In order to give more weight to the studies' findings, it is useful to associate them with beneficiary surveys, which can give an interesting perspective on the evaluation, even though they do not always have statistical validity. These studies concentrate on observed effects by relating differences between sites to explain links and causes between the observed changes and the evaluated programmes. |
The cumulative case study |
This type of case study brings together the findings from case studies or more general studies undertaken at different times. It aggregates information from one or more sites collected over extended period of time. Clearly, such studies are of a great interest and relevance but there are substantial difficulties in finding sufficient historical material to constitute the starting point for cumulative case studies. The opportunity to implement these case studies during country/region evaluations is very limited, and the generalisation of the findings may not be a straightforward task. Thus, they are only mentioned here for the record. . |
WHY AND WHEN? |
||||||||||||||||||||||||||||||
. In which situation is this tool appropriate?The case study is an outstanding tool for addressing complex situations, mostly:
Case studies can be used in all types of evaluation (ex-ante, intermediary and ex-post). However, even if they can be designed for ex-ante evaluations, in practice implementation is difficult because ex-ante evaluations have to be conducted rapidly. Usually, only exploratory case study can be adapted to ex-ante evaluations. Multiple case studies provide the evaluators with the opportunity to observe and analyse situations that are of special interest for the evaluation (for example, aspects that are the most problematic, success stories, etc…). Therefore, this tool can be applied using cases as close to the evaluation questions as required. |
||||||||||||||||||||||||||||||
What are its advantages and its limitations? |
||||||||||||||||||||||||||||||
The advantages of this tool |
||||||||||||||||||||||||||||||
Paradoxically, some case study advantages are cited as drawbacks by opponents to this tool, for example:
The limitations of this toolAlthough this tool has many advantages, it also has limitations beyond which it should not be used. Apart from the individual limits inherent in each type of case study (Illustrative, Exploratory, Programme implementation case study, etc…), a number of common limitations exist, such as:
|
||||||||||||||||||||||||||||||
Can the case study be combined with other tools? |
||||||||||||||||||||||||||||||
The case study has very flexible implementation requirements and evaluators can combine it with other tools and methods. Among them are tools which are fully incorporated within case studies, such as:
Other tools can only be applied to specific studies, such as:
Some tools can be added to the study, such as:
The case study could overlap with existing investigations if the contexts were similar, but as a general rule, the case study has a specific purpose and would not be made redundant by other interviews or focus group activities carried out during the evaluation. |
||||||||||||||||||||||||||||||
What are the pre-conditions for the use of the case study in evaluation ?The time spanThe preparation stage of case studies can take a relatively long time, especially during the identification of sites, finalisation of the content and determining the logistics. Typically, an elapsed period of 2 to 3 months should be planned for a multiple sites study, and 15 to 30 working days should be assigned to its preparation.
As one expert cannot implement a large number of case studies during an evaluation (due to a lack of time), several evaluators should work concurrently on-site. In this case, a period of training of 1 to 2 days should be planned. For multiple sites case studies, the benefits of a pilot case study, implemented before the other case studies should be strongly considered. Its time span would normally be a little longer than average. The analysis and conclusions from multiple case studies can take 10 to 30 days, depending on the complexity of the cases and on their number. Human resourcesCase studies must be carried out by qualified people, who:
Financial resourcesThe full cost of case studies is very variable and depends, among other things, on:
A minimum budget of €15,000 should be fixed and allocated to the case studies preparation stage. |
||||||||||||||||||||||||||||||
When should a case study be used in project evaluations ?Three types of case studies can be useful in project evaluations:
. |
||||||||||||||||||||||||||||||
IN COUNTRY / REGION EVALUATION |
||||||||||||||||||||||||||||||
. Why are case studies used in country/region evaluation?The role of case studies in country/region evaluationCase studies have a role to play in country/region evaluation at 4 levels:
The specific use conditions in country/region evaluationTo make the utilisation of the results of case studies as credible and useful as possible, and to derive generalised recommendations that work for all case studies' applications, the evaluators should:
In addition, in country/region evaluation, the main problem consists of determining what kind of detailed information to seek and who should be interviewed, in a relatively short time and on loosely defined questions. Case studies have similar problems. |
||||||||||||||||||||||||||||||
What kind of use for each type of case study in coutnry/region evaluation?6 Types of Case Study and Their Possible Use in Country/Region Evaluation
. |
||||||||||||||||||||||||||||||
HOW IS A CASE STUDY CARRIED OUT? |
||||||||||||||||||||||||||||||
. What are the stages to follow when carrying out a case study?STAGE 1: THE CASE STUDY PREPARATIONHow is the methodology designed?The design of the methodology includes the basis for analysis and conclusions, the importance of the study in the overall organisation, the general design and the sources of information. During the design of the methodology, 5 successive steps should be taken before deciding how many case studies and which type to implement. First step![]() Second stepDetermine the case study role within the evaluation process (designing the methodology or giving data on the implementation process, etc.). Third stepDetermine the guidelines for the study's general approach. Stipulate the primary hypotheses (where they exist), the selected analysis units (for example, regions), the kind of information to be collected, etc. These guidelines must determine a precise process, the more so when multiple sites studies are planned. The final stage of this general approach and the case study process will be achieved in the pre-operational stage, once adequate data has been collected. Fourth stepPropose the sources of information that have been identified before the implementation of the case study. These will be used to collect each type of information. The documentation:Documentation can be very varied and may include, for example:
This documentation can be collected on-site but may also be obtainable before the departure of the evaluators from the administration or the operators working in Europe and/or in the country to be evaluated. Some documentation can also be acquired from the Internet (for example, other tender evaluations). The interviews:This is the most important source of information in case studies and should be systematically planned. In country/region evaluation, however, one of the most difficult tasks is to decide on whom to focus. If some targets are obvious, others are less so. Thus, the list of the target group will need to be reviewed on-site and after preliminary interviews. This should also be done to avoid being restricted to interviewing officials nominated by the local authorities. Focus groups can also be conducted, especially with the beneficiaries. Direct observation:Direct observation is an absolute necessity in case studies because the evaluator has to have first-hand knowledge of the events he is studying. Most of the time, the observation includes:
The other sources of information:Many other sources of information are at the disposal of the evaluator in case studies. Among them, the most common is the carrying out of concurrent surveys (in particular, beneficiary surveys) in order to add quantitative data to the qualitative data. Thus, a larger number of surveys within case studies gives the evaluators a better quantitative understanding of the observed events. |
||||||||||||||||||||||||||||||
Instance selectionThis selection is a crucial stage because a wrong basis for selecting an instance can lead to a flawed evaluation outcome and can jeopardise its generalisation.
The instance selection varies with the combination of the 3 keys and also with the number of case studies. ConvenienceUsing this key, the selection is justified because the data collection is straightforward (for example, a site near the capital where most of the fieldwork is implemented; a site of a limited size that will only need a short mission, etc…). Purposive samplesThis second key is the most used, and is sub-divided into several options, depending on the evaluation objectives:
ProbabilityUnder this third key, instances are selected from a list with the help of traditional criteria for purposive sampling, with an equal chance of being included. This selection method is not widely used, mostly because the number of cases is insufficient to lead to a representative sample. It can also be argued that case studies focus by definition on very specific problems. Combination of the 3 keysSome combinations of different instance selections are possible. For example, representative elements can be coupled with a bracketing purpose, or specific cases. Number of case studiesThere is no minimum number of case studies; however, the more included, the more detailed the future analysis will be, along with the increased likelihood of making useful generalisations. For most of the time, case studies will not normally permit statistical analyses, therefore unduly multiplying the use of case studies should be avoided. Five to 10 studies for an evaluation should be enough to make generalisations. This figure can rise to 15 or so if a series of distinctive problems and situations need to be studied in various case studies. This is however seldom encountered. |
||||||||||||||||||||||||||||||
STAGE 2: THE PRE-OPERATIONAL STAGEOnce the decision to carry out a case study has been taken and guidelines for its contents defined, the pre-operational stage can begin. This stage is designed to facilitate the operational stage. Several activities have to be conducted before starting the case studies: Finalise the processA modus operandi defining how to carry out one or more case studies is always useful. It is part of the design of the evaluation methodology but goes beyond it by additionally including operational aspects. While the modus operandi is brief for a case study conducted by only one evaluator, it has to be very detailed where multi-site case studies are concurrently carried out by different evaluators. The modus operandi generally includes:
|
||||||||||||||||||||||||||||||
Choosing the teams to implement the case studiesConcurrent with the finalisation of the organisation stage (if it has not already been done in the consultant's answer to the call for tenders), a list of the evaluators in charge of the case studies and their supporting staff (for example, the investigators) should be agreed with the client.
Once this list has been established, the consultants should ask their client for a formal notification of mission, describing their intervention objectives and asking the people they are going to meet to assist them as much as they can. Scheduling the case studiesScheduling a case study is a very detailed operation that needs to be carried out well in advance in order to get the required appointments on-site. In country/region assistance, the use of a local representative is almost obligatory in order to:
give assistance with translations or with the conduct of some parts of the study (for example, surveys) where required If possible conduct a pilot studyThere are many reasons to justify the conducting of a pilot case study before starting a multiple sites case study. Indeed, this good practice allows, inter alia, the:
The pilot site may be chosen because of its practicability (accessibility, sufficiency of data or interlocutors' availability). However, as the evaluator who will supervise all the case studies often carries out the pilot case study, it will usually be selected on the basis of its potential contribution to subsequent case studies. These criteria could be:
Train the evaluation team when multiple sites case studies are scheduledArranging for one evaluator to implement all the case studies can be very difficult or impracticable, in particular because of delays in completing the studies. As a consequence, case studies are usually conducted by several evaluators, which inevitably risks an "evaluator effect" in the conduct of the case study and in the interpretation of the findings. This means that the evaluator's subjectivity should be controlled as much as possible. This training should be linked with permanent monitoring, with each evaluator being able to contact the person in charge of the case studies quickly to get information and advice concerning their conduct or reporting. If the studies have to be written in several languages (for example, where regional evaluations include case studies from many countries), this information should also cover the meaning of particular concepts which could be ambiguous or unclear in translation. |
||||||||||||||||||||||||||||||
STAGE 3: HOW ARE THE DATA COLLECTED?This stage, which can start before the fieldwork with the bibliographical data collection or preliminary meetings, is crucial, especially when the evaluator goes on-site. To ensure that the case study findings are reliable, a number of fundamental elements should be carefully taken into account:
The language in which the case study is to be implemented is also an important point. Thus, the evaluator can be locally recruited or assisted by an interpreter for the whole study or only a part of it (for example, during exchanges with the beneficiaries). |
||||||||||||||||||||||||||||||
STAGE 4: HOW ARE THE FINDINGS ANALYSED AND INTERPRETED?This is the most challenging stage of the case study. Its goal is to analyse the data that have been collected during the fieldwork and to link as far as possible the effects of the observed facts to their causes. This analysis is difficult to conduct because it is less structured than at the conception and the collection stages. The two important analysis patternsTwo main analysis patterns are linked to the commencement of the case study itself:
Process and link to other stagesThe analysis overlaps the data collection stage, and this is particularly true for case studies in which:
If the findings are to be investigated by other tools after the fieldwork stage, with a view to assembling and comparing data (for example, survey analysis, benchmarking, etc.), the analysis must always be based on the construction of a chain of evidence.
The particular case of the multiple sites case studiesThe analysis methods for multiple sites case studies use different types of tools which make the management of various qualitative data easier. Among these techniques are:
This list of tools is not exhaustive. Each evaluator must be free to show the findings from several of his case studies by using any tools providing a chain of evidence, the ultimate goal. With the benefit of studies carried out at various sites, the chains of evidence are reinforced when similar effects happen and are identified in different sites at the same time. |
||||||||||||||||||||||||||||||
STAGE 5: THE REPORTINGThis stage is crucial because the reader of the report must get the same insight from the evaluation as the evaluator. Thus, the importance of this section of the case study should not be underestimated. For the presentation, the report must include:
Ideally, the in-mission report should be distinguished from the final report in order to avoid any bias from empathy. Thus, the author of the final report should not normally be responsible for the production of the in-mission report. However, this can be difficult to achieve in country/region evaluation, where time and budget constraints are significant. The editing style of case study reports largely depends on the role that the evaluators want them to play. This role should be defined in advance, for example, in a context of:
The target group (to whom the case study findings are directed) must also be identified in advance. In country/region evaluation, the target group will mostly be the evaluation users. Report production will therefore be essentially focused on the answers to the evaluation questions, and on the conclusions and recommendations directly linked to programme management. If the case study is addressed to a general audience, reporting should reflect this by taking a less technical approach to the content. Giving a non-technical summary at the beginning of the report can also be helpful to non-specialists. |
What are the pitfalls to avoid?Point 1: inapropriate design of the studyWithin the general design of the study, the evaluator must always be able to answer to the question "How?" and, wholly or partially, "Why?". Its validity must be tested once the design stage begins and must include:
If one of these criteria is poorly conceived, doubts will be raised about the validity of the whole case study. Thus, checks must be made as to whether:
It is important to bear in mind that in such a difficult field as in country/region evaluation:
|
Point 2: poorly selected sitesThe sheet "preparation for case studies" gives details of the various choices for site selection assisting evaluators to implement case studies.
|
Point 3: insufficient information collection and weak argumentationCase studies are mostly used because they offer the opportunity for a detailed examination of a situation. That is their main justification. To be decisive, the information collection must be focused. As a consequence, there must be a systematic search for exhaustiveness and quality when planning and carrying out a case study. The collection and the analysis stages detail the points that have to be met in order to reach this objective. If the evaluation design has been well conducted, there is less chance of insufficient information collection and weak argumentation; however, some precautions may improve the case studies content, such as:
Starting the process of production of the report, by drafting appropriate sections during the collection stage, is an excellent way to identify any lack of information. This is because the evaluator's judgement is of the utmost importance. In any case, an assertion without any evidence is a pitfall to be avoided at all costs in a case study evaluation. The analysis and production stages must always reflect impartiality, which is the only guarantee of the findings' credibility. This is because the only element eventually taken into account is the evaluator's judgement of the evidence. This judgement can be tested, particularly by submitting the case study report to the people from whom data were collected to obtain their views. This leads to:
|
Point 4: Excessive generalisationGeneralisation from findings is one of the most valuable achievements of an evaluation. But, it can also be the case studies' weakest point if they are badly conducted, especially when case studies are contrasted with tools based on statistical analysis. Thus, being able to generalise from case studies implies bringing enough evidence to be convincing. Excessive generalisation must be avoided. This is also often linked to the following points:
Usually, generalisation is possible when case studies have an external validity, for example, as demonstrate by a survey (if the interviewees are representative). This external validity is easier to deduce from several case studies rather than from one. Questions on the external validity may also concern only specific points of the study, and not the whole case study (for example, conclusions about implementation problems can be generalised but not those about the programme impacts). The pitfall to avoid in terms of generalisation is to "imitate" surveys when implementing case studies. In a case study, generalisation is derived from analysis, not statistics, as with experimentation. Generalisation can only be used if the case studies' selection, and the way they are carried out, supports it. In particular, multiple sites case studies can produce useful, comparative analyses, from which common conclusions are derived. In order to know if a generalisation is valid, it is possible to use the replication technique (as in the scientific field), which tests whether replication is possible on a second site, or a second series of sites, and if a general theory is emerging which can be universally applied. |
EXAMPLES |
BIBLIOGRAPHY |
|
Check-list
CHECK-LIST FOR EVALUATORS |
||||||||||||||||||||||||||||||||||||||||||
.
. |
||||||||||||||||||||||||||||||||||||||||||
CHECK-LIST FOR MANAGERS |
||||||||||||||||||||||||||||||||||||||||||
.
|
Context indicators
SUMMARY |
|||||||||||||||||||||
. DefinitionA context indicator is a datum which provides simple and reliable information describing a variable relative to the context. It gives information about a situation and its evolution in a country, or an area relevant to the assistance policy. Other types of indicators:
Indicators are designed by national statistical services and in the context of specific programmes. National statistical services, many donors and international organisations have co-operated to establish standard indicators, to ease comparisons over time and between geographical areas.
These indicators are often designed to highlight the specificities of a local context without, however, enabling the evaluator to make a comparison between countries (for example, information by comparison of the level of the indicator in another country) or get a global and normative view of the country's situation. What use can be made of context indicators?To present the countryContext indicators are usually displayed in the introductory chapter of the evaluation. They deal with:
In certain cases and countries, the emphasis may be put on particular sectors or issues (for example, poverty analysis, conflict analysis). To portray the country's level of assistanceContext indicators also describe the nature of the assistance provided to the country. They should indicate the type of assistance, the amounts disbursed, the sector-based allocation of assistance and the European Commission's activity, compared with that of other donors. To answer evaluation questionsContext indicators can also be used to answer evaluation questions which need a preliminary presentation. They facilitate the understanding of the country's situation for the readers. The following table shows a selection of indicators which are internationally comparable. Tanzania's situation is compared with the situation of a group of 7 African countries considered to be similar: Burkina Faso, Gambia, Ghana, Madagascar, Mali, Nigeria and Zambia.
Source: World Bank * Progress is assessed from the most positive (++) to the most negative (--) evolution with respect to the indicator's evolution over the last 3 to 10 years How are context indicators selected, collected and used during an evaluation?A good indicator should:
European Commission indicatorsDG Dev indicators for the appraisal of country assistance performanceThe indicators used for the appraisal of country assistance performance by the European Commission's DG Dev have been constructed with the assistance of various donors, including the European Union Member States, the World Bank, the UNDP and the OECD/DAC). Their construction is mostly based on the following typology, and particularly on impact indicators.
The construction and implementation of such indicators target two goals and require the monitoring of two distinct series of indicators, in order to:
EUROSTAT dataEUROSTAT provides Member States with numerous data on several topics which are partly expressed by indicators and ranked into short-term (balance of payments, consumer prices, etc.), long-term (economy and ecology, the business structures, etc.) and structural factors (employment, general economic background, environment, etc.). Moreover, EUROSTAT holds data on trade flows between European Union Members and the rest of the world. United Nations indicatorsIndicators related to the Millennium Development GoalsFor the Millennium Goals, the levels to be targeted and 48 Indicators Development Goals have been identified. They are available on the websites of the United Nations and the World Bank. They deal with more than 200 countries, and their methodologies and their precise definitions are presented to ease their understanding and use. These indicators focus more on outcomes than inputs.
Common Country Assessment indicatorsIn 1997, the United Nations decided to create a system of Common Country Assessment (CCA). In this system, the CCA is used as a tool for analysing the country's development situation and identifying the main development challenges. Development indexThe United Nations database is one of the most developed in the world. Approximately 200 indicators have been developed. This database is less developing country-oriented than other databases; yet it presents indicators and development indices, which are designed by the United Nations. Human Development ReportEach year, the UNDP publishes its Human Development Report, which includes a large proportion of the United Nations indicators. World Bank indicatorsInternational Development Association indicatorsThe International Development Association has recently presented in "IDA Results Measurements System: Progress and Proposal, April 2003" a series of indicators aiming at improving the monitoring of the countries' development outcomes, and particularly for countries benefiting from a Poverty Reduction Strategic Programme. IDA indicators cover the following fields: income and poverty, malnutrition, maternal and child health, HIV, gender, primary education, drinking water, infrastructure, private sector development, public sector management, and economic growth. International Comparaison ProgramFounded in 1968, the International Comparison Program is a statistical system used to produce data by country. These data facilitate international comparisons based on prices, expenses value and purchasing power parities. Because of the information about the purchasing power, this statistical system provides the evaluators with comparable data that are valuable for economic and social topics. Purchasing Power Parities (PPP) are monetary conversion rates which express the purchasing powers of various currencies in a common unit. In other words, PPPs can determine how much national currency is needed to buy the same amount of goods and services in any countries. In that sense, PPPs are monetary conversion rates which erase price differences between countries. OECD indicatorsIn its statistic portal, OECD offers a range of precise and updated information about its member countries, from which indicators can be constructed. This descriptive information covers sector-based, social and economic fields. Precise data (geographical, such as country-, regional-, or sector-based) dealing with non-member economies and their development are available on the OECD website. They describe the context in which assistance to a country is carried out. Other sourcesTransparency International indicatorsTransparency International seeks to provide reliable quantitative diagnostic tools regarding levels of transparency and corruption, both at global and local levels. The best known of Transparency International tools is the annual Corruption Perceptions Index. It ranks more than 150 countries in terms of perceived levels of corruption, as determined by expert assessments and opinion surveys. The Global Corruption Barometer (GCB) and the Bribe Payers' Index (BPI) complete the CPI. NGOs such as Human Rights Watch and Amnesty International publish reports on a large number of countries dealing with human rights and other important issues. What are the preconditions for its use?
What are the advantages and limitations of the tool?
|