What are Impact Evaluations?

Impact Evaluations

Impact Evaluations are used in the international humantiarian aid and development sector to determine and attribute the impact of an intervention to a community or individuals.

Impact Evaluations are scientifically robust evaluations using complex sampling and statistical tools. They are the most robus method for determining impact of a project or program.

Types of Evaluations

When evaluating a project or program there are a number of options to choose from:

⇒ Process Evaluation

⇒ Participatory Evaluation

⇒ Impact Evaluation

⇒ Performance Evaluation

⇒ Implementing Partner Evaluation

The above design approaches can be used Formatively (at the design and development stage of the program for example with a baseline assessment), and/or Summatively (at the conclusion of the program where an end-line evaluation is conducted.

Because each of the above design approaches achieve different types of verification, they can also be conducted together. For example, one monitoring, evaluation and learning (MEAL) plan can contain a Process Evaluation, a Performance Evaluation and an Impact Evaluation.

Impact Evaluation Benefits

Impact Evaluations provide two major benefits.

The first is that they provide robust statistical confirmation (or not) that the program has achieved the intended impact.

This provides confidence that the work that you are doing is having real impact for the communities you are serving.

The second major benefit is that this also allows implementers and donors to scale successful programs with confidence.

 

 Process Evaluations

Process Evaluations are used to determine if the program is being implementing as agreed. For example, in a health project to check that the health commodities are being distributed in the manner, to the groups, according to the timeline and in the condition agreed.

It verifies that activities happened, not the impact of those activities.

Participatory Evaluation

Participatory Evaluations determine whether the beneficiaries have been involved throughout the program’s design and implementation. Usually Participatory Evaluations are used on capacity building projects to determine if the capabilities of the beneficiaries has been increased.

Performance Evaluations

Performance Evaluations aim to build a strong empiracle case for whether the program has achieved its goals. Performance Evaluations are less robust and less defensible than an Impact Evaluation and will rarely be able to make claims about impact or outcomes.

Performance Evaluations are more common than Impact Evaluations because they are much easier to conduct.

Performance Evaluations almost always use the OECD Development Assistance Committee (DAC) Criteria. Using the DAC Criteria requires the formulation of evaluation questions, which the evaluation methods are then designed around.

Impact Evaluation Key Concepts

Impact

We define impact as the ultimate goal of the program.

Attribution

Impact Evaluations aim to determine attribution.

Attribution is the extent to which the observable change is a result of the project.

Another way to express this is… How much better is a beneficiary who has participated in the program?

The key is to determine the program’s contribution to this. For example, if a change in income is USD100, then how much of that increased income has been brought about by the project?

Counterfactual

One of the main design reasons for why Impact Evaluations are different to Performance Evaluations is because Impact evaluations use Counterfactuals.

The purpose of a Counterfactual is to determine what would have happened if the program had not been implemented (the Counterfactual). This is the core of the Impact Evaluation – it seeks to compare the impact of the program against what would have happened had it not been implemented.

There are two approaches to determining the Counterfactual – either Quasi Experimental Design or an Experimental Design, also known as a Randomised Control Trial (RCT). All Impact Evaluations use one of these.

Sampling

Sampling for Impact Evaluations is a very complex process involving the use of statistics and depending on the type of Impact Evaluation, ca use Precision Based Sample Calculations or Comparative Power-Based Sample Calculations.  

Types of Impact Evaluation

Experimental Design / Randomised Control Trial (RCT)

RCTs are compare to groups (a treatment group and a control group) that are randomly selected from the target population. This can be done using Simple Random Assignment/Selection, Cluster Randomisation, Phased-in Selection or Random Sampling.

These two groups need to be as similar as possible and because of this managing the Spillover Effect is very difficult.  

Because of this, RCTs need to be built into the design of the program from the very beginning. They cannot be implemented retrospectively.

RCTs require a list of the eligible population and then draw on large sample sizes from this eligible population.

ESDO will help partners design RCTs into their programs and advise what is possible from an Effect Size and Confidence Level. This will inform the sample sizes and data collection effort required.

RCTs are the gold standard for evidence of impact. If implemented correctly, at the end of the program ESDO will be able to determine the attribution and impact of the program to an agreed confidence level.

This can then be used as evidence to either scale or improve program design.

Quasi Experimental Design

Quasi Experimental Design (QED) uses a non-random technique for assigning treatment and control groups.

QED PSM analysis of beneficiary migrant households in Europe

The big advantage of QED is that it can be used at the end of a project without having it built in at the beginning. This can be done if:

⇒ Sufficient comparable data is available from the non-treatment group from baseline and end-line quantitative studies.

⇒ The comparison groups are very similar.

⇒ The eligible population is large enough.

QED can be conducted using either Propensity Score Matching (PSM), Difference in Difference (DiD) or Regression Discontinuity. ESDO will advise which method is the most appropriate in your situation.

The output of a QED Impact Evaluation is very similar to an RCT and can still determine if the P value is statistically significant to enable a determination of attributable impact from the program.

We care deeply that all our work is

We reduce our climate impact in every way possible

Inclusion & equality are central to our work

Communities own, drive & validate our work

Where We Work

© European Sustainable Development Organisation, Ireland