Flagship Course • Free Forever

Monitoring, Evaluation & Learning

Building Evidence Systems for Development Impact

A comprehensive course on MEL systems for development professionals. From theory of change to adaptive management, learn to design evidence systems that drive real impact—with rigorous academic foundations and South Asian case studies.

Frameworks & Methods Case Studies South Asia Focus Interactive Lexicon
14 Modules + 3 Interactive Labs
40+ Academic Papers
South Asia Focus
Practitioner-Focused

Why Study MEL?

Development programs collectively spend billions of dollars annually, yet many lack robust evidence of what works. MEL—Monitoring, Evaluation, and Learning—is the discipline of building evidence systems that drive accountability and adaptive improvement.

MEL in development differs from academic research in crucial ways: it must balance rigor with practicality, serve both accountability and learning purposes, and operate within complex institutional environments where power dynamics shape what gets measured and how.

Evidence-Based Practice

Move beyond anecdotes to systematic evidence. Learn to design indicator systems, collect quality data, and analyze results that drive real decisions.

Methodological Range

Master the full spectrum—from RCTs to contribution analysis, from household surveys to Most Significant Change. Know when to use which approach.

South Asian Context

Deep focus on India, Bangladesh, and the region—with case studies from MGNREGA, Pratham, BRAC, and J-PAL South Asia that show MEL in action.

"What gets measured gets managed—but what gets measured is often what's easy to measure, not what matters most." — Carol Weiss, Evaluation Theorist

Module 1: MEL—History, Variations & Foundations

Before diving into MEL techniques, we must understand where this field came from and why it exists in so many variations. The alphabet soup of MEL, MEAL, MLE, DME, and RBM reflects decades of institutional evolution and ongoing debates about what evidence systems should prioritize.

A Brief History of Development MEL

Results-based thinking in development emerged from surprisingly military origins. Understanding this history illuminates many of today's tensions.

The Logframe's Military Origins: The Logical Framework was developed in 1969 by Practical Concepts Inc. for USAID, adapting systems engineering approaches from the U.S. Department of Defense. Its linear, hierarchical structure reflects Cold War-era faith in rational planning—an approach that has proven both durable and deeply contested.

Era Key Developments Dominant Paradigm
1960s-1970s Logframe developed (1969); Program evaluation emerges as field; USAID mandates logframes (1971) Rational planning; systems engineering
1980s Participatory approaches emerge; PRA/RRA methods; critique of top-down planning Participation vs. measurement tension emerges
1990s Theory of Change emerges (Aspen Institute, Weiss); Results-Based Management adopted by UN; OECD-DAC criteria established (1991) Results-orientation; outcome focus
2000s Paris Declaration (2005) emphasizes mutual accountability; RCT revolution begins; J-PAL founded (2003) Evidence-based policy; randomized evaluation
2010s Adaptive management gains prominence; USAID CLA framework; Doing Development Differently manifesto (2014) Complexity thinking; learning focus
2020s Decolonizing evaluation movement; real-time data; AI in MEL; OECD-DAC criteria revised (2019) Equity, localization, adaptive learning
"The logframe has been compared to the curate's egg—good in parts. It can be a useful thinking and planning tool, but its rigid application has done real damage to development effectiveness." — Robert Chambers, IDS Sussex

The Alphabet Soup: MEL, MEAL, MLE, and More

Different organizations use different acronyms, reflecting distinct institutional histories and priorities. Understanding these variations is essential for navigating the field:

M&E / MEL

Monitoring & Evaluation / Monitoring, Evaluation & Learning

The classic formulation. Adding "L" (Learning) in recent years reflects recognition that collecting data without using it is pointless. Common in bilateral donors (USAID, FCDO).

MEAL

Monitoring, Evaluation, Accountability & Learning

Dominant in humanitarian sector (UNHCR, WFP, NGOs). The explicit "A" for Accountability reflects humanitarian principles of accountability to affected populations (AAP).

MLE

Monitoring, Learning & Evaluation

Used by some foundations (e.g., Hewlett Foundation). Sequence emphasizes that learning should be continuous, not just post-evaluation.

DME

Design, Monitoring & Evaluation

Emphasizes that MEL should be built into program design from the start, not added as an afterthought. Common in international NGOs.

RBM

Results-Based Management

The overarching framework within which MEL sits. Focuses on managing toward results rather than just activities. Dominant at UN agencies and multilaterals.

KM / KML

Knowledge Management / Knowledge Management & Learning

Some organizations separate knowledge functions from evaluation. Focuses on capturing, organizing, and sharing institutional knowledge.

Coach Vandana
Vandana
Don't get too hung up on acronyms—they often reflect institutional politics more than substantive differences. What matters is whether your evidence system serves both accountability and learning, whoever your stakeholders are.

Humanitarian vs. Development MEL

One of the most important distinctions in the field is between humanitarian and development contexts. Each has distinct evidence needs, timeframes, and ethical considerations:

Dimension Humanitarian MEL Development MEL
Timeframe Days to months; rapid onset, rapid response Years to decades; long-term systems change
Primary accountability Affected populations first (AAP) Often donors first (upward accountability)
Success metric Lives saved; suffering reduced; dignity preserved Sustainable outcomes; systemic change
Counterfactual Often impossible/unethical to establish Can often design comparison groups
Key frameworks Sphere Standards; CHS; AAP OECD-DAC; logframes; Theory of Change
Common acronym MEAL (with explicit Accountability) MEL / M&E / RBM
Data challenges Population movement; security; rapid context change Attribution; long causal chains; sustainability

South Asia Case: The Nexus Challenge

In contexts like Bangladesh (Rohingya response), Nepal (post-earthquake), and India (disaster-prone regions), organizations increasingly work across the humanitarian-development "nexus." This requires hybrid MEL approaches: rapid humanitarian monitoring that can transition into longer-term development evaluation.

BRAC's response to the Rohingya crisis exemplifies this: initial MEAL focused on immediate service delivery, gradually transitioning to development-oriented indicators around livelihoods and social cohesion.

The Logframe: Enduring Tool, Enduring Critiques

The Logical Framework Approach (LFA) remains the most widely used planning and monitoring tool in development, despite decades of criticism. Understanding both its utility and limitations is essential:

Logframe Strengths

  • Forces clarity about objectives and assumptions
  • Creates common language across stakeholders
  • Enables systematic tracking of progress
  • Facilitates donor reporting requirements
  • Simple, widely understood format

Logframe Critiques

  • Assumes linear causality in complex systems
  • Discourages adaptation (lock-in effect)
  • Often becomes compliance exercise
  • Struggles with emergent outcomes
  • Can reduce local ownership

The Rise of Theory of Change

Theory of Change (ToC) emerged partly as a response to logframe limitations. Pioneered by Carol Weiss and further developed through the Aspen Institute's work on community initiatives in the 1990s, ToC emphasizes understanding how and why change happens, not just what changes.

Logframe vs. Theory of Change:

Logframe asks: What will we do? What will we deliver? How will we know?

Theory of Change asks: Why do we believe this will work? What must be true for success? What could invalidate our assumptions?

In practice, most organizations now use both: ToC for strategic thinking and program design, logframes for operational tracking and reporting.

The Accountability-Learning Tension

Every MEL system faces a fundamental challenge: the same evidence that demonstrates success to donors might also reveal failures that could threaten funding. This creates powerful incentives that can distort what gets measured, how it's measured, and what gets reported.

"Evaluation serves many masters... The same data that an agency uses to improve its programs may be used by others to criticize and cut the agency's budget. The tension between accountability and improvement is fundamental, and there is no easy resolution." — Carol Weiss, Evaluation: Methods for Studying Programs and Policies (1998), p. 29
Dimension Accountability Focus Learning Focus
Primary question Did we achieve targets? How can we improve?
Audience Donors, board, external stakeholders Program staff, partners, beneficiaries
Timing Retrospective, at project end Real-time, ongoing
Incentive Show success, minimize problems Surface problems early, adapt
Risk Gaming, selective reporting Paralysis by analysis

Results-Based Management (RBM)

The dominant paradigm in development MEL is Results-Based Management, which organizes interventions around a "results chain" connecting resources to impact:

The Results Chain: Inputs → Activities → Outputs → Outcomes → Impact

  • Inputs: Resources invested (funding, staff, materials)
  • Activities: What the program does (training, service delivery)
  • Outputs: Direct products (people trained, services delivered)
  • Outcomes: Changes in behavior, knowledge, practice
  • Impact: Long-term changes in well-being

Major Donor MEL Frameworks

Different development agencies have evolved distinct approaches to MEL, reflecting their institutional priorities and cultures:

USAID CLA

Collaborating, Learning, Adapting—emphasizes organizational learning and real-time adaptation. Focuses on building learning cultures within implementing partners.

FCDO (UK)

Theory-based approaches—strong emphasis on theory of change, contribution analysis, and understanding mechanisms of change rather than just outcomes.

World Bank IEG

Rigorous impact evaluation—gold standard experimental designs, quasi-experimental methods, focus on attributable impact and cost-effectiveness.

EU ROM

Results-Oriented Monitoring—structured monitoring visits, standardized assessment criteria, emphasis on relevance, efficiency, effectiveness, sustainability.

Case Study: MGNREGA Social Audits

South Asia Case: India's MGNREGA (Mahatma Gandhi National Rural Employment Guarantee Act) includes a legally mandated social audit system. In Andhra Pradesh, these community-led audits have recovered over ₹350 crore (~$42 million) in misappropriated funds since 2006.

The Andhra Pradesh model shows how MEL can be redesigned for accountability to beneficiaries rather than just upward to donors—with village assemblies publicly reviewing records, workers verifying payments, and independent audit teams facilitating the process.

Coach Varna
Varna
The tension between donor-centric and community-centric MEL is one I've navigated throughout my career. Want to discuss how to build MEL systems that genuinely serve beneficiaries while meeting funder requirements? Let's talk.
Exercise: Mapping the MEL Landscape

Part A: Acronym Analysis

Research three organizations you've worked with or are interested in. For each, identify:

  1. What MEL-related acronym do they use (M&E, MEL, MEAL, DME, etc.)?
  2. What does this choice suggest about their priorities?
  3. How is their MEL function positioned organizationally (separate unit, embedded in programs, etc.)?

Part B: Historical Analysis

Find a logframe from a project in your sector (many are available in evaluation databases like 3ie or USAID DEC). Analyze:

  1. What assumptions are explicitly stated? What assumptions seem implicit?
  2. How well does the linear results chain capture the complexity of the intervention?
  3. What aspects of the intervention might the logframe miss or undervalue?
Sources: INTRAC M&E Universe; USAID CLA Framework; Aiyar & Mehta (2015) "Spectacles & Scorecard: The Politics of MGNREGA"; World Bank IEG; ALNAP (2016) Evaluation of Humanitarian Action Guide

Check Your Understanding

Test your comprehension of the key concepts from this module.

1 The Logical Framework Approach (LFA) was originally developed for: Multiple Choice
2 What does the "A" in MEAL specifically emphasize that MEL does not? Multiple Choice
3 In the results chain, which level captures changes in beneficiary behavior or practices? Multiple Choice
4 A key difference between humanitarian and development MEL is: Multiple Choice
5 Reflection: How does your organization balance the accountability-learning tension? What pressures shape what gets measured and reported? Reflection

Consider: Who are the primary audiences for your MEL data? What incentives exist to show success versus surface problems? How might you create more space for genuine learning?

Module 2: Theory of Change & Logic Models

A Theory of Change (ToC) articulates the causal pathways through which an intervention is expected to produce outcomes. It makes explicit the assumptions underlying program design and provides a framework for evaluation.

"A theory of change approach to evaluation examines the theoretical assumptions underlying the program and tests them against reality. It's not enough to know whether a program works; we need to know how and why it works, and under what conditions." — Carol Weiss, "Nothing as Practical as Good Theory," American Journal of Evaluation (1995), p. 69

Components of a Theory of Change

A robust Theory of Change includes several interconnected elements that together tell the story of how change happens:

The IF-THEN-BECAUSE Framework:

IF we provide [activities/interventions], THEN we will see [outcomes], BECAUSE [causal mechanism/assumption].

Example: IF we train community health workers in newborn care, THEN neonatal mortality will decrease, BECAUSE trained CHWs will identify danger signs early and refer appropriately.

Logframe vs. Theory of Change

The Logical Framework (logframe) and Theory of Change are related but distinct tools. Understanding their differences helps you use each appropriately:

Dimension Logframe Theory of Change
Format 4x4 matrix Narrative + diagram
Causal logic Linear, sequential Complex, branching pathways
Assumptions Listed separately Embedded in pathways
Flexibility Static, hard to revise Living document
Primary use Project management, reporting Strategy, evaluation design
Handles complexity Poorly Better (but still limited)

The "Missing Middle" Problem

Many Theories of Change suffer from a "missing middle"—a gap between activities and outcomes where the mechanisms of change are assumed rather than articulated. This makes it impossible to understand why a program works or doesn't work.

Weak ToC (Missing Middle)

Training → [black box] → Behavior Change

No explanation of HOW training leads to behavior change

Strong ToC (Filled Middle)

Training → Knowledge → Confidence → Practice → Behavior Change

Explicit intermediate outcomes with testable assumptions

Theory Failure vs. Implementation Failure

When a program doesn't achieve expected outcomes, there are two fundamentally different explanations—and the distinction matters enormously for what you do next:

Dimension Theory Failure Implementation Failure
Definition The causal logic was wrong—even perfect implementation wouldn't have worked The theory was sound, but execution fell short
Example Providing textbooks doesn't improve learning if children can't read Teacher training was effective but only 30% of teachers attended
Diagnosis Intermediate outcomes achieved but final outcomes not Activities delivered but intermediate outcomes not achieved
Response Redesign the intervention logic entirely Strengthen delivery, address bottlenecks
MEL implication Need to measure along the full causal chain Need process monitoring and fidelity tracking

Why This Matters: Many programs are abandoned prematurely because evaluators couldn't distinguish theory failure from implementation failure. A well-designed ToC with indicators at each link in the causal chain allows you to diagnose where the breakdown occurred—and respond appropriately.

Alternative Frameworks

Outcome Mapping

Focuses on changes in "boundary partners" (people the program works with directly) rather than ultimate impacts. Useful when direct attribution is impossible.

Outcome Harvesting

Works backward from observed outcomes to understand what contributed. Useful for complex, emergent change processes.

Contribution Analysis

Builds evidence for contribution claims without requiring attribution. Uses "contribution stories" verified against evidence.

Case Study: Pratham's Teaching at the Right Level

South Asia Case: Pratham's Teaching at the Right Level (TaRL) program evolved through explicit Theory of Change iteration. The original theory—"build more schools → more children in school → better learning"—failed when data showed enrollment increasing but learning stagnating.

The revised theory: "Assess children's actual level → Group by learning level (not grade) → Targeted instruction → Foundational skills acquisition → Improved learning outcomes."

This ToC revision, informed by ASER data, led to the TaRL methodology now replicated in 14 countries through the PAL Network, with RCT evidence showing learning gains of 0.6-0.7 standard deviations.

Coach Vandana
Vandana
Theory of Change should be a living hypothesis, not a static document filed away after proposal submission. I've seen teams transform their programs by treating ToC as a learning tool. Let me help you develop one that actually drives decisions.
Data Exercise: Mapping a Theory of Change

Using the provided template, develop a Theory of Change for a hypothetical maternal health program in rural Bihar:

  1. Context analysis: The dataset includes baseline indicators for institutional delivery (38%), ANC coverage (62%), and maternal mortality ratio (221/100,000). What does this data suggest about where the "breaks" in the causal chain might be?
  2. Map the causal pathways: Identify at least 3 distinct pathways through which your intervention might lead to reduced maternal mortality. For each pathway, list the intermediate outcomes.
  3. Surface assumptions: For each causal link, articulate the underlying assumption. Which assumptions are you most uncertain about?
  4. Identify indicators: For each intermediate outcome, propose one indicator that would help you test whether that link in the causal chain is working.
Download ToC Mapping Template (Excel)
Try it in the Lab
Build your Theory of Change interactively with our MLE Framework Lab. Map causal pathways, surface assumptions, and generate a visual ToC diagram you can share with stakeholders.
Open MLE Framework Lab
Sources: Weiss (1995) "Nothing as Practical as Good Theory"; Mayne (2015) "Contribution Analysis"; Pratham/ASER Centre; J-PAL TaRL evaluations

Check Your Understanding

Test your comprehension of Theory of Change concepts.

1 What is the "missing middle" in many Theories of Change? Multiple Choice
2 Which approach is most appropriate when you cannot directly attribute outcomes to your intervention? Multiple Choice

Module 3: Indicator Design & Frameworks

Indicators are the specific, measurable signals we use to track progress. A good indicator translates abstract concepts like "empowerment" or "resilience" into observable, verifiable data points.

"The choice of indicators is a political act. Indicators determine what gets attention, what gets resources, and ultimately what gets done. The seemingly technical question of 'what should we measure?' is actually a normative question about 'what do we value?'" — Robert Chambers, Whose Reality Counts? (1997), p. 58

Five Indicator Frameworks

Different frameworks offer different criteria for what makes a good indicator. The choice depends on your context, values, and methodological approach:

SMART

Specific, Measurable, Achievable, Relevant, Time-bound

Most widely used. Good for standard project management.

CREAM

Clear, Relevant, Economic, Adequate, Monitorable

World Bank framework. Adds cost-effectiveness consideration.

SPICED

Subjective, Participatory, Interpreted, Cross-checked, Empowering, Diverse

For participatory approaches. Values community perspectives.

RACER

Relevant, Accepted, Credible, Easy, Robust

EU framework. Emphasizes stakeholder buy-in.

When to Use Which Framework

Context Recommended Framework Why
Standard bilateral projects SMART Familiar to donors, manageable
World Bank/large institutions CREAM Emphasizes verifiability, cost
Community-driven development SPICED Values local knowledge, empowerment
Complex systems change FABRIC (emerging) Handles uncertainty, emergence

Types of Indicators

Indicators can be classified along several dimensions:

  • By results level: Input, output, outcome, impact indicators
  • By type: Process vs. performance indicators
  • By timing: Leading (predictive) vs. lagging (confirmatory) indicators
  • By directness: Direct vs. proxy indicators

Disaggregation

Disaggregation breaks down aggregate data to reveal who is (or isn't) benefiting. Common disaggregation dimensions include:

  • Sex/Gender: Male, female, other (as culturally appropriate)
  • Age: Age bands relevant to intervention (0-5, 6-14, 15-24, etc.)
  • Geography: Urban/rural, district, block, village
  • Wealth: Quintiles, BPL status, asset indices
  • Disability: Washington Group questions
  • Caste/Ethnicity: SC, ST, OBC, minority status (context-specific)

Common Indicator Pitfalls

Indicator Proliferation

Too many indicators burden data collection and dilute focus. Rule of thumb: 8-12 key indicators per project.

Measuring Easy, Not Important

Outputs are easier to measure than outcomes. Training attendance is easier than behavior change.

Perverse Incentives

Indicators can drive gaming. "Number of referrals" may incentivize inappropriate referrals.

Case Study: ASER's Simple Indicators

South Asia Case: The Annual Status of Education Report (ASER) in India uses remarkably simple indicators: "Can a child read a short story?" and "Can a child do two-digit subtraction?" These binary, foundational measures have proven more powerful for driving policy than complex standardized tests.

By 2023, the ASER methodology had been replicated in 14 countries through the PAL Network, demonstrating that simple, well-chosen indicators can have more impact than sophisticated but opaque measures.

Coach Varna
Varna
The most powerful indicators are often the simplest. I've seen organizations drown in 50-indicator logframes while missing the 3-4 measures that actually matter. Let's work together to identify what really counts in your context.
Data Exercise: Indicator Quality Assessment

The Excel file contains a draft indicator framework for a women's economic empowerment program with 35 indicators. Your task:

  1. Apply CREAM criteria: Score each indicator 1-5 on each CREAM dimension (Clear, Relevant, Economic, Adequate, Monitorable). Which indicators score lowest? Why?
  2. Identify missing levels: Map indicators to the results chain. Are there gaps at any level (input/output/outcome/impact)?
  3. Disaggregation analysis: Which indicators can be meaningfully disaggregated by wealth quintile? By age? What sample size would you need?
  4. Prioritization: If the program could only track 10 indicators, which would you keep? Justify your selection.
Download Indicator Assessment Exercise (Excel)
Try it in the Lab
Design your indicator framework with the MEL Design Lab. Apply SMART/CREAM criteria interactively, map indicators to results levels, and export a ready-to-use indicator matrix.
Open MEL Design Lab
Sources: Kusek & Rist (2004) "Ten Steps to a Results-Based M&E System"; World Bank CREAM framework; ASER Centre methodology; Roche (1999) SPICED indicators

Check Your Understanding

1 Which indicator framework is most appropriate for community-driven development where local perspectives matter most? Multiple Choice

Module 4: Baselines, Targets & Benchmarks

A baseline establishes the starting point for measuring change. Without a valid baseline, you cannot know whether observed conditions at endline represent improvement, decline, or no change.

Why Baselines Matter

Baselines serve four critical functions in MEL systems:

  1. Establish starting conditions before intervention begins
  2. Enable change measurement by comparison with later data
  3. Inform program design by revealing actual (vs. assumed) conditions
  4. Set realistic targets based on evidence rather than aspiration

Target-Setting Approaches

Approach Description When to Use
Historical trend Project forward from past performance When historical data exists
Evidence-based benchmark Use effect sizes from rigorous evaluations When comparable evidence exists
Expert judgment Consult technical specialists When quantitative evidence limited
Stakeholder negotiation Agree targets with funders/partners When buy-in matters more than precision

The Problem with Over-Ambitious Targets: CGD research on World Bank projects found that projects with more ambitious targets were more likely to be rated as failures—not because they achieved less, but because they promised more. Set targets based on evidence of what's achievable, not donor expectations.

Data Exercise: Baseline Analysis & Target Setting

Work with education baseline data from 5 districts in Madhya Pradesh. Analyze gaps between baseline values and national benchmarks, then set realistic 3-year targets using different methodologies (absolute, relative improvement, equity-weighted).

  1. Identify the district with greatest need and the one closest to benchmarks
  2. Calculate baseline-to-benchmark gaps for each indicator
  3. Set Year 1, 2, and 3 targets with clear rationale
  4. Compare trade-offs between uniform vs. differentiated targets
Download Baseline & Target Setting Exercise (Excel)
Sources: INTRAC Baselines guidance; CGD research on World Bank project ratings; ASER annual benchmarking methodology

Module 5: Data Collection Methods

Data collection methods range from large-scale household surveys to participatory methods that put communities in the driver's seat. The choice depends on your questions, resources, and values about whose knowledge counts.

Quantitative Methods

Method Strengths Limitations
Household surveys Representative, comparable Expensive, time-consuming
Administrative data Low cost, routine Quality varies, limited scope
Facility assessments Objective service quality Doesn't capture client experience
Direct observation Unbiased by reporting Hawthorne effect, labor-intensive

Qualitative Methods

Key Informant Interviews (KII)

In-depth conversations with knowledgeable individuals. Good for understanding context, mechanisms, and expert perspectives.

Focus Group Discussions (FGD)

Group conversations exploring shared experiences. Good for understanding social norms and generating hypotheses.

Most Significant Change (MSC)

Participatory story collection and selection. Good for capturing unexpected changes and diverse perspectives.

Digital Data Collection

Mobile data collection tools like KoBoToolbox, ODK, and SurveyCTO have transformed field data collection:

  • Skip logic: Questions appear based on previous answers
  • Validation: Built-in range checks and consistency rules
  • GPS capture: Automatic location tagging
  • Real-time monitoring: Daily data quality dashboards
Coach Varna
Varna
Match your methods to your actual decision needs. A quick qualitative study that informs real decisions is more valuable than a rigorous survey whose results arrive too late to matter.
Data Exercise: Sampling & Power Calculation

You are designing a baseline survey for a nutrition program in Odisha. The workbook contains:

  1. Power calculation: Determine the sample size needed to detect a 10 percentage point change in stunting prevalence (baseline 35%) with 80% power and 95% confidence. What if you need to disaggregate by 5 districts?
  2. Sampling frame: The sampling frame lists 500 villages with population data. Design a probability proportional to size (PPS) strategy to select 60 villages.
  3. Cost estimation: Using the unit cost data, estimate total survey cost. What trade-offs emerge if budget is cut by 30%?
Download Sampling Exercise (Excel)
Try it in the Lab
Plan your data collection strategy with the MEL Design Lab. Define data sources, sampling approaches, and collection methods for each indicator—with built-in guidance on method selection.
Open MEL Design Lab
Sources: INTRAC Data Collection guidance; J-PAL survey methodology; KoBoToolbox documentation

Module 6: Data Quality Assurance

Data quality determines whether your evidence is trustworthy enough to inform decisions. Poor quality data can lead to worse decisions than no data at all.

USAID's Five DQA Dimensions

Validity

Does the data measure what it claims to measure?

Reliability

Would the same measurement process produce consistent results?

Timeliness

Is the data available when needed for decisions?

Precision

Is the data sufficiently accurate for intended use?

Integrity

Is the data protected from manipulation or bias?

The 10% Back-Check Rule

Standard practice: re-contact 10% of respondents (randomly selected) to verify key responses. This deters fabrication and identifies systematic errors. Back-checks should occur within 48 hours of original data collection.

Data Exercise: Data Quality Audit

The Excel file contains raw survey data from a health program baseline. Conduct a quality audit:

  1. Missing data analysis: Calculate the percentage of missing values for each variable. Which variables exceed the 5% threshold?
  2. Outlier detection: Identify statistical outliers in continuous variables (age, income, distance). Are these likely data errors or real extreme values?
  3. Logic checks: Apply the validation rules (e.g., children's ages must be less than mothers' ages). How many records fail?
  4. Enumerator effects: Compare key indicators across enumerators. Do any show suspiciously different patterns?
Download Data Quality Exercise (Excel)
Sources: USAID Data Quality Assessment framework; J-PAL survey protocols; INTRAC data quality guidance

Module 7: Quantitative Analysis

Quantitative analysis transforms raw data into actionable insights. The goal is not statistical sophistication but clear answers to decision-relevant questions.

Descriptive Statistics

Start with the basics: means, medians, percentages, and distributions. These simple measures often provide 80% of the insight.

When to Use Median vs. Mean: For income, wealth, and other skewed distributions, the median (middle value) better represents the "typical" case than the mean (average), which is pulled by outliers.

Comparing Groups

  • Percentage point change: From 40% to 60% = 20 percentage point increase
  • Percent change: From 40% to 60% = 50% increase
  • Difference-in-differences: Change in treatment minus change in comparison
  • Effect size: Standardized difference (Cohen's d, standard deviations)

Common Misunderstandings in MEL Data

Even experienced practitioners fall into these conceptual traps. Understanding these distinctions will save you from costly misinterpretations:

1. Internal Validity vs. External Validity

Concept Internal Validity External Validity
Question Did X cause Y in this specific study? Would X cause Y in other contexts?
Threat Confounding, selection bias, attrition Context-specificity, sample selectivity
RCT strength High (randomization controls confounds) Often low (artificial conditions)
Trade-off Designs optimizing internal validity (lab experiments) often sacrifice external validity, and vice versa

Practitioner Warning: A study with high internal validity showing a program "works" in one context may not replicate elsewhere. Before scaling, ask: What was special about the study context? Were implementers unusually motivated? Was the population selected for likely success?

2. Accuracy vs. Precision

Accurate but Not Precise

Estimates cluster around the true value but with high variance

Example: Survey finds poverty rate "between 20-40%"—correct on average but too imprecise for targeting

Precise but Not Accurate

Estimates are tightly clustered but systematically wrong

Example: Administrative data reports 5.2% stunting when true rate is 35%—consistent but biased

MEL Implication: Small sample sizes give you imprecision (wide confidence intervals). Systematic measurement errors give you inaccuracy (bias). Both can coexist—you can be both imprecise AND inaccurate.

3. Statistical Significance vs. Effect Size vs. Practical Magnitude

The Three Questions You Must Ask:

  • Statistical significance (p-value): Could this result have occurred by chance? (p < 0.05 means <5% chance)
  • Effect size (Cohen's d): How large is the effect in standardized units? (d=0.2 small, 0.5 medium, 0.8 large)
  • Practical magnitude: Does this effect matter for real-world decisions?
Scenario Significant? Effect Size Interpretation
Large sample, tiny effect Yes (p<0.001) d=0.05 Statistically detectable but practically meaningless
Small sample, large effect No (p=0.12) d=0.7 Potentially important but underpowered study
Adequate sample, medium effect Yes (p=0.02) d=0.4 Credible evidence of meaningful impact
"Statistical significance is not the same as practical importance. A treatment effect can be highly significant statistically but trivial in magnitude, or large in magnitude but statistically insignificant due to small sample size." — Gertler et al., Impact Evaluation in Practice (2016), p. 89

4. Why Averages Can Mislead

The mean (average) is heavily influenced by outliers and skewed distributions. Consider this example:

Example: Village Income Data

10 households with incomes: ₹8K, ₹10K, ₹12K, ₹12K, ₹15K, ₹15K, ₹18K, ₹20K, ₹25K, ₹500K

  • Mean: ₹63,500 (pulled up by one wealthy household)
  • Median: ₹15,000 (middle value—better represents "typical" household)
  • Standard Deviation: ₹153,000 (huge spread indicates outliers)

Reporting "average income is ₹63,500" would grossly misrepresent village welfare.

Rule of Thumb: For income, wealth, land holdings, and other variables with right-skewed distributions, always report the median alongside the mean, and consider the standard deviation to understand spread.

5. Cumulative vs. Current Year Numbers

This is one of the most common sources of confusion in program reporting:

Metric Type Definition Example When to Use
Cumulative Total since program start "500,000 farmers trained since 2018" Overall program reach, final reports
Current year This reporting period only "75,000 farmers trained in 2024" Annual performance, trends
Unique beneficiaries Distinct individuals (deduplicated) "320,000 unique farmers reached" True coverage, avoiding double-counting

Watch For: Programs may report cumulative numbers to inflate apparent reach, or fail to deduplicate beneficiaries who receive multiple services. Always ask: "Is this cumulative or current year?" and "Are these unique individuals or service contacts?"

Coach Varna
Varna
Prioritize actionable insight over statistical sophistication. A clear bar chart that shows the problem is more useful than a regression table that no one reads.
Data Exercise: Program Performance Analysis

Analyze baseline-endline data from an agricultural extension program:

  1. Descriptive statistics: Calculate mean, median, and standard deviation for yield at baseline and endline by treatment group.
  2. Difference-in-differences: Compute the DiD estimate of program impact using the formula provided.
  3. Statistical significance: Test whether the difference is statistically significant using the t-test calculator.
  4. Effect size: Calculate Cohen's d. How does this compare to typical education interventions (d=0.3-0.5)?
  5. Summary table: Prepare a results table suitable for a program report.
Download Analysis Exercise (Excel)
Sources: INTRAC Quantitative Analysis; Gertler et al. "Impact Evaluation in Practice"

Module 8: Qualitative Analysis

Qualitative analysis moves from raw text to patterns and meaning. Done well, it reveals the "how" and "why" behind quantitative findings.

Major Approaches

Thematic Analysis

Identify recurring themes and patterns across data. Most common approach in development MEL.

Framework Analysis

Apply a pre-defined framework to organize data. Good when theory guides analysis.

Grounded Theory

Build theory from data through iterative coding. Good for exploratory research.

Narrative Analysis

Analyze stories and their structure. Good for understanding lived experience.

The Coding Process

  1. Familiarization: Read through all data multiple times
  2. Initial coding: Label meaningful segments
  3. Code consolidation: Group similar codes
  4. Theme development: Identify higher-level patterns
  5. Theme review: Check themes against data
  6. Interpretation: Explain what themes mean
Coach Vandana
Vandana
Good qualitative analysis is systematic, not just "picking good quotes." Let me show you how to build credible interpretations that stand up to scrutiny.
Sources: Braun & Clarke (2006) "Thematic Analysis"; INTRAC Qualitative Analysis; Ritchie & Spencer Framework Analysis

Module 9: Evaluation Design & the OECD-DAC Framework

Evaluation design determines what questions you can answer and how credibly. This module covers both the dominant OECD-DAC evaluation criteria and the methodological choices between experimental, quasi-experimental, and theory-based approaches.

The OECD-DAC Evaluation Criteria: A Deep Dive

Established in 1991 and revised in 2019, the OECD Development Assistance Committee (DAC) criteria provide the most widely used framework for structuring development evaluations. The 2019 revision added Coherence as a sixth criterion and refined definitions to emphasize equity and human rights.

Why These Criteria Matter: The DAC criteria shape how billions of dollars in development spending are evaluated. Understanding them isn't just academic—they structure donor expectations, evaluation ToRs, and funding decisions. Mastering these criteria is essential for any MEL professional.

Criterion Core Question Sub-Questions to Consider
1. Relevance Is the intervention doing the right things? Does it respond to beneficiary needs? Is it aligned with partner country priorities? Does it address root causes or symptoms?
2. Coherence How well does it fit? Internal coherence: Is the intervention logic sound? External coherence: Does it complement other actors' work? Does it align with international norms (human rights, SDGs)?
3. Effectiveness Is the intervention achieving its objectives? Are outputs and outcomes being achieved? What factors enable or constrain success? Are there differential effects across groups?
4. Efficiency How well are resources being used? Could the same results be achieved with fewer resources? Are there more cost-effective alternatives? Is implementation timely?
5. Impact What difference does the intervention make? What are higher-level effects (intended and unintended)? What would have happened without the intervention? Are there transformative effects on systems or norms?
6. Sustainability Will the benefits last? Are results likely to continue after funding ends? Are there financial, institutional, technical, and environmental risks to sustainability?

Operationalizing Each Criterion

Relevance

Methods: Needs assessments; stakeholder consultations; context analysis; policy alignment review.

Watch for: Interventions designed in headquarters without local input; changing contexts that make original design obsolete.

Coherence

Methods: Stakeholder mapping; portfolio analysis; policy coherence review; SDG alignment assessment.

Watch for: Duplication with other programs; contradictions with other donor initiatives; misalignment with national policies.

Effectiveness

Methods: Outcome monitoring; contribution analysis; process tracing; beneficiary feedback.

Watch for: Outputs achieved but not translating to outcomes; success in pilots but failure to scale; elite capture.

Efficiency

Methods: Cost-effectiveness analysis; value for money assessment; implementation review; benchmark comparison.

Watch for: High overhead ratios; delays in disbursement; procurement inefficiencies; scope creep.

Impact

Methods: Experimental/quasi-experimental designs; contribution analysis; Most Significant Change; systems mapping.

Watch for: Attribution challenges; displacement effects; unintended negative consequences; time lag in impact realization.

Sustainability

Methods: Exit strategy review; institutional capacity assessment; financial sustainability analysis; ownership assessment.

Watch for: Dependency on continued external funding; weak local ownership; capacity gaps in local institutions.

Coach Vandana
Vandana
The DAC criteria are a starting point, not a straitjacket. Good evaluators adapt and prioritize based on the evaluation purpose and stakeholder needs. Don't try to answer every sub-question for every criterion—focus on what matters most for decision-making.

Beyond DAC: Emerging Evaluation Principles

The 2019 DAC revision also emphasized cross-cutting principles that should inform all evaluation work:

  • Equity and inclusion: Who benefits and who is left behind? Are results disaggregated?
  • Gender equality: How does the intervention affect gender relations and women's empowerment?
  • Human rights: Does the intervention respect and advance human rights?
  • Environmental sustainability: What are the environmental implications?
  • Leave no one behind: Are the most marginalized populations reached?

Evaluation Design Choices

Experimental Designs

Randomized Controlled Trials (RCTs) randomly assign units to treatment and control groups. When well-implemented, they provide the most credible evidence of causal impact.

RCT Strengths

  • Eliminates selection bias
  • Clear counterfactual
  • Credible causal inference
  • Statistical power for effect detection

RCT Limitations

  • Ethical constraints on randomization
  • External validity concerns
  • Expensive and time-consuming
  • May miss mechanisms and context
  • Hawthorne and John Henry effects

Quasi-Experimental Designs

Difference-in-Differences

Compare changes over time between treatment and comparison groups. Requires parallel trends assumption—both groups would have followed similar trajectories without intervention.

Regression Discontinuity

Exploit eligibility cutoffs (e.g., poverty line, geographic boundary). Very credible when applicable, but only estimates local average treatment effect near the cutoff.

Propensity Score Matching

Match treatment units with similar non-treatment units based on observable characteristics. Requires rich data and assumption of no unmeasured confounders.

Instrumental Variables

Use a variable that affects treatment but not outcomes directly. Powerful when valid instrument exists, but instruments are often hard to find and defend.

Theory-Based Evaluation

Contribution Analysis

Builds evidence for contribution claims through "contribution stories" verified against evidence. Developed by John Mayne. Particularly useful for complex programs where experimental designs are infeasible.

Realist Evaluation

Asks "what works, for whom, in what circumstances?" Uses Context-Mechanism-Outcome (CMO) configurations. Developed by Pawson & Tilley. Excellent for understanding heterogeneous effects.

Process Tracing

Detailed examination of causal mechanisms linking intervention to outcomes. Uses evidence like "smoking guns" and "hoops tests." Strong for single-case causal inference.

Most Significant Change

Participatory method collecting and selecting stories of change. Reveals unexpected outcomes and values of different stakeholders. Developed by Rick Davies for complex programs.

Choosing an Evaluation Design

Design Selection Framework:

Ask these questions in sequence:

  1. What decisions will this evaluation inform? (Determines rigor needed)
  2. Is random assignment ethically and practically feasible? (If yes, consider RCT)
  3. Is there a natural comparison group or cutoff? (If yes, consider quasi-experimental)
  4. Do we need to understand mechanisms and context? (If yes, add theory-based elements)
  5. What's the budget and timeline? (Constrains options)

Case Study: J-PAL South Asia

South Asia Case: J-PAL South Asia (based at IFMR) has conducted over 200 RCTs in the region since 2007. Their work on Teaching at the Right Level showed learning gains of 0.6-0.7 standard deviations—evidence strong enough to influence education policy across India, Africa, and beyond.

But J-PAL also demonstrates the limits of RCTs: their evaluations require 2-3 years, limiting their usefulness for rapid program adaptation. They've increasingly complemented RCTs with process evaluations and cost-effectiveness analyses.

Coach Varna
Varna
Don't let the perfect be the enemy of the good. A well-designed quasi-experimental study that answers your actual questions is more useful than an RCT that answers questions you didn't need to ask. Match your design to your decision needs.
Exercise: DAC Criteria Application

Part A: Evaluation Matrix Design

Select a program you're familiar with (or use a case from the course). For each DAC criterion:

  1. Write 2-3 evaluation questions specific to this program
  2. Identify data sources you would use to answer each question
  3. Note which criteria are most important for this particular evaluation and why

Part B: Design Selection

For the same program, determine the most appropriate evaluation design by considering:

  1. Is randomization feasible? Why or why not?
  2. What natural comparison opportunities exist?
  3. What theory-based methods would add value?
  4. Given your constraints, what design would you recommend?
Sources: OECD-DAC Evaluation Criteria (2019 revision); Gertler et al. "Impact Evaluation in Practice"; J-PAL Research Resources; Mayne (2001) Contribution Analysis; Pawson & Tilley "Realistic Evaluation"

Check Your Understanding

Test your comprehension of the key concepts from this module.

1 Which DAC criterion was added in the 2019 revision? Multiple Choice
2 Regression Discontinuity Design is most appropriate when: Multiple Choice
3 The DAC criterion "Coherence" assesses: Multiple Choice
4 Reflection: Think of an evaluation you've been involved in or read about. Which DAC criteria received the most attention? Which were underemphasized? What might explain this pattern? Reflection

Consider: How did donor requirements shape the evaluation focus? Were equity considerations adequately addressed? What would you do differently?

Module 10: Learning Systems & Adaptive Management

The purpose of MEL is not to produce reports but to drive learning and adaptation. This module explores how to build organizational systems that actually use evidence to improve programs.

USAID CLA Framework

USAID's Collaborating, Learning, and Adapting framework emphasizes three interconnected components:

Collaborating

Working with stakeholders to share knowledge and coordinate efforts. Internal collaboration (within teams) and external collaboration (with partners, communities).

Learning

Generating and accessing knowledge through technical evidence, monitoring data, and tacit knowledge. Includes learning questions and knowledge management.

Adapting

Using evidence to make decisions and course corrections. Requires enabling conditions: resources, culture, processes that support adaptation.

Adaptive Management Practices

  • Pause-and-reflect sessions: Regular (quarterly) team sessions to review data and discuss implications
  • Decision logs: Document what was decided, why, and based on what evidence
  • Fail-fast mechanisms: Early warning indicators that trigger review
  • Stakeholder feedback loops: Regular input from beneficiaries and partners
  • Flexible funding: Budget mechanisms that allow reallocation based on learning

Case Study: BRAC's Research & Evaluation Division

South Asia Case: BRAC established its Research and Evaluation Division (RED) in 1975—one of the earliest examples of an NGO building in-house research capacity. RED operates with significant independence, enabling honest assessment rather than just validation of programs.

This institutional design has enabled genuine learning: RED's research on the Ultra-Poor Graduation program led to iterative refinements that increased effectiveness. The program now operates in over 50 countries.

Coach Vandana
Vandana
Learning systems require more than good intentions—they need institutional design. Let me help you identify the structural changes needed to make learning happen in your organization.
Sources: USAID CLA Framework; BRAC Research and Evaluation Division; ODI Doing Development Differently; Andrews et al. "Building State Capability"

Module 11: MEL Technology & Data Systems

Technology can dramatically improve MEL efficiency and quality—but only if it serves your actual needs. This module covers the technology landscape from mobile data collection to visualization dashboards.

Mobile Data Collection Tools

KoBoToolbox

Free, open-source. Widely used in humanitarian contexts. Strong offline capabilities. Good for surveys and assessments.

ODK / ODK-X

Open-source foundation for many tools. Highly customizable. Large developer community.

SurveyCTO

Commercial tool built on ODK. Strong data quality features. Good technical support. Popular with research organizations.

CommCare

Designed for frontline workers. Strong case management features. Good for health and social programs.

Data Visualization & Dashboards

Good visualization makes data accessible to decision-makers. Key principles:

  • Start with decisions: What actions might the dashboard inform?
  • Less is more: Focus on 5-7 key metrics, not 50
  • Update frequency: Match refresh rate to decision cycles
  • Mobile-first: Many users will access on phones
Try it in the Lab
Build your data management plan with the MEL Plan Lab. Define data flows, storage systems, quality protocols, and reporting rhythms—with templates you can customize for your context.
Open MEL Plan Lab
Sources: KoBoToolbox; ODK; SurveyCTO; DHIS2; Principles for Digital Development

Module 12: MEL Leadership & Sustainability

MEL systems are only as strong as the people and institutions behind them. This module addresses the human and organizational dimensions of building sustainable MEL capacity.

Building MEL Culture

Technical systems alone don't create learning organizations. Building MEL culture requires:

  • Leadership commitment: Senior leaders must model evidence use
  • Psychological safety: People must feel safe surfacing problems
  • Time and space: Reflection requires protected time
  • Incentives alignment: Reward learning, not just success stories
  • Capacity building: Invest in people, not just systems

Sustainability Considerations

Financial Sustainability

Budget MEL at 5-10% of program costs. Build MEL into core funding, not just project budgets.

Technical Sustainability

Use appropriate technology that can be maintained locally. Avoid dependency on external consultants.

Institutional Sustainability

Embed MEL in organizational structure, not just individual projects. Build internal capacity.

Essential Reading

Sources: USAID CLA Maturity Tool; Senge "Fifth Discipline"; INTRAC MEL Systems

Module 13: Capstone Project

Apply what you've learned by designing a complete MEL system for a real or hypothetical development program. This capstone project integrates all course concepts into a practical deliverable.

Choose Your Project

Select one of three project options:

Option A: MEL Plan Design

Design a complete MEL plan for a program in your context. Include theory of change, indicator framework, data collection plan, and learning agenda.

Option B: Evaluation Design

Design an evaluation for an existing program. Include evaluation questions, methodology, sampling strategy, and analysis plan.

Option C: MEL System Assessment

Assess an existing MEL system and propose improvements. Include gap analysis, recommendations, and implementation roadmap.

Grading Rubric

Criterion Weight Description
Conceptual Understanding 40% Demonstrates mastery of MEL concepts and frameworks
Practical Application 30% Applies concepts appropriately to real-world context
Critical Thinking 20% Shows awareness of trade-offs, limitations, and context
Communication 10% Clear, well-organized, professional presentation

Deliverables

  • Written Report: 10-15 pages covering all required components
  • Visual Framework: Theory of change diagram or MEL system map
  • Indicator Matrix: Complete with disaggregation and data sources
  • Reflection: 500 words on key learnings and future development

Module 14: Building Your Own MEL System

This module brings together everything you've learned into a practical, step-by-step guide for building a complete MEL system from scratch. Whether you're starting a new program or redesigning an existing system, this framework will help you create MEL infrastructure that actually serves your organization's learning and accountability needs.

"A good MEL system is like good plumbing—when it works, no one notices it. When it fails, everyone suffers. The goal is invisible infrastructure that makes evidence-based decision-making the path of least resistance." — adapted from USAID Learning Lab

When to Build vs. Adapt

Before building a new MEL system, ask whether you need to build from scratch or can adapt existing infrastructure:

Build New When...

  • Starting a new program or organization
  • Existing systems fundamentally don't serve your needs
  • Major strategic shift requires new evidence architecture
  • Current system creates more burden than value

Adapt Existing When...

  • Current system captures useful data but underutilizes it
  • Staff are trained on existing tools
  • Historical data provides valuable baselines
  • Incremental improvements can address gaps

The Six-Step MEL System Building Framework

1

Stakeholder Mapping & MEL Purpose Definition

Before designing any indicators or data collection, clarify who needs what evidence for which decisions. This shapes everything else.

Stakeholder Primary Evidence Needs Decision Types Frequency
Donors/Funders Outcome achievement, value for money Continue/modify funding Quarterly/Annual
Program Managers Implementation fidelity, bottlenecks Tactical adjustments Weekly/Monthly
Frontline Staff Individual case progress Service delivery Daily/Weekly
Beneficiaries Service quality, responsiveness Engagement; feedback Ongoing
The Accountability-Learning Balance: Every MEL system must serve both accountability (proving value to funders) and learning (improving programs). Explicitly decide your balance—a 70/30 split toward accountability is common, but learning-focused organizations might aim for 50/50.
2

Theory of Change → Logframe Translation

Transform your strategic theory into operational monitoring architecture. The Theory of Change captures why you believe the program will work; the logframe operationalizes what you'll track.

  • Articulate the Theory of Change: Map causal pathways, surface assumptions, identify intermediate outcomes
  • Identify Critical Assumptions: Which assumptions, if wrong, would invalidate the entire theory?
  • Translate to Results Framework: Structure into Impact → Outcomes → Outputs → Activities
  • Build the Logframe: Add indicators, means of verification, assumptions column
Try it in the Lab
Use the MLE Framework Lab to build your Theory of Change interactively. Map pathways, add assumptions, and export a visual diagram plus logframe template.
Open MLE Framework Lab
3

Indicator Selection & Operationalization

Select indicators that balance rigor, feasibility, and utility. The goal is the minimum set that enables the decisions identified in Step 1.

The Indicator Selection Funnel
  1. Long List — Brainstorm all possible indicators (typically 30-50)
  2. Apply CREAM/SMART — Score each indicator on quality criteria
  3. Map to Decisions — Which decisions does each indicator inform?
  4. Assess Feasibility — Cost, data source availability, staff capacity
  5. Final Selection — 8-15 key performance indicators
Try it in the Lab
Build your indicator framework interactively in the MEL Design Lab. Apply quality criteria, define operationalization details, and export a complete indicator reference sheet.
Open MEL Design Lab
4

Data Architecture & Collection Protocols

Design the infrastructure for data to flow from collection points to decision-makers. This includes tools, systems, roles, and quality assurance protocols.

Data Flow Architecture

Your MEL data system should move through four layers:

  1. Collection Layer: Field surveys (KoBoToolbox), administrative records, participant feedback, partner reports
  2. Storage & Processing Layer: Central database, data cleaning protocols, quality assurance checks
  3. Analysis & Visualization Layer: Dashboards, standard analysis templates, report generation
  4. Decision Layer: Pause-and-reflect sessions, management reviews, stakeholder reporting
Try it in the Lab
Design your data management system with the MEL Plan Lab. Define data flows, storage requirements, quality protocols, and responsible parties—with exportable templates.
Open MEL Plan Lab
5

Reporting Rhythms & Feedback Loops

Design the rhythm of evidence use—how data moves from collection to decision-making. Without intentional design, data gets collected but never used.

Frequency Product Audience Decision Forum
Weekly Activity tracker dashboard Program team Team standup
Monthly Output summary + issues log Management Management meeting
Quarterly Progress report + outcome data Donors, Board Pause-and-reflect session
Annual Annual report + learning synthesis All stakeholders Annual review
6

Costing & Sustainability Planning

MEL systems require resources to function. Build sustainability into the design from the start.

The 5-10% Rule: MEL should typically represent 5-10% of total program budget. Below 5%, systems tend to be under-resourced and produce poor quality data. Above 10%, you may be over-investing relative to program scale.
MEL Budget Components
  • Personnel (50-60%): MEL staff salaries, data collectors, enumerator fees
  • Data collection (20-25%): Survey costs, travel, participant incentives
  • Technology (10-15%): Software licenses, devices, server costs
  • External support (5-15%): Evaluations, technical assistance
  • Learning activities (5-10%): Workshops, dissemination, knowledge products

South Asia Case: BRAC's Evolution to Learning Organization

BRAC, the world's largest development NGO, offers a compelling case of MEL system evolution. Founded in 1972, BRAC established its Research and Evaluation Division (RED) in 1975—remarkably early for an NGO.

Key design features that enabled genuine learning:

  • Structural independence: RED reports to the Executive Director, not program managers—enabling honest assessment
  • Mixed methods commitment: Combines rigorous quantitative evaluation with deep qualitative understanding
  • Iteration culture: Programs are explicitly designed as "prototypes" to be refined based on evidence

The Ultra-Poor Graduation program exemplifies this approach: RED's research identified why initial designs weren't working for the poorest households, leading to iterative refinements. The resulting model has now been replicated in over 50 countries.

Coach Varna
Varna
Building a MEL system is one thing—making it actually work in your organizational context is another. The technical design is often the easy part; navigating the politics, building buy-in, and creating a genuine learning culture requires different skills. Let's discuss your specific context.
Sources: Kusek & Rist (2004) "Ten Steps to a Results-Based M&E System"; USAID CLA Framework; BRAC Research and Evaluation Division; INTRAC M&E Universe

Papers & Resources

This course draws on decades of MEL practice and research. Below are the key resources organized by topic.

Key Academic Resources

MEL Lexicon

Master the language of MEL with our interactive glossary of 65+ essential terms. From "Theory of Change" to "Contribution Analysis"—the concepts you need to navigate the field.

Sample Terms

Theory of Change

A comprehensive description of how and why a desired change is expected to happen, articulating causal pathways and underlying assumptions.

Counterfactual

What would have happened in the absence of the intervention—the comparison scenario for measuring impact.

Contribution Analysis

Approach building a credible story of how a program contributed to observed outcomes, without claiming exclusive attribution.