Monitoring and Evaluation
for Complicated and Complex Aspects of
Programmes and Policies
1 - 3 March 2011
Crowne Plaza, Rosebank, Johannesburg
The issues faced by those involved in development in southern Africa
are not simple. And yet the logic models and monitoring and
evaluation frameworks used are often based on over-simplified
descriptions of programmes and policies.
How do practitioners develop ways that adequately evaluate aspects
of programmes and policies that are complicated or complex?
This course is aimed at practitioners in eastern and southern Africa who find
that their current monitoring and evaluation does not adequately describe or
support effective practice or policy. Drawing on work on systems approaches
to planning and evaluation by Glouberman and Zimmerman, Kurtz and
Snowden, and Patton, the course is designed to help participants distinguish
between aspects of their programmes and policies that are useful to think
about as simple, and those that need to be considered as complicated or
Complicated aspects of programmes and policies include multiple
components, multiple agencies, or might work in different ways for different
people or in different situations. Complex aspects of programmes and
policies are those which are inherently unpredictable, where activities are
flexible and adaptive, causal relationships are disproportionate (where at
critical levels, a small change can make a big difference – a ‘tipping point’)
and outcomes are emergent.
The programme is aimed at:
Practitioners designing programmes, policies or M&E processes for complex
human service interventions.
Civil society organisations, academic programmes, HIV & development
responses, international partner organisations and government members.
COURSE DETAILS Duration:
1-3 March 2011
Venue: Crowne Plaza, Rosebank, Johannesburg (SA)
Cost: R4 000 per Day delegate
Resident delegate cost available on application
Closing Date: Registration closes 18 February
Scholarships: Scholarships are available on application
motivation from sending organization)
1. We can accommodate a maximum of 20 delegates and therefore
early registration is advised to secure your booking.
2. Although closing date is 18 February, bookings will close when 20
applications have been received.
3. Day delegate cost includes tuition costs, lunches & teas.
4. Resident delegate cost includes tuition costs, accommodation
and all meals.
• Using logic models and program theory for planning interventions and
KEY their evaluations - this goes well beyond listing
inputs/processes/outputs/outcomes to articulating program theories that
COMPONENTS incorporate both a theory of change and a theory of action
COVERED • Using evidence from research, evaluations and practice to assess the
plausibility of the intervention design
• Using programme theory to structure the monitoring and evaluation,
including investigating how interventions work differently for different
people in different settings
• Options for impact evaluation - overview of methodological options in
terms of the six tasks involved in impact evaluation:
1. Identifying and negotiating the values that will be used
2. Conceptualising how the program or policy is intended to work
3. Describing what is happening in terms of processes, results and
4. Analysing the causal contribution of the intervention to the
results in conjunction with other contributing factors
5. Synthesising evidence and reporting the findings
6. Supporting utilization
• Using non-experimental methods, designs and approaches to the issue of
causal attribution/contribution for situations where experimental or quasi-
experimental designs are not feasible or appropriate.
ABOUT THE Patricia Rogers is Professor in Public Sector Evaluation and leader of the
research program in Evidence Based Policy and Practice in the Sustainable Health
PRESENTER and Well-Being Research Institute at the Royal Melbourne Institute of Technology.
Patricia was previously Spencer Postdoctoral Fellow at the Harvard Project on
Schooling at Children (1997-8), Research Analyst at the City of Melbourne (1986-
1988) and Research Co-ordinator at the City of Fairfield (NSW) (1984-1986).
She has worked in public sector evaluation and research for more than 20 years,
across a wide range of programmes (including health, early childhood, education,
community development, Indigenous housing, criminal justice, international
development, and agriculture) and levels of government (national, state and local).
After more than 25 years of experience in evaluation, she now focuses particularly on
evidence-based policy and practice in complicated and complex human service areas.
REGISTRATION 1. Please complete the course registration form - click here to access the
registration form on the HEARD website*.
2. Arrange payment – see registration form for bank details.
3. Return completed registration form and proof of payment to
email@example.com, by 18 February 2011.
4. After submission delegates will receive confirmation.
* should you have difficulty accessing the registration form, simply email firstname.lastname@example.org
to request that a form be emailed to you.
HEARD | www.heard.org.za
The course is being HEARD conducts applied research and runs development interventions aimed at
offered by: mobilising evidence for impact in health and HIV in the SADC and east Africa
region. HEARD has been situated since 1998 at the University of KwaZulu-Natal
in Durban, South Africa and collaborates with a range of institutional and
individual partners spanning the globe.
t: +27 31 260-2592 | f: +27 31 260-2587|
a: University of KwaZulu-Natal, Westville Campus, Private Bag X54001 Durban 4000, South Africa