scispace - formally typeset
Open AccessJournal ArticleDOI

Evaluation designs for adequacy, plausibility and probability of public health programme performance and impact.

Reads0
Chats0
TLDR
It is argued that the answer to this question is essential for choosing an appropriate evaluation design, drawing upon examples from the fields of health and nutrition, and a framework is proposed for deciding upon appropriate evaluation designs.
Abstract
The question of why to evaluate a programme is seldom discussed in the literature. The present paper argues that the answer to this question is essential for choosing an appropriate evaluation design. The discussion is centered on summative evaluations of large-scale programme effectiveness, drawing upon examples from the fields of health and nutrition but the findings may be applicable to other subject areas. The main objective of an evaluation is to influence decisions. How complex and precise the evaluation must be depends on who the decision maker is and on what types of decisions will be taken as a consequence of the findings. Different decision makers demand not only different types of information but also vary in their requirements of how informative and precise the findings must be. Both complex and simple evaluations, however, should be equally rigorous in relating the design to the decisions. Based on the types of decisions that may be taken, a framework is proposed for deciding upon appropriate evaluation designs. Its first axis concerns the indicators of interest, whether these refer to provision or utilization of services, coverage or impact measures. The second axis refers to the type of inference to be made, whether this is a statement of adequacy, plausibility or probability. In addition to the above framework, other factors affect the choice of an evaluation design, including the efficacy of the intervention, the field of knowledge, timing and costs. Regarding the latter, decision makers should be made aware that evaluation costs increase rapidly with complexity so that often a compromise must be reached. Examples are given of how to use the two classification axes, as well as these additional factors, for helping decision makers and evaluators translate the need for evaluation--the why--into the appropriate design--the how.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Combination prevention: a deeper understanding of effective HIV prevention.

TL;DR: Applying the principles of combination prevention systematically and consistently in HIV programme planning, with due attention to context, can increase HIV programme effectiveness.
Journal ArticleDOI

Assessing impact and impact pathways of a homestead food production program on household and child nutrition in Cambodia

TL;DR: Cambodia's homestead food production program increased household production and consumption of micronutrient-rich foods and maternal and child intake (or frequency of intake) of some of these foods, but household-level benefits from the program did not translate into significant improvements in mothers and child health and nutrition.

How effective are community health workers? An Overview of Current Evidence with Recommendations for Strengthening Community Health Worker Programs to Accelerate Progress in Achieving the Health-related Millennium Development Goals.

Perry H, +1 more
TL;DR: This document serves to provide an update and supplement to the extensive review carried out by Bhutta and colleagues under the auspices of the World Health Organization and the Global Health Workforce Alliance in 2010.
References
More filters
Book

The Theory and Practice of Econometrics

TL;DR: The Classical Inference Approach for the General Linear Model, Statistical Decision Theory and Biased Estimation, and the Bayesian Approach to Inference are reviewed.
Book

Case-Control Studies: Design, Conduct, Analysis

TL;DR: Case-control studies, often called 'retrospective' studies, provide a research method for investigating factors that may prevent or cause disease.
Journal ArticleDOI

Why we need observational studies to evaluate the effectiveness of health care

Nick Black
- 11 May 1996 - 
TL;DR: The view is widely held that experimental methods (randomised controlled trials) are the "gold standard" for evaluation and that observational methods have little or no value, but this ignores the limitations of randomised trials.
Related Papers (5)