Difference between revisions of "Randomized Evaluations: Principles of Study Design"

Jump to: navigation, search
Line 20: Line 20:


==Step 2: Randomization==
==Step 2: Randomization==
Broadly speaking, this process involves allocating the sample selected (based on calculations in Step 1) into one of two groups: '''''treatment group''''', and ''control group''. This is the basis for establishing the '''''causal effect''''', which is the cornerstone of a randomized evaluation.
Broadly speaking, this process involves allocating the sample selected (based on calculations in Step 1) into one of two groups: '''''treatment group''''', and '''''control group'''''. This is the basis for establishing the '''''causal effect''''', which is the cornerstone of a randomized evaluation.
(See [[Randomization in Stata]] for technical explanation)
(See [[Randomization in Stata]] for technical explanation)



Revision as of 05:48, 22 February 2020

Randomized evaluations are field experiments involving the assignment of subjects randomly to one of two groups: one, the treatment group, which is receiving the policy intervention being evaluated, and two, the control group, which is in status-quo.

The results of the trial are used to answer questions about effectiveness of an intervention, and can prevent inefficient allocation of resources to programs that might not be effective.

This section covers the key principles of study design to guide researchers on best-practices in conducting field evaluations.

Read First

Step 1: Comprehensive protocol for the evaluation

This involves selecting a hypothesis (assumption) that specifies the anticipated link between the predictor variables and the outcomes, that is, the null hypothesis.

Key Concerns

  • The sample to be studied must be clearly specified, including exclusion/inclusion criteria.
  • Pilot studies can help identify ideal target population, as well as ascertain take-up rates, that can help with sample-size and power calculations.
  • The sample size must be selected in a manner that provides a high probability.(see Sample Size and Power Calculations)
  • A good study can be designed by consulting experienced researchers.

Step 2: Randomization

Broadly speaking, this process involves allocating the sample selected (based on calculations in Step 1) into one of two groups: treatment group, and control group. This is the basis for establishing the causal effect, which is the cornerstone of a randomized evaluation. (See Randomization in Stata for technical explanation)

Key Concerns

  • Effective randomization is important to tackle the issue of confounding, that is, when a characteristic is associated with the intervention, as well as the outcome.
  • This process must be concealed from the investigator. (see also Research Ethics)
  • Initial baseline characteristics must be measured across the two groups, and these should not be significantly different. One solution to this problem is randomization inference), a concept that is gaining ground in the field of randomized evaluations.
  • Care must be taken to ensure minimum attrition, that is, dropping out of some subjects after assignment.
  • Regardless, outcomes should still be compared against initial members of the control group - this is called intention-to-treat.

Step 3: Intervention, followed by measuring the outcomes

The next step is to apply the intervention, and then measure outcomes, called endline characteristics, after the pre-determined time-period has passed since the intervention.

Key Concerns

  • Sufficient time should be given for the intervention to have its intended effect. Premature calculations of outcomes can indirectly affect the power of the evaluation by affecting the minimum detectable effect size(MDES)
  • Blinding of the investigator to the intervention is crucial. It is also important for the subject to be blind to both, the assignment as well as the intervention, to prevent spillovers. This is called double blinding.
  • Also refer to measuring abstract concepts, and questionnaire design.

Final Step: Quality Control

Quality control is not just a step that needs to be exercised when measuring outcomes. It is a constant, rigorous process that needs to be carried out at various stages to ensure the integrity of the evaluation.

It includes dealing with concerns about design, measurement of outcomes, as well as handling data, and ensuring anonymity of subjects. (See data quality)

Key Concerns

  • Lack of quality control can lead to erroneous conclusions, for instance, evidence of ineffective treatment even though the problem really was ineffective evaluation.
  • This is where training manuals can help, by setting out rigorous standards for investigators, and providing ways to enforce these standards.
  • Training can also include information on standardizing data-collection and reporting.
  • See Data for Development Impact for a more comprehensive discussion on quality control.

Back to Parent

This article is part of Experimental Methods. However, most of the principles highlighted above can be applied in general to all kinds of evaluations, and therefore act as a crucial pointer to anyone looking to foray into the world of evaluations.

Additional Resources