Difference between revisions of "Reproducible Research"

Jump to: navigation, search
Line 3: Line 3:
 
*Even in such cases, however, we should till require reproducibility: this means that different researchers, when running the same analysis in the same data should find the same results. That may seem obvious, but unfortunately is not as widely observed as we would like.
 
*Even in such cases, however, we should till require reproducibility: this means that different researchers, when running the same analysis in the same data should find the same results. That may seem obvious, but unfortunately is not as widely observed as we would like.
 
*The bottom line of research reproducibility is that the path used to get to your results are as much a research output as the results themselves, making the research process fully transparent. This means that not only the final findings should be made available by researchers, but data, codes and documentation are also of great relevance to the public.
 
*The bottom line of research reproducibility is that the path used to get to your results are as much a research output as the results themselves, making the research process fully transparent. This means that not only the final findings should be made available by researchers, but data, codes and documentation are also of great relevance to the public.
 
== Pre-registration ==
 
Trial registries offer researchers the chance to upload and timestamp their study designs before they have been conducted. The aim of these registries is to build research transparency by reducing selective reporting and provide researchers with an overview of ongoing studies in their field. While trial registration is commonplace in the clinical health trials (see, for example, https://clinicaltrials.gov/), their use in development economics is more recent.
 
 
===Where can I register?===
 
The American Economic Association (AEA) hosts a trial registry specifically for randomized controlled trials[https://www.socialscienceregistry.org/]. The international Initiative for Impact Evaluation (3ie) provides a registry for experimental and quasi-experimental research in developing countries [http://www.ridie.org/].
 
 
===What information should be included?===
 
The information required for registering a trial typically includes the country and title, a brief description of the project, timeline, outcomes, sample size, study design, and ethical approval details. Some of the details provided can be uploaded and time stamped, but hidden from public view prior to study completion. A pre-analysis plan can be uploaded providing a detailed description of how the analysis will be conducted, but this is typically not mandatory for registration.
 
 
===When should I register?===
 
While clinical trials in health are expected to be registered before patient enrolment [http://icmje.org/recommendations/browse/publishing-and-editorial-issues/clinical-trial-registration.html], there is currently no formal requirement for development economics trials to be registered by a particular stage of the research. In cases where intervention delivery is uncertain, development economics researchers wait to register their trials after baseline and interventions have been completed, but before any follow up data collection or analysis [http://blogs.worldbank.org/impactevaluations/trying-out-new-trial-registries].
 
  
 
== Code replication ==
 
== Code replication ==

Revision as of 23:56, 9 February 2018

Read First

  • In most scientific fields, results are validated through replication: that means that different scientists run the same experiment independently in different samples and find similar conclusions. That standard is not always feasible in development research. More often than not, the phenomena we analyze cannot be artifically re-created. Even in the case of field experiments, different populations can respond differently to a treatment, and the costs involved are high.
  • Even in such cases, however, we should till require reproducibility: this means that different researchers, when running the same analysis in the same data should find the same results. That may seem obvious, but unfortunately is not as widely observed as we would like.
  • The bottom line of research reproducibility is that the path used to get to your results are as much a research output as the results themselves, making the research process fully transparent. This means that not only the final findings should be made available by researchers, but data, codes and documentation are also of great relevance to the public.

Code replication

  • Git is a free version-control software. Files are stored in Git Repositories, most commonly on GitHub. To learn GitHub, there is an introductory training available through GitHub Services, and multiple tutorials available through GitHub Guides

Data publication

Dynamic documents

  • R-markdown is a widely adopted tool for creating fully reproducible documents. It allows users to write text and code simultaneously, running analyses in different programming languages and printing results in the final document along with the text. Stata 15 also allows users to create dynamic documents using dyndoc.
  • Jupyter Notebook is used to create and share code in different programming languages, including Python, R, Julia, and Scala. It can also create dynamic documents in HTML, LaTeX and other formats.
  • LaTeX is another widely used tool in the scientific community. It is a type-setting system that allows users to reference code outputs such as tables and graphs so that they can be easily updated in a text document. Overleaf is a web based platform for collaboration in TeX documents.
  • Open science framework is a web based project management platform that combines registration, data storage (through Dropbox, Box, Google Drive and other platforms), code version control (through GitHub) and document composition (through Overleaf).

Additional Resources

From Data Colada:

From the Abul Latif Jameel Poverty Action Lab (JPAL)

From Innovations for Policy Action (IPA)

Center for Open Science

Berkeley Initiative for Transparency in the Social Sciences

Reproducible Research in R

Reproducible Research in Stata

Additional Resources