How to Open Science: Debugging Reproducibility within the Educational Data Mining Conference
Contributors
Editors:
- 1. WestEd, USA
- 2. EPFL, Switzerland
- 3. Google Research and Indian Institute of Science, India
Description
Despite increased efforts to assess the adoption rates of open science and robustness of reproducibility in sub-disciplines of education technology, there is a lack of understanding of why some research is not reproducible. Prior work has taken the first step toward assessing reproducibility of research, but has assumed certain constraints which hinder its discovery. Thus, the purpose of this study was to replicate previous work on papers within the proceedings of the International Conference on Educational Data Mining to accurately report on which papers are reproducible and why. Specifically, we examined 208 papers, attempted to reproduce them, documented reasons for reproducibility failures, and asked authors to provide additional information needed to reproduce their study. Our results showed that out of 12 papers that were potentially reproducible, only one successfully reproduced all analyses, and another two reproduced most of the analyses. The most common failure for reproducibility was failure to mention libraries needed, followed by non-seeded randomness.
Files
2023.EDM-long-papers.10.pdf
Files
(362.2 kB)
Name | Size | Download all |
---|---|---|
md5:e5cc65d7a15cd55c199aea3775f62e73
|
362.2 kB | Preview Download |