Authors
Chun-Kai Huang, Cameron Neylon, Richard Hosking, Lucy Montgomery, Katie Wilson, Alkim Ozaygen, Chloe Brookes-Kenworthy
Publication date
2020/3/21
Journal
bioRxiv
Pages
2020.03. 19.998542
Publisher
Cold Spring Harbor Laboratory
Description
In the article “Evaluating institutional open access performance: Methodology, challenges and assessment” we develop the first comprehensive and reproducible workflow that integrates multiple bibliographic data sources for evaluating institutional open access (OA) performance. The major data sources include Web of Science, Scopus, Microsoft Academic, and Unpaywall. However, each of these databases continues to update, both actively and retrospectively. This implies the results produced by the proposed process are potentially sensitive to both the choice of data source and the versions of them used. In addition, there remain the issue relating to selection bias in sample size and margin of error. The current work shows that the levels of sensitivity relating to the above issues can be significant at the institutional level. Hence, the transparency and clear documentation of the choices made on data sources (and their versions) and cut-off boundaries are vital for reproducibility and verifiability.
Total citations
2020202121
Scholar articles