Sly Saint
Senior Member (Voting Rights)
Despite agreeing to make raw data available, some authors fail to comply. The right strategies and platforms can ease the task.
https://www.nature.com/articles/d41586-022-03133-5
Journals and funding bodies increasingly require manuscript authors to share data on request or make the information publicly available. It’s a big ask from a technical standpoint, but some straightforward strategies can simplify the process.
Scientific papers rarely include all the data used to justify the conclusions, even in the supplementary material. Authors might fear getting scooped, or that other researchers will use the raw data to make fresh discoveries, or they might wish to protect the privacy of study participants. Or, more probably, authors have neither the time nor the expertise to package the data for others to view and understand.
Such reticence costs the research community. Data transparency allows others to repeat analyses and catch mistakes or fraudulent claims. It allows for new findings through the reanalysis of existing data sets, and it increases trust in the scientific process. In August, the White House Office of Science and Technology Policy announced that, by 2025, scientific data from all new federally funded research must be made accessible to the US public. And when submitting papers, authors are increasingly required to provide raw data to editors, to place data online or to include data-sharing statements as to whether they will offer data on request. Unfortunately, such policies are not bulletproof, as the largest study of its kind starkly documents.
Tom Jefferson, an epidemiologist at the University of Oxford, UK, says authors should face consequences for making false data-availability statements. “The editors should take action, whether it’s a correction or retraction,” he says, adding that the excuse of no longer having the data to hand is like saying “the cat ate my filing cabinet”.
full articleValentin Danchev, a computational social scientist at Queen Mary University of London, calls the study a useful step towards understanding the actual state of data sharing. But, he adds, “we need more of those studies so that we can generalize across different areas and different survey designs”.
Last year, Danchev co-authored a study2 of 487 clinical trials that were published in JAMA, The Lancet or The New England Journal of Medicine. The authors of 89 of these articles said they’d stored data sets in online repositories, but Danchev’s team could find only 17 in the designated locations.
https://www.nature.com/articles/d41586-022-03133-5