Updating search results...

OSKB

This endorsement is the seal of approval for inclusion in the OSKB Library collections.

These resources have been vetted by the OSKB Team.

329 affiliated resources

Search Resources

View
Selected filters:
Open science challenges, benefits and tips in early career and beyond
Unrestricted Use
CC BY
Rating
0.0 stars

The movement towards open science is a consequence of seemingly pervasive failures to replicate previous research. This transition comes with great benefits but also significant challenges that are likely to affect those who carry out the research, usually early career researchers (ECRs). Here, we describe key benefits, including reputational gains, increased chances of publication, and a broader increase in the reliability of research. The increased chances of publication are supported by exploratory analyses indicating null findings are substantially more likely to be published via open registered reports in comparison to more conventional methods. These benefits are balanced by challenges that we have encountered and that involve increased costs in terms of flexibility, time, and issues with the current incentive structure, all of which seem to affect ECRs acutely. Although there are major obstacles to the early adoption of open science, overall open science practices should benefit both the ECR and improve the quality of research. We review 3 benefits and 3 challenges and provide suggestions from the perspective of ECRs for moving towards open science practices, which we believe scientists and institutions at all levels would do well to consider.

Subject:
Biology
Life Science
Material Type:
Reading
Provider:
PLOS Biology
Author:
Christopher Allen
David M. A. Mehler
Date Added:
08/07/2020
Optimizing Research Collaboration
Unrestricted Use
CC BY
Rating
0.0 stars

In this webinar, we demonstrate the OSF tools available for contributors, labs, centers, and institutions that support stronger collaborations. The demo includes useful practices like: contributor management, the OSF wiki as an electronic lab notebook, using OSF to manage online courses and syllabi, and more. Finally, we look at how OSF Institutions can provide discovery and intelligence gathering infrastructure so that you can focus on conducting and supporting exceptional research. The Center for Open Science’s ongoing mission is to provide community and technical resources to support your commitments to rigorous, transparent research practices. Visit cos.io/institutions to learn more.

Subject:
Applied Science
Computer Science
Information Science
Material Type:
Lecture
Provider:
Center for Open Science
Author:
Center for Open Science
Date Added:
08/07/2020
Optimizing Research Collaboration for Remote Teams
Unrestricted Use
CC BY
Rating
0.0 stars

Many in the global research community are adapting to conducting work remotely while exploring the best ways to maintain collaboration with colleagues across teams and institutions.

Join us as we discuss the OSF tools available for contributors, labs, centers, and institutions that support stronger collaborations.

We demonstrate: contributor management, the OSF wiki as an electronic lab notebook, how to affiliate research projects for institution-wide discovery, using OSF to manage online courses and syllabi, and more. Plus, see examples from the research teams optimizing their workflows for inclusive collaboration and efficient data management.

Subject:
Education
Material Type:
Lecture
Provider:
Center for Open Science
Date Added:
03/21/2021
Outcome reporting bias in randomized-controlled trials investigating antipsychotic drugs
Unrestricted Use
CC BY
Rating
0.0 stars

Recent literature hints that outcomes of clinical trials in medicine are selectively reported. If applicable to psychotic disorders, such bias would jeopardize the reliability of randomized clinical trials (RCTs) investigating antipsychotics and thus their extrapolation to clinical practice. We therefore comprehensively examined outcome reporting bias in RCTs of antipsychotic drugs by a systematic review of prespecified outcomes on ClinicalTrials.gov records of RCTs investigating antipsychotic drugs in schizophrenia and schizoaffective disorder between 1 January 2006 and 31 December 2013. These outcomes were compared with outcomes published in scientific journals. Our primary outcome measure was concordance between prespecified and published outcomes; secondary outcome measures included outcome modifications on ClinicalTrials.gov after trial inception and the effects of funding source and directionality of results on record adherence. Of the 48 RCTs, 85% did not fully adhere to the prespecified outcomes. Discrepancies between prespecified and published outcomes were found in 23% of RCTs for primary outcomes, whereas 81% of RCTs had at least one secondary outcome non-reported, newly introduced, or changed to a primary outcome in the respective publication. In total, 14% of primary and 44% of secondary prespecified outcomes were modified after trial initiation. Neither funding source (P=0.60) nor directionality of the RCT results (P=0.10) impacted ClinicalTrials.gov record adherence. Finally, the number of published safety endpoints (N=335) exceeded the number of prespecified safety outcomes by 5.5 fold. We conclude that RCTs investigating antipsychotic drugs suffer from substantial outcome reporting bias and offer suggestions to both monitor and limit such bias in the future.

Subject:
Applied Science
Health, Medicine and Nursing
Material Type:
Reading
Provider:
Translational Psychiatry
Author:
C. H. Vinkers
C. M. C. Lemmens
J. J. Luykx
M. Lancee
R. S. Kahn
Date Added:
08/07/2020
Peer Review: Decisions, decisions
Unrestricted Use
CC BY
Rating
0.0 stars

Journals are exploring new approaches to peer review in order to reduce bias, increase transparency and respond to author preferences. Funders are also getting involved. If you start reading about the subject of peer review, it won't be long before you encounter articles with titles like Can we trust peer review?, Is peer review just a crapshoot? and It's time to overhaul the secretive peer review process. Read some more and you will learn that despite its many shortcomings – it is slow, it is biased, and it lets flawed papers get published while rejecting work that goes on to win Nobel Prizes – the practice of having your work reviewed by your peers before it is published is still regarded as the 'gold standard' of scientific research. Carry on reading and you will discover that peer review as currently practiced is a relatively new phenomenon and that, ironically, there have been remarkably few peer-reviewed studies of peer review.

Subject:
Applied Science
Information Science
Material Type:
Reading
Provider:
eLife
Author:
Peter Rodgers
Date Added:
08/07/2020
Plotting and Programming in Python
Unrestricted Use
CC BY
Rating
0.0 stars

This lesson is part of Software Carpentry workshops and teach an introduction to plotting and programming using python. This lesson is an introduction to programming in Python for people with little or no previous programming experience. It uses plotting as its motivating example, and is designed to be used in both Data Carpentry and Software Carpentry workshops. This lesson references JupyterLab, but can be taught using a regular Python interpreter as well. Please note that this lesson uses Python 3 rather than Python 2.

Subject:
Applied Science
Computer Science
Information Science
Mathematics
Measurement and Data
Material Type:
Module
Provider:
The Carpentries
Author:
Adam Steer
Allen Lee
Andreas Hilboll
Ashley Champagne
Benjamin
Benjamin Roberts
CanWood
Carlos Henrique Brandt
Carlos M Ortiz Marrero
Cephalopd
Cian Wilson
Dan Mønster
Daniel W Kerchner
Daria Orlowska
Dave Lampert
David Matten
Erin Alison Becker
Florian Goth
Francisco J. Martínez
Greg Wilson
Jacob Deppen
Jarno Rantaharju
Jeremy Zucker
Jonah Duckles
Kees den Heijer
Keith Gilbertson
Kyle E Niemeyer
Lex Nederbragt
Logan Cox
Louis Vernon
Lucy Dorothy Whalley
Madeleine Bonsma-Fisher
Mark Phillips
Mark Slater
Maxim Belkin
Michael Beyeler
Mike Henry
Narayanan Raghupathy
Nigel Bosch
Olav Vahtras
Pablo Hernandez-Cerdan
Paul Anzel
Phil Tooley
Raniere Silva
Robert Woodward
Ryan Avery
Ryan Gregory James
SBolo
Sarah M Brown
Shyam Dwaraknath
Sourav Singh
Steven Koenig
Stéphane Guillou
Taylor Smith
Thor Wikfeldt
Timothy Warren
Tyler Martin
Vasu Venkateshwaran
Vikas Pejaver
ian
mzc9
Date Added:
08/07/2020
Poor replication validity of biomedical association studies reported by newspapers
Unrestricted Use
CC BY
Rating
0.0 stars

Objective To investigate the replication validity of biomedical association studies covered by newspapers. Methods We used a database of 4723 primary studies included in 306 meta-analysis articles. These studies associated a risk factor with a disease in three biomedical domains, psychiatry, neurology and four somatic diseases. They were classified into a lifestyle category (e.g. smoking) and a non-lifestyle category (e.g. genetic risk). Using the database Dow Jones Factiva, we investigated the newspaper coverage of each study. Their replication validity was assessed using a comparison with their corresponding meta-analyses. Results Among the 5029 articles of our database, 156 primary studies (of which 63 were lifestyle studies) and 5 meta-analysis articles were reported in 1561 newspaper articles. The percentage of covered studies and the number of newspaper articles per study strongly increased with the impact factor of the journal that published each scientific study. Newspapers almost equally covered initial (5/39 12.8%) and subsequent (58/600 9.7%) lifestyle studies. In contrast, initial non-lifestyle studies were covered more often (48/366 13.1%) than subsequent ones (45/3718 1.2%). Newspapers never covered initial studies reporting null findings and rarely reported subsequent null observations. Only 48.7% of the 156 studies reported by newspapers were confirmed by the corresponding meta-analyses. Initial non-lifestyle studies were less often confirmed (16/48) than subsequent ones (29/45) and than lifestyle studies (31/63). Psychiatric studies covered by newspapers were less often confirmed (10/38) than the neurological (26/41) or somatic (40/77) ones. This is correlated to an even larger coverage of initial studies in psychiatry. Whereas 234 newspaper articles covered the 35 initial studies that were later disconfirmed, only four press articles covered a subsequent null finding and mentioned the refutation of an initial claim. Conclusion Journalists preferentially cover initial findings although they are often contradicted by meta-analyses and rarely inform the public when they are disconfirmed.

Subject:
Applied Science
Health, Medicine and Nursing
Material Type:
Reading
Provider:
PLOS ONE
Author:
Andy Smith
Estelle Dumas-Mallet
François Gonon
Thomas Boraud
Date Added:
08/07/2020
Poor statistical reporting, inadequate data presentation and spin persist despite editorial advice
Unrestricted Use
CC BY
Rating
0.0 stars

The Journal of Physiology and British Journal of Pharmacology jointly published an editorial series in 2011 to improve standards in statistical reporting and data analysis. It is not known whether reporting practices changed in response to the editorial advice. We conducted a cross-sectional analysis of reporting practices in a random sample of research papers published in these journals before (n = 202) and after (n = 199) publication of the editorial advice. Descriptive data are presented. There was no evidence that reporting practices improved following publication of the editorial advice. Overall, 76-84% of papers with written measures that summarized data variability used standard errors of the mean, and 90-96% of papers did not report exact p-values for primary analyses and post-hoc tests. 76-84% of papers that plotted measures to summarize data variability used standard errors of the mean, and only 2-4% of papers plotted raw data used to calculate variability. Of papers that reported p-values between 0.05 and 0.1, 56-63% interpreted these as trends or statistically significant. Implied or gross spin was noted incidentally in papers before (n = 10) and after (n = 9) the editorial advice was published. Overall, poor statistical reporting, inadequate data presentation and spin were present before and after the editorial advice was published. While the scientific community continues to implement strategies for improving reporting practices, our results indicate stronger incentives or enforcements are needed.

Subject:
Applied Science
Health, Medicine and Nursing
Material Type:
Reading
Provider:
PLOS ONE
Author:
Annie A. Butler
Joanna Diong
Martin E. Héroux
Simon C. Gandevia
Date Added:
08/07/2020
The Post-Embargo Open Access Citation Advantage: It Exists (Probably), It’s Modest (Usually), and the Rich Get Richer (of Course)
Unrestricted Use
CC BY
Rating
0.0 stars

Many studies show that open access (OA) articles—articles from scholarly journals made freely available to readers without requiring subscription fees—are downloaded, and presumably read, more often than closed access/subscription-only articles. Assertions that OA articles are also cited more often generate more controversy. Confounding factors (authors may self-select only the best articles to make OA; absence of an appropriate control group of non-OA articles with which to compare citation figures; conflation of pre-publication vs. published/publisher versions of articles, etc.) make demonstrating a real citation difference difficult. This study addresses those factors and shows that an open access citation advantage as high as 19% exists, even when articles are embargoed during some or all of their prime citation years. Not surprisingly, better (defined as above median) articles gain more when made OA.

Subject:
Applied Science
Health, Medicine and Nursing
Material Type:
Reading
Provider:
PLOS ONE
Author:
Jim Ottaviani
Date Added:
08/07/2020
Practical Considerations for Navigating Registered Reports
Unrestricted Use
CC BY
Rating
0.0 stars

Practical Considerations for Navigating Registered Reports

Subject:
Education
Material Type:
Reading
Author:
Albers C
Button K S
Chambers C D
Helen Wills Neuroscience Institute University Of California Berkeley Ca Usa
Jason M
Jscimeca Berkeley Edu
Kiyonaga Berkeley Edu
Lakens D
Nosek B A
Poldrack R A
Date Added:
12/21/2021
Practical steps for increasing opennes and reproducibility: an introduction to OSF, Workshop Manual
Unrestricted Use
CC BY
Rating
0.0 stars

Reproducibility, Research Management Planning, Structuring a study, Preregistration + Analysis Plan, Files and Version Control, Sharing on the OSF, Incentives (Badges, RR)

Subject:
Education
Material Type:
Reading
Author:
Center For Open Science
Courtney K
Ian Sullivan
Jennifer Freeman Smith
Jolene Esposito
Natalie Meyers
Nicole Pfeiffer
Sara Bowman
Date Added:
12/21/2021
Pre-analysis Plans: A Stocktaking
Read the Fine Print
Rating
0.0 stars

The evidence-based community has championed the public registration of pre-analysis plans (PAPs) as a solution to the problem of research credibility, but without any evidence that PAPs actually bolster the credibility of research. We analyze a representative sample of 195 pre-analysis plans (PAPs) from the American Economic Association (AEA) and Evidence in Governance and Politics (EGAP) registration platforms to assess whether PAPs are sufficiently clear, precise and comprehensive to be able to achieve their objectives of preventing “fishing” and reducing the scope for post-hoc adjustment of research hypotheses. We also analyze a subset of 93 PAPs from projects that have resulted in publicly available papers to ascertain how faithfully they adhere to their pre-registered specifications and hypotheses. We find significant variation in the extent to which PAPs are accomplishing the goals they were designed to achieve

Subject:
Economics
Social Science
Material Type:
Reading
Author:
Daniel Posner
George Ofosu
Date Added:
08/07/2020
Preparing code and data for computationally reproducible collaboration and publication: a hands-on workshop
Unrestricted Use
CC BY
Rating
0.0 stars

Computational analyses are playing an increasingly central role in research. Journals, funders, and researchers are calling for published research to include associated data and code. However, many involved in research have not received training in best practices and tools for sharing code and data. This course aims to address this gap in training while also providing those who support researchers with curated best practices guidance and tools.This course is unique compared to other reproducibility courses due to its practical, step-by-step design. It is comprised of hands-on exercises to prepare research code and data for computationally reproducible publication. Although the course starts with some brief introductory information about computational reproducibility, the bulk of the course is guided work with data and code. Participants move through preparing research for reuse, organization, documentation, automation, and submitting their code and data to share. Tools that support reproducibility will be introduced (Code Ocean), but all lessons will be platform agnostic.Level: IntermediateIntended audience: The course is targeted at researchers and research support staff who are involved in the preparation and publication of research materials. Anyone with an interest in reproducible publication is welcome. The course is especially useful for those looking to learn practical steps for improving the computational reproducibility of their own research.

Subject:
Applied Science
Life Science
Physical Science
Social Science
Material Type:
Activity/Lab
Author:
April Clyburne-Sherin
Date Added:
08/08/2019
The Preregistration Challenge: A How To Guide
Unrestricted Use
CC BY
Rating
0.0 stars

This video shows interested researchers how to get started on their own preregistration as part of the Preregistration Challenge. Learn how to create a new draft, find example preregistrations from different fields, respond to comments from the preregistration review team, and turn your final draft into a formal preregistration. For more information, check out https://www.cos.io/initiatives/prereg-more-information.

Subject:
Education
Material Type:
Lesson
Provider:
Center for Open Science
Date Added:
03/31/2021
Preregistration: Improve Research Rigor, Reduce Bias
Unrestricted Use
CC BY
Rating
0.0 stars

In this webinar Professor Brian Nosek, Executive Director of the Center for Open Science (https://cos.io), outlines the practice of Preregistration and how it can aid in increasing the rigor and reproducibility of research. The webinar is co-hosted by the Health Research Alliance, a collaborative member organization of nonprofit research funders. Slides available at: https://osf.io/9m6tx/

Subject:
Applied Science
Computer Science
Information Science
Material Type:
Lecture
Provider:
Center for Open Science
Author:
Center for Open Science
Date Added:
08/07/2020
Preregistration Overview page
Unrestricted Use
CC BY
Rating
0.0 stars

What is Preregistration? When you preregister your research, you're simply specifying your research plan in advance of your study and submitting it to a registry. Preregistration separates hypothesis-generating (exploratory) from hypothesis-testing (confirmatory) research. Both are important. But the same data cannot be used to generate and test a hypothesis, which can happen unintentionally and reduce the credibility of your results. Addressing this problem through planning improves the quality and transparency of your research. This helps you clearly report your study and helps others who may wish to build on it.

Subject:
Applied Science
Life Science
Physical Science
Social Science
Material Type:
Reading
Provider:
Center for Open Science
Author:
Center for Open Science
Date Added:
06/18/2020
Preregistration in Complex Contexts: A Preregistration Template for the Application of Cognitive Models
Unrestricted Use
CC BY
Rating
0.0 stars

In recent years, open science practices have become increasingly popular in psychology and related sciences. These practices aim to increase rigour and transparency in science as a potential response to the challenges posed by the replication crisis. Many of these reforms -- including the highly influential preregistration -- have been designed for experimental work that tests simple hypotheses with standard statistical analyses, such as assessing whether an experimental manipulation has an effect on a variable of interest. However, psychology is a diverse field of research, and the somewhat narrow focus of the prevalent discussions surrounding and templates for preregistration has led to debates on how appropriate these reforms are for areas of research with more diverse hypotheses and more complex methods of analysis, such as cognitive modelling research within mathematical psychology. Our article attempts to bridge the gap between open science and mathematical psychology, focusing on the type of cognitive modelling that Crüwell, Stefan, & Evans (2019) labelled model application, where researchers apply a cognitive model as a measurement tool to test hypotheses about parameters of the cognitive model. Specifically, we (1) discuss several potential researcher degrees of freedom within model application, (2) provide the first preregistration template for model application, and (3) provide an example of a preregistered model application using our preregistration template. More broadly, we hope that our discussions and proposals constructively advance the debate surrounding preregistration in cognitive modelling, and provide a guide for how preregistration templates may be developed in other diverse or complex research contexts.

Subject:
Applied Science
Life Science
Physical Science
Social Science
Material Type:
Reading
Author:
Nathan Evans
Sophia Crüwell
Date Added:
12/07/2019
Preregistration of Qualitative Research
Unrestricted Use
CC BY
Rating
0.0 stars

In this webinar, Tamarinde Haven provides an overview of the process of preregistration in qualitative research. We review the process of preregistration, how we partnered with a community of qualitative researchers to develop a template for qualitative research through a Delphi study, and a guide to the fields that were included in the final form.

Subject:
Education
Material Type:
Lecture
Provider:
Center for Open Science
Author:
Tamarinde Haven
Date Added:
03/21/2021
Pre-results Review in Economics: Lessons Learned from Setting up Registered Reports
Unrestricted Use
CC BY
Rating
0.0 stars

Hear from Andrew Foster, editor at the Journal of Development Economics, and Irenaeus Wolff, a guest editor for Experimental Economics, as they discuss their experiences with implementing the Registered Reports format, how it was received by authors, and the trends they noticed after adoption. Aleksandar Bogdanoski of BITSS also joins us to explore pre-results review, how to facilitate the process at journals, and best practices for supporting authors and reviewers.

Subject:
Education
Material Type:
Lesson
Provider:
Center for Open Science
Author:
Aleksandar Bogdanoski
Andrew Foster
Irenaeus Wolff
Date Added:
03/31/2021