This article provides a roadmap to assist graduate students and their advisors to engage in open science practices. We suggest eight open science practices that novice graduate students could begin adopting today. The topics we cover include journal clubs, project workflow, preprints, reproducible code, data sharing, transparent writing, preregistration, and registered reports. To address concerns about not knowing how to engage in open science practices, we provide a difficulty rating of each behavior (easy, medium, difficult), present them in order of suggested adoption, and follow the format of what, why, how, and worries. We give graduate students ideas on how to approach conversations with their advisors/collaborators, ideas on how to integrate open science practices within the graduate school framework, and specific resources on how to engage with each behavior. We emphasize that engaging in open science behaviors need not be an all or nothing approach, but rather graduate students can engage with any number of the behaviors outlined.
215 Results
A Data Carpentry curriculum for Economics is being developed by Dr. Miklos Koren at Central European University. These materials are being piloted locally. Development for these lessons has been supported by a grant from the Sloan Foundation.
- Subject:
- Applied Science
- Computer Science
- Economics
- Information Science
- Mathematics
- Measurement and Data
- Social Science
- Material Type:
- Module
- Provider:
- The Carpentries
- Author:
- Andras Vereckei
- Arieda Muço
- Miklós Koren
- Date Added:
- 08/07/2020
Low reproducibility rates within life science research undermine cumulative knowledge production and contribute to both delays and costs of therapeutic drug development. An analysis of past studies indicates that the cumulative (total) prevalence of irreproducible preclinical research exceeds 50%, resulting in approximately US$28,000,000,000 (US$28B)/year spent on preclinical research that is not reproducible—in the United States alone. We outline a framework for solutions and a plan for long-term improvements in reproducibility rates that will help to accelerate the discovery of life-saving therapies and cures.
- Subject:
- Biology
- Life Science
- Material Type:
- Reading
- Provider:
- PLOS Biology
- Author:
- Iain M. Cockburn
- Leonard P. Freedman
- Timothy S. Simcoe
- Date Added:
- 08/07/2020
In animal experiments, animals, husbandry and test procedures are traditionally standardized to maximize test sensitivity and minimize animal use, assuming that this will also guarantee reproducibility. However, by reducing within-experiment variation, standardization may limit inference to the specific experimental conditions. Indeed, we have recently shown in mice that standardization may generate spurious results in behavioral tests, accounting for poor reproducibility, and that this can be avoided by population heterogenization through systematic variation of experimental conditions. Here, we examined whether a simple form of heterogenization effectively improves reproducibility of test results in a multi-laboratory situation. Each of six laboratories independently ordered 64 female mice of two inbred strains (C57BL/6NCrl, DBA/2NCrl) and examined them for strain differences in five commonly used behavioral tests under two different experimental designs. In the standardized design, experimental conditions were standardized as much as possible in each laboratory, while they were systematically varied with respect to the animals' test age and cage enrichment in the heterogenized design. Although heterogenization tended to improve reproducibility by increasing within-experiment variation relative to between-experiment variation, the effect was too weak to account for the large variation between laboratories. However, our findings confirm the potential of systematic heterogenization for improving reproducibility of animal experiments and highlight the need for effective and practicable heterogenization strategies.
- Subject:
- Applied Science
- Health, Medicine and Nursing
- Material Type:
- Reading
- Provider:
- PLOS ONE
- Author:
- Benjamin Zipser
- Berry Spruijt
- Britta Schindler
- Chadi Touma
- Christiane Brandwein
- David P. Wolfer
- Hanno Würbel
- Johanneke van der Harst
- Joseph P. Garner
- Lars Lewejohann
- Niek van Stipdonk
- Norbert Sachser
- Peter Gass
- S. Helene Richter
- Sabine Chourbaji
- Vootele Võikar
- Date Added:
- 08/07/2020
Software Carpentry lección para control de versiones con Git Para ilustrar el poder de Git y GitHub, usaremos la siguiente historia como un ejemplo motivador a través de esta lección. El Hombre Lobo y Drácula han sido contratados por Universal Missions para investigar si es posible enviar su próximo explorador planetario a Marte. Ellos quieren poder trabajar al mismo tiempo en los planes, pero ya han experimentado ciertos problemas anteriormente al hacer algo similar. Si se rotan por turnos entonces cada uno gastará mucho tiempo esperando a que el otro termine, pero si trabajan en sus propias copias e intercambian los cambios por email, las cosas se perderán, se sobreescribirán o se duplicarán. Un colega sugiere utilizar control de versiones para lidiar con el trabajo. El control de versiones es mejor que el intercambio de ficheros por email: Nada se pierde una vez que se incluye bajo control de versiones, a no ser que se haga un esfuerzo sustancial. Como se van guardando todas las versiones precedentes de los ficheros, siempre es posible volver atrás en el tiempo y ver exactamente quién escribió qué en un día en particular, o qué versión de un programa fue utilizada para generar un conjunto de resultados en particular. Como se tienen estos registros de quién hizo qué y en qué momento, es posible saber a quién preguntar si se tiene una pregunta en un momento posterior y, si es necesario, revertir el contenido a una versión anterior, de forma similar a como funciona el comando “deshacer” de los editores de texto. Cuando varias personas colaboran en el mismo proyecto, es posible pasar por alto o sobreescribir de manera accidental los cambios hechos por otra persona. El sistema de control de versiones notifica automáticamente a los usuarios cada vez que hay un conflicto entre el trabajo de una persona y la otra. Los equipos no son los únicos que se benefician del control de versiones: los investigadores independientes se pueden beneficiar en gran medida. Mantener un registro de qué ha cambiado, cuándo y por qué es extremadamente útil para todos los investigadores si alguna vez necesitan retomar el proyecto en un momento posterior (e.g. un año después, cuando se ha desvanecido el recuerdo de los detalles).
- Subject:
- Applied Science
- Computer Science
- Information Science
- Mathematics
- Measurement and Data
- Material Type:
- Module
- Provider:
- The Carpentries
- Author:
- Alejandra Gonzalez-Beltran
- Amy Olex
- Belinda Weaver
- Bradford Condon
- Casey Youngflesh
- Daisie Huang
- Dani Ledezma
- Francisco Palm
- Garrett Bachant
- Heather Nunn
- Hely Salgado
- Ian Lee
- Ivan Gonzalez
- James E McClure
- Javier Forment
- Jimmy O'Donnell
- Jonah Duckles
- K.E. Koziar
- Katherine Koziar
- Katrin Leinweber
- Kevin Alquicira
- Kevin MF
- Kurt Glaesemann
- LauCIFASIS
- Leticia Vega
- Lex Nederbragt
- Mark Woodbridge
- Matias Andina
- Matt Critchlow
- Mingsheng Zhang
- Nelly Sélem
- Nima Hejazi
- Nohemi Huanca Nunez
- Olemis Lang
- P. L. Lim
- Paula Andrea Martinez
- Peace Ossom Williamson
- Rayna M Harris
- Romualdo Zayas-Lagunas
- Sarah Stevens
- Saskia Hiltemann
- Shirley Alquicira
- Silvana Pereyra
- Tom Morrell
- Valentina Bonetti
- Veronica Ikeshoji-Orlati
- Veronica Jimenez
- butterflyskip
- dounia
- Date Added:
- 08/07/2020
The information provided on this website is designed to assist the extramural community in addressing rigor and transparency in NIH grant applications and progress reports. Scientific rigor and transparency in conducting biomedical research is key to the successful application of knowledge toward improving health outcomes.
Definition Scientific rigor is the strict application of the scientific method to ensure unbiased and well-controlled experimental design, methodology, analysis, interpretation and reporting of results.
Goals The NIH strives to exemplify and promote the highest level of scientific integrity, public accountability, and social responsibility in the conduct of science. Grant applications instructions and the criteria by which reviewers are asked to evaluate the scientific merit of the application are intended to:
• ensure that NIH is funding the best and most rigorous science,
• highlight the need for applicants to describe details that may have been previously overlooked,
• highlight the need for reviewers to consider such details in their reviews through updated review language, and
• minimize additional burden.
- Subject:
- Applied Science
- Health, Medicine and Nursing
- Material Type:
- Reading
- Author:
- NIH
- Date Added:
- 08/07/2020
Scientists should be able to provide support for the absence of a meaningful effect. Currently, researchers often incorrectly conclude an effect is absent based a nonsignificant result. A widely recommended approach within a frequentist framework is to test for equivalence. In equivalence tests, such as the two one-sided tests (TOST) procedure discussed in this article, an upper and lower equivalence bound is specified based on the smallest effect size of interest. The TOST procedure can be used to statistically reject the presence of effects large enough to be considered worthwhile. This practical primer with accompanying spreadsheet and R package enables psychologists to easily perform equivalence tests (and power analyses) by setting equivalence bounds based on standardized effect sizes and provides recommendations to prespecify equivalence bounds. Extending your statistical tool kit with equivalence tests is an easy way to improve your statistical and theoretical inferences.
- Subject:
- Psychology
- Social Science
- Material Type:
- Reading
- Provider:
- Social Psychological and Personality Science
- Author:
- Daniël Lakens
- Date Added:
- 08/07/2020
Despite the importance of patent landscape analyses in the commercialization process for life science and healthcare technologies, the quality of reporting for patent landscapes published in academic journals is inadequate. Patents in the life sciences are a critical metric of innovation and a cornerstone for the commercialization of new life-science- and healthcare-related technologies. Patent landscaping has emerged as a methodology for analyzing multiple patent documents to uncover technological trends, geographic distributions of patents, patenting trends and scope, highly cited patents and a number of other uses. Many such analyses are published in high-impact journals, potentially allowing them to gain high visibility among academic, industry and government stakeholders. Such analyses may be used to inform decision-making processes, such as prioritization of funding areas, identification of commercial competition (and therefore strategy development), or implementation of policy to encourage innovation or to ensure responsible licensing of technologies. Patent landscaping may also provide a means for answering fundamental questions regarding the benefits and drawbacks of patenting in the life sciences, a subject on which there remains considerable debate but limited empirical evidence.
- Subject:
- Applied Science
- Biology
- Engineering
- Life Science
- Material Type:
- Reading
- Provider:
- Nature Biotechnology
- Author:
- Andrew J. Carr
- David A. Brindley
- Hannah Thomas
- James A. Smith
- Zeeshaan Arshad
- Date Added:
- 08/07/2020
A focus on novel, confirmatory, and statistically significant results leads to substantial bias in the scientific literature. One type of bias, known as “p-hacking,” occurs when researchers collect or select data or statistical analyses until nonsignificant results become significant. Here, we use text-mining to demonstrate that p-hacking is widespread throughout science. We then illustrate how one can test for p-hacking when performing a meta-analysis and show that, while p-hacking is probably common, its effect seems to be weak relative to the real effect sizes being measured. This result suggests that p-hacking probably does not drastically alter scientific consensuses drawn from meta-analyses.
- Subject:
- Biology
- Life Science
- Material Type:
- Reading
- Provider:
- PLOS Biology
- Author:
- Andrew T. Kahn
- Luke Holman
- Megan L. Head
- Michael D. Jennions
- Rob Lanfear
- Date Added:
- 08/07/2020
And so, my fellow scientists: ask not what you can do for reproducibility; ask what reproducibility can do for you! Here, I present five reasons why working reproducibly pays off in the long run and is in the self-interest of every ambitious, career-oriented scientist.A complex equation on the left half of a black board, an even more complex equation on the right half. A short sentence links the two equations: “Here a miracle occurs”. Two mathematicians in deep thought. “I think you should be more explicit in this step”, says one to the other.This is exactly how it seems when you try to figure out how authors got from a large and complex data set to a dense paper with lots of busy figures. Without access to the data and the analysis code, a miracle occurred. And there should be no miracles in science.Working transparently and reproducibly has a lot to do with empathy: put yourself into the shoes of one of your collaboration partners and ask yourself, would that person be able to access my data and make sense of my analyses. Learning the tools of the trade (Box 1) will require commitment and a massive investment of your time and energy. A priori it is not clear why the benefits of working reproducibly outweigh its costs.Here are some reasons: because reproducibility is the right thing to do! Because it is the foundation of science! Because the world would be a better place if everyone worked transparently and reproducibly! You know how that reasoning sounds to me? Just like yaddah, yaddah, yaddah …It’s not that I think these reasons are wrong. It’s just that I am not much of an idealist; I don’t care how science should be. I am a realist; I try to do my best given how science actually is. And, whether you like it or not, science is all about more publications, more impact factor, more money and more career. More, more, more… so how does working reproducibly help me achieve more as a scientist.
- Subject:
- Applied Science
- Life Science
- Physical Science
- Social Science
- Material Type:
- Reading
- Author:
- Florian Markowetz
- Date Added:
- 12/08/2015
Scientific research relies on computer software, yet software is not always developed following practices that ensure its quality and sustainability. This manuscript does not aim to propose new software development best practices, but rather to provide simple recommendations that encourage the adoption of existing best practices. Software development best practices promote better quality software, and better quality software improves the reproducibility and reusability of research. These recommendations are designed around Open Source values, and provide practical suggestions that contribute to making research software and its source code more discoverable, reusable and transparent. This manuscript is aimed at developers, but also at organisations, projects, journals and funders that can increase the quality and sustainability of research software by encouraging the adoption of these recommendations.
- Subject:
- Applied Science
- Computer Science
- Information Science
- Material Type:
- Reading
- Provider:
- F1000Research
- Author:
- Alejandra Gonzalez-Beltran
- Allegra Via
- Andrew Treloar
- Bernard Pope
- Björn GrüningJonas Hagberg
- Brane Leskošek
- Bérénice Batut
- Carole Goble
- Daniel S. Katz
- Daniel Vaughan
- David Mellor
- Federico López Gómez
- Ferran Sanz
- Harry-Anton Talvik
- Horst Pichler
- Ilian Todorov
- Jon Ison
- Josep Ll. Gelpí
- Leyla Garcia
- Luis J. Oliveira
- Maarten van Gompel
- Madison Flannery
- Manuel Corpas
- Maria V. Schneider
- Martin Cook
- Mateusz Kuzak
- Michelle Barker
- Mikael Borg
- Monther Alhamdoosh
- Montserrat González Ferreiro
- Nathan S. Watson-Haigh
- Neil Chue Hong
- Nicola Mulder
- Petr Holub
- Philippa C. Griffin
- Radka Svobodová Vařeková
- Radosław Suchecki
- Rafael C. Jiménez
- Rob Hooft
- Robert Pergl
- Rowland Mosbergen
- Salvador Capella-Gutierrez
- Simon Gladman
- Sonika Tyagi
- Steve Crouchc
- Victoria Stodden
- Xiaochuan Wang
- Yasset Perez-Riverol
- Date Added:
- 08/07/2020
Research funders are requiring or strongly encouraging open and reproducible methods at increased rates, leading researchers to rely on more data management tools while institutions continue to provide services to support them. Research support staff adapt quickly to guide their stakeholders and provide resources, while administrators must find methods to determine adoption and success across the community.
In this webinar, COS Director of Policy David Mellor shares an update on funder expectations like preregistration, data sharing, and open access outputs, as well as strategies to highlight these practices in funding proposals. COS Director of Product Nici Pfeiffer also discusses OSF features that enable researchers to meet and exceed these expectations, as well as provide unique activity insights for administrators, and how COS continues to work with the funder and institution communities to facilitate transparent practices across the lifecycle.
- Subject:
- Education
- Higher Education
- Material Type:
- Primary Source
- Provider:
- Center for Open Science
- Date Added:
- 12/04/2020
Preclinical studies using animals to study the potential of a therapeutic drug or strategy are important steps before translation to clinical trials. However, evidence has shown that poor quality in the design and conduct of these studies has not only impeded clinical translation but also led to significant waste of valuable research resources. It is clear that experimental biases are related to the poor quality seen with preclinical studies. In this chapter, we will focus on hypothesis testing type of preclinical studies and explain general concepts and principles in relation to the design of in vivo experiments, provide definitions of experimental biases and how to avoid them, and discuss major sources contributing to experimental biases and how to mitigate these sources. We will also explore the differences between confirmatory and exploratory studies, and discuss available guidelines on preclinical studies and how to use them. This chapter, together with relevant information in other chapters in the handbook, provides a powerful tool to enhance scientific rigour for preclinical studies without restricting creativity.
- Subject:
- Applied Science
- Health, Medicine and Nursing
- Material Type:
- Reading
- Provider:
- Good Research Practice in Non-Clinical Pharmacology and Biomedicine
- Author:
- Andrew S. C. Rice
- Jan Vollert
- Nathalie Percie du Sert
- Wenlong Huang
- Date Added:
- 08/07/2020
Workshop overview for the Data Carpentry genomics curriculum. Data Carpentry’s aim is to teach researchers basic concepts, skills, and tools for working with data so that they can get more done in less time, and with less pain. This workshop teaches data management and analysis for genomics research including: best practices for organization of bioinformatics projects and data, use of command-line utilities, use of command-line tools to analyze sequence quality and perform variant calling, and connecting to and using cloud computing. This workshop is designed to be taught over two full days of instruction. Please note that workshop materials for working with Genomics data in R are in “alpha” development. These lessons are available for review and for informal teaching experiences, but are not yet part of The Carpentries’ official lesson offerings. Interested in teaching these materials? We have an onboarding video and accompanying slides available to prepare Instructors to teach these lessons. After watching this video, please contact team@carpentries.org so that we can record your status as an onboarded Instructor. Instructors who have completed onboarding will be given priority status for teaching at centrally-organized Data Carpentry Genomics workshops.
- Subject:
- Applied Science
- Computer Science
- Genetics
- Information Science
- Life Science
- Mathematics
- Measurement and Data
- Material Type:
- Module
- Provider:
- The Carpentries
- Author:
- Amanda Charbonneau
- Erin Alison Becker
- François Michonneau
- Jason Williams
- Maneesha Sane
- Matthew Kweskin
- Muhammad Zohaib Anwar
- Murray Cadzow
- Paula Andrea Martinez
- Taylor Reiter
- Tracy Teal
- Date Added:
- 08/07/2020
Data Carpentry’s aim is to teach researchers basic concepts, skills, and tools for working with data so that they can get more done in less time, and with less pain. Interested in teaching these materials? We have an onboarding video available to prepare Instructors to teach these lessons. After watching this video, please contact team@carpentries.org so that we can record your status as an onboarded Instructor. Instructors who have completed onboarding will be given priority status for teaching at centrally-organized Data Carpentry Geospatial workshops.
- Subject:
- Applied Science
- Geology
- Information Science
- Mathematics
- Measurement and Data
- Physical Geography
- Physical Science
- Social Science
- Material Type:
- Module
- Provider:
- The Carpentries
- Author:
- Anne Fouilloux
- Arthur Endsley
- Chris Prener
- Jeff Hollister
- Joseph Stachelek
- Leah Wasser
- Michael Sumner
- Michele Tobias
- Stace Maples
- Date Added:
- 08/07/2020
This webinar provides an overview of TOP Factor: its rationale, how it is being used, and how each of the TOP standards relate to individual scores. We also cover how to get involved with TOP Factor by inviting interested community members to suggest journals be added to the database and/or evaluate journal policies for submission.
- Subject:
- Education
- Material Type:
- Lecture
- Provider:
- Center for Open Science
- Date Added:
- 03/21/2021
Computers are now essential in all branches of science, but most researchers are never taught the equivalent of basic lab skills for research computing. As a result, data can get lost, analyses can take much longer than necessary, and researchers are limited in how effectively they can work with software and data. Computing workflows need to follow the same practices as lab projects and notebooks, with organized data, documented steps, and the project structured for reproducibility, but researchers new to computing often don't know where to start. This paper presents a set of good computing practices that every researcher can adopt, regardless of their current level of computational skill. These practices, which encompass data management, programming, collaborating with colleagues, organizing projects, tracking work, and writing manuscripts, are drawn from a wide variety of published sources from our daily lives and from our work with volunteer organizations that have delivered workshops to over 11,000 people since 2010.
- Subject:
- Biology
- Life Science
- Material Type:
- Reading
- Provider:
- PLOS Computational Biology
- Author:
- Greg Wilson
- Jennifer Bryan
- Justin Kitzes
- Karen Cranston
- Lex Nederbragt
- Tracy K. Teal
- Date Added:
- 08/07/2020
"Harry Potter and the Methods of Reproducibility -- A brief Introduction to Open Science" gives a brief overview of Open Science, particularly reproducibility, for newcomers to the topic. It introduces the concept of questionable research practices (QRPs) and Open Science solutions to these QRPs, such as preregistrations, registered reports, Open Data, Open Code, and Open Materials.
- Subject:
- Applied Science
- Life Science
- Physical Science
- Social Science
- Material Type:
- Lesson
- Author:
- Mariella Paul
- Date Added:
- 10/17/2019
This webinar outlines how to use the free Open Science Framework (OSF) as an Electronic Lab Notebook for personal work or private collaborations. Fundamental features we cover include how to record daily activity, how to store images or arbitrary data files, how to invite collaborators, how to view old versions of files, and how to connect all this usage to more complex structures that support the full work of a lab across multiple projects and experiments.
- Subject:
- Applied Science
- Computer Science
- Information Science
- Material Type:
- Lecture
- Provider:
- Center for Open Science
- Author:
- Center for Open Science
- Date Added:
- 08/07/2020
This lesson shows how to use Python and skimage to do basic image processing. With support from an NSF iUSE grant, Dr. Tessa Durham Brooks and Dr. Mark Meysenburg at Doane College, Nebraska, USA have developed a curriculum for teaching image processing in Python. This lesson is currently being piloted at different institutions. This pilot phase will be followed by a clean-up phase to incorporate suggestions and feedback from the pilots into the lessons and to make the lessons teachable by the broader community. Development for these lessons has been supported by a grant from the Sloan Foundation.
- Subject:
- Applied Science
- Computer Science
- Information Science
- Mathematics
- Measurement and Data
- Material Type:
- Module
- Provider:
- The Carpentries
- Author:
- Mark Meysenberg
- Date Added:
- 08/07/2020