Abstract Training materials. The DATUM for Health training programme covers both generic …
Abstract Training materials. The DATUM for Health training programme covers both generic and discipline-specific issues, focusing on the management of qualitative, unstructured data, and is suitable for students at any stage of their PhD. It aims to provide students with the knowledge to manage their research data at every stage in the data lifecycle, from creation to final storage or destruction. They learn how to use their data more effectively and efficiently, how to store and destroy it securely, and how to make it available to a wider audience to increase its use, value and impact.
Background All clinical research benefits from transparency and validity. Transparency and validity …
Background All clinical research benefits from transparency and validity. Transparency and validity of studies may increase by prospective registration of protocols and by publication of statistical analysis plans (SAPs) before data have been accessed to discern data-driven analyses from pre-planned analyses. Main message Like clinical trials, recommendations for SAPs for observational studies increase the transparency and validity of findings. We appraised the applicability of recently developed guidelines for the content of SAPs for clinical trials to SAPs for observational studies. Of the 32 items recommended for a SAP for a clinical trial, 30 items (94%) were identically applicable to a SAP for our observational study. Power estimations and adjustments for multiplicity are equally important in observational studies and clinical trials as both types of studies usually address multiple hypotheses. Only two clinical trial items (6%) regarding issues of randomisation and definition of adherence to the intervention did not seem applicable to observational studies. We suggest to include one new item specifically applicable to observational studies to be addressed in a SAP, describing how adjustment for possible confounders will be handled in the analyses. Conclusion With only few amendments, the guidelines for SAP of a clinical trial can be applied to a SAP for an observational study. We suggest SAPs should be equally required for observational studies and clinical trials to increase their transparency and validity.
Struggling with data at work? Wasting valuable time working in multiple spreadsheets …
Struggling with data at work? Wasting valuable time working in multiple spreadsheets to gain an overview of your business? Find it hard to gain sharp insights from piles of data on your desktop?
If you are looking to enhance your efficiency in the office and improve your performance by making sense of data faster and smarter, then this advanced data analysis course is for you.
If you have already sharpened your spreadsheet skills in Data Analysis: Take It to the MAX(), this course will help you dig deeper. You will learn advanced techniques for robust data analysis in a business environment. This course covers the main tasks required from data analysts today, including importing, summarizing, interpreting, analyzing and visualizing data. It aims to equip you with the tools that will enable you to be an independent data analyst. Most techniques will be taught in Excel with add-ons and free tools available online. We encourage you to use your own data in this course but if not available, the course team can provide.
These course materials are part of an online course of TU Delft. Do you want to experience an active exchange of information between academic staff and students? Then join the community of online learners and enroll in this MOOC. This course is part of the Data Analysis XSeries.
Python is a general purpose programming language that is useful for writing …
Python is a general purpose programming language that is useful for writing scripts to work effectively and reproducibly with data. This is an introduction to Python designed for participants with no programming experience. These lessons can be taught in one and a half days (~ 10 hours). They start with some basic information about Python syntax, the Jupyter notebook interface, and move through how to import CSV files, using the pandas package to work with data frames, how to calculate summary information from a data frame, and a brief introduction to plotting. The last lesson demonstrates how to work with databases directly from Python.
Data Carpentry lesson from Ecology curriculum to learn how to analyse and …
Data Carpentry lesson from Ecology curriculum to learn how to analyse and visualise ecological data in R. Data Carpentry’s aim is to teach researchers basic concepts, skills, and tools for working with data so that they can get more done in less time, and with less pain. The lessons below were designed for those interested in working with ecology data in R. This is an introduction to R designed for participants with no programming experience. These lessons can be taught in a day (~ 6 hours). They start with some basic information about R syntax, the RStudio interface, and move through how to import CSV files, the structure of data frames, how to deal with factors, how to add/remove rows and columns, how to calculate summary statistics from a data frame, and a brief introduction to plotting. The last lesson demonstrates how to work with databases directly from R.
Python is a general purpose programming language that is useful for writing …
Python is a general purpose programming language that is useful for writing scripts to work effectively and reproducibly with data. This is an introduction to Python designed for participants with no programming experience. These lessons can be taught in a day (~ 6 hours). They start with some basic information about Python syntax, the Jupyter notebook interface, and move through how to import CSV files, using the pandas package to work with data frames, how to calculate summary information from a data frame, and a brief introduction to plotting. The last lesson demonstrates how to work with databases directly from Python.
Understanding the types, processes, and frameworks of workflows and analyses is helpful …
Understanding the types, processes, and frameworks of workflows and analyses is helpful for researchers seeking to understand more about research, how it was created, and what it may be used for. This lesson uses a subset of data analysis types to introduce reproducibility, iterative analysis, documentation, provenance and different types of processes. Described in more detail are the benefits of documenting and establishing informal (conceptual) and formal (executable) workflows.
Data Carpentry trains researchers in the core data skills for efficient, shareable, …
Data Carpentry trains researchers in the core data skills for efficient, shareable, and reproducible research practices. We run accessible, inclusive training workshops; teach openly available, high-quality, domain-tailored lessons; and foster an active, inclusive, diverse instructor community that promotes and models reproducible research as a community norm.
Data Carpentry's aim is to teach researchers basic concepts, skills, and tools …
Data Carpentry's aim is to teach researchers basic concepts, skills, and tools for working more effectively with data. The lessons below were designed for those interested in working with Genomics data in R.
The Biology Semester-long Course was developed and piloted at the University of …
The Biology Semester-long Course was developed and piloted at the University of Florida in Fall 2015. Course materials include readings, lectures, exercises, and assignments that expand on the material presented at workshops focusing on SQL and R.
Data citation is a key practice that supports the recognition of data …
Data citation is a key practice that supports the recognition of data creation as a primary research output rather than as a mere byproduct of research. Providing reliable access to research data should be a routine practice, similar to the practice of linking researchers to bibliographic references. After completing this lesson, participants should be able to define data citation and describe its benefits; to identify the roles of various actors in supporting data citation; to recognize common metadata elements and persistent data locators and describe the process for obtaining one, and to summarize best practices for supporting data citation.
FRED® (Federal Reserve Economic Data) provides access to a wide range of …
FRED® (Federal Reserve Economic Data) provides access to a wide range of data from multiple sources. Describing the source of data used in a presentation, written report, or research project with a citation makes that work more thorough and easier to replicate. This article describes the elements of a good data citation for new data users and serves as a reference for advanced data users.
A part of the data workflow is preparing the data for analysis. …
A part of the data workflow is preparing the data for analysis. Some of this involves data cleaning, where errors in the data are identified and corrected or formatting made consistent. This step must be taken with the same care and attention to reproducibility as the analysis. OpenRefine (formerly Google Refine) is a powerful free and open source tool for working with messy data: cleaning it and transforming it from one format into another. This lesson will teach you to use OpenRefine to effectively clean and format data and automatically track any changes that you make. Many people comment that this tool saves them literally months of work trying to make these edits by hand.
Data curation primers are peer-reviewed, living documents to provide practical and concise …
Data curation primers are peer-reviewed, living documents to provide practical and concise guides on curating a specific data type or format, or addressing a particular challenge in data curation work. All the primers are developed by Data Curation Network (DCN) which is a seed funding project from the Alfred P Sloan Foundation. The target audiences of primers are data curators and/or data librarians. To date, DCN has published more than 25 primers on database, Excel, netCDF, NVivo, R, SPSS, etc.
The lessons posted on this site were designed to engage students with …
The lessons posted on this site were designed to engage students with real-world data relevant to content taught in middle school and high school science courses, and to foster an understanding of ways in which they might gather organize, analyze and interpret the data in order to draw scientifically valid inferences, interpretations and conclusions. Most of the labs use computer-based technology of spreadsheet programs or the Python programming interface. The Python lessons guide students in computational thinking to create simple programs to manipulate data. The lessons also provide students (and teachers) with instructions and guidance in the use of these technologies. Teacher and Student worksheets, as well as any supporting files, are linked to from links at the top of each lesson webpage as well as from the downloads page ("downloads" link on the scrolling menu to the left).
Learn how instructional designers use data to inform the creation of a …
Learn how instructional designers use data to inform the creation of a Learning Persona. Learning Personas help determine the needs of the training and help ...
When entering data, common goals include creating data sets that are valid, …
When entering data, common goals include creating data sets that are valid, have gone through an established process to ensure quality, are organized, and reusable. This lesson outlines best practices for creating data files. It will detail options for data entry and integration, and provide examples of processes used for data cleaning, organization and manipulation.
No restrictions on your remixing, redistributing, or making derivative works. Give credit to the author, as required.
Your remixing, redistributing, or making derivatives works comes with some restrictions, including how it is shared.
Your redistributing comes with some restrictions. Do not remix or make derivative works.
Most restrictive license type. Prohibits most uses, sharing, and any changes.
Copyrighted materials, available under Fair Use and the TEACH Act for US-based educators, or other custom arrangements. Go to the resource provider to see their individual restrictions.