All Learning Resources

  • Ocean Health Index Toolbox Training

    The single biggest motivation of the Ocean Health Index is to use science to inform marine management. And not just any science, the best available science, data, methods, and tools. OHI assessments use collaborative open software so that they are transparent and reproducible; we call this software the OHI Toolbox.

    Openness is an important part of how we work; we describe how and why in Lowndes et al. 2017, Nature Ecology & Evolution: Our path to better science in less time using open data science tools. Using the OHI Toolbox requires coding and using data science software; you can learn this in OHI’s Intro to Open Data Science training book.

    This Toolbox Training book will train you to prepare for and use the OHI Toolbox. It can be used to teach workshops as a curriculum or workshop guide or for self-paced learning.

    Chapters include:
    Introduction
    Planning and Gathering Data
    OHI Planner
    OHI Toolbox
    Toolbox Ecosystem
    Preparing data
    Calculations:  basic workflow
    Pressures and Resilience
    Calculations
    Reporting
    Communication:  OHI+ websites

     

  • Intro to Data Management

    This guide will provide general information about data management, including an overview of Data Management Plans (DMPs), file naming conventions, documentation, security, backup, publication, and preservation. We have included the CMU data life cycle to put the pieces in context in the Data 101 section.
    The CMU Libraries provides research data management resources for guidance on data management, planning, and sharing for researchers, faculty, and students.

  • MANTRA Research Data Management Training

    MANTRA is a free, online non-assessed course with guidelines to help you understand and reflect on how to manage the digital data you collect throughout your research. It has been crafted for the use of post-graduate students, early career researchers, and also information professionals. It is freely available on the web for anyone to explore on their own.

    Through a series of interactive online units you will learn about terminology, key concepts, and best practice in research data management.

    There are eight online units in this course and one set of offline (downloadable) data handling tutorials that will help you:

    Understand the nature of research data in a variety of disciplinary settings
    Create a data management plan and apply it from the start to the finish of your research project
    Name, organise, and version your data files effectively
    Gain familiarity with different kinds of data formats and know how and when to transform your data
    Document your data well for yourself and others, learn about metadata standards and cite data properly
    Know how to store and transport your data safely and securely (backup and encryption)
    Understand legal and ethical requirements for managing data about human subjects; manage intellectual property rights
    Understand the benefits of sharing, preserving and licensing data for re-use
    Improve your data handling skills in one of four software environments: R, SPSS, NVivo, or ArcGIS

  • RDMRose Learning Materials

    RDMRose was a JISC funded project to produce, and teach professional development learning materials in Research Data Management (RDM) tailored for Information professionals. The Slideshare presentations and documents include an overview of RDM, research in higher education, looking at research data, the research data lifecycle, data management plans, research data services, metadata, and data citation.  

    RDMRose developed and adapted learning materials about RDM to meet the specific needs of liaison librarians in university libraries, both for practitioners’ CPD and for embedding into the postgraduate taught curriculum. Its deliverables included open educational resources materials suitable for learning in multiple modes, including face to face and self-directed learning.

    Session topics include:
    Introductions, RDM, and the Role of LIS
    The Nature of Research and the Need for RDM
    The digital curation lifecycle
    Key Institutions and Projects in RDM
    What is data?
    Managing data
    Case Studies of Resaerch Projects
    Case Study:  Institutional Context, and Conclusions

     

  • Research Data Management and Open Data

    This was a presentation during the Julius Symposium 2017 on Open Science and in particular on Open data and/or FAIR data.  Examples are given of medical and health research data.

  • The Service Oriented Toolkit for Research Data Management

    The Service Oriented Toolkit for Research Data Management project was co-funded by the JISC Managing Research Data Programme 2011-2013 and The University of Hertfordshire. The project focused on the realisation of practical benefits for operationalising an institutional approach to good practice in RDM. The objectives of the project were to audit current best practice, develop technology demonstrators with the assistance of leading UH research groups, and then reflect these developments back into the wider internal and external research community via a toolkit of services and guidance. The overall aim was to contribute to the efficacy and quality of research data plans, and establish and cement good data management practice in line with local and national policy.

    The toolkit offers blog entries, survey results and analysis, case studies, reviews of service, test data, services, artifacts such as research project file plans, workflow recommendations, datasets, presentations, example data management plans,  and training on topics such as data encryption.  Information is also provided on best practice assessments in Astronomy, Physics, Maths, Robotics, and Atmospheric sciences based on formal and informal interviews with researchers.

  • 'Good Enough' Research Data Management: A Brief Guide for Busy People

    This brief guide presents a set of good data management practices that researchers can adopt, regardless of their data management skills and levels of expertise.

  • De bonnes pratiques en gestion des données de recherche: Un guide sommaire pour gens occupés (French version of the 'Good Enough' RDM)

    Ce petit guide présente un ensemble de bonnes pratiques que les chercheurs peuvent adopter, et ce, indépendamment de leurs compétences ou de leur niveau d’expertise. 

  • Coffee and Code: Content Platform

    UNM RDS Content Platform for the Coffee & Code Workshop Series

    This repository contains the needed code to replicate the presentation and playground environments used for the UNM Research Data Services (RDS) Coffee & Code workshop series. The materials in this repository leverage Docker as a platform for developing and deploying portable containers that support individual applications. In the case of the Coffee & Code instruction platform, the applications that are integrated into the system include:

    • Jupyter Notebooks as a presentation, demonstration, and experimentation environment (based on the datascience-notebook container with the addition of Pandoc and LaTeX)
    • A web-based RStudio environment (based on the rocker/rstudio with the addition of the R dplyr, ggplot2, ggrepel)
    • Installed tools within the Jupyter Notebook platform include:
      • Git
      • Pandoc & LaTeX
      • BASH shell
      • Python
      • R
  • Digital Preservation Workshop Module 5: Post Submission

    Module 5 of The Digital Preservation Network's Digital Preservation Workflow Curriculum examines the relationship of the content holders to the preservation service on an ongoing basis following the submission of content. Regardless of whether the preservation environment is internal to the organization, an external service providing organization, or a collaborative consortium, ongoing preservation is a shared responsibility. This module will lay out the various roles, responsibilities, and tasks, and the service framework that will support the establishment of a sustainable preservation program.

     

  • Digital Preservation Workshop Module 6: Sustainability

    Module 6 of The Digital Preservation Network's Digital Preservation Workflow Curriculum introduces some of the standards and best practices for digital preservation program assessment, tools and activities for performing assessments, and developing a business case for digital preservation. Finally, the module provides practical next steps for applying knowledge gained through the workshop. 

     

  • Digital Preservation Workshop Module 1: Programmatic Digital Preservation

    Module 1 of The Digital Preservation Network's Digital Preservation Curriculum provides an overview of the workshop contents and approach. It begins with a discussion of the goal of the workshop — providing participants with the capacity to ensure valuable content is stored in a managed environment over the long-term, and enact digital preservation programs at their organizations — and provides an opportunity for participants to discuss what this might look like within different organizational contexts. Participants will look at the factors involved in operationalizing a digital preservation program, and the pathway that content travels along as it nears a long-term storage environment. This module introduces the problem-solving and decision-making framework that will run throughout all subsequent modules.

     

  • Digital Preservation Workshop Module 2: Selection

    Module 2 of The Digital Preservation Network's Digital Preservation Workflow Curriculum introduces the concept of selection for digital preservation and how understanding an organization’s collections landscape can help with planning, selection, and prioritization of digital content for preservation. Lectures will discuss planning, and offer criteria and introduce tools to track and document collections and evaluate their readiness to prioritize submission to a digital preservation service. Participants will consider factors such as legal status, “done-ness” (when is an asset ready to be preserved?), and roles and responsibilities for decision making. They will be asked to look at how the sources of content (whether from digitization or borndigital) affect decision making, and will apply what they have learned through discussions and a case study on evaluation. 

     

  • Digital Preservation Workshop Module 4: Submission and Ingest

    Module 4 of the Digital Preservation Network's Digital Preservation Workflow Curriculum introduces the concept of transferring submission packages to preservation environments. It underscores the importance of logging transfer, upload, and verification events during ingest for the establishment (or continuation) of an audit trail that will track digital assets throughout their life in the preservation environment. Lecture will provide an overview of best practices for submission and the capture of information produced by the related events. Participants will gain experience with tools that support package transfer and will upload submission packages into a local environment and a cloud or preservation service. 

     

  • Digital Preservation Workshop Module 3: Preparing for Submission

    Module 3 of The Digital Preservation Network's Digital Preservation Workflow Workshop Curriculum focuses on preparing content for submission to a long-term storage service, whether in-house or external to the organization. It will emphasize requisite tasks such as understanding and conforming to submission requirements, local file management prior to submission, and tracking asset status. This module will explore common challenges encountered during this stage in the workflow, such as determining how and when to capture metadata, deciding what is “good enough” to submit, dealing with different content sources (e.g., born-digital vs. digitized), and work through ways of resolving these. A case study will be used to provide participants with experience creating a plan for this stage. A hands-on exercise creating a preservation package according to the specifications of a long-term storage service will expose participants to common tools and approaches for compliance with requirements. It will conclude with a discussion of how the processes reviewed during this module can be implemented in a program that will support all organizational content regardless of type, source, or owner.

     

  • Digital Preservation Workflow Curriculum Development

    The purpose of this workshop curriculum is to provide attendees with: A. An understanding of the goals, processes, and responsibilities involved in the creation of a digital preservation program B. Problem-solving and decision-making skills to enable ongoing, collaborative digital preservation throughout technological, organizational, and content changes.

    This workshop will equip participants with a set of skills and knowledge that will enable them to enact programmatic digital preservation within their organization. It is focused on equipping organizations with the capability to implement and manage a digital preservation program. The workshop modules present the requirements of a digital preservation ecosystem from two parallel viewpoints: 1) governance and program management, including the creation of a unified strategy and the need for cross-organizational coordination, balancing the competing priorities of innovation and maintenance, and 2) asset management, including the selection and submission of content to a managed preservation environment, and ongoing post-submission responsibilities.

    Module topics include:
    1.  Enabling Programmatic Digital Preservation
    2.  Selection
    3.  Preparing for Submission
    4.  Submission and Ingest
    5.  Post-Submission
    6.  Sustainability
     

  • Coffee and Code: Natural Language Processing with Python

    Github repository for this workshop: https://github.com/unmrds/cc-nlp https://github.com/unmrds/cc-nlp

    The processing and analysis of natural languages is a core requirement for extracting structured information from spoken, signed, or written language and for feeding that information into systems or processes that generate insights from, or responses to provided language data. As languages that are naturally evolved and not designed for a specific purpose natural languages pose significant challenges when developing automated systems.

    Natural Language Processing - the class of activities in which language analysis, interpretation, and generation play key roles - is used in many disciplines as is demonstrated by this random sample of recent papers using NLP to address very different research problems:

    "Unsupervised entity and relation extraction from clinical records in Italian" (1)
    "Candyflipping and Other Combinations: Identifying Drug–Drug Combinations from an Online Forum" (2)
    "How Can Linguistics Help to Structure a Multidisciplinary Neo Domain such as Exobiology?" (3)
    "Bag of meta-words: A novel method to represent document for the sentiment classification" (4)
    "Information Needs and Communication Gaps between Citizens and Local Governments Online during Natural Disasters" (5)
    "Mining the Web for New Words: Semi-Automatic Neologism Identification with the NeoCrawler" (6)
    "Distributed language representation for authorship attribution" (7)
    "Toward a computational history of universities: Evaluating text mining methods for interdisciplinarity detection from PhD dissertation abstracts" (8)
    "Ecological momentary interventions for depression and anxiety" (9)

  • Data and Software Skills Training for Librarians

    Library Carpentry is an open education volunteer network and lesson organization dedicated to teaching librarians data and software skills. The goal is to help librarians better engage with constituents and improve how they do their work. This presentation will serve as an introduction on how Library Carpentry formed in 2015, evolved as a global community of library professionals and will continue as a future sibling of the Carpentries, an umbrella organization of distinct lesson organizations, such as Data and Software Carpentry. We’ll cover existing collaborative lesson development, curricula coverage, workshop activities and the global instructor community. We’ll then talk about the future coordinating activities led by the UC system to align and prepare for a merging with Data and Software Carpentry.

  • Train the Trainer Workshop: How do I create a course in research data management?

    Presentations and excercises of a train-the-trainer Workshop on how to create a course in research data management, given at the International Digital Curation Conference 2018 in Barcelona.

  • Developing Data Management Education, Support, and Training

    These presentations were part of an invited guest lecture on data management for CISE graduates students of the CAP5108: Research Methods for Human-centered Computing course at the University of Florida (UF) on April 12, 2018. Graduate students were introduced to the DCC Checklist for a Data Management Plan, OAIS Model (cessda adaptation), ORCiD, IR, high-performance computing (HPC) storage options at UF, data lifecycle models (USGS and UNSW), data publication guides (Beckles, 2018) and reproducibility guidelines (ACM SIGMOD 2017/2018). This was the first guest lecture on data management for UF computer & information science & engineering (CISE) graduate students in CAP 5108: Research Methods for Human-centered Computing - https://www.cise.ufl.edu/class/cap5108sp18/.  A draft of a reproducibility template is provided in version 3 of the guest lecture.  

  • Coffee and Code: Write Once Use Everywhere (Pandoc)

    Pandoc at http://pandoc.org  is a document processing program that runs on multiple operating systems (Mac, Windows, Linux) and can read and write a wide variety of file formats. In many respects, Pandoc can be thought of as a universal translator for documents. This workshop focuses on a subset of input and output document types, just scratching the surface of the transformations made possible by Pandoc.

    Click 00-Overview.ipynb on the provided GitHub page or go directly to the overview, here:
    https://github.com/unmrds/cc-pandoc/blob/master/00-Overview.ipynb

  • U.S. Fish and Wildlife Service National Conservation Training Center

    The National Conservation Training Center (NCTC) of  the U.S. Fish and Wildlife (USFWS) provides a search service on top of a catalog of the courses offered at the NCTC physical location and online that are related to data skills, and data management.  The courses include instructor led,  online self study,  online instructor led courses, and webinars.  Some courses are free;  others have a fee associated with them.  Many of the courses use various GIS data sources and systems including USFWS datasets that can be found at:  https://www.fws.gov/gis/data/national/index.html  The NCTC provides a searching interface on its home page.

  • LP DAAC Data Recipes

    A collection of tutorials that describe how to use Earth science data from NASA's Land Processes Distributed Active Archive Center (LP DAAC) using easily available tools and commonly used formats for Earth science data.  These tutorials are available to assist those wishing to learn or teach how to obtain and view these data. 

  • Genomics Workshop

    Getting Started
    This lesson assumes no prior experience with the tools covered in the workshop. However, learners are expected to have some familiarity with biological concepts, including nucleotide abbreviations and the concept of genomic variation within a population. 
    Workshop Overview.  Workshop materials include a recommendation for a dataset to be used with the lesson materials.Project organization and management:
    Learn how to structure your metadata, organize and document your genomics data and bioinformatics workflow, and access data on the NCBI sequence read archive (SRA) database.Introduction to the command line:
    Learn to navigate your file system, create, copy, move, and remove files and directories, and automate repetitive tasks using scripts and wildcards.Data wrangling and processing:
    Use command-line tools to perform quality control, align reads to a reference genome, and identify and visualize between-sample variation.Introduction to cloud computing for genomics:
    Learn how to work with Amazon AWS cloud computing and how to transfer data between your local computer and cloud resources.
     

  • The Horizon 2020 Open Research Data Pilot: Introduction to the Requirements of the Open Research Data Pilot

    This course provides an introduction to the European Commission's Open Research Data Pilot in Horizon 2020. It includes two sections: Introduction to the Requirements of the Open Research Data Pilot and How to Comply with the Requirements of the Open Research Data Pilot. Each section may include videos, presentation slides, demonstrations, associated readings, and quizzes which can be found at the URL to the home page for this course.
    Learning objectives:

    • Understand what is required of participants in the Horizon 2020 Open Research Data pilot
    • Learn about the concepts of open data, metadata, licensing and repositories
    • Identify key resources and services that can help you to comply with requirements
    • Undertake short tests to check your understanding
  • Software Preservation Network 2019 Webinar Series Episode 3: Making Software Available Within Institutions and Networks

    This episode is one of 7 in the Software Preservation Network's 2019 Webinar Series on Code of Best Practices and Other Legal Tools for Software Preservation.  Each episode is recorded;  both presentation slides and webinar transcript as well as links to supplementary resources are also available.   Information about the full series can be found at:  https://www.softwarepreservationnetwork.org/events 

    In this third episode in a seven-part series about using the Code of Best Practices in Fair Use, you’ll learn about:

    • How fair use enables institutions to provide access to software for use in research, teaching, and learning settings while minimizing any negative impact on ordinary commercial sales
    • How to provide broader networked access to software maintained and shared across multiple institutions, including off-premise access under some circumstances
    • Safeguards to minimize potential risks, such as the establishment of a mechanism to register concerns by stakeholders
  • Software Preservation Network 2019 Webinar Series Episode 6: Making the Code Part of Software Preservation Culture

    This episode is one of 7 in the Software Preservation Network's 2019 Webinar Series on Code of Best Practices and Other Legal Tools for Software Preservation.  Each episode is recorded;  both presentation slides and webinar transcript as well as links to supplementary resources are also available.   Information about the full series can be found at:  https://www.softwarepreservationnetwork.org/events 

    In this sixth episode in a seven-part series about using the Code of Best Practices in Fair Use, you’ll learn:

    • The difference between a document and a shift in practice
    • How other communities have incorporated fair use into their professional practice
    • How to talk to gatekeepers and to allies in your network, to strengthen field-wide practice
  • Software Preservation Network 2019 Webinar Series Episode 5: Understanding the Anti-circumvention Rules and the Preservation Exemptions

    This episode is one of 7 in the Software Preservation Network's 2019 Webinar Series on Code of Best Practices and Other Legal Tools for Software Preservation.  Each episode is recorded;  both presentation slides and webinar transcript as well as links to supplementary resources are also available.   Information about the full series can be found at:  https://www.softwarepreservationnetwork.org/events 

    In this fifth episode in a seven-part series about using the Code of Best Practices in Fair Use, you’ll learn :

    • What the DMCA anti-circumvention provisions are and how they relate to copyright, fair use, and the Code
    • How the triennial exemption rulemaking works and how SPN obtained an exemption for software (and video game) preservation
    • How to apply the new exemption to your own activities
  • Software Preservation Network 2019 Webinar Series Episode 4: Working with Source Code and Software Licenses

    This episode is one of 7 in the Software Preservation Network's 2019 Webinar Series on Code of Best Practices and Other Legal Tools for Software Preservation.  Each episode is recorded;  both presentation slides and webinar transcript as well as links to supplementary resources are also available.   Information about the full series can be found at:  https://www.softwarepreservationnetwork.org/events 

    In this fourth episode in a seven-part series about using the Code of Best Practices in Fair Use, you’ll learn:

    • How the Code treats preservation and access to source code in your collections
    • How software licenses interact with fair use
    • What kinds of software license provisions might prevent using fair use
    • When licenses bind (and do not bind) owners of physical copies of software
    • Non-copyright concerns associated with software licenses.
  • Software Preservation Network 2019 Webinar Series Episode 7: International Implications

    This episode is one of 7 in the Software Preservation Network's 2019 Webinar Series on Code of Best Practices and Other Legal Tools for Software Preservation.  Each episode is recorded;  both presentation slides and webinar transcript as well as links to supplementary resources are also available.   Information about the full series can be found at:  https://www.softwarepreservationnetwork.org/events 

    In this seventh and final episode of a series about using the Code of Best Practices in Fair Use, you’ll learn: 

    • Why licensing isn’t a viable solution to copyright issues in preservation projects with global reach
    • How U.S. fair use law applies to initiatives that involve foreign materials
    • How preservationists in other countries can take advantage of local law (and the Code) to advance their work and the roles they can play in advocacy for better and more flexible copyright exceptions
  • Data Management Lifecycle and Software Lifecycle Management in the Context of Conducting Science

    This paper examines the potential for comparisons of digital science data curation lifecycles to software lifecycle development to provide insight into promoting sustainable science software. The goal of this paper is to start a dialog examining the commonalities, connections, and potential complementarities between the data lifecycle and the software lifecycle in support of sustainable software. We argue, based on this initial survey, delving more deeply into the connections between data lifecycle approaches and software development lifecycles will enhance both in support of science.

  • Seismic Data Quality Assurance Using IRIS MUSTANG Metrics

    Seismic data quality assurance involves reviewing data in order to identify and resolve problems that limit the use of the data – a time-consuming task for large data volumes! Additionally, no two analysts review seismic data in quite the same way. Recognizing this, IRIS developed the MUSTANG automated seismic data quality metrics system to provide data quality measurements for all data archived at IRIS Data Services. Knowing how to leverage MUSTANG metrics can help users quickly discriminate between usable and problematic data and it is flexible enough for each user to adapt it to their own working style.
    This tutorial presents strategies for using MUSTANG metrics to optimize your own data quality review. Many of the examples in this tutorial illustrate approaches used by the IRIS Data Services Quality Assurance (QA) staff.
     

  • Project Open Data

    This description of Project Open Data Data includes information about the code, tools, and case studies that have been gathered to help agencies adopt the Open Data Policy and unlock the potential of government data. Project Open Data will evolve over time as a community resource to facilitate broader adoption of open data practices in government. Anyone – government employees, contractors, developers, the general public – can view and contribute. Learn more about Project Open Data Governance and dive right in and help to build a better world through the power of open data.is a valuable national resource and a strategic asset to the U.S. Government, its partners, and the public.

    Sections include information about: definitions, implementation guidance, tools, resources, case ctudies, suggestions for community engagement, and open data stakeholder contacts.  The page also provides links to the associated code and related discussions on a github repository at:  https://github.com/project-open-data.

  • Do-It-Yourself Research Data Management Training Kit for Librarians

    Online training materials on topics designed for small groups of librarians who wish to gain confidence and understanding of research data management.  The DIY Training Kit is designed to contain everything needed to complete a similar training course on your own (in small groups) and is based on open educational materials. The materials have been enhanced with Data Curation Profiles and reflective questions based on the experience of academic librarians who have taken the course.

    The training kit includes:  
     - Promotional slides for the RDM Training Kit
    - Training schedule
    - Research Data MANTRA online course by EDINA and Data Library, University of Edinburgh
    - Reflective writing questions
    - Selected group exercises (with answers) from UK Data Archive, University of Essex - Managing and sharing data: Training resources. September, 2011 (PDF).
    - Podcasts for short talks by the original Edinburgh speakers if running course without ‘live’ speakers (Windows or Quicktime versions).
    - Presentation files (pptx) if learners decide to take turns presenting each topic.
    - Evaluation forms
    - Independent study assignment: Interview with a researcher, based on Data Curation Profile, from D2C2, Purdue University Libraries and Boston University Libraries.

  • Software Preservation Network 2019 Webinar Series Episode 1: The Code of Best Practices for Fair Use in Software Preservation, Why and How?

    This episode is one of 7 in the Software Preservation Network's 2019 Webinar Series on Code of Best Practices and Other Legal Tools for Software Preservation.  Each episode is recorded;  both presentation slides and webinar transcript as well as links to supplementary resources are also available.   Information about the full series can be found at:  https://www.softwarepreservationnetwork.org/events 
    Episode 1: The Code of Best Practices for Fair Use in Software Preservation, Why and How
    In this introduction to the webinar series, you’ll learn: 

    • What the Code is 
    • What the copyright doctrine of fair use is 
    • How it addresses problems such as making preservation copies, patron access onsite and remotely, sharing software with other institutions, and providing access to source code 
    • Why best practices codes are a robust, reliable guide to practice
  • Software Preservation Network 2019 Webinar Series Episode 2: Beginning the Preservation Workflow

    This episode is one of 7 in the Software Preservation Network's 2019 Webinar Series on Code of Best Practices and Other Legal Tools for Software Preservation.  Each episode is recorded;  both presentation slides and webinar transcript as well as links to supplementary resources are also available.   Information about the full series can be found at:  https://www.softwarepreservationnetwork.org/events 
    In the second episode in the series, speakers cover how to read the Situations in the Code (parsing descriptions, Principles, and Limitations), and explore how fair use applies to the foundational steps in a preservation workflow including stabilizing, describing, evaluating, and documenting software.

  • Singularity User Guide

    Singularity is a container solution created by necessity for scientific and application driven workloads.  .
    Over the past decade and a half, virtualization has gone from an engineering toy to a global infrastructure necessity and the evolution of enabling technologies has flourished. Most recently, we have seen the introduction of the latest spin on virtualization… “containers”. 
    Many scientists, especially those involved with the high performance computation (HPC) community, could benefit greatly by using container technology, but they need a feature set that differs somewhat from that available with current container technology. This necessity drives the creation of Singularity and articulated its four primary functions:

    • Mobility of compute
    • Reproducibility
    • User freedom
    • Support on existing traditional HPC 


    This user guide introduces Singularity, a free, cross-platform and open-source computer program that performs operating-system-level virtualization also known as containerization.

  • The Oxford Common File Layout

    The Oxford Common File Layout (OCFL) specification describes an application-independent approach to the storage of digital information in a structured, transparent, and predictable manner. It is designed to promote long-term object management best practices within digital repositories.  This presentation was given at under the topic of Preservation Tools, Techniques and Policies for the Research Data Alliance Preserving Scientific Annotation Working Group on April 4, 2017. 
     

  • Workshop: Research Data Management in a Nutshell

    The workshop Research Data in a Nutshell was part of the Doctoral Day of the Albertus Magnus Graduate Center (AMGC) at the University of Cologne on January 18 2018.

    The workshop was intended as a brief, interactive introduction into RDM for beginning doctoral students.

  • Data Warehouse Tutorial For Beginners

    This Data Warehouse Tutorial For Beginners will give you an introduction to data warehousing and business intelligence. You will be able to understand basic data warehouse concepts with examples. The following topics have been covered in this tutorial:
    1. What Is The Need For BI?
    2. What Is Data Warehousing?
    3. Key Terminologies Related To DWH Architecture: a. OLTP Vs OLAP b. ETL c. Data Mart d. Metadata
    4. DWH Architecture
    5. Demo: Creating A DWH