Skip to content

steffibu/nfdi4biodiversity-sle

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

76 Commits
 
 
 
 
 
 

Repository files navigation

NFDI4Biodiversity Self-Study Unit - Research Data Management for Biodiversity Data

{{|>}}

Creators:

Marlen Fischer, Juliane Röder, Johannes Signer, Daniel Tschink, Tanja Weibulat, Ortrun Brand.

For more details about NFDI4Biodiversity and research data management for biodiversity data, visit www.nfdi4biodiversity.org.


How to use this course

{{|>}}


How to use this course

This is a LiaScript course. To follow the course in its intended format, please follow this link!

Only if you use the link, you will be able use all features of this course:

    {{1}}

After each chapter, you can test your knowledge with questions. These questions only work in the LiaScript version, not in simple Markdown.

LiaScript question

    {{2}}

You can listen to the text by clicking on the little PLAY button on top of each page. Please note that this feature is used as commentary, this is not a tool to increase accessibility.

location play button

    {{3}}

You can automatically translate the course with one click. However, please be aware that any automatic translation may contain mistakes and mistranslations of terms and concepts.

where to click for translation

    {{4}}

Please note that if you want to listen to a translation of the text, you should change the narrator voice in the html header in the raw version of this file.

picture of header with narrator English Female

For more details on the Markdown dialect LiaScript, see the LiaScript documentation.


Description

{{|>}}


The NFDI4Biodiversity Self-Study Unit provides in-depth knowledge for both students and researchers specializing in biodiversity and environmental sciences. This course is based on the online course Selbstlerneinheit Forschungsdatenmanagement - eine Online-Einführung (HeFDI Data Learning Materials) of the Hessian Research Data Infrastructures (HeFDI). Developed in collaboration between several partners within NFDI4Biodiversity, this course offers essential domain-specific knowledge in research data management.

  • Prerequisites: No previous knowledge is required for this module. The chapters build on each other thematically, but can also be worked through individually. If information from other chapters is needed, they are linked.

  • Target audience: Master and PhD students and researchers in the field of biology and environmental sciences, especially those working with biodiversity data, who are looking for a first introduction to research data management.

  • Learning objectives: After completing this unit you will be able to understand the content and purpose of research data management and apply it in the field of biology. Learning objectives are given at the beginning of each chapter.

This course will be updated regularly, responding to feedback from the community for continuous improvement. For inquiries or additional information, please feel free to reach out to us by opening an issue or via our contact form: NFDI4Biodiversity Contact.


Published versions

{{|>}}



1. Introduction

{{|>}}


Biodiversity data encompass a vast and interdisciplinary collection of information, spanning all living species and the entire spectrum of life that has ever existed. This remarkable diversity within biodiversity data makes it a highly heterogeneous field. These datasets range from laboratory-generated data, such as results from chemical assays, biological tests, and DNA sequencing, to taxonomic records, as well as spatial and temporal data collected during field experiments, and comprehensive insights into entire ecosystems123 . Furthermore, biodiversity data extend their reach into time, including data collected from the distant past, like fossils, and continue into the present while even venturing into predictive models of the future. The complexity of biodiversity data doesn't stop there; these datasets manifest in various formats, including images, photographs, sounds, sensor data, and more. The confluence of this diversity, alongside the increasing automation processes and digitalization of data acquisition, has resulted in a deluge of heterogeneous data, necessitating progressively complex organisation, coordination, and analysis methods.

To address these challenges, Germany has established the National Research Data Infrastructure (NFDI), a pivotal initiative. Within this framework, the consortium "NFDI4Biodiversity" plays a central role. This collaborative effort brings together experts and resources to develop innovative strategies and tools for managing, coordinating, and extracting meaningful insights from the wealth of biodiversity data available today. These efforts are essential in supporting research and conservation endeavours in this field.

In the ever-evolving realms of ecology and environmental science, data play a pivotal role for generating new knowledge. From collecting data in the field to conducting experiments in the lab, researchers generate vast amounts of data. This data holds the key to understanding our natural world, from the intricate ecosystems that surround us to the impact of human activities on the environment. However, the value of this data is fully realised only when it is properly managed, organised, and shared, according to the FAIR4 and CARE-principles5. This self-learning unit is designed to empower you with the knowledge and skills you need to learn independently and effectively. Throughout this document, we will explore topics such as data collection best practices, data organisation and documentation, data storage and security, ethical considerations, and the importance of data sharing and collaboration in advancing ecological and environmental research. Whether you are a seasoned researcher looking to enhance your data management practices or a student taking your first steps into this exciting domain, you'll find valuable insights and practical guidance here, to promote data diversity for biodiversity.

As you progress through this unit, you'll become proficient in managing data effectively and contribute to the advancement of ecological and environmental science by ensuring that your research data is findable, accessible, interoperable, and reusable. These skills are not only valuable for your work but also for the broader scientific community and society as a whole. Ecology and environmental science rely heavily on data to understand the natural world and address environmental challenges - data diversity for biodiversity.


2. Basics in Research Data Management

{{|>}}


After completing this chapter, you will be able to...

{{1}}

...explain the relevance of research data management for biodiversity data

{{2}}

...name the four FAIR principles

{{3}}

...contextualise how biodiversity research may raise ethical concerns related to indigenous data


2.1. Relevance of research data management: FAIR and CARE

{{|>}}


Collecting biodiversity data takes a lot of energy and resources. This makes biodiversity data not only valuable for the scientist who collected it, but also to support further research and to inform policy on conservation, natural resources, land use, agriculture and more12. It is important to create a maximum output of this valuable data to protect nature, wildlife and endangered species, to prevent overfishing and extinctions as a way to counteract the climate crisis and biodiversity crisis3. Thus, biodiversity data need to be carefully handled, preserved, and shared, by making them Findable, Accessible, Interoperable and Reusable (FAIR)4, and considering CARE-Principles (Collective Benefit, Authority to Control, Responsibility, and Ethics)5.

The pictures shows the current global extinction risk in different species groups. The horizontal axis shows how endangered the species of each category are. The numerical scale (from 0 to 100) indicates the percentage of species in each category, while the colour indicates how threatened this percentage is.

To put good research practice6 and the FAIR principles into practice effective research data management (RDM) is needed. This requires a comprehensive approach that encompasses the planning, collection, storage, analysis, and collaborative sharing of diverse biodiversity data through a well-structured and coordinated research data management strategy7. For research in the biodiversity field, research data management is even more important as non-repeatability and the need to rely on old data, e.g. for trend analysis, is of utmost relevance.

!?This video by GFBio (2020) explains the challenges and solutions of dealing with heterogeneous data.


2.1.1. FAIR principles

{{|>}}


The FAIR principles play a pivotal role in Research Data Management, emphasising the broad and diverse utilisation of research data while striving to minimise redundant research efforts. Consequently, research data should remain accessible without undue restrictions for an extended duration. This applies to the use of research data collected by the researchers themselves, but also to research data that researchers make available to each other.

The FAIR principles were developed in 2014 in a workshop at the Lorentz Center in the Netherlands and published for the first time in March 2016 in the journal Scientific Data1.

Data has to adhere to certain properties to be Findable, Accessible, Interoperable and Reusable, for example machine-readability, being deposited in a trusted repository etc. The FAIR principles have been significantly refined recently2; in addition, tools have been developed to evaluate your data according to FAIR3. Ultimately, the entire research data management is geared towards fulfilling these concrete verge requirements of the FAIR principles.


2.1.2 CARE principles

{{|>}}


The CARE principles (Collective Benefit, Authority to Control, Responsibility, and Ethics), an addition to the FAIR principles, were introduced by the Global Indigenous Data Alliance (GIDA) in 2019 to address ethical concerns related to indigenous data1. They emphasise the rights of Indigenous Peoples in relation to their data, including information about their language, customs, and territories. It is true that the application of the CARE principles in Germany is still in its infancy. The Government has established the so called 3-way-strategy to provide for access, transparency and cooperation with regard to artefacts of indigenous provenance2. At the same time, they are certainly relevant for biodiversity data. For example, specimens from natural history museums have sometimes been acquired in such a way that they might contradict the CARE principles. The process of coming to terms with these relationships is just beginning and will certainly become more relevant to the biodiversity community in the future. These CARE Principles collectively aim to promote the responsible and ethical use of indigenous data, safeguarding the rights and interests of Indigenous Peoples.


2.2. The data life cycle

{{|>}}


The image is a circular, color-coded diagram resembling a pie chart with six segments, each representing a different step in the research data management cycle, as indicated by the acronym "RDMkit" in the bottom right corner. Starting from the top right and moving clockwise, each segment has an icon and a label:"Plan" - Illustrated with a pencil icon and colored in orange."Collect" - Depicted with a flask icon and colored in yellow."Process" - Represented by three circles connected by lines, suggesting a process flow, and colored in green "Analyse" - Shown with a bar chart icon and colored in darker green. "Preserve" - Indicated with a database or stack icon and colored in blue. "Share" - Represented by a share icon, two connected dots with a line, and colored in purple. "Reuse" - Illustrated with a circular arrow, implying recycling or reuse, and colored in red. The segments create a cycle around a central white empty circle, suggesting that each step leads to the next in a continuous loop. This diagram likely serves as a visual guide for the stages of managing research data, from planning to reusing data.

The data life cycle (DLC) is a conceptual tool which helps to understand the different steps that data follow from data generation to knowledge creation and specifically focuses on the role of data. It strongly suggests that a professional approach to research data management involves more than just collection and analysis, but begins with detailed planning. In order to cope with the heterogeneity of biological datasets, data users, and data producers, a large variety of DLC´s are commonly used within the community. DLC´s often include between 5 and 10 steps, depending on the institution and respective research mission. However, the message and content are quite similar across DLC´s and disciplines. All DLCs share the commonality of adhering to the FAIR principles at each individual step. The differentiation between research domains relates to specific tools and services that are used within the respective community. Figure 2 displays the exemplary DLC RDM-Kit developed by a diverse community of experts and educators spanning various disciplines within the life sciences. The RDM-Kit is a suitable DLC for introducing RDM to a wider audience. The seven phases of a data cycle according to the RDM-Kit are planning, collecting, processing, analysing, preserving, sharing and reusing; an unlimited number of subsequent cycles can follow. Each step of the DLC contains further explanations and specifications in order to cope with the challenges in biodiversity research. Ideally, the FAIR principles should be considered throughout the life cycle of research data. However, the realisation might vary depending on the type of conducted research, the type of collected data and the type of researchers. Here, we focus on the steps that are specifically important for researchers, which work with biodiversity and environmental data.

As a researcher, it is worthwhile to always consider all phases when making decisions and to find out at an early stage which tools and options are available to optimise your practice in dealing with research data. It starts with the planning, then the data is collected, processed and analysed. The life cycle continues with the preservation and sharing of data, until it re-enters the process by data reuse.

!?This video by the Ghent University Data Stewards (2020) explains the research data lifecycle.


2.3 Questions

{{|>}}


Which of the following statements is a goal of research data management? Choose the correct answer.

  • [( )] Reducing the amount of research data
  • [( )] Preventing active re-use of one's own research data by other researchers
  • [( )] Making any research data available to all without technical restrictions
  • [(X)] Minimising data loss

Correct! Minimising data loss is a goal of data security, which aims at the general protection of data and belongs to the area of data organisation. Data security doesn’t only include measures on the digital level (e.g. using anti-virus programmes or regular backups), but also on the analogue level (e.g. storing data media with very important data in fireproof boxes).


What is not part of this unit? Choose the correct answer.

  • [( )] Principles to make data findable, accessible, interoperable and reusable
  • [( )] Contents of a data management plan
  • [(X)] Legally binding information on your research data
  • [( )] Metadata and the use of metadata standards

Correct! No trained lawyers or research data officers were involved in the creation of this self-study unit, which is why legally binding information on research data cannot be given. It is merely information on the topic of “research data and law”, but it reflects common practice in research data management and is also taught by lawyers. If you need legally binding information, you should contact the legal department or/and the data protection officer of your institution.


Select the correct abbreviations.

  • [[DLC] (ELN) (GBIF) (ROR) (NFDI) [FAIR] (EOSC)]
  • [ (X) ( ) ( ) ( ) ( ) ( ) ( ) ] Data Life Cycle
  • [ ( ) (X) ( ) ( ) ( ) ( ) ( ) ] Electronic Laboratory Notebook
  • [ ( ) ( ) (X) ( ) ( ) ( ) ( ) ] Global Biodiversity Information Facility
  • [ ( ) ( ) ( ) (X) ( ) ( ) ( ) ] Research Organisation Registry
  • [ ( ) ( ) ( ) ( ) (X) ( ) ( ) ] Nationale Forschungsdateninfrastruktur
  • [ ( ) ( ) ( ) ( ) ( ) (X) ( ) ] Findable, Accessible, Interoperable, Re-usable
  • [ ( ) ( ) ( ) ( ) ( ) ( ) (X) ] European Open Science Cloud

3. Planning Research Projects

3.1. Introduction and learning objectives

{{|>}}

Let's have a look at the first step in the data life cycle: planning is essential for good results. This requires careful consideration, consultation, and research. A data management plan helps to organise and define the key issues.

After completing this chapter, you will be able to...

{{1}}

...define the key aspects of your planning process

{{2}}

...decide how to structure your data management plan

{{3}}

...find a suitable tool to help you create a data management plan


3.2. Relevant Aspects for Planning

{{|>}}


When it comes to research data management, many research funders already require a so-called data management plan (DMP) when the application is submitted. However, even without explicit requirements, it is beneficial to document in advance exactly how the data are to be handled. This creates commitment and uniformity (especially in projects with several participants or partners) and can serve as a reference, a checklist, and as documentation.

In general, the following aspects could be relevant for planning:

  1. Determine study design
  2. Assemble project team and clarify roles
  3. Set up a schedule
  4. Plan data management (formats, storage locations, file naming, collaborative platforms, etc.)
  5. Review existing literature and data
  6. Reuse of existing data, if applicable and accessible
  7. Clarify authorship and data ownership
  8. Coordinate access possibilities and conditions

Creating a DMP for your research project is the initial step of the data life cycle and the starting point of your research data management12.


3.3. Data management plan (DMP)

{{|>}}


In your data management plan (DMP), you define the strategy on how to manage your data and documentation during the whole project.

A DMP is a comprehensive document that outlines the activities to be performed and their implementation during all phases of the data life cycle to ensure the data remain accessible, usable, and comprehensible (understandable). Of course, this also includes basic information such as the project name, third-party funders, project partners, among others.

The DMP thus records how the research data are handled during and after the research project. To create a substantial DMP, it is necessary to systematically address topics such as data management, metadata (structured data that describe the data), data retention and data analysis. It makes sense to create the DMP before starting the data collection because it forms the basis for decisions concerning data storage, backup, and processing, among other aspects. Nevertheless, a DMP is not a static document, but rather a dynamic document that is constantly updated and provides assistance throughout your whole project.

The DMP contains general information about the project and the way of documentation, descriptions of the used and collected data, used metadata, and ontologies (Figure 3). Furthermore, it includes information about data storage, data security, and preservation during and after the project, sharing of the data, and publication. The responsibilities, costs, and resources needed are listed as well as information about ethical and legal issues1. This is also important to meet the Nagoya protocol needs, e.g. permits for collecting samples.

!?This video by GFBio (2020) explains why you should create a DMP at the beginning of your project.


3.3.1. Content of a data management plan

{{|>}}


The DMP contains information about the data, the data format, how the data are handled and how the data are to be interpreted. Recently, a DFG checklist has provided further guidance on these aspects1. To decide which aspects should be included, the following questions can be helpful:

  • What data are created?
  • How and when do you collect the data? And, which legal regulations are relevant for your research project? For example, are permits required for collecting samples, is the Nagoya protocol relevant and permits to import or export samples are needed?
  • How do you process the data?
  • What format do you have for data storage and why?
  • Do you use file naming standards?
  • How do you ensure the quality of the data? This refers to the collection as well as to the analysis and processing
  • Do you use existing data? If so, where does it come from? How will existing and newly collected data be combined and what is the relationship between them?
  • Who is responsible for data management?
  • Are there any obligations, e.g. by third-party funding bodies or other institutions, regarding the sharing of the data created? (legal requirements also play a role here)
  • How will the research data be shared? From when on, and for how long will it be available?
  • How will the research data and objects be archived and is it mandatory (e.g. DFG requires storing data)? (When you plan to add your data and objects to a collection, curators should be already contacted during the planning phase, also keep the storage regulations for physical objects in mind).
  • What costs arise for the RDM (these include e.g. personnel costs, hardware and software costs, possible costs for a repository) and how are these costs covered?
  • What ethical and data protection issues need to be taken into account?
  • Is it necessary for political, commercial, or patent reasons to make the research data accessible only after a certain blocking period (Embargo)?
  • How will the data be used in the future?
  • In what way should the data be cited? Can the data be made unambiguous and permanently traceable using a persistent identifier?

There are different DMP templates available which you can use. Before writing your DMP check with your funding agency or institution whether a specific DMP template is required. If not, here are two examples:

  1. GFBio services provide expert guidance and assistance on standards and suitable archives. You can use the GFBio DMP Tool online, or refer to the GFBio Model Data Management Plan (DMP) for basic structuring.
  2. The HU Berlin offers standardised DMP templates for different funders (Horizon 2020, BMBF, DFG, Volkswagenstiftung).

3.3.2. Benefits of a DMP

{{|>}}


A DMP will make your project more efficient, simplify your own reuse of the data, and help you make your data findable, accessible, interoperable and reusable, also called FAIR. It will help you to plan the costs of equipment and resources, identify problems and find solutions at an early stage. It will also clarify responsibilities and roles in the project team and facilitate data preservation, sharing and reuse. It is good research practice to write a data management plan and thus take care of your research data. Currently, it is frequently a requirement by research organisations and funders12.

Overall, the DMP will save you time later on in the publishing process and prevents data loss - the main winner of creating a DMP is you! If you consider in advance how the data should be processed, stored, and filed, you probably won't have to reorganise your data. If, for example, it is already clear during data collection how the data is to be archived later, it can be formatted and stored right away in such a way that the transfer to the later archive is as simple as possible. Some domain-specific data centers strongly encourage you to contact them early on in a project to determine concrete requirements for taking on your data. There is a strong benefit of going towards FAIRness with a DMP: A DMP will allow you to gather in one document how to deal with the heterogeneous biodiversity data from collections, research, etc. and will serve as a model for understanding this information and disseminating it.


3.4. Tools for data management

{{|>}}

There is now a whole range of tools to support researchers during each step of the data lifecycle. Data Management Planning tools, for example, facilitate the creation of data management plans either using text modules (RDMO Organiser on GitHub, RDMO University Marburg) or one is guided through a catalogue of questions. There are usually different templates for different funders and purposes that often provide web-based DMP tools that help you to draft your own suitable DMP1. GFBio, for instance, offers a Data Management Plan Tool (GFBio DMP-Tool, DMP Tool, DMP online) that is tailored to create customised DMPs according to DFG guidelines based on the requirements of biodiversity, ecological, and environmental projects. Further, GFBio offers professional support. Thus, your DMP will be cross-checked by experts in the field.

Figure 3a. shows the user interface of the GFBio DMP Tool with the entry fields "Name", "Category", "Reproducible", "Additional information" and "Project Type" Figure 3b. shows the final DMP created with the GFBio DMP tool. Besides showing the information the user just provided in the "general project information" like explained in fgire 3a., it also delivers information about the data collection, like the type of data or the number of datasets e.g.

Furthermore, there are data management tools which support data management during the active project phase. Data management tools are software applications or platforms designed to efficiently collect, store, organise, process, and analyse data, but they often have to be set up and hosted by your institution or project partners. Here's a brief explanation of key aspects of data management tools:

Data Collection: Data management tools often provide capabilities to collect data from various sources, including databases, sensors, forms, and external APIs. They streamline the process of gathering information and ensure data is standardised for consistency.

Data Storage: These tools often provide storage solutions, such as databases, data warehouses, or cloud-based storage. Data is typically organised into structured formats for easy retrieval.

Data Organisation: Data management tools allow users to categorise and label data, assign metadata (information about the data), and create data dictionaries to maintain order and accessibility.

Data Processing: Many tools include data transformation and processing features for tasks like cleaning, aggregating, and merging data. This ensures data is accurate and ready for analysis.

Data Analysis: Advanced data management tools often integrate with analytics and visualisation tools, enabling users to derive insights, generate reports, and make data-driven decisions.

Data Security: Security features like encryption, access control, and authentication are critical to protect sensitive data from unauthorised access or breaches.

Data Backup and Recovery: Reliable tools include data backup and recovery mechanisms to prevent data loss due to accidents or system failures.

Data Compliance: Data management tools often assist in compliance with data regulations by providing features for data anonymization, audit trails, and consent management.

Data Collaboration: Some tools offer collaboration features, allowing teams to work together on data-related tasks, share insights, and maintain version control.

Data Governance: Data governance features help organisations establish policies, standards, and procedures for managing data, ensuring data quality, and maintaining data integrity.

Integration: Data management tools often integrate with other software and systems, enhancing their capabilities and enabling data flow between different parts of an organisation.

GFBio recommends two major systems with tools for biodiversity data management: Diversity Workbench (DWB) and BEXIS2 and offers training for both. DWB facilitates efficient data entry, storage, and management for biodiversity research, encompassing various data types. It enforces data standardisation to maintain consistency. BEXIS2 specialises in managing large-scale biodiversity data, integrating diverse sources for easier analysis. It emphasises metadata and documentation to provide context for datasets. Both systems can further help you during data collection. In addition, various tools exist which help you collecting and managing your data in the lLab as well as in the field such as Smatrix, RightField and Laboratory Information Management Systems (LIMS) or Electronic Lab Notebooks2345. Smatrix and RightField aid in data collection and structuring, while LIMS systems provide comprehensive solutions for data organisation and management within laboratory environments. ELNs, on the other hand, offer digital platforms for recording, storing, and sharing experimental findings, streamlining the research process. Together, these tools play a pivotal role in advancing data management capabilities.


3.5 Questions

{{|>}}


Which of these aspects are part of a data management plan?

  • [[X]] Naming those responsible for data management
  • [[ ]] Vote of an ethics committee
  • [[X]] Information on data archiving
  • [[X]] Costs of data management and data storage
  • [[X]] Information on storage and backup during the project
  • [[ ]] Description of the benefits of the data management plan
  • [[X]] Information on data publication

Right! There is of course much more information that should also be mentioned in a DMP, such as information on the data formats produced, data types and the estimated amount of data, in projects with several participants also information on how the data is shared within the project, or information on legal features, if existing.


Which statement is true?

  • [(X)] A data management plan saves time and helps me to minimize data loss.
  • [( )] I only need to create a DMP if I expect large amounts of data.
  • [( )] The DMP is primarily a control instrument for funding institutions.

Right! Also, a DMP created before or with the start of the project will help you organise the research data management for you and your team. Remember, however, that it is a living document that might need adjustments at regular intervals. A well-maintained DMP can also facilitate the onboarding of new project staff with regard to research data management.



4. Collecting Data

4.1. Introduction and learning objectives

{{|>}}


Data collection is an important step of the data life cycle as it is the basis of knowledge creation. Data is collected regarding specific methods, settings and instruments. To generate comprehensible and reusable data, it is necessary to document data acquisition using metadata.

It is also possible to reuse already collected data and to integrate it into your own project. However, be aware that data integration can be very tedious. This can be data from your previous research project or data from repositories12.

Since in biology, taxonomic designations can vary depending on the source and concept used, harmonisation is necessary in some cases.

After completing this chapter, you will be able to...

{{1}}

...select appropriate tools to assist you in the data collection process

{{2}}

...recognise metadata and the benefits of metadata

{{3}}

...name important categories of metadata

{{4}}

...name selected metadata standards

{{5}}

...create your own metadata

{{6}}

...select a suitable tool for handling metadata

{{7}}

...describe your research data via metadata so that your research data can be used in the future and by machine-reading systems

{{8}}

...name the important elements of taxonomic information that are necessary to merge different species data sets


4.2. Data search

{{|>}}


Often, already existing data can help answer your research question. Searching for such data is easier with well-maintained and annotated (= enriched with metadata) data. This applies for both data providers and subsequent users. Making research data available beyond a research project allows other researchers and research groups to retrieve the data when it has become relevant for research again.

If you want to reuse already existing data, you can use different search tools, including:

It is always important to check the copyrights and licences before reusing data and seek permission when necessary.


4.3. Data collection

{{|>}}


Overall, the data collection should cover the following aspects:

  • Standardised sampling protocols

  • Carrying out the experiments, observations, measurements, simulations, etc.

    • at the same time: metadata collection and creation

    • at the same time: documentation of the data collection

  • Generation of digital raw data (e.g. by digitising or transcribing)

  • Storage of the data in a uniform format

  • Systematic and consistent folder structure

  • Backup and management of data

It is important to create an experimental design in advance with a collection strategy on which data will be collected (What? Where? When? Who? How?) including calibration, controls, repetitions, randomisation, etc. You need to determine suitable metadata standards and methodologies, and to consider how and where to store the data and how to capture provenance (e.g. of samples and instruments). Be always aware of the data quality. If you work with sensitive, confidential, or human-related data, consider aspects such as permissions or consent, data protection, and data security.

In addition, errors occurring during data collection affect the downstream research process and in the worst-case lead to incorrect results without notice. This makes it all the more important to be careful during the data collection phase.


4.3.1. Electronic lab notebooks (ELNs)

{{|>}}


An Electronic Lab Notebook (ELN) is a software tool, which replicates a page in a paper lab notebook. You can write your observations, notes, protocols and more using your computer. An ELN has several advantages: Your notes are conserved and your colleagues or followers can repeat the same experiments, data can easily be linked with the experiment, are more traceable, more secure as well as reusable. Hence, good data management and implementation of the FAIR principles is facilitated. Some ELNs can also benefit your teamwork, for example, handle equipment use and booking, or manage inventories of supplies, chemicals or samples. There are different ELNs, with advantages and disadvantages. Some are generic, some are specialised for certain disciplines or provide specific functions e.g. drawing of chemical structures1. However, there is no ELN specifically for biodiversity and environmental data. The establishment of an ELN depends to a large extent on the individual possibilities in the working group/institution and, above all, on its own data types. The ELN Finder can help you decide which ELN fits your requirements2. A free and very flexible solution that has proven itself in the natural sciences is the cloud-based ELN eLabFTW.

!?This video published by NCSU BIT (2017) shows you how to use electronic lab notebooks.

eLabFTW is a free, secure and open source ELN for research teams. You can document your experiments and manage your lab. There is a big github community, where you can find a lot of information.

!?This video by The NOMAD Laboratory (2022) shows you how to set up eLabFTW.


4.3.2. Metadata and standards

{{|>}}


A crucial step in RDM is to describe your data in a way that you and other researchers can understand, find, reuse and properly cite your data. Making your data FAIR means that you need to describe the data in a structured and detailed way. A structured description of the data is called metadata. It contains information about the subject and the creator of the data, such as why, how, where and when the data was generated and what the content of the data is. To avoid errors and to create compatible and interoperable metadata, specific metadata standards can be used12.


4.3.2.1. Metadata

{{|>}}


Metadata describes your data in detail. It should contain the technical details (names of datasets and data files, the file formats) and information about the data structure and processing (versioning, hard- and software, methods, tools, instruments, etc.). The descriptive metadata should further contain information about scientists/collectors and contact persons involved as well as the scientific background of the data: the hypothesis, standards, calibrations, spatial and temporal information, units, formats, codes and abbreviations12. Furthermore, each data set is accompanied by administrative information such as stakeholders, funding, access rights, etc.

Metadata ensures that research data can continue to be used today and in the future, even if the people involved in the experiments at the time are now busy with other research priorities or no longer available to provide more detailed information about the earlier experiments. Without metadata, such research data is often worthless, as it is incoherent and incomprehensible.

In order to assign metadata correctly and to be able to continue to use your data correctly and in an orderly manner, it is best to document metadata right from the start of the research project. However, metadata must be created at the latest when your research data is to be deposited in a repository, published, or archived for the long term.

Often, however, it is no longer possible to create certain metadata retrospectively. This can be the case, for example, in a long project when it is necessary to explain the provenance (origin) of the data precisely for others.


4.3.2.2. Metadata structure

{{|>}}


Metadata always have a certain internal structure, even though the actual application can take different forms (e.g. from a simple text document to a table form to a very formalised form as an XML file that follows a certain metadata standard). The structure itself depends on the described data (for example, use of headers and legends in Excel spreadsheets versus a formalised description of a literary work in an OPAC), the intended use and the standards used. Generally speaking, metadata describe (digital) objects in a formalised and structured way. Such digital objects also include research data.

It makes sense, but is not absolutely necessary, for metadata to be readable not only by humans, but also by machines, so that research data can be processed by machines and automatically. Machines are primarily computers in this case, which is why one can also speak more precisely of readability for a computer. To achieve this, the metadata must be available in a machine-readable markup language. Research-specific standards in the markup language XML (Extensible Markup Language) are often used for this, but there are also others such as JSON (JavaScript Object Notation). When submitting (research data) publications, in most cases there is the option of entering the metadata directly into a prefabricated online form. Often it can take a substantial amount of time until a publication is submitted. Thus, it is important that metadata are collected and stored right when data are collected. A detailed knowledge of XML, JSON or other markup languages is therefore not necessarily required when creating metadata for your own project, but it can contribute to understanding how the research data is processed.

Machine readability is an essential point and becomes important, for example, when related research data are to be found by keyword search or compared with each other. A machine-readable file can be created using special programs. In the section "How do I create my metadata" you will be introduced to appropriate programs.


4.3.2.3. Metadata categories

{{|>}}


There are many categories that can and often need to be described by metadata (Figure 4). Depending on the field and research data, these categories can differ greatly, but some are considered standard categories for all disciplines.

In the context of a citable publication, the inclusion of a 'persistent identifier' (PID) within the metadata is essential. An identifier is used for permanent and unmistakable identification (see chapter 6.4). Furthermore, the metadata should indicate who the author of the data is. In the case of research groups, all those involved in the work or who may have rights to the research data should be named. The latter should, of course, include third-party funders like the DFG, BMBF or companies that may have contributed to the funding of the research. Always make sure that the names are complete and unambiguous. If a researcher ID (e.g. ORCID) is available, this should be mentioned (see chapter 6.4).

The research topic should be described in as much detail as necessary. For best findability of the research data, it can also be useful to mention keywords that can then be used in a digital database search to achieve better results. Keywords should complement the publication title, not repeat it.

Furthermore, for the traceability of the research data, clear information is needed for parameters such as place/time/temperature/social setting, and any other conditions that make sense for the data. This also includes instruments and devices used with their exact configurations.

If specific software was used to create the research data, the name of the software must also be mentioned in the metadata. Of course, this also includes naming the software version used, as this makes it easier for researchers to repeat analyses e.g. with older versions of R, or to understand later why this data can no longer be reproduced or opened in the case of very old data.

Some metadata requirements are always the same. This also applies to the categories just listed, which are very generic. For such cases, there are subject-independent metadata standards. Other requirements can differ greatly between different disciplines. Therefore, there are subject-specific standards that cover these requirements.

![The image is a colorful diagram with an octagonal shape made up of several smaller segments, centered around the concept of "Research data," which is written in the middle of a yellow hexagonal core. Around this core are six orange hexagons with the questions "What," "Who," "Where," "When," "Why," and "With what," each pointing to the central concept. Surrounding these question hexagons are larger blue segments with more specific details related to the questions: "What" points to "Research subject" and "Unambiguous attribution of research data." "Who" points to "Researcher with persistent ID" and "DOI for research data and metadata." "Where" points to "Geographical features," "Country," "Society," and "Laboratory conditions." "When" points to "Date of the survey," "Day/night," and "Season." "Why" points to "Research question and hypothesis." "With what" points to "Devices and measurement technology used," "Documentation of the procedure," and "Settings of devices and techniques used." Each of these blue segments also contains smaller text detailing various aspects of research data management, such as rights of use, citation guidelines, possible uses of data, co-authors, and parameters. The overall diagram serves as a visual guide to the considerations and components of managing and understanding research data, emphasizing the interconnectedness of these elements.](images/figure4_metadata_categories_FOKUSproject.png

"Figure 4. Different categories of metadata to describe the actual research data. Source: FOKUS, by Henrike Becker, CC BY 4.0, English translation by HeFDI."

4.3.2.4. Metadata standards

{{|>}}


One very important aspect of metadata already mentioned at the beginning is its readability for humans and machines. The large number of different metadata needed to describe research data can become a problem in view of the additional large number of different scientific communities, each with their own needs. On the one hand, there is metadata that is necessary across scientific fields (e.g., name of author, title, date of creation, etc.), but on the other hand there is also subject-specific metadata that depends on the research area or even the research subject.

Imagine that research group 1 has created a lot of research data over several experiments of the same kind with different room temperatures. Research group 2 has conducted the same experiment with the same substances at the same room temperature and different levels of oxygen in the air and has also created research data. Research group 1 refers to the parameter "room temperature" as "rtemp" in their metadata, but research group 2 only refers to it as "temp". How do the researchers of research group 1 and how does a computer system know that the value "temp" of research group 2 is the value "rtemp" of research group 1? It's just not easily possible and thus reduces the usefulness of the data.

So how can it be ensured that both research groups use the same vocabulary (= terminology) when describing their metadata, so that in the end it is not only readable but also interpretable? For such cases, metadata standards have been and are being developed by various research communities to ensure that all researchers in a scientific discipline use the same descriptive vocabulary. This ensures interoperability between research data, which plays a crucial role in expanding knowledge when working with data.

Metadata standards thus enable a uniform design of metadata. They are a formal definition, based on the conventions of a research community, about how metadata should be collected and recorded. Despite this claim, metadata standards do not represent a static collection of rules for collecting metadata. They are dynamic and adaptable to individual needs. This is particularly necessary because research data in projects with new research methods can be very project-specific and therefore the demands on their metadata are just as strongly project-specific.

There are many different metadata standards, some are more generic (e.g. Dublin Core), others are for specific disciplines (e.g. ABCD, Darwin Core, EML, MIxS or SDD for biological data) or data types (e.g. MIABIS for biosamples, MIAME for microarrays, MIAPPE for plant phenotyping experiments). To decide which standard to use, you can have a look at the repository where you want to deposit your data, to see if they have guidelines, checklists or best practices. Some repositories (e.g. GenBank). If you are unsure which repository you are going to use, you can check the following websites for more information about the right metadata1:


4.3.2.5. Controlled vocabularies and authority files

{{|>}}


As you have seen so far, metadata standards define the categories with which data can be described in more detail. On the one hand, these include interdisciplinary categories such as title, author, date of publication, type of study, etc., but on the other hand they also include subject-specific categories such as substance temperature in chemistry or materials science. However, there is no definition or validity check of how you fill the respective categories with information.

What date format do you use? Is the temperature given in Celsius or Fahrenheit and with "°" or "degrees"? Is it a "survey" or a "questionnaire"? These questions seem superficial at first glance, but predefined and uniform terms and formats are closely related to machine processing, search results and linkage with other research data. For example, if the date format does not correspond to the format a search system works with, the research data with the incompatible format will not be found. If questionnaires are searched for, but the term "survey" is used in the metadata, it is not certain that the associated research data will also be found.

For the purpose of linguistic standardisation in the description of metadata, so-called controlled vocabularies have been developed. In the simplest form, these can be pure word lists that regulate the use of language in the description of metadata, but also complex, structured thesauri. Thesauri are word networks that contain words and their semantic relations to other words. This makes it possible, among other things, to unambiguously resolve polysemous (ambiguous) terms (see e.g. Thesaurus of Plant Traits). A semantic thesaurus, often referred to as a semantic network, is a structured vocabulary or knowledge representation system that goes beyond traditional thesauri by capturing not only hierarchical relationships (such as broader terms and narrower terms) between concepts or terms but also semantic relationships, associations, and meaning. It is designed to provide a richer and more contextually meaningful representation of the terms or concepts within a specific domain or knowledge area. Thesauri primarily focus on managing and facilitating the retrieval of words or terms used in documents or databases. Their main purpose is to help users find synonyms, related terms, and broader or narrower terms to improve information retrieval and search. In contrast, ontologies, such as ENVO or the unit ontologies (e.g. UO) are used to represent and define complex relationships and concepts within a specific domain. They aim to capture the semantics and meaning of concepts, enabling a more comprehensive understanding within a specific domain or subject area. They serve as a formal and standardised framework for organising and categorising information, making data more accessible, interoperable, and understandable across various applications and disciplines. By providing a common vocabulary and a set of defined relationships, ontologies are invaluable tools in fields such as information retrieval, data integration, artificial intelligence, and the Semantic Web. They enable researchers and systems to achieve a shared understanding of complex domains, fostering more efficient data management and knowledge exchange.

!?This video by GFBio (2017) explains the importance of semantic search engines to find relevant data.

Ontologies and thesauri are distinct systems for structuring and categorising information. Ontologies serve to represent complex relationships and concepts within a specific domain. In contrast, thesauri primarily aid in word or term retrieval, offering synonyms, related terms, and hierarchical relationships to enhance information search and indexing. Ontologies provide a comprehensive, detailed representation of knowledge, while thesauri are more focused on managing and organising terms for search and categorization. The choice between the two depends on the specific needs and goals of a given project or application.

As a researcher or research group, how can you ensure the use of consistent terms and formats? As an individual in a scientific discipline, it is worth asking about controlled vocabularies within that discipline at the beginning of a research project. A simple search on the internet is usually enough. Even in a research group with a research project lasting several years, a controlled vocabulary should be searched for before the project begins and before the first analysis. If none can be found, it is worthwhile, depending on the number of researchers involved in the project and the number of sites involved, to create an internal project document for the uniform coordination of the terms and technical terms used, which should be used in the respective metadata categories.

GFBio provides a Terminology Service containing a collection of terminologies that are relevant in the field of biology and environmental sciences. This service acts as a resource, providing researchers with a common framework for defining and categorising various aspects of biodiversity, including species names, ecological interactions, and ecosystem processes. By utilising GFBio's Terminology Service, scientists can ensure consistency and precision in their research, promoting clarity and enabling seamless integration and comparison of biodiversity data across different studies. This unified approach to vocabulary and terminology strengthens collaboration and understanding within the scientific community, advancing our knowledge and conservation efforts in the diverse and intricate world of biodiversity.

In addition to controlled vocabularies, there are also many authority files which, in addition to uniform naming, make many entities uniquely referenceable. For example ORCID, short for Open Researcher and Contributor ID, which identifies academic and scientific authors via a unique code. The specification of such an ID distinguishes any frequently occurring and therefore ambiguous names and should hence be used as a matter of preference.

Probably the best-known standards file in Germany is the Gemeinsame Normdatei (GND, eng. Integrated Authority File), which is maintained by the Deutsche Nationalbibliothek (DNB, eng. German National Library), among others. It describes not only persons, but also "corporate bodies, conferences, geographies, subject terms and works related to cultural and scientific collections". Each entity in the GND is given its own GND ID, which uniquely references that entity. For example, Charles Darwin has the ID 118523813 and Alfred Russel Wallace has the ID 118806009 in the GND. These IDs can be used to reference Darwin and Wallace unambiguously in metadata with reference to the GND (e.g. Darwin in the DNB, Wallace in Wikidata). Another example from research is the so-called Research Organization Registry (ROR), that provides an unique identifier for each research organisation.

GeoNames is an online encyclopaedia of places, also called a gazetteer. It contains all countries and more than 11 million place names that are assigned a unique ID. This makes it possible, for example, to directly distinguish between places with the same name without knowing the officially assigned municipality code (in Germany, the postcode). For example, Manchester in the UK (2643123), Manchester in the state of New Hampshire in the US (5089178) and Manchester in the state of Connecticut in the US (4838174) can be clearly distinguished.

In general, investigate about specific requirements as soon as you know where you want to store or publish your research data. Once you know these requirements, you can create your own metadata. When referring to specific commonly known entities, always try to use a unique ID, specifying the thesaurus used.

If you want to know whether a controlled vocabulary or ontology already exists for your scientific discipline or a specific subject area, you can carry out a search at BARTOC, the "Basic Register of Thesauri, Ontologies & Classifications" as a first step. As a second step, BioPortal and the EMBL-EBI Ontology Lookup Service are very suitable resources for our domain.


4.3.2.6. Creation of your metadata

{{|>}}


Metadata can be created manually or using specific tools. Tools, also for subject-specific metadata, are available online and often free to use. However, ask colleagues or data managers at your institution about established routines or licences for proprietary software commonly used in your research area. The following tools for creating metadata are only a selection (also Table 2). Maintaining tools and software packages is a lot of work, so some tools may be discontinued after a while. The important step is to add metadata to your data. Which tools you use is secondary.

A simple way to create metadata is to use a table with the elements of your chosen metadata standard as columns. You can use any spreadsheet software to set this up. For example, GBIF provides templates for occurrence data, checklist data, sampling event data and resource metadata following the metadata standard Darwin Core, an extension of Dublin Core that facilitates the sharing of information about biological diversity. GFBio provides more specific templates which also prepare sample depositions in German data centers for different sample types and taxa (organisms, tissue, molecular data, etc.). All of these templates are spreadsheets with details on minimum metadata requirements.

RightField is an open-source tool which sets up or helps you fill in simple spreadsheet tables depending on the metadata standard you choose. Ontologies can either be imported from local file systems, the web, or from the BioPortal ontology repository. RightField translates these ontologies into a spreadsheet table, and you only have to choose from the predefined terms to create your standardised metadata. You can build your own templates, too, but we highly recommend using an existing, widespread metadata standard to make your metadata findable and interoperable.

Tools with simple graphical user interfaces are not available for all metadata standards. Therefore, if you want or need to work directly with an existing XML metadata standard, you should either use the free editor Notepad++ or the paid software oXygen, if licences are available at your institution. Both editors offer better usage and display options to make content and element labels visible separately.

To create metadata in the XML metadata standard EML (Ecological Metadata Language), you can use the online tool ezEML. ezEML is free of charge and only requires a registration. It can be used as a "wizard" to guide you step-by-step through the creation of your EML document. However, ezEML also offers a range of other workflows, such as data table upload, and checking your EML document for correctness and completeness.

The open source online tool CEDAR Workbench allows online templates based on metadata standards to be created via a graphical user interface, filled in and also shared with other users. At the same time, templates created by other users can also be used for one's own research. All you need to do is register free of charge. If you are working with biodiversity survey or monitoring data, you can use the ADVANCE template in CEDAR based on a controlled vocabulary integrating terrestrial, freshwater, and marine monitoring metadata to facilitate data interoperability and analyses across realms.

The ISA framework (Investigation, Study, Assay) is suitable for experiments in the life sciences and environmental research. It is open source and consists of several programmes that can help in the management of experiments from planning and conduction to the final description. You can start with the ISA Creator, which is used to create files in the ISA-TAB format. This format is explicitly required, for example, by the Scientific Data Journal of the Nature publishing house.

Which programmes are suitable for your metadata depends very much on the type of research data and your wishes for use (also see Table 2 & Table 3). It is therefore worthwhile to talk to other researchers in advance to find the best way to create metadata for yourself. Familiarising yourself with the metadata standard relevant to you and searching for programmes that use this standard can have an advantage in terms of automatic processing of the data and later publication. At the very least, the use of a simple, subject-independent metadata standard such as Dublin Core should be considered. Of course, you should try to use the most suitable metadata standard, especially with respect to the recommendations of the chosen repository; however, there is no such thing as a perfect choice. Overall, huge effort is invested in making metadata standards interoperable in the future.


4.4. Taxonomy - a special challenge for data integration

{{|>}}


Biodiversity encompasses different approaches that are related to the underlying data and organisational type such as genetic diversity, species diversity, and ecosystem diversity. In practice, biodiversity data is predominantly associated with species, whether you are working with proteins, genes, or ecosystem functions. Species, as fundamental biological entities, and their nomenclature serve as the backbone for the integration and synthesis needed for biodiversity research and the decision-making processes in nature conservation1.

However, species names and concepts are not static; they evolve across both space and time. Harmonising species names from various datasets is an essential aspect of data integration in the field of biological sciences2. The main reason for the diversity of taxonomies lies in the existence of multiple scientifically sound methods for defining a species. Diverse institutions and academic traditions curate a multitude of taxonomic frameworks for a wide range of taxa. These differing species concepts can diverge markedly when it comes to determining which populations should be included or excluded in the characterization of a species. It's essential to recognize that no single concept comprehensively accounts for all facets of diversity across all organisms. Equally important is understanding which taxonomic reference databases are suitable for specific taxa.


4.4.1. What is a species?

4.4.1.1. Species concepts

{{|>}}


The concept of a species is a multifaceted and often controversial matter, as illustrated by Charles Darwin's observation in 1859 that the distinction between species and varieties can be somewhat vague and arbitrary1. Main concepts on defining species include:

  1. Biological Species Concept: This approach defines a species as a group of natural populations that either interbreed or have the potential to do so2, while being reproductively isolated from other such groups. It suits sexually reproducing organisms but is insufficient in cases like asexual reproduction, fossils, hybridization, species complexes.

  2. Morphological Species Concept: Species are identified based on similar physical traits3. While intuitive, this method can be subjective due to varying trait expressions, phenotypic plasticity, and cryptic species with nearly identical appearances.

  3. Phylogenetic Species Concept: It defines a species as the smallest aggregation of populations or lineages distinguishable by unique genetic characteristics4. It is well-suited for asexual species but might lead to "taxonomic inflation" and conservation challenges.

  4. DNA Barcoding: DNA barcoding relies on genetic sequences, using markers like cytochrome c oxidase for animals, 16S rRNA for bacteria, and internal transcribed spacer (ITS) 28S rRNA for fungi. It's a powerful tool for species identification, including eDNA5. However, it relies on reference sequences and can suffer from incomplete databases and misidentified species.

Other noteworthy concepts include the evolutionarily significant unit (conservation purposes), recognition and cohesion species (behavioural isolation), chronospecies in paleobiology, and ecological species occupying the same niche.

These concepts illuminate the diverse ways species are defined and have significant implications in various fields of biology and conservation.


4.4.1.2. Species names

{{|>}}


A species name is a hypothesis about the evolutionary relationships and distinctive characteristics between groups of organisms. The binomial nomenclature, i.e. the scientific name of a species, carries a lot of information, provided it is reported in full: Genus epithet Authority Year, e.g. Canis lupus Linnaeus, 1758 (ASM Mammal Diversity Database #1005943) fetched 2023-10-09. Mammal Diversity Database. 2023.

The authority and year point to the publication of the respective species description. This publication reports a distinct set of morphological, (phylo-)genetic, geographical and/or behavioural characteristics to define a new species. Furthermore, it specifies the type, i.e. the actual specimen or group of specimens that was used to describe the new species. The type specimens are used as physical reference and usually incorporated into institutional scientific collections to facilitate long-term preservation, curation and public access. There are institutions that govern the process of naming species such as International Code of Nomenclature for algae, fungi, and plants (ICN) or International Code of Zoological Nomenclature (ICZN), but these institutions do not regulate what constitutes a species (see species concepts).

Understanding species and subspecies is crucial in conservation efforts, aiding in the organisation of breeding programs, protection of endangered species (IUCN), and wildlife trade regulation through conventions like Convention on International Trade in Endangered Species of Wild Fauna and Flora (CITES). For instance, the discovery of the critically endangered Tampanuli orangutan underscored the immediate need for preservation due to habitat loss from industrial projects. CITES plays a pivotal role in regulating the international trade of over 40,000 species. Customs officers rely on databases with species names, references, synonyms, and vernacular names to enforce these regulations. The German Federal Agency for Nature Conservation (BfN) provides a similar online database (WISIA) covering all species that are protected under German national or international law. Again, the database contains species names, the nomenclature references for each species or taxon, synonyms and vernacular names. Species names and respective references are also considered in many laws, ranging from fishing quotas in the Baltic Sea, compulsory environmental impact studies and mitigation efforts accompanying large building projects, to the prevention, minimisation and mitigation of the enormous detrimental effects of invasive alien species to both biodiversity and economy. In biological sciences, the integration of taxonomic data is vital for ecological research, global biodiversity trends, and decision-making. It supports initiatives like the Intergovernmental Panel on Biodiversity and Ecosystem Services (IPBES), which provides policy-relevant knowledge, aligning with the Sustainable Development Goals (SDG). This integration enhances the scientific foundation of such reports and contributes to essential functions like food security, disease prevention, carbon storage, and climate regulation12

In summary, species data integration is pivotal in safeguarding biodiversity and making informed decisions with far-reaching implications.


4.4.2. Pitfalls in taxonomic data integration

{{|>}}


What could possibly go wrong when you try to integrate two data sets with species names? This GBIF blog post highlights some pitfalls of matching species names from new data sets to the GBIF taxonomic backbone.

In the following there are common pitfalls:

    {{1}}

Variable spelling of species name

  • Data set 1: Carabus arvensis Herbst 1784 (valid name)

  • Data set 2: Carabus arcensis Herbst 1784 (Synonym)

  • Data set 3: Carabus aruensis Herbst 1784 (misspelt epithet)

    {{2}}
    

Variable spelling of authority

  • Data set 1: Carabus arvensis Herbst 1784

  • Data set 2: Carabus arvensis Hbst., 1784 (official abbreviation of authority)

  • Data set 3: Carabus arvensis Hrst. 1784 (misspelt abbreviation of authority)

    {{3}}
    

Missing authority


4.4.3. Solutions and recommendations for taxonomic data integration

{{|>}}


To integrate taxonomic data sets for your own purposes, you should match the two sets of species names against a taxonomic reference database. Many databases are using alphanumeric unique taxon identifiers (UTI) additionally to species names to facilitate taxonomic data integration. A multitude of software tools and R packages helps to access and use these taxonomic databases, matching both UTIs and species names. Many research teams are working hard to improve taxonomic integration at all levels of data curation: Schellenberger Costa et al. (2023) compared four global authoritative checklists for vascular plants and proposed workflows to better integrate these important databases in the future1, Grenié et al. (2021) compiled a living database of taxonomic databases and R packages2 and gave an overview of tools, databases and best practices for matching species names3, and Sandall et al. (2023) proposed key elements of a global integrated structure of taxonomy (GIST) to reconcile different aspects, approaches and cultures of taxonomy4.

This image is a diagram explaining the concept of "Taxonomic Harmonization" between two datasets, A and B. On the left side, labeled "Dataset A," there are circles with different colors and labels representing the conservation status of species: green with 'LC' for Sp1, yellow with 'VU' for Sp2, teal with 'NT' for Sp3, and red with 'CR' for Sp5. In the center, there is a column with the title "Taxonomic Reference Database" with a list of species, Sp1 to Sp6. The species from Dataset A (Sp1, Sp2, Sp3, and Sp5) are matched with their counterparts in the database. Sp5 from Dataset A is dashed, indicating an uncertain match with Sp6 in the database. On the right side, labeled "Dataset B," there are colored bars representing the traits of species Sp2, Sp3, Sp4, and Sp6, with each species having a unique combination of colors. Arrows show the correspondence between the datasets and the taxonomic reference database, illustrating the process of harmonization where species are cross-referenced and aligned between different datasets based on a standard taxonomic concept.

Taxonomic databases serve specific tasks and encompass different aspects of biodiversity. The suitability of a particular database for your project depends on the type of data you are looking for 4:

  • Databases creating new taxonomic frameworks: Examples include FishBase, Avibase, Amphibian Species of the World (ASW). These databases develop novel taxonomic backbones.

  • Databases aggregating primary taxonomic lists and linking databases: Notable databases in this category are the Catalogue of Life (COL+), Encyclopedia of Life (EoL). They bring together primary taxonomic information and link various databases.

  • Databases combining new frameworks and taxonomic lists: Databases like ITIS and WoRMS fall into this category. They not only produce new taxonomic backbones but also integrate primary taxonomic lists.

  • Databases focused on biodiversity data aggregation: This group includes databases like GBIF, OBIS, Map of Life (MOL), INSDC and IUCN. While not primarily taxonomic, their main purpose is to aggregate and provide comprehensive biodiversity data.

Understanding the nature and focus of these databases is essential in choosing the right one for your research or project. If you are re-using taxonomic data collected by others, you should therefore take care for

  • Find out which taxonomic reference was used to identify the species

  • Consider the possibility of misidentification (e.g. if a genetic sequence is not linked to a voucher specimen, or if the authority is missing)

  • Consider the possibility of misspelling of both species names and authority names

  • Select a taxonomic backbone that is complete and suitable for your taxon of interest for taxonomic harmonisation

If you are collecting new taxonomic data yourself, or in collaboration with taxonomic experts, you should additionally take care to

  • Note the scientific name and authority of a species

  • Note the identification literature that was used to identify specimens and the person(s) who identified specimens (when working with taxonomic experts: ask them which reference literature they used)

  • Note the taxonomic references for each taxon, i.e. which taxonomic references the used identification literature is based on

  • Note the place and date when you found a specimen

  • Note details about the method used to record an occurrence

Species names connect physical, functional, spatial, and genetic data sets. Providing detailed taxonomic metadata will help to keep your data alive and re-usable for your own and for others research.


4.5 Questions

{{|>}}


How do I collect metadata?

  • [(X)] I enquire about suitable metadata standards before starting research and collect as much metadata as possible to ensure that the data can also be understood by other researchers.
  • [( )] I simply transfer the metadata from another project.
  • [( )] I always follow the same scheme, no matter what type of research data it is.

Correct! But it really won't hurt to ask colleagues of the same discipline whether there are any known suitable metadata standards and how they should be applied.


When should the associated metadata be deleted if research data is deleted from a repository?

  • [(X)] Preferably not at all. If research data is withdrawn, the metadata provides an overview of the project and possibly why the data was withdrawn.
  • [( )] The metadata is deleted together with the research data.
  • [( )] For legal reasons, the metadata must remain in the repository for at least one year.

Correct! One reason, above all, is the obligation to provide proof of research data. If, for example, a text was published in 2017 that examined data from 2015, but in 2019 it turns out that the data collected at that time violated certain legal requirements, the text must still retain evidence that this research data actually still existed and was accessible at the time of publication, and also information about what this data contained.


What exactly is metadata?

  • [(X)] Metadata describes other data of any kind in a structured way and helps to understand this data.
  • [( )] Metadata is data that stores information about files in encrypted form.
  • [( )] Metadata is used to make the actual research data available, as it would otherwise not be accessible.

Correct! Metadata is data about data. Metadata therefore contain descriptive information to understand another or the described data set. Understanding the described data is mostly dependent on already having expertise in dealing with this type of data through one's own research or studies.


How should metadata ideally be stored?

  • [( )] I create metadata according to my own schema according to the requirements of my own research data.
  • [(X)] Metadata should be created according to specific metadata standards.
  • [( )]Metadata should only be readable by myself in order to protect my research data.

Correct. Ideally, metadata should be available as subject-specific standards. However, this is not always the case, so depending on the research project, metadata sometimes has to be created.


What is the correct order of a species name?

[[lowii | J. E. Gray | (Ptilocercus) | 1848 ]][[ 1848 | Ptilocercus | J. E. Gray | (lowii) ]][[ 1848 | (J. E. Gray) | lowii| Ptilocercus ]][[ J. E. Gray | Ptilocercus | lowii | (1848) ]]


Correct! Species names consist of Genus, epithet, authority, and year of publication of the species description.

The pen-tailed treeshrew (Ptilocercus lowii J. E. Gray, 1848) is a treeshrew of the family Ptilocercidae native to southern Thailand, the Malay Peninsula, Borneo, and some Indonesian islands. In a study of wild pen-tailed treeshrews, the animals frequently consumed large amounts of fermented nectar, equivalent of 10–12 glasses of wine adjusted to body weight with an alcohol content up to 3.8%. The pen-tailed treeshrews did not show any signs of intoxication, probably because they use a different pathway to metabolize alcohol compared to humans.


What is the correct order of a species name?

[[ Fresen | oligospora | (Arthrobotrys) | 1850 ]] [[ Arthrobotrys | 1850 | Fresen | (oligospora) ]] [[ 1850 | oligospora |Arthrobotrys | (Fresen) |]] [[ Arthrobotrys | (1850) | Fresen | oligospora ]]


Correct! Species names consist of Genus, epithet, authority, and year of publication of the species description.

Arthrobotrys oligospora Fresen. (1850) is a nematode-capturing fungus. In low nitrogen environments, it grows sticky nets from hyphae and attracts nematodes with semiochemicals1 that nematodes use to communicate with each other. Small nematodes can be caught with a single loop. The more the nematodes struggle to escape, the more they are glued to the net. Finally, the nematodes are paralysed and digested by hyphae growing into their bodies.


What is the correct order of a species name?

[[ (Cymothoa) | exigua | Schiødte & Meinert | 1884 ]] [[ Cymothoa | (exigua) | Schiødte & Meinert | 1884 ]][[ Cymothoa | exigua | (Schiødte & Meinert) | 1884 ]] [[ Cymothoa | exigua | Schiødte & Meinert | (1884) ]]


Correct! Species names consist of Genus, epithet, authority, and year of publication of the species description.

Cymothoa exigua (Schiødte & Meinert, 1884), the tongue-eating louse, is a parasitic isopod of the family Cymothoidae. Females attach to the tongue and sever the blood vessels. When the tongue falls off, the female attaches itself to the remaining stub of tongue and the parasite itself effectively serves as the fish's new "tongue". The parasite feeds on the host's blood and mucus. Juveniles likely first attach to the gills of a fish and become males. As they mature, they become females, with mating likely occurring on the gills.


What is the correct order of a species name?

[[ (Ailanthus) | altissima | Mill. | Swingle ]][[ Ailanthus | (altissima) | Mill. | Swingle ]][[ Ailanthus | altissima | (Mill.) | Swingle ]][[ Ailanthus | altissima | Mill. | (Swingle) ]]


Correct! Species names of plants consist of Genus, epithet, and authority.

Ailanthus altissima (Mill.) Swingle is a deciduous tree native to northeast and central China, and Taiwan. The tree grows rapidly and is very tolerant of pollution. It is considered to be one of the worst invasive species in Europe and North America. It spreads aggressively both by seeds and vegetatively by very long root sprouts, often "tunneling" even below wide streets and crossroads. This "tree of heaven" inhibits the growth of other plants in their sourroundings. Oh, and the male flowers have a very distinctive smell...



5. Analysing Data

5.1. Introduction and learning objectives

{{|>}}


Data analysis enables you to uncover hidden insights and correlations within the data, ultimately leading to the creation of knowledge. The nature of this analysis can vary significantly, depending on the heterogeneity of the data at hand. You, as the data analyst, are best equipped to determine the most appropriate methods for your specific dataset. However, it's imperative to adhere to and document the standards and methods commonly accepted within your field. You may employ a range of tools and software, including application-specific software or programming languages like Python and R123.

Given that data analysis yields critical research outcomes, it's of utmost importance that the entire analysis workflow is thoroughly documented, published, and aligns with FAIR principles. This approach ensures the reproducibility of your results456. In our data-driven world, effective data analysis is paramount for informed decision-making across diverse domains. However, access to data alone is insufficient; it must be managed and analysed in a manner that guarantees it meets the criteria of being FAIR. This can be assured by providing code you used/produced for your analysis, execution environments, and data analysis workflows.

After completing this chapter, you will be able to...

    {{1}}

...name useful tools to document your data analysis workflow

    {{2}}

...explain why code management helps to make your data analyses more transparent


5.2 Tools for code management, execution environemts and data analysis workflows

5.2.1. Code management

{{|>}}


Code Management: A fundamental aspect of FAIR analysis is the organisation, accessibility, and version control of your code. Git repositories are invaluable for effective code management. Here are key considerations:

  • Git Basics: Familiarise yourself with Git, including core concepts like commits, branches, and pull requests. You could e.g. start with the introductory course on version control by the Software Carpentries or this tutorial.

  • Repository Structure: Maintain a clear and structured directory layout for your codebase. Suggestions and examples for useful structures are here, here and here.

  • Documentation: Thoroughly document your code to ensure it's comprehensible to others. Here are some basic suggestions. Advanced R users can e.g. use package roxygen2.


5.2.2. Execution environment

{{|>}}


Execution Environment: Enhance your code's accessibility and reproducibility by embracing package and environment management systems. These specialised tools empower fellow researchers to seamlessly install the precise tool versions, including older ones, within a controlled environment. This ensures that they can execute your code in an equivalent computational setup, complete with all necessary dependencies, such as the correct version of R and specific libraries and packages:

  • Bioconda: Bioconda is a specialised distribution of bioinformatics software that simplifies the installation and management of tools and packages for biological data analysis. Here is a tutorial to get started.

  • mybinder.org: MyBinder enables the creation of interactive and executable environments directly from your code repositories, simplifying the process for others to replicate your analyses. Here is a tutorial to get started. You can also follow the novice Python lesson by the Software Carpentry which uses binder.


5.2.3 Data analysis workflows

{{|>}}


Data Analysis Workflows: To conduct FAIR analysis, it's essential to select appropriate data analysis workflows. Tools such as Jupyter Notebook and R Markdown facilitate transparent and reproducible analyses:

  • Jupyter Notebooks: Jupyter provides an interactive environment for data exploration and analysis, allowing you to seamlessly blend code (such as R, Julia, Python), visualisations, and narrative explanations in a single document. You can try Jupyter Notebooks in your browser with a programming language of your choice.

  • R Markdown: R Markdown combines R code with markdown text, empowering you to create dynamic reports and documents that are easily shareable and reproducible.


5.3 Summary

{{|>}}


By embracing these principles and utilising tools like Git, Bioconda, mybinder.org, Jupyter, and R Markdown, you'll be better equipped to perform data analysis that not only adheres to FAIR principles but also fosters transparency, accessibility, and collaboration. This approach contributes to the advancement of research and science in our data-driven era. An illustrative example of FAIR analysis was produced in the context of the NFDI4Biodiversity and GfÖ Winter School 2022 in the form of a Jupyter Notebook.


5.4 Questions

{{|>}}


Which of these are tools to organize your data analysis?

  • [[Tools for data analysis] (Not a tool for data analysis)]
  • [ (X) ( ) ] Jupyter
  • [ (X) ( ) ] Binder
  • [ (X) ( ) ] Git
  • [ (X) ( ) ] RMarkdown
  • [ (X) ( ) ] Python
  • [ (X) ( ) ] Bioconda
  • [ (X) ( ) ] roxygen2
  • [ ( ) (X) ] Saturn
  • [ ( ) (X) ] Boa
  • [ ( ) (X) ] Reddit
  • [ ( ) (X) ] Klaxxon
  • [ ( ) (X) ] Miro board
  • [ ( ) (X) ] Jamboard

Which of these statements are true? Select all correct answers.

  • [[X]] Git is a free software for version control.
  • [[X]] Git means "unpleasant person".
  • [[ ]] Code documentation is unnessesary - code is self-explaining.
  • [[ ]] GitHub, GitLab and Git are just different names for the same platform.

6. Preserving Data

6.1. Introduction and learning objectives

{{|>}}


Data Preservation is equal to long-term archival (LTA) and aims at the integrity of data meaning that data are accessible, authentic, and reusable for the long-term, independently of software and hardware decay. Preservation is done by designated data repositories with expertise in data curation. Thus, the following information on LTA has mainly a theoretical value for you and only a limited action value. If you publish your data in a certified repository, you are well advised. Above all, make sure that you do so at a trustworthy institution and obtain information from this institution in advance about possibilities or plans regarding an LTA. You can use the aspects listed here for a good LTA to formulate possible questions for the facilities. This should provide sufficient preconditions for the LTA.

Why is it so important to preserve data?

!?In this video by the European Space Agency (2019) you will meet Professor Smith, who was lucky to find well preserved data.

Long-term archival of data is important for its availability and reusability while ensuring high data quality, integrity, safety and accessibility. Archives, data centers or data managers often conduct the preservation process. Nevertheless, the data producer is responsible for the data quality and can thus facilitate the process by using the best (open access) format, the right preservation metadata and documentation of the workflow.

Check your options and domain-specific requirements for long-term archiving research data, and inform yourself in advance if the archive or data centers of your choice have special requirements for data preservation12.

Data preservation is important for its reuse, but also for the possibility to reproduce the data after time or if the data is of great value to an organisation, a country, the environment or the society as a whole. However, since it takes a lot of effort and cost, only an appropriate selection of data should be preserved. If you want to preserve materials, such as microorganisms or biomaterials, contact your institution's data management support services to find appropriate centres or biobanks1.

However, in order to enable actual reuse, a number of conditions must be met.

After completing this chapter, you will be able to...

    {{1}}

... preserve your data in a comprehensible way

    {{2}}

...choose durable, preferably non-proprietary (i.e. free and open source) file formats for your data

    {{3}}

...select durable storage media

    {{4}}

...assure the findability of your data

    {{5}}

...choose a repository where to publish your data


6.2. Long-Term Data archiving

{{|>}}


Besides data storage, long-term data archiving is a necessary step in the research data life cycle ensuring that your data are well preserved for you and other researchers. While data storage primarily involves the storage of data during the ongoing work process in the project period, data archiving is concerned with how the data can be made available in as reusable a way as possible after the project has been completed. A distinction is often made between data storage in a repository and data archiving in the sense of LTA. However, in many places, including the DFG's "Guidelines for Safeguarding Good Research Practice" from 2019 ("Guideline 17: Archiving"), both terms are used equivalently. When we speak of preservation or data retention in the following, we mean the storage of data in a research data repository and long-term archiving is meant. The differences between the two variants are the subject of this chapter.

With today's rapidly evolving digital possibilities, the older data becomes, the more likely it is that this data can no longer be opened, read, or understood in the future. There are several reasons for this: The necessary hardware and/or software is missing, or scientific methods have changed so much that data is now collected in other ways with other parameters. Modern computers and notebooks, for example, now almost always do without a CD or DVD drive, which means that these storage media can no longer be widely used. Long-term archiving therefore aims to ensure the long-term use of data over an unspecified period of time beyond the limits of media wear and technical innovations. This includes both the provision of the technical infrastructure and organisational measures. In doing so, LTA pursues the preservation of the authenticity, integrity, accessibility, and comprehensibility of the data.

In order to enable long-term archiving of data, it is important that the data are provided with meta-information relevant to LTA, such as the collection method used, hardware of the system used to collect the data, software, coding, metadata standards including version, possibly a migration history, etc. In addition, the datasets should comply with FAIR principles as far as possible. This includes storing data preferably in non-proprietary, openly documented data formats and avoiding proprietary data formats. Open formats need to be migrated less often and are characterised by a longer lifespan and higher dissemination. Also, make sure that the files to be archived are unencrypted, patent-free and non-compressed. In principle, file formats can be converted lossless, lossy, or according to the meaning. Lossless conversion is usually preferable, as all information is retained. However, if smaller file sizes are preferred, information losses must often be accepted. For example, if you convert audio files such as WAV to MP3, information is lost through compression and the sound quality decreases. However, the conversion results in a smaller file size. The following table gives a first basic overview of which formats are suitable and which are rather unsuitable for a certain data type:

Table 1. Recommended and non-recommended data formats by file name (source: HeFDI self-study unit, CC BY 4.0).

Data type Recommended formats Less suitable or unsuitable formats
Audio .wav / .flac .mp3
Computer-aided design (CAD) .dwg / .dxf / .x3d / .x3db / .x3dv -
Databases .sgl / .xml .accdb / .mdb
Raster graphics & images .tif (uncompressed) / .jp2 / .jpg2 / .png .gif / .jpeg / .jpg / .psd
Statistical data .por / .csv .sav (SPSS)
Tables .csv / .tsv / .tab .xls / .xlsx / .xlx
Texts .odf / .rtf / .txt / PDF/A .docx / .doc / PDF
Vector graphics .svg / .svgz .cdr
Video .mp4 / .mkv / .mj2 / .avi (uncompressed) .mov / .wmv

The list of formats in the column labelled "less suitable or unsuitable" does not imply that you cannot utilise these formats for long-term data storage. It is rather a matter of being sensitised to questions of long-term availability in a first start. Make it clear which format offers which advantages and disadvantages. If you want to delve further, you will find what you are looking for on the website of NESTOR -- the German competence network for long-term archiving and long-term availability of digital resources. Under NESTOR - Topic you will find current short articles from the field, e.g. on tiff or pdf formats. If you put these and other overviews side by side, you will notice that the recommendations on file formats differ from each other. We do not yet have enough experience in this field. Another good way to find out if you are uncertain about formats is to ask a specialised data center or a research data network, if one exists. If you want to store your data there, this approach is even more advisable. You may then find that your data will be taken even if the chosen data format is not the first choice from an LTA perspective. Operators of repositories or research data centers work close to science and always try to find a way of dealing with formats that are widely used in the respective fields, e.g. Excel files. As an example of this, you can take a look at the guidelines of the Association for Research Data Education.

To be able to decide for yourself which formats are suitable for your project, there are many criteria that you should consider when making your selection (according to Harvey/Weatherburn 2018: 131):

  • Extent of dissemination of the data format

  • Dependence on other technologies

  • Public accessibility of the file format specifications

  • Transparency of the file format

  • Metadata support

  • Reusability/Interoperability

  • Robustness/complexity/profitability

  • Stability

  • Rights that can complicate data storage

LTA currently uses two strategies for long-term data preservation: emulation and migration.

Emulation means that on a current, modern system, an often older system is emulated, which imitates the old system in as many aspects as possible. Programmes that do this are called emulators. A prominent example of this is DOSBox, which makes it possible to emulate an old MS DOS system including almost all functionalities on current computers and thus to use software for this system, which is most likely no longer possible with a more current system.

Migration or data migration means the transfer of data to another system or another data carrier. In the area of LTA, the aim is to ensure that the data can still be read and viewed on the system to be transferred. For this, it is necessary that the data are not inseparably linked to the data carrier on which they were originally collected. Remember that metadata must also be migrated!

When choosing a suitable storage location for long-term archiving, you should consider the following points:

  • Technical requirements -- The service provider should have a data conversion, migration and/or emulation strategy. In addition, a readability check of the files and a virus check should be carried out at regular intervals. All steps should be documented. Plus copies of the data in several locations (on-site, near-side, off-site)

  • Seals for trustworthy long-term archives -- Various seals have been developed to assess whether a long-term archive is trustworthy, e.g. the nestor seal, which was developed on the basis of DIN 31644 "Criteria for trustworthy digital long-term archives", the ISO 16363 standard based on the Reference Model for an Open Archival Information System (OAIS), or the CoreTrustSeal.

  • Costs -- The operation of servers as well as the implementation of technical standards are associated with costs, which is why some service providers charge for their services. The price depends above all on the amount of data.

  • Making the data accessible -- Before choosing the storage location, you should decide whether the data should be accessible or only stored.

  • Service provider longevity -- Economic and political factors influence the longevity of service providers.


6.3. Data preservation using repositories and data centers

{{|>}}


Data preservation in a research data repository is usually accompanied by publication of the data produced. Access to such publications can and, in the case of sensitive data such as personal data, must be restricted. In accordance with good scientific practice, repositories must ensure that the published research data are stored and made available for at least ten years, after which time availability is no longer necessarily guaranteed, but is nevertheless usually continued. If data are removed from the repository after this minimum retention period at the decision of the operator, the reference to the metadata must remain available.

By using data centers or even repositories, it is possible to restrict access to confidential and sensitive data and at the same time enable data sharing for research and educational purposes. The data held in data centers and archives are generally not publicly accessible. Their use after user registration is restricted to specific purposes. Users sign an end-user licence in which they agree to certain conditions, such as not using data for commercial purposes or not identifying potentially identifiable individuals. The type of data access permitted is determined in advance with the originator. Furthermore, data centers can impose additional access regulations for confidential data1.

Repositories are usually divided into three different types: Institutional repositories, subject repositories and interdisciplinary or generic repositories. A fourth, more specific variant are so-called software repositories, in which software or pure software code can be published. These are usually designed for one programming language at a time (e.g. PyPI for the programming language "Python").

Institutional repositories include all those repositories that are provided by mostly state-recognised institutions. These may include universities, museums, research institutions or other institutions that have an interest in making research results or other documents of scientific importance available to the public. As part of the DFG's "Guidelines for Safeguarding Good Research Practice", there is an official requirement that the research data on which a scientific work is based must be "archived in an accessible and identifiable manner for a period of ten years at the institution where the data were produced or in cross-location repositories"2. The archiving period begins on the date when the results are made publicly available.

Subject-specific repositories are equipped to handle domain-specific data and hold valuable community resources for research. Publishing in a renowned subject-specific repository can greatly contribute to enhancing your scientific reputation, but this needs to be prepared carefully. To find out whether a suitable subject-specific repository is available for your research area, it is useful to search via the repository index re3data and FAIRsharing.org.

If there is no suitable repository, the last option is to publish in a large interdisciplinary repository. A free option is offered by the service Zenodo, funded by the European Commission. If your university is a member of Dryad, you can also publish there, free of charge. RADAR offers a fee-based service for publishing data in Germany and is also a good option for larger projects with centralised data management. The most frequently used option in Europe is probably Zenodo. When publishing on Zenodo, make sure that you also assign your research data to one or more communities that in some way reflect a subject-specificity within this generic service.

Repositories are typically domain-agnostic, accommodating a wide range of research fields. They may lack standardised protocols and structures, leading to variability in data organisation and metadata. Additionally, repositories may not always provide guarantees for long-term archiving, as their sustainability can depend on institutional or project-specific factors, potentially making them less reliable for the extended preservation of research data when compared to the more comprehensive and standardised services offered by data centers. When choosing your preferred repository, also look to see if it is certified in any way (e.g. CoreTrustSeal). Whether a repository is certified can be checked at re3data.

Recommended selection criteria for repositories:

  • Repositories should be trustworthy (recommended: Core Trust Seal or DINI-certified)

  • As discipline-specific and suitable for data type as possible

  • Published data should be fully citable via persistent identifier (e.g. DOI, URN)

  • Data can be cross-referenced with journal articles

  • Searchable across publishers & institutions, by human or machine

  • Metadata are machine harvestable, contents machine-retrievable

  • Provides steps for data preparation and quality control

  • Transparent usage and licensing conditions

  • Open terms of reuse, no paywalls

  • Transparent cost/business mode

  • The repository provides professional support during the preparation and archiving of your data

  • Preservation services, including migration of formats

Check re3data.org and further resources listed in Table 4 to get information on what kind of services repositories provide.

Data centers, in contrast to repositories, are well-equipped to ensure long-term preservation of data, often adhering to domain-specific standards and best practices for data management and archiving. They are typically connected to data portals that systematically harvest and disseminate metadata, enhancing data discoverability and accessibility on a broader scale. This structured approach and their commitment to domain-specific standards make data centers a reliable choice for researchers seeking robust and standardised long-term data preservation solutions. NFDI4Biodiversity collaborates with a network of specialised data centers, the GFBio Data Centers (Table 5), each with distinct expertise. These centres include those focused on nucleotide, plant, and environmental data such as e!DAL-PGP, ENA, and PANGAEA, as well as data centers associated with natural science collections like the Botanical Garden and Botanical Museum Berlin (BGBM), DSMZ-German Collection of Microorganisms and Cell Cultures, Leibniz-Institut zur Analyse des Biodiversitätswandels (LIB), Museum für Naturkunde Berlin (MfN), Senckenberg, Staatliches Museum für Naturkunde Stuttgart (SMNS) and Staatliche Naturwissenschaftliche Sammlungen Bayerns (SNSB). These centres can manage a wide range of data types crucial for biodiversity research, from occurrence data and environmental variables to molecular sequences, multimedia files, experimental measurements, and even geospatial data like orthophotos and digital surface models, offering comprehensive support for biodiversity-related research endeavours. The method for data submission varies depending on the data center you choose. You have the option to submit data directly or utilise tools like Diversity Workbench. Additionally, the GFBio Data Submission Service offers professional assistance, particularly beneficial if you are dealing with heterogeneous datasets or are uncertain about the most suitable data center for your specific requirements. Regardless of where you end up publishing your data, always make sure to include a descriptive "metadata file" in addition to the data, describing the data and setting out the context of the data collection.


6.4. Persistent identifiers

{{|>}}


Ensuring the findability of research data is crucial for the reusability of the data. An important step towards making data retrievable/findable is the assignment of so-called persistent identifiers (PID), which globally ensure the unique and permanent identification of a digital resource. They permanently refer to a specific content (e.g. data set) and link the associated metadata to the actual data set. A web link can change, a PID always remains the same. There are two types of PIDs:

  1. Identifiers for digital objects, e.g.

    1. DOI = Digital Object Identifier

    2. URN = Uniform Resource Name

  2. Identifier for persons or organisations (clear scientific identity), e.g.

    1. ORCID = Open Researcher Contributor Identification

    2. ResearcherID

    3. ROR = Research Organisation Registry

DOIs serve as crucial tools for uniquely identifying and accessing digital resources. They are primarily assigned by authoritative bodies like DataCite, which ensure their official registration and management. Through the allocation of DOIs, research data and documents gain a distinctive identity. Moreover, DOIs enable linking between metadata and research data. This linkage facilitates effective data management and citation practices, making it easier for researchers to acknowledge and reference specific data in their work. Additionally, other identifiers like URNs (Uniform Resource Names) contribute to this ecosystem by providing a standardised nomenclature for digital resources. Meanwhile, identifiers such as ORCID, ResearcherID, and ROR establish clear and unambiguous scientific identities for individuals and organisations, further enhancing the accuracy and traceability of scholarly contributions. Repositories and journals automatically assign corresponding identifiers for the data/contributions submitted. If you also have a personal identifier (such as ORCID), your work can be automatically linked to your profile.


6.5. Questions

{{|>}}


Which of these statements are true? Select all correct answers.

  • [[ ]] Published data is always open access.
  • [[ ]] Metadata should be accessible, even if data access is restricted.
  • [[X]] Publishing data in a subject-specific repository will increase data visibility and data citations.
  • [[X]] Publishing data is always free of charge.

What is a repository?

  • [(X)] A digital, curated data center for the storage and reuse of research data
  • [( )] A search engine that makes any research data freely available to everyone without restrictions
  • [( )] A research community that deals with the reutilization of research data
  • [( )] A local copy of all research data collected in a project

That is correct. At the University of Marburg, data_UMR is available to you as a repository.


Which of the following repositories are generic, and which are subject-specific? Assign the repositories to the two categories.

  • [[Generic] [Subject-specific]]
  • [ (X) ( ) ] Zenodo
  • [ (X) ( ) ] figshare
  • [ (X) ( ) ] data_UMR
  • [ (X) ( ) ] RADAR
  • [ (X) ( ) ] Dryad
  • [ (X) ( ) ] Plant Genomics and Phenomics Research Data Repository (e!DAL-PGP)
  • [ ( ) (X) ] PANGAEA
  • [ ( ) (X) ] European Nucleotide Archive (ENA)
  • [ ( ) (X) ] Movebank
  • [ ( ) (X) ] Ocean Biodiversity Information System (OBIS)
  • [ ( ) (X) ] FishBase

7. Sharing Data

Learning objectives

{{|>}}


After completing this chapter, you will be able to...

    {{1}}

...decide if you can publish your data and which permits you might still need

    {{2}}

...know relevant legal frameworks

    {{3}}

...decide what to do with sensitive data

    {{4}}

...choose a licence for your publication


7.1 Introduction

{{|>}}


Why is it important to share and publish data? Let's have a look at this video from GBIF (2022):

!?GBIF - the library of life (2022)

In addition to (text) publication in scientific journals, the data on which publications are based are also increasingly in demand. Many research funders and journals now require explicit data publication. This can provide additional quality assurance and, if other researchers work with your data, you gain reputation through citations.

There are basically three ways of publishing research data:

  1. As an independent publication in a repository (= long-term storage and publication location for data) - preferred choice

  2. As an article in a data journal

    1. These are (usually) peer-reviewed papers that present and describe datasets with a high value for reuse in more detail

    2. The data itself is usually published in a research data repository.

  3. As a supplement to a scientific article (= data supplement) - not the best option

Sharing of data means publishing it and making it accessible for the research community as a whole. Publishing data does not mean making it available for everybody. There are different licences or restrictions, and you can decide to which extent you want to publish your data1. For example, under some circumstances, data may be kept back and are embargoed for a defined amount of time until primary publications are published. The European Code of Conduct for Research Integrity advises making data "as open as possible, as closed as necessary"2.

When a paper is published, it is also a good time to publish the associated data. However, data should be published throughout the whole data life cycle 1, almost independent of a scholarly publication.

Sharing data for reproducing results is essential to science and practice. However, only by publishing data becomes FAIR. Some funders or institutions already have a data sharing commitment (data policy) that is usually impossible to avoid unless there are legitimate reasons 1.

To publish your data, make sure you have the rights for it, keep all ethical, legal or intellectual property restrictions in mind, check the requirements of your funder or institution, use persistent identifiers, so that the data is citable and consider applying for a specific licence 13.


7.2. Legal aspects of research data management

{{|>}}


Legal issues in dealing with research data arise at every stage of the research data life cycle. A very short summary: you are most likely not the only one who has some form of ownership and/or right of use for the data you collected. Depending on e.g.

  • Your employment status

  • Types and sources of data

  • Where and how the data is stored

  • Rights of collaborators

  • Funding and institutional guidelines

  • Federal, national legislation

  • Additional legal agreements between you and other project partner

When handling research data, it's essential to navigate various legal aspects, including but not limited to data protection law, copyright and neighbouring rights, contract law, and federal, national, and international regulations. Depending on the nature of your research project, other legal aspects may come into play. For example, if your research involves inventions, patent law must be considered. Collaboration with companies or contract research may entail contractual agreements, including confidentiality agreements. Additionally, when collecting data on privately owned land, such as agricultural fields, you might encounter personal data concerns, necessitating data anonymization. In some cases, specific projects may require an ethics committee's expert opinion, typically maintained by universities to assess ethical matters in scientific investigations. Professional societies may also have their own ethics committees, although expenses might be associated with the evaluation process.

Before data can be made publicly available, there are many legal aspects to consider -- because not all data can or should be made public. The most important legal aspects are considered in the following decision aid in the form of a flow chart. Answering the questions will guide you through the decision-making process to a recommendation:

Figure 6 is showing the decision process for data publication. Most of the questions that have to be answered with yes or no in order to reach a decision relate to the legal usability of the data or datasets. This concerns, for example, whether the data was collected by third parties or whether it can be anonymised

Essentially, but not exclusively, questions of data protection and copyright must be clarified before publication. The decisive course for the possibility of publishing research data in a repository is therefore often already set when the data are collected, and the corresponding declarations of consent are obtained. If you are unsure, contact the legal department of your university or institute.


7.2.1. What can inhibit publication?

{{|>}}


Not all research data may or should actually be published. Before you decide not to publish your data at all, you should always check whether you can take measures to enable legally and ethically unobjectionable publication.

In addition, you should also take research ethics aspects into account when deciding whether to publish your research data. The following points should give you some criteria without claiming to be complete:

  • Can the data be used in a way that is harmful to society?

  • For example, does publication pose risks to the researched individuals or to collaborators (even if they have consented to the use of their data)?

  • Do participating working group members have legitimate interests in preventing or delaying data publication (e.g. for the completion of qualification work)?

Comprehending these legal obligations is crucial to determine what actions are permissible with the data you've collected. While you need not become an expert in all the intricacies of these legal matters, aligning with good scientific practice and research ethics necessitates a fundamental understanding of potential legal scenarios and challenges. This understanding encompasses national and international obligations and contracts, such as the German Federal Nature Conservation Act (Bundesnaturschutzgesetz), the German environmental information act (Umweltinformationsgesetz, UIG), and international agreements like the Nagoya Protocol on access to genetic resources.


7.2.2. Data protection

{{|>}}


Data protection rights must be observed when collecting, storing, processing, and passing on research data relating to individuals. For instance, if you work as a researcher at a German university with such data, it is advisable to know the main features of the following legal texts in particular:

  • General Data Protection Regulation of the European Union (GDPR) - The GDPR is the world's most stringent privacy and security law, defining fundamental rights for individuals in the digital age, specifying obligations for data processors, and outlining compliance measures and sanctions for rule violations.

  • German Federal Data Protection Act (BDSG) - The BDSG is a crucial legislation in Germany that governs the processing of personal data and upholds data privacy rights. It implements key aspects of the European Union's GDPR and sets out rules for data protection in various sectors within the country.

  • Federal Data Protection Acts - There is a regional law in Germany, specific to the federal state. It regulates data protection and access to government information, ensuring the rights of individuals and promoting transparency in the respective public institutions.

This video series (1-3) published by DARIAH Ethics & Legality (2023) explains data protection in research practice:

!?Data protection in research practice I - Key terms of the GDPR !?Data protection in research practice II -- Principles, rights and responsibilities !?Data protection in research practice III -- The GDPR in science and research

Data without personal reference or anonymised information, on the other hand, do not fall under data protection law and can usually be processed freely, taking into account other rights (e.g. copyrights). What exactly distinguishes personal data from other (anonymous) research data is explained in detail in the following chapter. In case of doubt, you should assume a personal reference with the legal department or data protection officer at your institution to avoid liability risks.


7.2.3. Personal data and special categories of personal data

{{|>}}


According to Art. 4 (1) of the GDPR, personal data is any information relating to an identified or identifiable living person. Examples of personal research data include survey data in the social sciences or health data in medical research. In biological research, the use of camera trapping, automatic audio recordings in bird and bat surveys, and the combination of research data and spatial data of privately owned land are examples where data protection has to be considered (e.g. stratified randomization of the locations of wells).

An identifiable person is one who can be identified directly or indirectly by means of attribution:

  • in particular to an identifier such as a name, an identification number, location data, an online identifier or

  • to one or more particular characteristics that are an expression of the physical, physiological, genetic, mental, economic, cultural, or social identity of that natural person.

The following cases in particular have recently been decided in case law:

  • Images, film, and sound recordings if there is a reference to a person

  • IP addresses

  • Written answers of a candidate in a vocational examination

  • Examiner's comments on the assessment of these answers

To determine whether a person is identifiable, the GDPR requires that account be taken of all the means likely to be used by the data protection officer or by any other person, in normal circumstances (in terms of cost and time), to identify the person (Recital 26 GDPR).

There are also categories of data in case law that are considered particularly sensitive. These include, for example, data on a person's state of health, sexual orientation, and political or religious views. A list of these special categories of personal data can be found in Article 9 of the GDPR.

This data is subject to special protection and special due diligence obligations during processing. This means, for example, that participants in scientific studies must explicitly consent to the processing of these special categories of personal data before the data is collected.


7.2.4. Decision-making authority

{{|>}}


In addition to data protection, another important question is who can decide on the handling of the research data, especially its publication. As a rule, the person to whom the research data are "assigned" can also decide on their handling, such as their publication. Such an "assignment" can result from copyright law, service contract law or patent law, for example.


7.2.4.1 Copyright and ancillary copyright law

{{|>}}


Copyright is a legal concept that grants creators the exclusive right to control the reproduction and distribution of their original works for a specified period. It aims to protect the intellectual property of authors and creators by preventing others from using their work without permission. As a rule, the protectability of individual research data under copyright law can only be assessed on a case-by-case basis regarding the underlying level of creativity and even then, not with sufficient legal certainty. Nevertheless, different case groups of research data can be distinguished according to the concrete type of content and, above all, how it was obtained:

  • Quantitative, raw data, for example unprocessed measurement results or statistical data, are not copyright protected.

  • Processed (quantitative) research data, the arrangement and compilation of which has the effect of establishing individuality, is a so-called database work (Section 4 UrhG1. Only its structure and not the information as such is subject to copyright protection. Furthermore, compilations of research data within the framework of a database can be protected by the database producer right (§87a UrhG). This ancillary copyright requires a substantial investment in terms of collecting, organising, and making research data accessible. The owner of the database producer rights is usually the person who makes the essential investments, e.g. pays the researchers' remuneration and bears the economic risk. Generally, this is also the employing university or research institution. In some cases, a third-party commissioning or funding institution may also be the owner.

  • Metadata: often are relatively short, purely descriptive representations. They are usually not protected by copyright. In principle, they can only be protected in the rare cases where they contain, for example, longer sections of text or photographs.

  • Creative work, such as manuscripts, drawings, plans, maps, sketches, software and tables, are usually subject to copyright protection.

  • Photographs and other images from imaging procedures, such as microscopic, fluorescence images are included, as well as photographs and individual images from films are usually subject to copyright protection. Furthermore, photographs and other photographic images may also be protected by a neighbouring right under Section 72 UrhG.

In the case of non-protected research data (e.g. measurement results), it can be unclear from a legal point of view who has the decision-making authority over the data in a specific individual case. Whether a possible personal right of the scientist also allows an assignment of the research data to a person in these cases is disputed. There may also be rights of the employing institution. Please make sure you discuss this topic early on and openly with your co-workers and supervisors to avoid conflict later on. A data management plan helps to define each contributor's rights of use of project data before any data is collected.


7.2.5 Licensing of research data

{{|>}}


Licences play a pivotal role in enhancing transparency regarding the ways in which copyright-protected materials can be used, fostering clarity and facilitating lawful and collaborative sharing. To enable others to use your copyright-protected material, it's essential to establish clear conditions of use, often through the use of a licensing framework. Without a licence, copyrighted protected material may only be utilised with the explicit permission of the copyright holder. In the fields of biodiversity and environmental science, interdisciplinary collaboration plays a central role, resulting in a multitude of creators or contributors. It's important to ensure the active participation of all creators when licensing copyright protected materials. In case of any uncertainties, do not hesitate to seek guidance from your institution's designated point of contact. In the following, their are some recommendations based on common, internationally approved licences for various data types:


7.2.5.1 Creative content

{{|>}}


Creative Commons licences provide a flexible framework for sharing and using creative work (e.g. text, articles, images, figures; see Figure 7). Just as the European Commission in its project Horizon 2020, the DFG recommends the use of these licence types. When deciding on a specific licence, the guiding principle is "as open as possible, as restrictive as necessary". The CC0 licence offers complete freedom, allowing users to use the work without any restrictions, often referred to as "No rights reserved." On the other hand, CC BY permits free use but requires attribution to the original creator, ensuring proper credit for their work. These licences enable a balance between openness and recognition in the sharing of creative content.

The use of further licence modules is not recommended. For example, the Creative Commons (CC) licenses with the attribute "ND" (e.g. CC-BY-ND) rule out the distribution of "modified" material. This would make it impossible to make a new database publicly available that was created from parts of other databases12. The use of a "NC" licence may seem attractive, because many want to prevent commercial exploitation of their research data. However, the term "commercial" is defined in a very broad sense in the CC licences, resulting in major legal uncertainties about which uses are allowed and which are prohibited34. For example, if you are exclusively employed by an exclusively publicly funded institution (third-party funding, even from public sources, is already a grey area), you may use CC BY-NC licensed material. However, if you want to publish research that is based on or involves CC-BY-NC licensed material, you cannot publish in any commercial journal (all publishers are commercial, even if they are not-for-profit). Even publishing that research in a private blog that e.g. uses GoogleAds to cover server costs is prohibited. Even any increase in reputation may be considered commercial - and reputation is a major factor in scientific careers. For details, see Klimpel (2012)3 and Kreutzner & Deutsche UNESCO-Kommission (2014)4.

The image is an infographic that illustrates a spectrum of copyright licensing, from the most open to the least open. At the top of the spectrum, indicated as "most open," there is a Public Domain (PD) symbol, followed by icons representing different Creative Commons (CC) licenses. The Public Domain mark (PD) is a circle with a slanted line through it, symbolizing unrestricted use. Below that are the various Creative Commons licenses, each represented by their specific symbols: CC0 (No copyright restrictions) CC BY (Attribution required) CC BY-SA (Attribution required, ShareAlike) CC BY-NC (Attribution required, Non-commercial) CC BY-NC-SA (Attribution required, Non-commercial, ShareAlike) CC BY-ND (Attribution required, No derivatives) CC BY-NC-ND (Attribution required, Non-commercial, No derivatives) These licenses are shown in circles with the respective abbreviations and symbols indicating the conditions of use: a person for attribution (BY), a dollar sign for no commercial use (NC), two arrows in a circle for ShareAlike (SA), and an equal sign in a circle for no derivatives (ND). At the bottom of the spectrum, labeled as "least open," is the traditional copyright symbol, a "C" within a circle, with the words "all rights reserved" next to it, and a bar indicating a gradient from green at the top (most open) to red at the bottom (least open). The background is divided into sections of light green, cream, and grey, each section darker as the licenses become less open. The graphic serves as a guide to understanding the levels of permissions and restrictions associated with different types of copyright licenses.)


7.2.5.2 Database material

{{|>}}


Database material generally refers to the data and content stored within a database. This can include a wide range of information, such as text, numbers, images, audio files, or any other type of structured or unstructured data from different sources that is organised and stored within a database system. Open Data Commons licences offer an alternative framework for sharing database material. The Public Domain Dedication & License (ODC-PDDL) is similar to CC0, granting complete freedom to use data without any restrictions, essentially declaring it "No rights reserved." For those seeking attribution, the Attribution Licence (ODC-BY) allows free use but requires acknowledgment of the data source. An additional recommendation by the DFG (German Research Foundation) is the use of CC BY 4.0, aligning with the international CC BY framework, which encourages open data sharing with proper attribution. These licences serve to promote transparency and accessibility in data sharing while accommodating different preferences for attribution.


7.2.5.3 Software

{{|>}}


Software, unlike much other research data, require separate licences that are specifically designed for sharing software. The use of Creative Commons licences is not recommended for this. Various open-source licenses offer different terms for sharing and using software. The General Public License (GNU GPLv3) allows free use while mandating that any reusing or publishing must occur under the same license. It also requires the indication of changes made to the code. The Massachusetts Institute of Technology License (MIT) permits free use, but attribution is necessary. The Mozilla Public License (MPL) goes a step further, necessitating the disclosure of changes to the source code and sharing of a narrower set of code under the same license, with all three licenses emphasizing transparency and proper credit in software sharing.


7.2.5.4 Raw data and metadata

{{|>}}


These data are not typically protected by copyright. By using tools like CC0 and the Public Domain Mark, you not only release these resources for unrestricted use but also underscore the absence of copyright restrictions. This practice serves to emphasise and promote transparency, facilitating the exchange and collaborative use of data and metadata among researchers and the wider community. CC0 is a legal tool that allows creators to waive all copyright and related rights in a work, essentially granting it to the public domain. This means anyone can use the work for any purpose without restrictions. On the other hand, the Public Domain Mark signifies that a work is already in the public domain and free from copyright. While both CC0 and the Public Domain Mark serve the purpose of promoting open access to creative works, CC0 is a proactive choice by creators, whereas the Public Domain Mark is applied to works that have already entered the public domain.

In a scientific context, the combination of data from different sources is a common and important part of research. Making data interoperable is supposed to facilitate re-use and re-combination. However, not all CC licences are compatible (Figure 8). Therefore, it is important to choose an open licence like CC0 or CC BY 4.0.

Figure 8. is the CC licence compatibility chart by Kennisland, CC0. It is a comparison of the different types of open source licences, as described in Figure 7. And thus shows which types are compatible with each other and which are not


7.2.6. Granting of rights of use within the framework of service and employment contracts

{{|>}}


If the creation of copyright-protected work is one of the duties or central tasks of the employment contract such as writing journal articles, the employer is granted rights of use to these so-called "compulsory works" on the basis of the employment contract or employment relationship 1.

The following "mapping" of research data results from the balance of interests with the freedom of research (Art. 5 para. 3 GG (German Basic Law)):

  • As a rule, university teachers are entitled to all rights of exploitation, use and publication of the works they have created, unless there are express contractual agreements (e.g. third-party funding, non-disclosure agreements). § Section 43 UrhG (so-called "compulsory works") does not apply here.

  • Scientific assistants and employees have certain rights outlined in Article 5(3) of the German Basic Law (GG) when they carry out their research independently, without being told what to do. If they conduct their research under specific instructions, it's assumed that they've given permission for the use of the research data they create.

  • In the case of students and external doctoral candidates, no rights of use are granted to the university, as they are not employees. However, different contractual agreements can be made, e.g. in the case of third-party funded projects, through which the university is granted rights of use.

The following Figure 9 illustrates the issues of the transfer of exploitation rights to the employer ("compulsory work" under Section 43 UrhG) and the balancing of interests with the freedom of research (Article 5(3) GG) according to roles as they are to be weighed in the scientific field in individual cases:

Figure 9. describes the ownership of research data throughout the title question "transfer of rights to the employer?". It tells us for example, that the basis is the employment contract and that its only the use of rights that are transferrable. Source: Brettschneider (2020), CC BY 4.0, translation by HeFDI).

It should be noted that the granting of rights of use within the framework of service and employment contracts may also be tacit if the granting of rights of use is not expressly regulated in the contract. Within the framework of the (tacit) granting, the scientist also leaves the right to determine to the employer whether and how the work is published. On the other hand, each scientist retains their right to be named.


7.2.7 National and international legal frameworks

{{|>}}


When conducting field work, it's important to secure the necessary permissions to collect various types of samples and, if needed, transport them out of the country. These permissions can be obtained from local authorities, such as the local environmental agency or landowners, as well as national authorities like the German environmental protection agency. Additionally, when dealing with protected species or internationally regulated activities, compliance with laws like the Naturschutzgesetz (Nature Conservation Act) and CITES (Convention on International Trade in Endangered Species of Wild Fauna and Flora) is crucial. Moreover, even laboratory-based research may require attention to international laws, such as the Nagoya Protocol, if the samples originate from another country.

We will give a short overview of some of the most important laws in the following chapter.


7.2.7.1 Protected species

{{|>}}


Protected species are subject to strict regulations aimed at conserving their populations. The Convention on International Trade in Endangered Species of Wild Fauna and Flora (CITES) is an international agreement that plays a pivotal role in safeguarding these species. In Germany, the German Federal Agency for Nature Conservation (BfN) offers the online database WISIA, which comprehensively covers all species protected under German national or international laws.

When considering how this relates to scientific work, it's important to note that even researchers and educators may need specific permits to work with protected species, as simply being a scientist or engaging in environmental education may not automatically exempt one from regulatory requirements.


7.2.7.2 The environmental information act

{{|>}}


According to the German Federal Environmental Information Act (Umweltinformationsgesetz, UIG), implemented 2005, and the environmental information laws of the federal states, every person has a right of access to information relating to the environment, e.g. held by public authorities. Access is granted upon request, but may be refused for certain reasons (§ 8 and 9 UIG) to protect public or other interests1. Researchers may therefore request access to environmental data collected by public authorities or private companies, but access may be refused e.g. when they concern trade secrets.

Environmental information includes data on the state of air, atmosphere, water, soil, landscape and natural habitats, as well as information on noise, energy, substances or radiation. But citizens can also obtain environmental information about plans and programs that actually or potentially affect the environment, about the implementation of environmental law or cost-benefit analyses of environmental projects12.


7.2.7.3 Nagoya protocol and digital sequence information (DSI)

{{|>}}


The Nagoya Protocol on Access to Genetic Resources and the Fair and Equitable Sharing of Benefits Arising from their Utilization is a legal framework under the United Nations Convention on Biological Diversity (CBD), which formally recognizes countries' sovereign rights over their genetic resources and the associated traditional knowledge. The Nagoya Protocol entered into force on 12 October 2014 and aims at sharing the benefits arising from the utilisation of genetic resources in a fair and equitable way.

!?This video by Naturvårdsverket (2018) explains the Nagoya Protocol.

Short, countries which ratified the Nagoya Protocol regulate access to their genetic resources by requiring users to obtain prior informed consent (PIC) and accept mutually agreed terms (normally referred to as MAT) that ensure fair and equitable benefit-sharing. This arrangement, known as Access and Benefit Sharing (ABS), aims that a portion of the advantages or profits derived from the use of genetic resources is shared with the providing country1.

For more information on the implications of the Nagoya Protocol in your research activities (i.e., when collecting samples abroad) and your obligations, we recommend you to access the German Nagoya Protocol Hub.

The genetic resources provided by a country can be decoded and digitally archived as DNA sequence data, known as digital sequence information (DSI) in policy circles2. DSI, and its open-access in databases, has revolutionised life sciences and became the foundation of modern-day biodiversity research and technological innovation in fields as diverse as medicine, food security, and green energy production3. Until recently, there were no agreements on the access to the DSI generated from the sovereign genetic resources because the use of DSI is unconnected from physical access and related permits, and financial advantages from DSI via commercialization do not return to the provider countries -- many of which are low- and middle-income countries (LMICs). DSI is seen as a flaw that prevents the fair and equal distribution of benefits, violating their sovereign rights3. During the Conference of the Parties under the Convention on Biological Diversity (COP 15), in December 2022, the parties endorsed an agreement on DSI benefit sharing. The decision calls for a multilateral mechanism which is decoupled from the access and use of DSI, aiming to ensure the open-access to DSI as we know. How the mechanism will actually be implemented, however, is still under debate and to be further discussed. In order to provide the scientific community a voice in worldwide policy discussions on DSI, the DSI Scientific Network was established. Their main goal is to bring awareness to policymakers and other stakeholders on the global advantages of open-access to DSI through public online databases and the importance to research, biodiversity conservation, and public health.


7.3 Questions

{{|>}}


Which of the following statements is NOT true? Choose the wrong answer.

  • [( )] The structure of a database is protected by copyright law in Germany.
  • [( )] The owner of the database producer rights is usually the institution employing a researcher.
  • [(X)] Quantitative, raw data, are copyright protected due to the investment to produce the data.
  • [( )] There are specific licenses to share database material.

What is the preferred choice for publishing research data? Sort the following data publication strategies from most accessible to least accessible

[[ (Publishing data as an independent publication in a subject-specific repository under an open license) | Publishing data as an article in a data journal | Publishing data in a generic repository |Publishing data as a supplement to a scientific article | Publishing data in a subject-specific repository under an indefinite embargo | Publishing data as a supplement to a scientific article as a text table in a pdf ]] [[ Publishing data as an independent publication in a subject-specific repository under an open license | (Publishing data as an article in a data journal) | Publishing data in a generic repository |Publishing data as a supplement to a scientific article | Publishing data in a subject-specific repository under an indefinite embargo | Publishing data as a supplement to a scientific article as a text table in a pdf ]] [[ Publishing data as an independent publication in a subject-specific repository under an open license | Publishing data as an article in a data journal | (Publishing data in a generic repository) |Publishing data as a supplement to a scientific article | Publishing data in a subject-specific repository under an indefinite embargo | Publishing data as a supplement to a scientific article as a text table in a pdf ]] [[ Publishing data as an independent publication in a subject-specific repository under an open license | Publishing data as an article in a data journal | Publishing data in a generic repository | (Publishing data as a supplement to a scientific article) | Publishing data in a subject-specific repository under an indefinite embargo | Publishing data as a supplement to a scientific article as a text table in a pdf ]] [[ Publishing data as an independent publication in a subject-specific repository under an open license | Publishing data as an article in a data journal | Publishing data in a generic repository |Publishing data as a supplement to a scientific article | (Publishing data in a subject-specific repository under an indefinite embargo) | Publishing data as a supplement to a scientific article as a text table in a pdf ]] [[ Publishing data as an independent publication in a subject-specific repository under an open license | Publishing data as an article in a data journal | Publishing data in a generic repository |Publishing data as a supplement to a scientific article | Publishing data in a subject-specific repository under an indefinite embargo | (Publishing data as a supplement to a scientific article as a text table in a pdf) ]]

What role do persistent identifiers (PIDs) play in data retrieval and findability? Choose the correct answer.

  • [( )] They provide encryption for data.
  • [(X)] They link metadata to the actual data set.
  • [( )] They limit access to confidential data.
  • [( )] They are used for short-term archiving.

Why are open and documented data formats preferred for long-term data storage? Choose the correct answer.

  • [( )] They are less compatible with various systems.
  • [( )] They require frequent migrations.
  • [(X)] They facilitate interoperability and have a longer lifespan.
  • [( )] They have higher compression ratios.

8. Resources

8.1 Central resources

{{|>}}



8.2 References

{{|>}}



8.3 Lists of metadata standards, tools and repositories

Table 2. List of metadata standards

{{|>}}


Please note that this list only shows a small selection of available resources.

Abbreviation Name Description
ABCD Access to Biological Collection Data For: biodiversity data
Dublin Core For: any data
Darwin Core For: biodiversity data
EML Ecological Metadata Language For: ecological data; Data format: XML
INSPIRE Infrastructure for Spatial Information in Europe For: spatial data sets
ISO 19115 For: geosciences; Data format: Dublin Core
MIABIS Minimum Information About BIobank data Sharing For: biobanks, research on samples and associated data
MIAME Minimum Information About a Microarray Experiment For: microarrays; Data format: MAGE-TAB format (MicroArray Gene Expression Tabular)
MIAPPE Minimum Information About Plant Phenotyping Experiments For: plant phenotyping experiments
MIAPPE checklist and data model v.1.1
MIxS Minimum Information about any (x) Sequence For: sequence data
SDD Structured Descriptive Data For: descriptive data

Table 3. List of tools for data annotation and metadata creation

{{|>}}


Please note that this list only shows a small selection of available resources.

Name Metadata standards Description
CEDAR workbench divers, incl. ADVANCE project biological data (ADVANCE: biodiversity survey and monitoring data); GUI; online platform; registration necessary; free of charge
ezEML EML ecological data; GUI; online platform; registration necessary; free of charge; developed and maintained by the Environmental Data Initiative (EDI); developed for data submissions to the EDI repository
GBIF spreadsheet templates Darwin Core biodiversity data; spreadsheet (.xlsx); download; free of charge
templates for occurrence data, checklist data, sampling event data and resource metadata
GFBio spreadsheet templates Darwin Core biodiversity data, incl molecular data; spreadsheet (.txt, .xls, .xlsx); download; free of charge; by German data centers (BGBM, DSMZ, MfN, SGN, SMNS, SNSB, LIB)
ISA framework ISA-TAB, ISA-JSON, ISA Python API life science, environmental and biomedical experiments; online plattform & local installation; registration necessary; open source; free of charge; for Windows, Linux, macOS
MIAPPE MIAPPE plant phenotyping experiments; templates & tools; online platform; registration necessary; open source; free of charge
Notepad++ any, XML any data; simple text; local installation; open source; free of charge; for Windows; developed and maintained by Don Ho & user community
oXygen any, XML any data; simple text; local installation; proprietary; monthly pricing; for Windows, macOS, Linux, Eclipse; developed and maintained by SyncRO Soft SRL
RightField divers ontologies; spreadsheets; local installation; open source; free of charge; for Windows, Linux, macOS; developed and maintained by the University of Manchester and HITS gGmbH

Table 4. List of repositories and search tools for biological data

{{|>}}


Information on GFBio Data Centers can be found in table 5. Please note that this list only shows a small selection of available resources.

Name Type of data Role Description
German Federation for Biological Data (GFBio) subject-specific search & submit, aggregator For biodiversity data; data curation; free of charge
cooperation of German and European archives, collections and repositories
Global Biodiversity Information Facility (GBIF) subject-specific search & submit, aggregator For biodiversity data; data submission only through GBIF partner institutions; no curation for individual data sets; free of charge
Ocean Biodiversity Information System (OBIS) subject-specific search & submit For marine biodiversity data; free of charge
Plant Genomics and Phenomics Research Data Repository (e!DAL-PGP) subject-specific search & submit For plant genomics & phenomics research data; data curation; free of charge
Movebank subject-specific search & submit,plattform For data sets and analysis workflows for animal movement data; data curation and review; free of charge
Python Package Index (PyPI) software search & submit For Python code; free of charge
Zenodo generic search & submit For any type of data; no curation; free of charge
Dryad generic search & submit For any type of data; data curation; costs depending on status of home institution
RADAR generic search & submit For generic data; no curation; costs depending on home institution, services, space;DINI certificate
re3data Index of research data repositories
FAIRsharing Index of research data repositories
Amphibian Species of the World (ASW) subject-specific search For amphibian taxonomic data; free of charge
Avibase subject-specific search, aggregator For bird data; database of taxonomic concepts rather than species, mapped and cross-referenced major taxonomies
BacDive
The Bacterial Diversity Metadatabase (by DSMZ)
subject-specific search, aggregator For bacterial data; free of charge
Catalogue of Life (COL+) subject-specific search, aggregator For taxonomic data, all kingdoms; registration necessary for some functions; free of charge
ITIS and Species 2000/Catalogue of Life (CoL) provide the taxonomic backbone to the Encyclopedia of Life (EOL)
ChecklistBank (CLB) subject-specific search & submit, aggregator For taxonomic data; registration necessary for some functions; free of charge
Maintained by Catalogue of Life (COL+) and GBIF
Encyclopedia of Life (EoL) subject-specific search, aggregator For biodiversity data, incl. traits and interactions; registration necessary for some functions; free of charge
FishBase subject-specific search & submit For data on finfish, incl. traits, biology, distribution, etc.; free of charge
observations can be uploaded directly via FishWatcher
Integrated Taxonomic Information System (ITIS) subject-specific search For taxonomic data; free of charge
ITIS and Species 2000/Catalogue of Life (CoL) provide the taxonomic backbone to the Encyclopedia of Life (EOL)
International Nucleotide Sequence Database Collaboration (INSDC) subject-specific search, aggregator The International Nucleotide Sequence Database Collaboration (INSDC) is a long-standing foundational initiative that operates between DDBJ, EMBL-EBI and NCBI.
IUCN subject-specific search For risk assessments of species; registration necessary for some functions; free of charge
Map of Life (MOL) subject-specific search, aggregator For georeferenced biodiversity data; registration necessary for some functions; free of charge
World Register of Marine Species (WoRMS) subject-specific search For taxonomic data of marine species; registration necessary for some functions; free of charge
official taxonomic reference list for OBIS

Table 5. List of GFBio Data Centers

{{|>}}


These ten GFBio Data Centers are associated with GFBio e.V. Data submission and search can be done with all of them. Physical objects can be given to some of them and only after consulting the responsible curators of the physical collections. GFBio defined five major types of biological data which are recognised by the Data Centers in NFDI4Biodiversity. They are used for the "Service Description" of the individual Data Centers as well as in the context of the technical documentations of processing tools:

  • Type 1: Biodiversity and Occurrence data
    These are the data from the classical collection and alpha-diversity research domain, i.e. digital objects with taxon name(s), georeferences, e.g. locality, date and often referenced resources as multimedia objects. We distinguish between:
  • Type 1a: Collection Data (with reference to physical object)

  • Type 1b: Observation Data (without reference to physical object)

  • Type 2: Taxon Data
    These are taxon-related data (e.g. in a catalogue, checklist or so-called red list).

  • Type 3: Environmental Biological and Ecological Data
    These are environmental biological and ecological study data including functional and phylogenetic trait data and other kind of analysis data.

  • Type 4: Non-Molecular Analysis Data
    These are non-molecular analysis data (data sets and/or data packages) in its original data file format (often RAW format).

  • Type 5: Molecular Sequence Data
    These are molecular sequence data including MIxS-compliant metadata.

Data Center Abbreviation Types of data Physical collections Description
Botanischer Garten und Botanisches Museum Berlin BGBM 1a+b, 2, 4 X For botanical and fungal data and (selected) specimens; data and sample curation; access to data free of charge
search in specific databases, or via GFBio, GBIF, etc.
Plant Genomics and Phenomics Research Data Repository e!DAL-PGP 3, 4 For plant genomics & phenomics research data; data curation; free of charge
European Nucleotide Archive ENA 5 For sequencing data; free of charge
Leibniz Institute DSMZ-German Collection of Microorganisms and Cell Cultures DSMZ 1a+b, 2 X For data and samples of microorganisms, human and animal cell lines, viruses; data and sample curation; access to data free of charge, costs for access to and for deposition of samples depending on types of samples, services, etc.
Leibniz Institute zur Analyse des Biodiversitäts Wandels LIB 1a+b, 2, 3 X For botanical, palaeontological and zoological data and (selected) specimens; data and sample curation; access to data free of charge
search in specific databases, or via GFBio, GBIF, etc.
Museum für Naturkunde Berlin Leibniz Institute for Research on Evolution and Biodiversity MfN 1a+b, 2 X For ecological, palaeontological and zoological data and (selected) specimens; data and sample curation; access to data free of charge
search in specific databases, or via GFBio, GBIF, etc.
PANGAEA Data Publisher for Earth & Environmental Science PANGAEA 3, 4 For earth science, environmental science, biodiversity data; data curation; free of charge;CoreTrustSeal,DINI certificate
Senckenberg Leibniz Institution for Biodiversity and Earth System Research SGN 1a+b, 3 X For botanical, fungal, palaeontological and zoological data and (selected) specimens; data and sample curation; access to data free of charge
search in specific databases, or via GFBio, GBIF, etc.
Staatliches Museum für Naturkunde Stuttgart SMNS 1a+b, 2 X For botanical, fungal, palaeontological and zoological data and (selected) specimens; data and sample curation; access to data free of charge
search in specific databases, or via GFBio, GBIF, etc.
Staatliche Naturwissenschaftliche Sammlungen Bayerns SNSB 1a+b, 2, 3, 4 X For botanical, fungal, palaeontological and zoological data and (selected) specimens; data and sample curation; access to data free of charge
search in specific databases, or via GFBio, GBIF, etc.

8.4 Questions

{{|>}}


Assign the terms to the FAIR principles.

Definition Term
[[ Findable Accessible
[[ (Findable) Accessible
[[ (Findable) Accessible
[[ Findable (Accessible)
[[ Findable Accessible

RDM1 introduces you to many new abbreviations. Which of the following abbrevations belongs to which term? 1 Research data management

  • [(DFG) (ORCID) (ABCD) (DwC) (GFBio) (GBIF) (DMP) (DSI) (DOI)]
  • [ (X) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ] Deutsche Forschungsgemeinschaft
  • [ ( ) (X) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ] Open Researcher Contributor Identification Initiative (full name obsolete)
  • [ ( ) ( ) (X) ( ) ( ) ( ) ( ) ( ) ( ) ] Access to Biological Collection Data
  • [ ( ) ( ) ( ) (X) ( ) ( ) ( ) ( ) ( ) ] Darwin Core
  • [ ( ) ( ) ( ) ( ) (X) ( ) ( ) ( ) ( ) ] Gesellschaft für Biologische Daten (German Federation for Biological Data)
  • [ ( ) ( ) ( ) ( ) ( ) (X) ( ) ( ) ( ) ] Global Biodiversity Information Facility
  • [ ( ) ( ) ( ) ( ) ( ) ( ) (X) ( ) ( ) ] Data Management Plan
  • [ ( ) ( ) ( ) ( ) ( ) ( ) ( ) (X) ( ) ] Digital Sequence Information
  • [ ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) (X) ] Digital Object Identifier

Footnotes

  1. American Museum of Natural History. (2023). What is Biodiversity? Available at: https://www.amnh.org/research/center-for-biodiversity-conservation/what-is-biodiversity. Last accessed 24 November 2023. 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34

  2. Biodiversity Data Journal. (2023). Biodiversity Data Journal. Available at: https://bdj.pensoft.net/. Last accessed 24 November 2023 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23

  3. National Geographic. (2023). Biodiversity. Available at: https://education.nationalgeographic.org/resource/biodiversity. Last accessed 24 November 2023.) 2 3 4 5 6 7 8 9 10 11 12

  4. Wilkinson, M.D., Dumontier, M., Aalbersberg, Ij.J., Appleton, G., Axton, M., Baak, A., et al. (2016). The FAIR Guiding Principles for scientific data management and stewardship. Sci Data, 3, 160018. http://"]https://doi.org/10.1038/sdata.2016.18 2 3 4 5 6 7 8 9

  5. Carroll, S.R., Garba, I., Figueroa-Rodríguez, O.L., Holbrook, J., Lovett, R., Materechera, S., et al. (2020). The CARE Principles for Indigenous Data Governance. Data Science Journal, 19, 43. https://doi.org/10.5334/dsj-2020-043 2 3 4 5

  6. DFG. (2022). Guidelines for Safeguarding Good Research Practice. Code of Conduct. https://doi.org/10.5281/zenodo.6472827 2

  7. Members of the Working Group on Data Management of the DFG Senate Commission on Biodiversity Research. (n.d.). Guidelines on the Handling of Research Data in Biodiversity Research. https://www.dfg.de/download/pdf/foerderung/grundlagen_dfg_foerderung/forschungsdaten/guidelines_biodiversity_research.pdf