STATUS: READY FOR REVIEW
Data locked away benefits no one, but when data is shared responsibly and carefully with bright minds everywhere, we get results that will give us all a healthier future. (UK Biobank is safely sharing health data to drive medical research)
Before data can be made FAIR, we first need to get it the right way. That means finding trustworthy sources, making sure we’re allowed to use the data, and keeping it safe and private. When we do this properly, it ensures that the data can be trusted, shared, and reused to support new research and innovation.
Short description
Data must be acquired responsibly and efficiently. This includes identifying how to access and retrieve the data and how to ensure the data meets legal, ethical, and technical requirements. This page outlines the steps and considerations involved in acquiring data.
Why is this step important
Correct and responsible data access and retrieval ensures you:
- Comply with regulations. You confirm that the data can be used legally and ethically, meeting requirements such as GDPR.
- Safeguard data integrity. Secure transfer methods and integrity checks (e.g. checksum) make sure the data are complete and unchanged.
- Ensure controlled access. By following agreed procedures, you make it possible for authorised users to obtain the data under the correct conditions.
- Enable reproducibility. Documenting where the data came from and how it was retrieved allows the process to be repeated if needed (e.g. queries, APIs, permissions or filtering steps used). This ensures that others can repeat the retrieval process accurately, making the data truly reproducible.
How to
Step 1 - Identify the data source
Locate a trusted source that holds the dataset of interest. This may include:
- Certified or trusted data repositories. These can be subject-specific, (e.g. BBMRI-ERIC Sample and Data Portal, institutional (e.g. Radboud Data Repository) or national (e.g. National Health Data Portal)
- Electronic health or patient record systems
- Biobanks or cohort studies with structured datasets (e.g. lifelines collection)
- Institutional research data archives (e.g. UMCG research data catalogue)
Verify that the source provides persistent identifiers, rich metadata, and clear licensing information.
When possible, choose sources that:
- Provide persistent identifiers (PIDs) for the dataset (FCB046 - Identify resolution services)
- Supply rich metadata describing the dataset’s content, provenance and license
- Are indexed in trusted registries or catalogues
- Have a clear record of data quality (e.g. accuracy, completeness, consistency and timelines, which indicate that the data are reliable for analysis) and make sure the source provides data in a FAIR-format (e.g. data have persistent identifiers, metadata completeness, standard formats).
If the source does not have a PID, for instance in case of extractions of electronic health or patient records, you can share the queries used to extract the data. In that case, others can use the queries to replicate the dataset.
Step 2 – Determine access requirements
Check if access is open, restricted, or controlled. The type of data often determines the access level: very sensitive or personally identifiable data are more likely to require restricted or controlled access.
- Open access. Data can be retrieved without authentication, often under open license (e.g. CC-BY, CC0). Typically includes aggregated, anonymised, or non-sensitive datasets.
- Restricted access. Requires an application or account registration, possibly with usage conditions (e.g. academic/non-commercial). Often applied to datasets containing sensitive information that has been anonymised or pseudonymised. Restricted datasets usually have a license and a clear application process (e.g. data request form or email contact). Costs or institutional approvals may apply, so plan for these in advance.
- Controlled access. Requires formal approval and legal agreements before use, often involving ethics board or data access committee clearance. This is typically for highly sensitive data, such as identifiable human health records or genomic data, where strict oversight ensures privacy and compliance with regulations.
Once you found a dataset, you need to check the license to see if you can actually reuse the data. For more information on data licenses, see here.
Depending on the access level and sensitivity, you may need:
- Institutional approval
- Ethics board or data access committee clearance
- Legal agreements such as Data Agreements (e.g. Processing, Transfer, Usage, Sharing)
If you would like to know more about Access conditions and how they are defined, please refer to Metroline Step: Define access conditions.
Step 3 – Choose a retrieval method
Select an appropriate method, this may depend on dataset’s size, sensitivity and intended use:
- Application Programming Interfaces (developing FAIR API’s for data access). Automated and scalable, ideal for frequent or programmatic retrieval. This can be REST, GraphQL or other API types, often returning JSON, CSV or RDF. APIs are ideal for automation and integrating retrieval into workflows.
Some data infrastructures support hybrid access, where metadata or filtered results are retrieved via an API, and the corresponding files or bulk data are downloaded separately (e.g. via HTTP, cloud storage or FTP). This approach is useful when querying large datasets to identify subsets before initiating bulk transfer, optimising both performance and bandwidth. - Query language access. Some repositories allow data to be retrieved by running queries directly against a database or knowledge graph (FCB070 - FAIR and Knowledge graphs).
- SQL. Used for structured data in relational databases.
- SPARQL. Used for querying RDF/linked data endpoints, enabling retrieval of highly specific variables, entities or relationships (FCB040 - Exploring data with SPARQL).
- Web interface. Manual downloads from portals, useful for small datasets or exploratory use. These direct downloads are typically suitable for small-to-medium datasets where scalability is not required. They offer a quick and intuitive way to inspect data structure and content before deciding whether a large-scale or automated retrieval is needed. Example portals which offer such an interface include:
- GEO (Gene Expression Omnibus) for genomic data
- BBMRI-ERIC Directory for Biobank metadata
- dbGaP for genomic + phenotype data
- Zenodo for general research data
- Secure file transfer. For large and sensitive datasets, use secure and authenticated transfer tools, e.g.:
Consider the retrieval method recommended by the data provider to ensure compatibility and compliance.
For datasets hosted on large-scale cloud platforms (e.g. Azure), data access may be provided through dedicated services like object storage (e.g.Azure Blob), cloud-hosted APIs, or cloud data warehouses. These platforms support scalable, high-performance data access and are often used for storing large genomics, imaging, or real-world data. Depending on the configuration, you may retrieve data using authenticated URLs, SDKs, or cloud-native tools (e.g. aws s3 cp, gsutil, or Azure CLI). Always ensure you understand the access permissions, egress costs and security settings when retrieving data from cloud environments.
Step 4 – Data retrieval and transfer considerations
Consider the following:
- File format and structure. Determine whether files are in CSV, JSON, RDF, XML, HDF5, imaging formats or other, as this will affect processing; Some formats (e.g. many small files vs. one large archive, plain text vs. compressed formats (.zip, .tar.gz)) can affect download time, transfer reliability, and whether special tools are needed to retrieve or unpack the data.
Representation may also vary:
- Flat tables (e.g. CSV, TSV) are commonly used for structured data and are easy to process in spreadsheets or statistical tools.
- Graph-based formats (e.g. RDF, JSON-LD) represent complex relationships between entities and are ideal for semantic data or knowledge graphs. The choice of representation will affect downstream integration and analysis.
- Size and speed of transfer. Large files may require dedicated bandwidth or scheduled transfer.
- Retry/resume capabilities. (for bulk transfer for large files e.g. imaging or omics datasets) - For multi-gigabyte datasets such as medical or omics datasets, use tools that can resume interrupted transfers instead of restarting (e.g. Aspera Connect, Globus, rsync, wget/curl with resume flags). This saves time and reduces the risk of incomplete downloads.
- Encryption in transit. Sensitive datasets should always be encrypted during transfer with access controls.
- Remote access without transfer. In some cases, data cannot be moved due to size or sensitivity. Access may instead occur via secure remote environments (e.g. virtual machines, data enclaves, or cloud-based analysis platforms) where analysis is performed without downloading the data locally. Logging and documentation of the retrieval process still apply.
Keep logs of the retrieval process, including timestamps, tools used, and any transfer issues.
Step 5 – Validate and store safely
Verify data integrity and store it in a secure, access-controlled environment with appropriate metadata.
After retrieval, you should:
- Verify integrity. Use checksums to confirm that the files match the originals.
- Ensure privacy safeguards. When handling sensitive personal data, confirm that appropriate privacy-preserving techniques such as pseudonymisation (removal of direct identifiers) or anonymisation (irreversible data masking) have been applied in compliance with GDPR or local regulations.
- Store securely. Place data in an access-controlled environment that complies with institutional and legal standards. If using sensitive information, make sure you use proper encryption.
- Preserve metadata. Place dataset documentation, provenance records and retrieval notes alongside the data.
- Back up appropriately. Follow institutional or project-level backup policies.
- Register (meta)data in public registries. To support findability, consider registrering the (meta)data in open platforms (FCB060 - Registering Datasets in Wikidata).
A checksum is a short digital code that works like a fingerprint for a file, letting you verify that the data hasn’t changed or been corrupted during transfer.
Expertise requirements for this step
You may need access to or support from:
- Data Stewards. Ensure proper metadata and FAIR practices.
- Legal and ethical advisors. Interpret GDPR and other ethical constraints).
- IT professionals. Manage secure storage and encrypted transfers.
- Domain experts. Assess data relevance and validity.
Refer to Metroline Step: Build the Team for role descriptions and team structure advice.
Practical examples from the community
We’re looking for practical examples from the community to illustrate this step. If you have an example to share, we’d love to hear from you. Visit our How to contribute page to get in touch.
Training
Training on data acquisition and access is crucial to meet FAIR principles and legal requirements. To acquire and retrieve data responsibly, researchers need to understand both the technical processes (e.g. secure file transfer, APIs, data validation) and the legal/ethical frameworks (e.g. data licenses, consent, institutional approvals). Several training resources are available to build these competencies:
- ELIXIR Luxembourg – Practicalities of Data Handling. A presentation covering key topics in secure data transfer, storage, encryption, and checksums.
- GOBLET – Bioinformatics Introductory Module. An introductory self-study course designed for life science students to learn about databases, sequence data, expression analysis, and basic bioinformatics tools.
- FAIR Cookbook. Offers practical recipes on data access, API development (e.g. FCB073 - Developing FAIR API for the Web) and secure transfer methods (e.g. FCB014 - Transferring data with SFTP, FCB015 - Downloading data with Aspera).
- RDMkit – Data Transfer Guide. Covers the practical and legal aspects of transferring research data, with emphasis on GDPR compliance and technical integrity checks. See Your tasks: Data transfer.
Additionally, some more specific trainings on regulations such as the GDPR in the European context are useful to have the overview of legal requirements for data access and retrieval.
- GDPR 4 Data Support - RDNL. The GDPR 4 Data Support (GDPR4DS) course is an introductory course designed for data supporters who want to understand more about the General Data Protection Regulation (GDPR) in the context of research and want to strengthen their role in protecting personal data.
Suggestions
This page is under construction. Learn more about the contributors here and explore the development process here. If you have any suggestions, visit our How to contribute page to get in touch.