<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.pangaea.de/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Lmoeller</id>
	<title>PANGAEA Wiki - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.pangaea.de/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Lmoeller"/>
	<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/wiki/Special:Contributions/Lmoeller"/>
	<updated>2026-04-28T17:33:53Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.6</generator>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16843</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16843"/>
		<updated>2026-04-22T10:46:41Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Access to Restricted Datasets */ changed formatting an position of the disclaimer&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided (i.e. meaningful) filename, follow redirects (to resolve the DOI), and fail with appropriate exit code on errors (easier to handle in scripts than the respective HTML errors).&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; &lt;br /&gt;
 &#039;&#039;&#039;Please note:&#039;&#039;&#039; The bearer token is private and should be treated like a password — it must not be shared with others or included in publicly accessible scripts or repositories. If a token has been accidentally disclosed, the user profile page at https://pangaea.de/user/ provides a &amp;quot;Log out from all devices&amp;quot; option, which immediately invalidates all active tokens. When using content negotiation via the DOI resolver at https://doi.org/, it is safe to include the Authorization header: PANGAEA trusts the DOI Foundation&#039;s infrastructure, and the request is redirected to PANGAEA&#039;s own servers before any protected content is served.&lt;br /&gt;
&lt;br /&gt;
The token can be passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 curl -OJLf -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required. In recent versions, curl can provide the token with an explicit flag (&amp;lt;code&amp;gt;--oauth2-bearer&amp;lt;/code&amp;gt;). The respective command example would then be: &lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf --oauth2-bearer xxx -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to your datasets that are subject to a moratorium (not of other authors!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access (to your own publications)&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|retrievable via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16842</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16842"/>
		<updated>2026-04-22T10:40:39Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Access to Restricted Datasets */ Added disclaimer notice on the security of tokens&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided (i.e. meaningful) filename, follow redirects (to resolve the DOI), and fail with appropriate exit code on errors (easier to handle in scripts than the respective HTML errors).&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 curl -OJLf -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&amp;lt;blockquote&amp;gt;&#039;&#039;&#039;Please note:&#039;&#039;&#039; The bearer token is private and should be treated like a password — it must not be shared with others or included in publicly accessible scripts or repositories. If a token has been accidentally disclosed, the user profile page at &amp;lt;nowiki&amp;gt;https://pangaea.de/user/&amp;lt;/nowiki&amp;gt; provides a &amp;quot;Log out from all devices&amp;quot; option, which immediately invalidates all active tokens. When using content negotiation via the DOI resolver at &amp;lt;nowiki&amp;gt;https://doi.org/&amp;lt;/nowiki&amp;gt;, it is safe to include the Authorization header: PANGAEA trusts the DOI Foundation&#039;s infrastructure, and the request is redirected to PANGAEA&#039;s own servers before any protected content is served.&amp;lt;/blockquote&amp;gt;In recent versions, curl can provide the token with an explicit flag (&amp;lt;code&amp;gt;--oauth2-bearer&amp;lt;/code&amp;gt;). The respective command example would then be: &lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf --oauth2-bearer xxx -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to your datasets that are subject to a moratorium (not of other authors!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access (to your own publications)&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|retrievable via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16841</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16841"/>
		<updated>2026-04-22T10:35:04Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* HTTP Content Negotiation */ Describe functions behind the curl flags&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided (i.e. meaningful) filename, follow redirects (to resolve the DOI), and fail with appropriate exit code on errors (easier to handle in scripts than the respective HTML errors).&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
In recent versions, curl can provide the token with an explicit flag (&amp;lt;code&amp;gt;--oauth2-bearer&amp;lt;/code&amp;gt;). The respective command example would then be: &lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf --oauth2-bearer xxx -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to your datasets that are subject to a moratorium (not of other authors!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access (to your own publications)&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|retrievable via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16840</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16840"/>
		<updated>2026-04-22T10:31:20Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: removed \ for line breaks in curl code examples&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail with appropriate exit code on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
In recent versions, curl can provide the token with an explicit flag (&amp;lt;code&amp;gt;--oauth2-bearer&amp;lt;/code&amp;gt;). The respective command example would then be: &lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf --oauth2-bearer xxx -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to your datasets that are subject to a moratorium (not of other authors!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access (to your own publications)&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|retrievable via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16839</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16839"/>
		<updated>2026-04-22T10:26:12Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Access to Restricted Datasets */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail with appropriate exit code on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; \&lt;br /&gt;
   -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
In recent versions, curl can provide the token with an explicit flag (&amp;lt;code&amp;gt;--oauth2-bearer&amp;lt;/code&amp;gt;). The respective command example would then be: &lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf --oauth2-bearer xxx   -H &#039;Accept: text/tab-separated-values&#039;   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to your datasets that are subject to a moratorium (not of other authors!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access (to your own publications)&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|retrievable via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16838</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16838"/>
		<updated>2026-04-22T10:25:43Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Access to Restricted Datasets */ Added alternative way to provide the bearer token&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail with appropriate exit code on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; \&lt;br /&gt;
   -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
In recent versions, curl can provide the token with an explicit flag (&amp;lt;code&amp;gt;--oauth2-bearer&amp;lt;/code&amp;gt;). The respective command example would then be: &lt;br /&gt;
&amp;lt;code&amp;gt;curl -OJLf --oauth2-bearer xxx   -H &#039;Accept: text/tab-separated-values&#039;   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to your datasets that are subject to a moratorium (not of other authors!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access (to your own publications)&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|retrievable via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16837</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16837"/>
		<updated>2026-04-22T10:10:25Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Downloading the Data File */ adjusted curl flag description&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail with appropriate exit code on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; \&lt;br /&gt;
   -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to your datasets that are subject to a moratorium (not of other authors!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access (to your own publications)&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|retrievable via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16836</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16836"/>
		<updated>2026-04-22T09:44:01Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Access to Restricted Datasets */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail silently on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; \&lt;br /&gt;
   -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to your datasets that are subject to a moratorium (not of other authors!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access (to your own publications)&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|retrievable via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16835</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16835"/>
		<updated>2026-04-22T09:37:32Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Summary of Access Methods */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail silently on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; \&lt;br /&gt;
   -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to moratorium datasets (Please note: this does not enable you to access restricted data of other authors but yours alone!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access (to your own publications)&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|retrievable via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16834</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16834"/>
		<updated>2026-04-22T09:20:07Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: repaired dysfunctional links&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail silently on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, still not possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; \&lt;br /&gt;
   -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to moratorium datasets (Please note: this does not enable you to access restricted data of other authors but yours alone!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|Via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA Community Workshop GitHub] — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* [https://pypi.org/project/pangaeapy/ pangaeapy on PyPI] — Python client&lt;br /&gt;
* [https://cran.r-project.org/package=pangaear pangaear on CRAN] — R client&lt;br /&gt;
* [https://ws.pangaea.de/oai/ PANGAEA web services] — REST and OAI-PMH endpoints&lt;br /&gt;
* [https://signposting.org/ Signposting standard] — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16833</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16833"/>
		<updated>2026-04-22T09:14:21Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [https://github.com/pangaea-data-publisher/community-workshop-material PANGAEA community workshop GitHub repository].&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail silently on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, to possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; \&lt;br /&gt;
   -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to moratorium datasets (Please note: this does not enable you to access restricted data of other authors but yours alone!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|Via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* PANGAEA Community Workshop GitHub — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* pangaeapy on PyPI — Python client&lt;br /&gt;
* pangaear on CRAN — R client&lt;br /&gt;
* PANGAEA web services — REST and OAI-PMH endpoints&lt;br /&gt;
* Signposting standard — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16832</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16832"/>
		<updated>2026-04-22T09:12:29Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Activate external links&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [PANGAEA community workshop GitHub repository](https://github.com/pangaea-data-publisher/community-workshop-material).&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at https://www.pangaea.de/, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at [[PANGAEA search]].&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (https://support.datacite.org/docs/api), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (https://signposting.org/), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail silently on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&lt;br /&gt;
 https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, to possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at https://pangaea.de/user/login.php. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at https://pangaea.de/user/ displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; \&lt;br /&gt;
   -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   https://doi.org/10.1594/PANGAEA.841672&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: https://pypi.org/project/pangaeapy/&lt;br /&gt;
* Source code: https://github.com/pangaea-data-publisher/pangaeapy&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to moratorium datasets (Please note: this does not enable you to access restricted data of other authors but yours alone!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at https://github.com/pangaea-data-publisher/community-workshop-material (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: https://cran.r-project.org/package=pangaear&lt;br /&gt;
* GitHub: https://github.com/ropensci/pangaear&lt;br /&gt;
* Introduction slides: https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds[[1]]$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds[[1]]$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at https://ws.pangaea.de/. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|https://www.pangaea.de/&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; https://doi.org/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I https://doi.pangaea.de/...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|Via https://pangaea.de/user/&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|https://ws.pangaea.de/oai/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|https://pypi.org/project/pangaeapy/&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|https://cran.r-project.org/package=pangaear&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|https://wiki.pangaea.de/wiki/Data_warehouse&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|https://support.datacite.org/docs/api&lt;br /&gt;
|}&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* [[PANGAEA search]] — documentation on search syntax and faceted filters&lt;br /&gt;
* [[Data warehouse]] — data warehouse documentation&lt;br /&gt;
* [[PANGAEA Community Workshops]] — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* PANGAEA Community Workshop GitHub — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* pangaeapy on PyPI — Python client&lt;br /&gt;
* pangaear on CRAN — R client&lt;br /&gt;
* PANGAEA web services — REST and OAI-PMH endpoints&lt;br /&gt;
* Signposting standard — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16831</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16831"/>
		<updated>2026-04-22T09:05:13Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [PANGAEA community workshop GitHub repository](https://github.com/pangaea-data-publisher/community-workshop-material).&lt;br /&gt;
&lt;br /&gt;
== Overview ==&lt;br /&gt;
Every dataset published in PANGAEA is assigned a globally unique and persistent Digital Object Identifier (DOI). The DOI is the single entry point for all forms of programmatic access: it resolves to a dataset landing page that, in addition to its human-readable representation, exposes all available metadata formats and data download options through standard HTTP mechanisms. There is no separate PANGAEA data API — all data and metadata access is built on top of the DOI and its landing page using standard web protocols. This design makes PANGAEA data directly accessible without any vendor-specific API key or proprietary client, while remaining fully compatible with FAIR principles.&lt;br /&gt;
&lt;br /&gt;
== Web-Based Search and Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== PANGAEA Search Interface ===&lt;br /&gt;
The primary discovery tool for PANGAEA data is the PANGAEA search engine at &amp;lt;nowiki&amp;gt;https://www.pangaea.de/&amp;lt;/nowiki&amp;gt;, based on Elasticsearch. It supports full-text search and faceted filtering across all published metadata. Faceted navigation allows users to constrain results by topic, device type, geographic region, and temporal coverage. Documentation on search functionality and syntax is available in the Wiki at &amp;lt;nowiki&amp;gt;[[PANGAEA search]]&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== External Portals and Registries ===&lt;br /&gt;
PANGAEA metadata is harvested by a large number of disciplinary and generic portals via OAI-PMH, making datasets discoverable beyond the PANGAEA website through services such as Google Dataset Search, OpenAIRE, DataCite Commons, DataONE, GBIF, EMODnet, GFBio, and others. PANGAEA is registered in re3data.org, FAIRsharing.org, RIsources, and the EOSC Marketplace.&lt;br /&gt;
&lt;br /&gt;
An alternative for search when a PANGAEA-specific search interface is not required is the &#039;&#039;&#039;DataCite Search API&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://support.datacite.org/docs/api&amp;lt;/nowiki&amp;gt;), which searches across a large inventory of research data from many repositories. It returns summary metadata including DOI names, and subsequent data access from any PANGAEA result can then proceed via content negotiation as described below.&lt;br /&gt;
&lt;br /&gt;
== DOI Landing Pages and Link Discovery ==&lt;br /&gt;
&lt;br /&gt;
=== The Landing Page as Access Hub ===&lt;br /&gt;
Each PANGAEA dataset is represented by a landing page accessible at its DOI. For example:&lt;br /&gt;
 &amp;lt;code&amp;gt;&amp;lt;nowiki&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672&amp;lt;/nowiki&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
The landing page serves a dual function: it presents dataset metadata and a data preview in human-readable form, and it simultaneously exposes all available machine-readable representations of the same resource through standard HTTP link headers and HTML &amp;lt;code&amp;gt;&amp;lt;link&amp;gt;&amp;lt;/code&amp;gt; elements. This implementation follows the &#039;&#039;&#039;Signposting standard&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://signposting.org/&amp;lt;/nowiki&amp;gt;), which allows any HTTP client to discover all alternate representations of a dataset — metadata in various formats, the data file itself, and the ORCID iDs of the authors — without any prior knowledge of PANGAEA-specific URL patterns.&lt;br /&gt;
&lt;br /&gt;
=== Discovering Available Representations ===&lt;br /&gt;
A simple HTTP HEAD request to the landing page returns the full set of typed link relations in the response header. The following example illustrates this using &amp;lt;code&amp;gt;curl&amp;lt;/code&amp;gt;:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -LI &amp;lt;nowiki&amp;gt;https://doi.org/10.1594/PANGAEA.841672&amp;lt;/nowiki&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
The response &amp;lt;code&amp;gt;Link:&amp;lt;/code&amp;gt; header will contain relations of the following types:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;cite-as&amp;lt;/code&amp;gt;&#039;&#039;&#039; — the canonical DOI citation URL&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;describedby&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to metadata in various formats (ISO 19139, DataCite XML, PANGAEA XML, BibTeX, RIS, JSON-LD)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;item&amp;lt;/code&amp;gt;&#039;&#039;&#039; — links to the data itself (tab-delimited text, HTML view)&lt;br /&gt;
* &#039;&#039;&#039;&amp;lt;code&amp;gt;author&amp;lt;/code&amp;gt;&#039;&#039;&#039; — ORCID iDs of the dataset authors&lt;br /&gt;
&lt;br /&gt;
This mechanism allows scripts, harvesters, and other machine clients to discover and retrieve any representation of a dataset from its DOI alone.&lt;br /&gt;
&lt;br /&gt;
== HTTP Content Negotiation ==&lt;br /&gt;
HTTP content negotiation allows a client to request a specific representation of a resource by specifying a MIME type in the &amp;lt;code&amp;gt;Accept&amp;lt;/code&amp;gt; header. All PANGAEA dataset landing pages support this mechanism, so both data and metadata can be downloaded programmatically by querying the DOI directly with the appropriate content type — no PANGAEA-specific URL construction is required.&lt;br /&gt;
&lt;br /&gt;
=== Downloading the Data File ===&lt;br /&gt;
The tabular data file for a dataset can be downloaded as tab-delimited text:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf &#039;&amp;lt;nowiki&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=textfile&#039;&amp;lt;/nowiki&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
Equivalently, using content negotiation directly against the DOI:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   &amp;lt;nowiki&amp;gt;https://doi.org/10.1594/PANGAEA.841672&amp;lt;/nowiki&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;-OJL&amp;lt;/code&amp;gt; flags instruct curl to save the file using the server-provided filename, follow redirects, and fail silently on errors.&lt;br /&gt;
&lt;br /&gt;
=== Retrieving Metadata in a Specific Format ===&lt;br /&gt;
The same mechanism applies to metadata. To retrieve a dataset&#039;s metadata in ISO 19139/19115 format:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -L -H &#039;Accept: application/vnd.iso19139.metadata+xml&#039; \&lt;br /&gt;
   &amp;lt;nowiki&amp;gt;https://doi.org/10.1594/PANGAEA.841672&amp;lt;/nowiki&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
The following MIME types are currently supported for metadata retrieval via content negotiation:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Format&lt;br /&gt;
!MIME type&lt;br /&gt;
|-&lt;br /&gt;
|PANGAEA internal XML&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.pangaea.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|DataCite XML (v4)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.datacite.datacite+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|ISO 19139 / 19115&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.iso19139.metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|NASA DIF&lt;br /&gt;
|&amp;lt;code&amp;gt;application/vnd.nasa.dif-metadata+xml&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|JSON-LD (Schema.org)&lt;br /&gt;
|&amp;lt;code&amp;gt;application/ld+json&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|BibTeX&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-bibtex&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|RIS&lt;br /&gt;
|&amp;lt;code&amp;gt;application/x-research-info-systems&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Plain text citation&lt;br /&gt;
|&amp;lt;code&amp;gt;text/x-bibliography&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Tab-separated data&lt;br /&gt;
|&amp;lt;code&amp;gt;text/tab-separated-values&amp;lt;/code&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
Alternatively, the same formats can be requested using explicit URL parameters:&lt;br /&gt;
 &amp;lt;code&amp;gt;&amp;lt;nowiki&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_datacite4&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_iso19139&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=metadata_jsonld&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
 &amp;lt;nowiki&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.841672?format=citation_bibtex&amp;lt;/nowiki&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Access to Restricted Datasets ==&lt;br /&gt;
A small fraction of PANGAEA datasets are under an active moratorium — access to the data is temporarily restricted while the metadata remains publicly visible. Access to your personal moratorium datasets requires authentication using a &#039;&#039;&#039;bearer token&#039;&#039;&#039;, which is the standard mechanism for API authorization (RFC 6750). Note that access to protected datasets of other authors is, of course, to possible.&lt;br /&gt;
&lt;br /&gt;
To obtain your bearer token, log in to PANGAEA at &amp;lt;nowiki&amp;gt;https://pangaea.de/user/login.php&amp;lt;/nowiki&amp;gt;. Login is supported both with a PANGAEA username and password, and via ORCID iD. After logging in, the user profile page at &amp;lt;nowiki&amp;gt;https://pangaea.de/user/&amp;lt;/nowiki&amp;gt; displays the current session token under &amp;quot;Your temporary login token.&amp;quot; This token can be renewed when logging out and logging in again, and passed as an &amp;lt;code&amp;gt;Authorization: Bearer&amp;lt;/code&amp;gt; header in any HTTP request:&lt;br /&gt;
 &amp;lt;code&amp;gt;curl -OJLf \&lt;br /&gt;
   -H &#039;Authorization: Bearer &amp;lt;your-token&amp;gt;&#039; \&lt;br /&gt;
   -H &#039;Accept: text/tab-separated-values&#039; \&lt;br /&gt;
   &amp;lt;nowiki&amp;gt;https://doi.org/10.1594/PANGAEA.841672&amp;lt;/nowiki&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
If a request is made without a token to a restricted dataset, the server responds with &amp;lt;code&amp;gt;401 Bearer Realm&amp;lt;/code&amp;gt;, signaling that authentication is required.&lt;br /&gt;
&lt;br /&gt;
== OAI-PMH Metadata Harvesting ==&lt;br /&gt;
For bulk metadata harvesting, PANGAEA provides an OAI-PMH endpoint at:&lt;br /&gt;
 &amp;lt;code&amp;gt;&amp;lt;nowiki&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/nowiki&amp;gt;&amp;lt;/code&amp;gt;&lt;br /&gt;
OAI-PMH allows systematic, incremental harvesting of all PANGAEA metadata in supported formats. The following metadata standards are available via OAI-PMH: Dublin Core, DataCite v3 and v4, ISO 19139, and DIF (Directory Interchange Format). This endpoint is used by the portals and registries listed in the discovery section above, and is equally available to any user who wishes to build their own index or integrate PANGAEA metadata into an institutional system.&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with Python: pangaeapy ==&lt;br /&gt;
&#039;&#039;&#039;pangaeapy&#039;&#039;&#039; is the official Python client library for PANGAEA, developed and maintained by the PANGAEA team. It provides a high-level interface for loading and analyzing PANGAEA datasets directly into native Python data structures, without requiring manual HTTP requests or format parsing.&lt;br /&gt;
&lt;br /&gt;
* PyPI package: &amp;lt;nowiki&amp;gt;https://pypi.org/project/pangaeapy/&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Source code: &amp;lt;nowiki&amp;gt;https://github.com/pangaea-data-publisher/pangaeapy&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Introduction slides: &amp;lt;nowiki&amp;gt;https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Introduction_to_pangaeapy.pdf&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;pip install pangaeapy&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
The central object in pangaeapy is &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt;, which takes a DOI or PANGAEA dataset ID and retrieves both the data and the associated metadata:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanDataSet&lt;br /&gt;
 &lt;br /&gt;
 ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a pandas DataFrame&lt;br /&gt;
 print(ds.data.head())&lt;br /&gt;
 &lt;br /&gt;
 # Access dataset metadata&lt;br /&gt;
 print(ds.title)&lt;br /&gt;
 print(ds.authors)&lt;br /&gt;
 print(ds.parameters)  # list of measured parameters with units and methods&amp;lt;/code&amp;gt;&lt;br /&gt;
The &amp;lt;code&amp;gt;data&amp;lt;/code&amp;gt; attribute is a pandas DataFrame in which each column corresponds to a measured parameter, and each row to a measurement. Parameter metadata — including units, standard names, and method descriptions — is accessible through the &amp;lt;code&amp;gt;parameters&amp;lt;/code&amp;gt; attribute.&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from Python ===&lt;br /&gt;
pangaeapy also provides a &amp;lt;code&amp;gt;PanQuery&amp;lt;/code&amp;gt; class for querying the PANGAEA search engine programmatically. Note that the underlying search API used by pangaeapy is currently internal and undocumented; an official, publicly documented Search REST API is under development. Until that is available, pangaeapy offers the most straightforward path to programmatic search for Python users:&lt;br /&gt;
 &amp;lt;code&amp;gt;from pangaeapy import PanQuery&lt;br /&gt;
 &lt;br /&gt;
 q = PanQuery(&#039;temperature salinity Atlantic&#039;, limit=10)&lt;br /&gt;
 for result in q.results:&lt;br /&gt;
     print(result[&#039;doi&#039;], result[&#039;title&#039;])&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Access to Restricted Datasets ===&lt;br /&gt;
A bearer token obtained from the PANGAEA user profile can be passed to &amp;lt;code&amp;gt;PanDataSet&amp;lt;/code&amp;gt; to enable access to moratorium datasets (Please note: this does not enable you to access restricted data of other authors but yours alone!):&lt;br /&gt;
 &amp;lt;code&amp;gt;ds = PanDataSet(&#039;10.1594/PANGAEA.841672&#039;, token=&#039;&amp;lt;your-bearer-token&amp;gt;&#039;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Further Training Materials ===&lt;br /&gt;
Jupyter notebooks with worked examples for data discovery, loading, and analysis with pangaeapy are available in the PANGAEA Community Workshop GitHub repository at &amp;lt;nowiki&amp;gt;https://github.com/pangaea-data-publisher/community-workshop-material&amp;lt;/nowiki&amp;gt; (see the &amp;lt;code&amp;gt;Python/&amp;lt;/code&amp;gt; directory).&lt;br /&gt;
&lt;br /&gt;
== Programmatic Access with R: pangaear ==&lt;br /&gt;
&#039;&#039;&#039;pangaear&#039;&#039;&#039; is a community-developed R client for PANGAEA, maintained as part of the rOpenSci ecosystem. It provides equivalent functionality to pangaeapy for R users.&lt;br /&gt;
&lt;br /&gt;
* CRAN: &amp;lt;nowiki&amp;gt;https://cran.r-project.org/package=pangaear&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* GitHub: &amp;lt;nowiki&amp;gt;https://github.com/ropensci/pangaear&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Introduction slides: &amp;lt;nowiki&amp;gt;https://github.com/pangaea-data-publisher/community-workshop-material/blob/master/Intro_PangaeaR.pdf&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Installation ===&lt;br /&gt;
 &amp;lt;code&amp;gt;install.packages(&amp;quot;pangaear&amp;quot;)&lt;br /&gt;
 # or the development version:&lt;br /&gt;
 # devtools::install_github(&amp;quot;ropensci/pangaear&amp;quot;)&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Loading a Dataset ===&lt;br /&gt;
 &amp;lt;code&amp;gt;library(pangaear)&lt;br /&gt;
 &lt;br /&gt;
 # Download and load a dataset by DOI&lt;br /&gt;
 ds &amp;lt;- pg_data(doi = &#039;10.1594/PANGAEA.841672&#039;)&lt;br /&gt;
 &lt;br /&gt;
 # Access the data as a data frame&lt;br /&gt;
 head(ds&amp;lt;nowiki&amp;gt;[[1]]&amp;lt;/nowiki&amp;gt;$data)&lt;br /&gt;
 &lt;br /&gt;
 # Metadata is accessible from the list object&lt;br /&gt;
 ds&amp;lt;nowiki&amp;gt;[[1]]&amp;lt;/nowiki&amp;gt;$metadata&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Searching PANGAEA from R ===&lt;br /&gt;
 &amp;lt;code&amp;gt;res &amp;lt;- pg_search(query = &#039;temperature salinity Atlantic&#039;, count = 10)&lt;br /&gt;
 print(res$doi)&amp;lt;/code&amp;gt;&lt;br /&gt;
R scripts and worked examples are available in the &amp;lt;code&amp;gt;R/&amp;lt;/code&amp;gt; directory of the PANGAEA Community Workshop GitHub repository.&lt;br /&gt;
&lt;br /&gt;
== The PANGAEA Data Warehouse ==&lt;br /&gt;
The PANGAEA data warehouse (based on Clickhouse) provides a powerful complementary access path for users who need to compile and aggregate data across large numbers of datasets — potentially across hundreds of thousands of individual publications — without having to download and merge files manually.&lt;br /&gt;
&lt;br /&gt;
The data warehouse supports spatially and chronologically constrained queries at the parameter level, returning data together with the DOI of each contributing dataset to maintain full provenance traceability. It is accessible in two ways:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Through the PANGAEA website:&#039;&#039;&#039; The data warehouse interface is integrated into the search results page. After performing a search, users can select &amp;quot;Data Warehouse&amp;quot; to configure and download a parameter-level aggregation from all datasets in the result set.&lt;br /&gt;
&lt;br /&gt;
Documentation: &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/Data_warehouse&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Programmatically via REST API:&#039;&#039;&#039; The data warehouse is also accessible through the PANGAEA web services at &amp;lt;nowiki&amp;gt;https://ws.pangaea.de/&amp;lt;/nowiki&amp;gt;. This allows automated, scripted aggregations and is supported by both pangaeapy and pangaear.&lt;br /&gt;
&lt;br /&gt;
Note that data warehouse exports represent compiled data products and do not replace the need to consult individual dataset landing pages to assess the fitness of individual studies for a specific scientific application. Every export includes the DOI name for each data point, ensuring that citation and provenance are preserved.&lt;br /&gt;
&lt;br /&gt;
== Summary of Access Methods ==&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Use case&lt;br /&gt;
!Method&lt;br /&gt;
!Entry point&lt;br /&gt;
|-&lt;br /&gt;
|Interactive discovery&lt;br /&gt;
|PANGAEA search&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://www.pangaea.de/&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Single dataset access&lt;br /&gt;
|Browser / landing page&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://doi.pangaea.de/10.1594/PANGAEA.xxxxx&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic data download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: text/tab-separated-values&#039; &amp;lt;nowiki&amp;gt;https://doi.org/&amp;lt;/nowiki&amp;gt;...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Programmatic metadata download&lt;br /&gt;
|HTTP content negotiation&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -H &#039;Accept: application/ld+json&#039; &amp;lt;nowiki&amp;gt;https://doi.org/&amp;lt;/nowiki&amp;gt;...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Link and format discovery&lt;br /&gt;
|HTTP HEAD / Signposting&lt;br /&gt;
|&amp;lt;code&amp;gt;curl -I &amp;lt;nowiki&amp;gt;https://doi.pangaea.de/&amp;lt;/nowiki&amp;gt;...&amp;lt;/code&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Restricted dataset access&lt;br /&gt;
|Bearer token authentication&lt;br /&gt;
|Via &amp;lt;nowiki&amp;gt;https://pangaea.de/user/&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Bulk metadata harvesting&lt;br /&gt;
|OAI-PMH&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://ws.pangaea.de/oai/&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (Python)&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://pypi.org/project/pangaeapy/&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Scripted data access (R)&lt;br /&gt;
|pangaear&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://cran.r-project.org/package=pangaear&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Cross-dataset aggregation&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/Data_warehouse&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
|Discovery across repositories&lt;br /&gt;
|DataCite Search API&lt;br /&gt;
|&amp;lt;nowiki&amp;gt;https://support.datacite.org/docs/api&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
|}&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Further Resources ==&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;nowiki&amp;gt;[[PANGAEA search]]&amp;lt;/nowiki&amp;gt; — documentation on search syntax and faceted filters&lt;br /&gt;
* &amp;lt;nowiki&amp;gt;[[Data warehouse]]&amp;lt;/nowiki&amp;gt; — data warehouse documentation&lt;br /&gt;
* &amp;lt;nowiki&amp;gt;[[PANGAEA Community Workshops]]&amp;lt;/nowiki&amp;gt; — hands-on training workshops on finding and using PANGAEA data&lt;br /&gt;
* PANGAEA Community Workshop GitHub — Jupyter notebooks, R scripts, and slide decks&lt;br /&gt;
* pangaeapy on PyPI — Python client&lt;br /&gt;
* pangaear on CRAN — R client&lt;br /&gt;
* PANGAEA web services — REST and OAI-PMH endpoints&lt;br /&gt;
* Signposting standard — link discovery for FAIR data&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16830</id>
		<title>Data Access and Reuse</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Access_and_Reuse&amp;diff=16830"/>
		<updated>2026-04-22T08:54:10Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Created page with &amp;quot;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.  Hands-on training materials covering m...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This article describes the different methods available for discovering, accessing, and reusing data published in PANGAEA. It is intended for researchers and data practitioners who want to interact with PANGAEA data beyond the standard web interface, including scripted and automated workflows. The methods range from interactive web-based search to fully programmatic access via HTTP content negotiation and dedicated client libraries.&lt;br /&gt;
&lt;br /&gt;
Hands-on training materials covering many of the topics below are available in the [[PANGAEA Community Workshops|PANGAEA Community Workshop Series]], including Jupyter notebooks, R scripts, and slide decks in the [PANGAEA community workshop GitHub repository](https://github.com/pangaea-data-publisher/community-workshop-material).&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16806</id>
		<title>Technology</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16806"/>
		<updated>2026-04-13T09:49:47Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Middleware: Processing and Transformation */ Made links work&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;PANGAEA is built on a three-tiered client/server architecture that has evolved continuously since the system&#039;s founding in the early 1990s. The architecture separates backend storage and database systems, middleware processing and transformation components, and frontend delivery and user interfaces into distinct layers that are individually maintainable and replaceable. This design philosophy has enabled PANGAEA to migrate core components — including its primary database engine, its editorial system, and its search infrastructure — without disruption to archived data or to users, and underpins the system&#039;s long-term stability as a trustworthy data repository.&lt;br /&gt;
&lt;br /&gt;
A detailed description of the PANGAEA information system and its workflows is provided in Felden et al. (2023) and Diepenbroek et al. (2017); this article provides an up-to-date overview oriented toward the general structure and key components.&lt;br /&gt;
&lt;br /&gt;
== System Architecture ==&lt;br /&gt;
&lt;br /&gt;
The technical architecture of PANGAEA follows a three-tiered model comprising a backend, a middleware layer, and a frontend. All hard- and software services are hosted and operated by the data and computing center of the Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research (AWI) in Bremerhaven. Most backend and middleware systems, as well as all frontend web servers and search engines, run on virtual machines (VMware) operated with Ubuntu Linux. Virtualization provides sufficient capacity and performance while enabling high availability and transparent hardware renewal on a typical cycle of three to four years.&lt;br /&gt;
[[File:PANGAEA overview architecture.png|alt=Schema illustrating the system architecture according to Felden et al., 2023|thumb|Schema illustrating the system architecture according to Felden et al., 2023|none|600x600px]]&lt;br /&gt;
&lt;br /&gt;
PANGAEA currently operates nine virtual machines at the AWI computing center, using a total of 53 CPUs, 162 GB RAM, and 28 TB of disk space. The relational database holdings currently occupy approximately 5 TB, with approximately 1 PB of data stored on tape.&lt;br /&gt;
&lt;br /&gt;
== Backend: Storage and Databases ==&lt;br /&gt;
&lt;br /&gt;
=== Relational Database ===&lt;br /&gt;
The primary store for all structured data and metadata in PANGAEA is a &#039;&#039;&#039;PostgreSQL&#039;&#039;&#039; relational database management system (RDBMS). The data model is highly normalized, reflecting the full observational context of each measurement: events (when and where data were collected), campaigns (cruises or field expeditions), methods and devices, parameters, references, and institutional provenance information are all stored in defined relational structures. This normalized model allows data descriptions to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived data.&lt;br /&gt;
&lt;br /&gt;
Database integrity is continuously maintained through &#039;&#039;&#039;PostgreSQL streaming replication&#039;&#039;&#039; to a dedicated backup system, which enables point-in-time recovery to any moment prior to a failure event. In addition, full database backup copies (`pg_base_backup`) are created each weekend and retained for three weeks, providing an independent recovery layer supplementary to the streaming replica.&lt;br /&gt;
&lt;br /&gt;
=== Data Warehouse ===&lt;br /&gt;
Fast access to large, aggregated compilations of data across the full PANGAEA holdings is provided by a &#039;&#039;&#039;Clickhouse&#039;&#039;&#039; data warehouse, which mirrors the numerical data inventory from the relational database. The data warehouse supports spatially and chronologically constrained queries at the parameter level across potentially hundreds of thousands of individual datasets, and is accessible both through the PANGAEA website and programmatically via a REST API. All data warehouse exports include the DOI for each contributing dataset, ensuring that provenance and citation remain intact in compiled data products.&lt;br /&gt;
&lt;br /&gt;
=== Archival Storage ===&lt;br /&gt;
For high-volume and binary data — including geophysical datasets, images, video, and community-specific formats such as NetCDF — data are stored in consistent formats on hard-disk arrays and robotic tape archives. The central archival storage system consists of two &#039;&#039;&#039;SpectraLogic TFinity ExaScale&#039;&#039;&#039; robotic tape libraries housed in separate buildings at AWI, with a combined capacity of up to 60 PB. Backup operations use high-capacity LTO-tape drives within this environment.&lt;br /&gt;
&lt;br /&gt;
All data are stored redundantly using &#039;&#039;&#039;erasure coding&#039;&#039;&#039; across disk and tape. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots. Write caches are battery-backed to ensure data integrity at the point of write.&lt;br /&gt;
&lt;br /&gt;
=== Off-Site Replica ===&lt;br /&gt;
Since 2025, PANGAEA operates a &#039;&#039;&#039;minimal viable repository service&#039;&#039;&#039; as an off-site replica at MARUM/University of Bremen, hosted in the Green IT Housing Center of Bremen University. This facility provides geographic and infrastructural separation from the primary AWI systems, enabling recovery of data delivery services within 24 hours following a technical breakdown or cyberattack. Disaster recovery and switchover procedures involving this facility are regularly exercised. The replica is kept current through snapshot-based replication several times per day and currently covers all dataset metadata — including individual landing pages and all harvesting endpoints — full representations of tabular data publications, the Elasticsearch search index, and the relevant web frontend. Extension of the replica to include binary data files is under active development, with formal commitment to be finalized through the ongoing MARUM hosting agreement negotiations.&lt;br /&gt;
&lt;br /&gt;
The off-site installation operates in a fully isolated environment with Layer 2 network separation. Access is restricted to VPN connections from a single designated gateway host at AWI, with access tokens and keys limited to the corresponding gateway and PANGAEA DevOps staff.&lt;br /&gt;
== Middleware: Processing and Transformation ==&lt;br /&gt;
The middleware layer manages the flow of data and metadata between the backend storage systems and the frontend interfaces. Its core functions are marshaling, indexing, transformation, and dissemination.&lt;br /&gt;
&lt;br /&gt;
Metadata are dynamically marshaled from the relational database to a PANGAEA-specific internal XML format, from which they are transformed via &#039;&#039;&#039;XSLT and XML-to-JSON pipelines&#039;&#039;&#039; into a range of content standards for delivery to users and harvesting services. Currently supported output standards include JSON-LD (Schema.org with Croissant extension), DataCite XML, Dublin Core XML, ISO 19115/ISO 19139, and DIF/FGDC. Dissemination occurs via OAI-PMH, HTTP content negotiation following the Signposting standard (https://signposting.org/), and other protocols. Metadata marshaling from the PostgreSQL database to the Elasticsearch search index is handled asynchronously by dedicated middleware components, which also manage automated DOI minting at DataCite upon dataset publication.&lt;br /&gt;
&lt;br /&gt;
The marshaled metadata are stored and indexed in &#039;&#039;&#039;Elasticsearch&#039;&#039;&#039;, which serves as the primary index for all public search and metadata access interfaces. This architecture separates the authoritative relational record from the search-optimized representation, allowing the search index to be rebuilt or updated without modifying archived data.&lt;br /&gt;
&lt;br /&gt;
The flexible metadata framework &#039;&#039;&#039;PanFMP&#039;&#039;&#039; (https://www.panfmp.org/) underpins the mapping between the PANGAEA internal schema and the various external output standards, ensuring that new or updated standards can be accommodated without structural changes to the underlying data.&lt;br /&gt;
&lt;br /&gt;
Data submissions, user requests, and bug reports are managed through a &#039;&#039;&#039;JIRA&#039;&#039;&#039; (Atlassian) issue tracking system, which serves as the primary communication channel between data providers and the PANGAEA editorial team throughout the submission and publication process.&lt;br /&gt;
== Frontend: Editorial System ==&lt;br /&gt;
The PANGAEA editorial system is the primary tool through which data editors review, curate, harmonize, and publish submitted data and metadata. It is a &#039;&#039;&#039;web-based client/server application&#039;&#039;&#039; developed entirely in-house, operating directly on the PostgreSQL databases.&lt;br /&gt;
&lt;br /&gt;
The backend of the editorial system is built on &#039;&#039;&#039;Java 17&#039;&#039;&#039; using the &#039;&#039;&#039;Dropwizard&#039;&#039;&#039; framework, exposing a REST API. The frontend is implemented in &#039;&#039;&#039;React&#039;&#039;&#039; with the &#039;&#039;&#039;Ant Design&#039;&#039;&#039; component library. Source code for both components, together with shared libraries and test suites, is maintained in an institutional &#039;&#039;&#039;GitLab&#039;&#039;&#039; instance hosted at AWI, structured as an &#039;&#039;&#039;Nx monorepo&#039;&#039;&#039;. The repository encompasses automated unit tests and &#039;&#039;&#039;Cypress&#039;&#039;&#039; end-to-end tests. All new versions are deployed exclusively through a &#039;&#039;&#039;GitLab CI/CD pipeline&#039;&#039;&#039;, with releases gated on successful completion of the full automated test suite.&lt;br /&gt;
&lt;br /&gt;
In production, the editorial system runs on Ubuntu virtual machines hosted on VMware infrastructure. &#039;&#039;&#039;Apache&#039;&#039;&#039; serves the React frontend and acts as a reverse proxy to the backend services. Four parallel instances are operated simultaneously, providing dedicated test and production environments for two editorial system versions at any given time. This setup supports controlled feature validation alongside uninterrupted service operation. Infrastructure components — including VMware, file systems, monitoring, GitLab, and the database platform — are managed by AWI, while the PANGAEA group is responsible for the Apache web server, Java services, and React frontend.&lt;br /&gt;
&lt;br /&gt;
The relational database model underlying the editorial system enforces structural and semantic consistency between submitted metadata and the existing data inventory. The editorial system also integrates a &#039;&#039;&#039;Terminology Catalogue (TC)&#039;&#039;&#039;, which manages controlled vocabularies and ontologies used to harmonize data and metadata during ingest. Supported terminologies include WoRMS, ITIS, ChEBI, EnvO, PATO, QUDT, and the NERC vocabulary server, among others. A detailed description of the Terminology Catalogue and its role in data archiving and publication is given in Diepenbroek et al. (2017).&lt;br /&gt;
== Frontend: Public Web Interface and Search ==&lt;br /&gt;
&lt;br /&gt;
=== Web Delivery Stack ===&lt;br /&gt;
The public web infrastructure is fronted by an &#039;&#039;&#039;NGINX&#039;&#039;&#039; reverse proxy that manages incoming traffic and supports &#039;&#039;&#039;HTTP/3 (QUIC)&#039;&#039;&#039; and &#039;&#039;&#039;HTTP/2&#039;&#039;&#039; alongside HTTP/1.1 for broad client compatibility. Behind this entry point, a dual-backend architecture separates concerns by function: a &#039;&#039;&#039;PHP 8&#039;&#039;&#039; environment handles general information pages and the PANGAEA Wiki, while a &#039;&#039;&#039;Java 17&#039;&#039;&#039; application layer running on &#039;&#039;&#039;Eclipse Jetty 12&#039;&#039;&#039; manages core repository services — including dataset landing pages, DOI and handle resolution, and content negotiation for both human- and machine-readable data access. The Jetty-based layer also manages delivery of high-volume and binary files from the tape archive by staging requested files to a local hard-disk cache before serving them to users.&lt;br /&gt;
&lt;br /&gt;
=== Search ===&lt;br /&gt;
The PANGAEA search engine supports full-text and faceted search across all published metadata. Faceted navigation is enabled by semantic metadata enrichment performed during the marshaling process: terminology-based annotations are added to metadata records, allowing consistent filtering by topic, device type, geographic region, and other dimensions. PANGAEA&#039;s approach to data discovery complies with the recent recommendations of the Research Data Alliance (RDA) Interest Group &amp;quot;Data Discovery Paradigms&amp;quot; (Wu, M. et al., 2026, &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;). Search documentation is available in the Wiki at https://wiki.pangaea.de/wiki/PANGAEA_search.&lt;br /&gt;
&lt;br /&gt;
=== Map Search ===&lt;br /&gt;
Geographic search and visualization are implemented using &#039;&#039;&#039;leaflet.js&#039;&#039;&#039;, serving map tiles from four configurable sources: the AWI basemap, OpenStreetMap, Google (hybrid), and ESRI.&lt;br /&gt;
&lt;br /&gt;
=== Dataset Landing Pages ===&lt;br /&gt;
Each published dataset is represented by a landing page resolved through its DOI. Landing pages present the full dataset metadata in human-readable form and are enriched with structured &#039;&#039;&#039;schema.org&#039;&#039;&#039; markup in JSON-LD extended with Croissant properties, ensuring machine-actionability and compatibility with generic web search engines and data registries. The schema.org metadata is also accessible via HTTP content negotiation following the Signposting standard. Minor editorial corrections to datasets that do not affect scientific content are documented in a &amp;quot;Change history&amp;quot; section of the landing page, recording the date and a summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
=== Programmatic Access ===&lt;br /&gt;
PANGAEA offers programmatic access to data and metadata through a range of web services (REST and SOAP). The OAI-PMH endpoint supports metadata harvesting in all supported standards. Client libraries for &#039;&#039;&#039;Python&#039;&#039;&#039; (pangaeapy, developed by PANGAEA) and &#039;&#039;&#039;R&#039;&#039;&#039; (pangaear, developed by the community) allow researchers to load and transform PANGAEA data directly into native data structures for analysis in environments such as Jupyter notebooks.&lt;br /&gt;
&lt;br /&gt;
== Monitoring and Analytics ==&lt;br /&gt;
Service health across the PANGAEA infrastructure is monitored through the &#039;&#039;&#039;AWI Grafana/Telegraf&#039;&#039;&#039; stack, covering service availability, application logs, and resource saturation for all production systems. This is supplemented by external availability checks via &#039;&#039;&#039;UptimeRobot&#039;&#039;&#039;, which provides independent verification of public-facing service endpoints from outside the AWI network. All external links in PANGAEA metadata records — references to related literature, other dataset versions, and external resources — are automatically checked for broken (HTTP 404) or permanently redirected (HTTP 301) responses on a weekly basis.&lt;br /&gt;
&lt;br /&gt;
User engagement and download metrics are captured by an integrated &#039;&#039;&#039;Matomo&#039;&#039;&#039; analytics instance configured to produce usage statistics compliant with the &#039;&#039;&#039;COUNTER&#039;&#039;&#039; (Counting Online Usage of Networked Electronic Resources) standard, ensuring that data impact is measured according to international scholarly norms. Usage statistics are publicly visible on each dataset landing page; details are documented in the Wiki at https://wiki.pangaea.de/wiki/Data_Usage_Statistics.&lt;br /&gt;
&lt;br /&gt;
== Security ==&lt;br /&gt;
Backend and middleware systems are protected behind a firewall; frontend systems operate in a demilitarized zone (DMZ) accessible from outside with restricted but still firewalled access. Frontend systems have no write access and only limited read access to the backend database and tape archives. Public access from the website and REST APIs is served from data replicas hosted in Elasticsearch or through read-only remote filesystem access. Data curators access production systems via virtual private networks (VPN), which enforce a basic check that client operating systems are up to date.&lt;br /&gt;
&lt;br /&gt;
Physical access to the AWI computing center is managed through an electronic access control system for all relevant entrances; key distribution is documented, and guest policies are formally established. The AWI maintains uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation. Fire alerts are automatically forwarded to the fire department with a contractual response time under 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Security of the technical infrastructure is further maintained through the use of asymmetric key infrastructures, mandatory minimum-length passwords for all user classes, short-cycle security patching for all hardware and software components, professional monitoring tools for hardware, firewall, software, services, performance, and attacks, and regular security training for all technical and non-technical staff. Security risks are assessed on an ongoing basis by AWI&#039;s institutional IT security team, which coordinates incident response and monitors threat landscapes relevant to research data infrastructure. PANGAEA technical staff participates in regular security reviews.&lt;br /&gt;
&lt;br /&gt;
Access controls for restricted or moratorium datasets are enforced at the application layer. Dataset metadata remains publicly accessible in all cases; access to the data itself is restricted to authorized users at the individual level for the duration of the moratorium, typically a maximum of two years from submission.&lt;br /&gt;
&lt;br /&gt;
The off-site replica at MARUM operates in a fully isolated network environment (Layer 2 separation) with access restricted to a firewalled VPN gateway at AWI. The Green IT Housing Center is monitored 24/7 by Bremen University staff located in very close proximity to the datacenter (15 m, adjacent building), with automated fire alarms, redundant power supply with battery backup, and multilevel physical access control.&lt;br /&gt;
&lt;br /&gt;
== Software Inventory ==&lt;br /&gt;
The following table summarizes the principal software components of the PANGAEA infrastructure and their development model.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Component !! Software !! Development Model&lt;br /&gt;
|-&lt;br /&gt;
| Primary database || PostgreSQL || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Data warehouse || Clickhouse || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Search and metadata index || Elasticsearch || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system backend || Java 17 / Dropwizard || In-house development&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system frontend || React / Ant Design || In-house development (open source framework)&lt;br /&gt;
|-&lt;br /&gt;
| Source control || GitLab || Open source (community-supported, self-hosted)&lt;br /&gt;
|-&lt;br /&gt;
| CI/CD pipeline || GitLab CI/CD || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web server / reverse proxy (editorial) || Apache || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web traffic management || NGINX || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| General web and Wiki frontend || PHP 8 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Repository services backend || Java 17 / Eclipse Jetty 12 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Map visualization || leaflet.js || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Metadata framework || PanFMP || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| Issue tracking || JIRA (Atlassian) || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Monitoring stack || Grafana / Telegraf || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| External uptime monitoring || UptimeRobot || Commercial SaaS&lt;br /&gt;
|-&lt;br /&gt;
| Analytics || Matomo || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Infrastructure virtualization || VMware || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Tape archive hardware || SpectraLogic TFinity ExaScale || Commercial hardware&lt;br /&gt;
|-&lt;br /&gt;
| Python data client || pangaeapy || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| R data client || pangaear || Community development (open source)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Documentation and Change Management ==&lt;br /&gt;
General documentation of PANGAEA systems and services is maintained in the public &#039;&#039;&#039;PANGAEA Wiki&#039;&#039;&#039; (https://wiki.pangaea.de/). A separate internal &#039;&#039;&#039;Confluence&#039;&#039;&#039; Wiki, kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes, contains detailed documentation of server configurations, installation and maintenance procedures, relationships between system components, VM snapshot procedures, backup routines, and service restart priorities for incident response.&lt;br /&gt;
&lt;br /&gt;
All changes to published data and metadata are recorded in the editorial system&#039;s version history. Substantive revisions to published datasets result in a new dataset version with a new DOI; all prior versions remain accessible and cross-referenced. Minor editorial corrections that do not affect scientific content are applied without creating a new identifier and are transparently documented in the &amp;quot;Change history&amp;quot; section of the dataset landing page, including the date and a brief summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science. &#039;&#039;Scientific Data&#039;&#039;, 10, 347. https://doi.org/10.1038/s41597-023-02269-x&lt;br /&gt;
* Diepenbroek, M., Schindler, U., Huber, R., Pesant, S., Stocker, M., Felden, J., Buss, M. &amp;amp; Weinrebe, M. (2017). Terminology supported archiving and publication of environmental science data in PANGAEA. &#039;&#039;Journal of Biotechnology&#039;&#039;, 261, 177–186. https://doi.org/10.1016/j.jbiotec.2017.07.016&lt;br /&gt;
* Wu, M. et al. (2026). RDA Interest Group on Data Discovery Paradigms. &#039;&#039;Data Science Journal&#039;&#039;. https://doi.org/10.5334/dsj-2026-006&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
*[[PANGAEA search]]&lt;br /&gt;
*[[Data submission]]&lt;br /&gt;
*[[Authors Guides]]&lt;br /&gt;
*[[Curation levels]]&lt;br /&gt;
*[[Processing levels]]&lt;br /&gt;
*[[Data Usage Statistics]]&lt;br /&gt;
*[[Format]]&lt;br /&gt;
*[[PANGAEA XML schema]]&lt;br /&gt;
*[[Best practice manuals and templates]]&lt;br /&gt;
*[[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
*[[Preservation Plan]] — https://www.pangaea.de/about/preservation.php&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16805</id>
		<title>Technology</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16805"/>
		<updated>2026-04-13T09:15:16Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Off-Site Replica */ cosmetic&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;PANGAEA is built on a three-tiered client/server architecture that has evolved continuously since the system&#039;s founding in the early 1990s. The architecture separates backend storage and database systems, middleware processing and transformation components, and frontend delivery and user interfaces into distinct layers that are individually maintainable and replaceable. This design philosophy has enabled PANGAEA to migrate core components — including its primary database engine, its editorial system, and its search infrastructure — without disruption to archived data or to users, and underpins the system&#039;s long-term stability as a trustworthy data repository.&lt;br /&gt;
&lt;br /&gt;
A detailed description of the PANGAEA information system and its workflows is provided in Felden et al. (2023) and Diepenbroek et al. (2017); this article provides an up-to-date overview oriented toward the general structure and key components.&lt;br /&gt;
&lt;br /&gt;
== System Architecture ==&lt;br /&gt;
&lt;br /&gt;
The technical architecture of PANGAEA follows a three-tiered model comprising a backend, a middleware layer, and a frontend. All hard- and software services are hosted and operated by the data and computing center of the Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research (AWI) in Bremerhaven. Most backend and middleware systems, as well as all frontend web servers and search engines, run on virtual machines (VMware) operated with Ubuntu Linux. Virtualization provides sufficient capacity and performance while enabling high availability and transparent hardware renewal on a typical cycle of three to four years.&lt;br /&gt;
[[File:PANGAEA overview architecture.png|alt=Schema illustrating the system architecture according to Felden et al., 2023|thumb|Schema illustrating the system architecture according to Felden et al., 2023|none|600x600px]]&lt;br /&gt;
&lt;br /&gt;
PANGAEA currently operates nine virtual machines at the AWI computing center, using a total of 53 CPUs, 162 GB RAM, and 28 TB of disk space. The relational database holdings currently occupy approximately 5 TB, with approximately 1 PB of data stored on tape.&lt;br /&gt;
&lt;br /&gt;
== Backend: Storage and Databases ==&lt;br /&gt;
&lt;br /&gt;
=== Relational Database ===&lt;br /&gt;
The primary store for all structured data and metadata in PANGAEA is a &#039;&#039;&#039;PostgreSQL&#039;&#039;&#039; relational database management system (RDBMS). The data model is highly normalized, reflecting the full observational context of each measurement: events (when and where data were collected), campaigns (cruises or field expeditions), methods and devices, parameters, references, and institutional provenance information are all stored in defined relational structures. This normalized model allows data descriptions to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived data.&lt;br /&gt;
&lt;br /&gt;
Database integrity is continuously maintained through &#039;&#039;&#039;PostgreSQL streaming replication&#039;&#039;&#039; to a dedicated backup system, which enables point-in-time recovery to any moment prior to a failure event. In addition, full database backup copies (`pg_base_backup`) are created each weekend and retained for three weeks, providing an independent recovery layer supplementary to the streaming replica.&lt;br /&gt;
&lt;br /&gt;
=== Data Warehouse ===&lt;br /&gt;
Fast access to large, aggregated compilations of data across the full PANGAEA holdings is provided by a &#039;&#039;&#039;Clickhouse&#039;&#039;&#039; data warehouse, which mirrors the numerical data inventory from the relational database. The data warehouse supports spatially and chronologically constrained queries at the parameter level across potentially hundreds of thousands of individual datasets, and is accessible both through the PANGAEA website and programmatically via a REST API. All data warehouse exports include the DOI for each contributing dataset, ensuring that provenance and citation remain intact in compiled data products.&lt;br /&gt;
&lt;br /&gt;
=== Archival Storage ===&lt;br /&gt;
For high-volume and binary data — including geophysical datasets, images, video, and community-specific formats such as NetCDF — data are stored in consistent formats on hard-disk arrays and robotic tape archives. The central archival storage system consists of two &#039;&#039;&#039;SpectraLogic TFinity ExaScale&#039;&#039;&#039; robotic tape libraries housed in separate buildings at AWI, with a combined capacity of up to 60 PB. Backup operations use high-capacity LTO-tape drives within this environment.&lt;br /&gt;
&lt;br /&gt;
All data are stored redundantly using &#039;&#039;&#039;erasure coding&#039;&#039;&#039; across disk and tape. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots. Write caches are battery-backed to ensure data integrity at the point of write.&lt;br /&gt;
&lt;br /&gt;
=== Off-Site Replica ===&lt;br /&gt;
Since 2025, PANGAEA operates a &#039;&#039;&#039;minimal viable repository service&#039;&#039;&#039; as an off-site replica at MARUM/University of Bremen, hosted in the Green IT Housing Center of Bremen University. This facility provides geographic and infrastructural separation from the primary AWI systems, enabling recovery of data delivery services within 24 hours following a technical breakdown or cyberattack. Disaster recovery and switchover procedures involving this facility are regularly exercised. The replica is kept current through snapshot-based replication several times per day and currently covers all dataset metadata — including individual landing pages and all harvesting endpoints — full representations of tabular data publications, the Elasticsearch search index, and the relevant web frontend. Extension of the replica to include binary data files is under active development, with formal commitment to be finalized through the ongoing MARUM hosting agreement negotiations.&lt;br /&gt;
&lt;br /&gt;
The off-site installation operates in a fully isolated environment with Layer 2 network separation. Access is restricted to VPN connections from a single designated gateway host at AWI, with access tokens and keys limited to the corresponding gateway and PANGAEA DevOps staff.&lt;br /&gt;
== Middleware: Processing and Transformation ==&lt;br /&gt;
The middleware layer manages the flow of data and metadata between the backend storage systems and the frontend interfaces. Its core functions are marshaling, indexing, transformation, and dissemination.&lt;br /&gt;
&lt;br /&gt;
Metadata are dynamically marshaled from the relational database to a PANGAEA-specific internal XML format, from which they are transformed via &#039;&#039;&#039;XSLT and XML-to-JSON pipelines&#039;&#039;&#039; into a range of content standards for delivery to users and harvesting services. Currently supported output standards include JSON-LD (Schema.org with Croissant extension), DataCite XML, Dublin Core XML, ISO 19115/ISO 19139, and DIF/FGDC. Dissemination occurs via OAI-PMH, HTTP content negotiation following the Signposting standard (&amp;lt;nowiki&amp;gt;https://signposting.org/&amp;lt;/nowiki&amp;gt;), and other protocols. Metadata marshaling from the PostgreSQL database to the Elasticsearch search index is handled asynchronously by dedicated middleware components, which also manage automated DOI minting at DataCite upon dataset publication.&lt;br /&gt;
&lt;br /&gt;
The marshaled metadata are stored and indexed in &#039;&#039;&#039;Elasticsearch&#039;&#039;&#039;, which serves as the primary index for all public search and metadata access interfaces. This architecture separates the authoritative relational record from the search-optimized representation, allowing the search index to be rebuilt or updated without modifying archived data.&lt;br /&gt;
&lt;br /&gt;
The flexible metadata framework &#039;&#039;&#039;PanFMP&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://www.panfmp.org/&amp;lt;/nowiki&amp;gt;) underpins the mapping between the PANGAEA internal schema and the various external output standards, ensuring that new or updated standards can be accommodated without structural changes to the underlying data.&lt;br /&gt;
&lt;br /&gt;
Data submissions, user requests, and bug reports are managed through a &#039;&#039;&#039;JIRA&#039;&#039;&#039; (Atlassian) issue tracking system, which serves as the primary communication channel between data providers and the PANGAEA editorial team throughout the submission and publication process.&lt;br /&gt;
== Frontend: Editorial System ==&lt;br /&gt;
The PANGAEA editorial system is the primary tool through which data editors review, curate, harmonize, and publish submitted data and metadata. It is a &#039;&#039;&#039;web-based client/server application&#039;&#039;&#039; developed entirely in-house, operating directly on the PostgreSQL databases.&lt;br /&gt;
&lt;br /&gt;
The backend of the editorial system is built on &#039;&#039;&#039;Java 17&#039;&#039;&#039; using the &#039;&#039;&#039;Dropwizard&#039;&#039;&#039; framework, exposing a REST API. The frontend is implemented in &#039;&#039;&#039;React&#039;&#039;&#039; with the &#039;&#039;&#039;Ant Design&#039;&#039;&#039; component library. Source code for both components, together with shared libraries and test suites, is maintained in an institutional &#039;&#039;&#039;GitLab&#039;&#039;&#039; instance hosted at AWI, structured as an &#039;&#039;&#039;Nx monorepo&#039;&#039;&#039;. The repository encompasses automated unit tests and &#039;&#039;&#039;Cypress&#039;&#039;&#039; end-to-end tests. All new versions are deployed exclusively through a &#039;&#039;&#039;GitLab CI/CD pipeline&#039;&#039;&#039;, with releases gated on successful completion of the full automated test suite.&lt;br /&gt;
&lt;br /&gt;
In production, the editorial system runs on Ubuntu virtual machines hosted on VMware infrastructure. &#039;&#039;&#039;Apache&#039;&#039;&#039; serves the React frontend and acts as a reverse proxy to the backend services. Four parallel instances are operated simultaneously, providing dedicated test and production environments for two editorial system versions at any given time. This setup supports controlled feature validation alongside uninterrupted service operation. Infrastructure components — including VMware, file systems, monitoring, GitLab, and the database platform — are managed by AWI, while the PANGAEA group is responsible for the Apache web server, Java services, and React frontend.&lt;br /&gt;
&lt;br /&gt;
The relational database model underlying the editorial system enforces structural and semantic consistency between submitted metadata and the existing data inventory. The editorial system also integrates a &#039;&#039;&#039;Terminology Catalogue (TC)&#039;&#039;&#039;, which manages controlled vocabularies and ontologies used to harmonize data and metadata during ingest. Supported terminologies include WoRMS, ITIS, ChEBI, EnvO, PATO, QUDT, and the NERC vocabulary server, among others. A detailed description of the Terminology Catalogue and its role in data archiving and publication is given in Diepenbroek et al. (2017).&lt;br /&gt;
== Frontend: Public Web Interface and Search ==&lt;br /&gt;
&lt;br /&gt;
=== Web Delivery Stack ===&lt;br /&gt;
The public web infrastructure is fronted by an &#039;&#039;&#039;NGINX&#039;&#039;&#039; reverse proxy that manages incoming traffic and supports &#039;&#039;&#039;HTTP/3 (QUIC)&#039;&#039;&#039; and &#039;&#039;&#039;HTTP/2&#039;&#039;&#039; alongside HTTP/1.1 for broad client compatibility. Behind this entry point, a dual-backend architecture separates concerns by function: a &#039;&#039;&#039;PHP 8&#039;&#039;&#039; environment handles general information pages and the PANGAEA Wiki, while a &#039;&#039;&#039;Java 17&#039;&#039;&#039; application layer running on &#039;&#039;&#039;Eclipse Jetty 12&#039;&#039;&#039; manages core repository services — including dataset landing pages, DOI and handle resolution, and content negotiation for both human- and machine-readable data access. The Jetty-based layer also manages delivery of high-volume and binary files from the tape archive by staging requested files to a local hard-disk cache before serving them to users.&lt;br /&gt;
&lt;br /&gt;
=== Search ===&lt;br /&gt;
The PANGAEA search engine supports full-text and faceted search across all published metadata. Faceted navigation is enabled by semantic metadata enrichment performed during the marshaling process: terminology-based annotations are added to metadata records, allowing consistent filtering by topic, device type, geographic region, and other dimensions. PANGAEA&#039;s approach to data discovery complies with the recent recommendations of the Research Data Alliance (RDA) Interest Group &amp;quot;Data Discovery Paradigms&amp;quot; (Wu, M. et al., 2026, &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;). Search documentation is available in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/PANGAEA_search&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Map Search ===&lt;br /&gt;
Geographic search and visualization are implemented using &#039;&#039;&#039;leaflet.js&#039;&#039;&#039;, serving map tiles from four configurable sources: the AWI basemap, OpenStreetMap, Google (hybrid), and ESRI.&lt;br /&gt;
&lt;br /&gt;
=== Dataset Landing Pages ===&lt;br /&gt;
Each published dataset is represented by a landing page resolved through its DOI. Landing pages present the full dataset metadata in human-readable form and are enriched with structured &#039;&#039;&#039;schema.org&#039;&#039;&#039; markup in JSON-LD extended with Croissant properties, ensuring machine-actionability and compatibility with generic web search engines and data registries. The schema.org metadata is also accessible via HTTP content negotiation following the Signposting standard. Minor editorial corrections to datasets that do not affect scientific content are documented in a &amp;quot;Change history&amp;quot; section of the landing page, recording the date and a summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
=== Programmatic Access ===&lt;br /&gt;
PANGAEA offers programmatic access to data and metadata through a range of web services (REST and SOAP). The OAI-PMH endpoint supports metadata harvesting in all supported standards. Client libraries for &#039;&#039;&#039;Python&#039;&#039;&#039; (pangaeapy, developed by PANGAEA) and &#039;&#039;&#039;R&#039;&#039;&#039; (pangaear, developed by the community) allow researchers to load and transform PANGAEA data directly into native data structures for analysis in environments such as Jupyter notebooks.&lt;br /&gt;
&lt;br /&gt;
== Monitoring and Analytics ==&lt;br /&gt;
Service health across the PANGAEA infrastructure is monitored through the &#039;&#039;&#039;AWI Grafana/Telegraf&#039;&#039;&#039; stack, covering service availability, application logs, and resource saturation for all production systems. This is supplemented by external availability checks via &#039;&#039;&#039;UptimeRobot&#039;&#039;&#039;, which provides independent verification of public-facing service endpoints from outside the AWI network. All external links in PANGAEA metadata records — references to related literature, other dataset versions, and external resources — are automatically checked for broken (HTTP 404) or permanently redirected (HTTP 301) responses on a weekly basis.&lt;br /&gt;
&lt;br /&gt;
User engagement and download metrics are captured by an integrated **Matomo** analytics instance configured to produce usage statistics compliant with the **COUNTER** (Counting Online Usage of Networked Electronic Resources) standard, ensuring that data impact is measured according to international scholarly norms. Usage statistics are publicly visible on each dataset landing page; details are documented in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/Data_Usage_Statistics&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Security ==&lt;br /&gt;
Backend and middleware systems are protected behind a firewall; frontend systems operate in a demilitarized zone (DMZ) accessible from outside with restricted but still firewalled access. Frontend systems have no write access and only limited read access to the backend database and tape archives. Public access from the website and REST APIs is served from data replicas hosted in Elasticsearch or through read-only remote filesystem access. Data curators access production systems via virtual private networks (VPN), which enforce a basic check that client operating systems are up to date.&lt;br /&gt;
&lt;br /&gt;
Physical access to the AWI computing center is managed through an electronic access control system for all relevant entrances; key distribution is documented, and guest policies are formally established. The AWI maintains uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation. Fire alerts are automatically forwarded to the fire department with a contractual response time under 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Security of the technical infrastructure is further maintained through the use of asymmetric key infrastructures, mandatory minimum-length passwords for all user classes, short-cycle security patching for all hardware and software components, professional monitoring tools for hardware, firewall, software, services, performance, and attacks, and regular security training for all technical and non-technical staff. Security risks are assessed on an ongoing basis by AWI&#039;s institutional IT security team, which coordinates incident response and monitors threat landscapes relevant to research data infrastructure. PANGAEA technical staff participates in regular security reviews.&lt;br /&gt;
&lt;br /&gt;
Access controls for restricted or moratorium datasets are enforced at the application layer. Dataset metadata remains publicly accessible in all cases; access to the data itself is restricted to authorized users at the individual level for the duration of the moratorium, typically a maximum of two years from submission.&lt;br /&gt;
&lt;br /&gt;
The off-site replica at MARUM operates in a fully isolated network environment (Layer 2 separation) with access restricted to a firewalled VPN gateway at AWI. The Green IT Housing Center is monitored 24/7 by Bremen University staff located in very close proximity to the datacenter (15 m, adjacent building), with automated fire alarms, redundant power supply with battery backup, and multilevel physical access control.&lt;br /&gt;
&lt;br /&gt;
== Software Inventory ==&lt;br /&gt;
The following table summarizes the principal software components of the PANGAEA infrastructure and their development model.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Component !! Software !! Development Model&lt;br /&gt;
|-&lt;br /&gt;
| Primary database || PostgreSQL || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Data warehouse || Clickhouse || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Search and metadata index || Elasticsearch || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system backend || Java 17 / Dropwizard || In-house development&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system frontend || React / Ant Design || In-house development (open source framework)&lt;br /&gt;
|-&lt;br /&gt;
| Source control || GitLab || Open source (community-supported, self-hosted)&lt;br /&gt;
|-&lt;br /&gt;
| CI/CD pipeline || GitLab CI/CD || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web server / reverse proxy (editorial) || Apache || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web traffic management || NGINX || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| General web and Wiki frontend || PHP 8 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Repository services backend || Java 17 / Eclipse Jetty 12 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Map visualization || leaflet.js || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Metadata framework || PanFMP || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| Issue tracking || JIRA (Atlassian) || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Monitoring stack || Grafana / Telegraf || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| External uptime monitoring || UptimeRobot || Commercial SaaS&lt;br /&gt;
|-&lt;br /&gt;
| Analytics || Matomo || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Infrastructure virtualization || VMware || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Tape archive hardware || SpectraLogic TFinity ExaScale || Commercial hardware&lt;br /&gt;
|-&lt;br /&gt;
| Python data client || pangaeapy || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| R data client || pangaear || Community development (open source)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Documentation and Change Management ==&lt;br /&gt;
General documentation of PANGAEA systems and services is maintained in the public &#039;&#039;&#039;PANGAEA Wiki&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/&amp;lt;/nowiki&amp;gt;). A separate internal &#039;&#039;&#039;Confluence&#039;&#039;&#039; Wiki, kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes, contains detailed documentation of server configurations, installation and maintenance procedures, relationships between system components, VM snapshot procedures, backup routines, and service restart priorities for incident response.&lt;br /&gt;
&lt;br /&gt;
All changes to published data and metadata are recorded in the editorial system&#039;s version history. Substantive revisions to published datasets result in a new dataset version with a new DOI; all prior versions remain accessible and cross-referenced. Minor editorial corrections that do not affect scientific content are applied without creating a new identifier and are transparently documented in the &amp;quot;Change history&amp;quot; section of the dataset landing page, including the date and a brief summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science. &#039;&#039;Scientific Data&#039;&#039;, 10, 347. &amp;lt;nowiki&amp;gt;https://doi.org/10.1038/s41597-023-02269-x&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Diepenbroek, M., Schindler, U., Huber, R., Pesant, S., Stocker, M., Felden, J., Buss, M. &amp;amp; Weinrebe, M. (2017). Terminology supported archiving and publication of environmental science data in PANGAEA. &#039;&#039;Journal of Biotechnology&#039;&#039;, 261, 177–186. &amp;lt;nowiki&amp;gt;https://doi.org/10.1016/j.jbiotec.2017.07.016&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Wu, M. et al. (2026). RDA Interest Group on Data Discovery Paradigms. &#039;&#039;Data Science Journal&#039;&#039;. &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
*[[PANGAEA search]]&lt;br /&gt;
*[[Data submission]]&lt;br /&gt;
*[[Authors Guides]]&lt;br /&gt;
*[[Curation levels]]&lt;br /&gt;
*[[Processing levels]]&lt;br /&gt;
*[[Data Usage Statistics]]&lt;br /&gt;
*[[Format]]&lt;br /&gt;
*[[PANGAEA XML schema]]&lt;br /&gt;
*[[Best practice manuals and templates]]&lt;br /&gt;
*[[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
*[[Preservation Plan]] — https://www.pangaea.de/about/preservation.php&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16804</id>
		<title>Technology</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16804"/>
		<updated>2026-04-10T16:39:04Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Removed construction flag&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;PANGAEA is built on a three-tiered client/server architecture that has evolved continuously since the system&#039;s founding in the early 1990s. The architecture separates backend storage and database systems, middleware processing and transformation components, and frontend delivery and user interfaces into distinct layers that are individually maintainable and replaceable. This design philosophy has enabled PANGAEA to migrate core components — including its primary database engine, its editorial system, and its search infrastructure — without disruption to archived data or to users, and underpins the system&#039;s long-term stability as a trustworthy data repository.&lt;br /&gt;
&lt;br /&gt;
A detailed description of the PANGAEA information system and its workflows is provided in Felden et al. (2023) and Diepenbroek et al. (2017); this article provides an up-to-date overview oriented toward the general structure and key components.&lt;br /&gt;
&lt;br /&gt;
== System Architecture ==&lt;br /&gt;
&lt;br /&gt;
The technical architecture of PANGAEA follows a three-tiered model comprising a backend, a middleware layer, and a frontend. All hard- and software services are hosted and operated by the data and computing center of the Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research (AWI) in Bremerhaven. Most backend and middleware systems, as well as all frontend web servers and search engines, run on virtual machines (VMware) operated with Ubuntu Linux. Virtualization provides sufficient capacity and performance while enabling high availability and transparent hardware renewal on a typical cycle of three to four years.&lt;br /&gt;
[[File:PANGAEA overview architecture.png|alt=Schema illustrating the system architecture according to Felden et al., 2023|thumb|Schema illustrating the system architecture according to Felden et al., 2023|none|600x600px]]&lt;br /&gt;
&lt;br /&gt;
PANGAEA currently operates nine virtual machines at the AWI computing center, using a total of 53 CPUs, 162 GB RAM, and 28 TB of disk space. The relational database holdings currently occupy approximately 5 TB, with approximately 1 PB of data stored on tape.&lt;br /&gt;
&lt;br /&gt;
== Backend: Storage and Databases ==&lt;br /&gt;
&lt;br /&gt;
=== Relational Database ===&lt;br /&gt;
The primary store for all structured data and metadata in PANGAEA is a &#039;&#039;&#039;PostgreSQL&#039;&#039;&#039; relational database management system (RDBMS). The data model is highly normalized, reflecting the full observational context of each measurement: events (when and where data were collected), campaigns (cruises or field expeditions), methods and devices, parameters, references, and institutional provenance information are all stored in defined relational structures. This normalized model allows data descriptions to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived data.&lt;br /&gt;
&lt;br /&gt;
Database integrity is continuously maintained through &#039;&#039;&#039;PostgreSQL streaming replication&#039;&#039;&#039; to a dedicated backup system, which enables point-in-time recovery to any moment prior to a failure event. In addition, full database backup copies (`pg_base_backup`) are created each weekend and retained for three weeks, providing an independent recovery layer supplementary to the streaming replica.&lt;br /&gt;
&lt;br /&gt;
=== Data Warehouse ===&lt;br /&gt;
Fast access to large, aggregated compilations of data across the full PANGAEA holdings is provided by a &#039;&#039;&#039;Clickhouse&#039;&#039;&#039; data warehouse, which mirrors the numerical data inventory from the relational database. The data warehouse supports spatially and chronologically constrained queries at the parameter level across potentially hundreds of thousands of individual datasets, and is accessible both through the PANGAEA website and programmatically via a REST API. All data warehouse exports include the DOI for each contributing dataset, ensuring that provenance and citation remain intact in compiled data products.&lt;br /&gt;
&lt;br /&gt;
=== Archival Storage ===&lt;br /&gt;
For high-volume and binary data — including geophysical datasets, images, video, and community-specific formats such as NetCDF — data are stored in consistent formats on hard-disk arrays and robotic tape archives. The central archival storage system consists of two &#039;&#039;&#039;SpectraLogic TFinity ExaScale&#039;&#039;&#039; robotic tape libraries housed in separate buildings at AWI, with a combined capacity of up to 60 PB. Backup operations use high-capacity LTO-tape drives within this environment.&lt;br /&gt;
&lt;br /&gt;
All data are stored redundantly using &#039;&#039;&#039;erasure coding&#039;&#039;&#039; across disk and tape. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots. Write caches are battery-backed to ensure data integrity at the point of write.&lt;br /&gt;
&lt;br /&gt;
=== Off-Site Replica ===&lt;br /&gt;
Since 2025, PANGAEA operates a **minimal viable repository service** as an off-site replica at MARUM/University of Bremen, hosted in the Green IT Housing Center of Bremen University. This facility provides geographic and infrastructural separation from the primary AWI systems, enabling recovery of data delivery services within 24 hours following a technical breakdown or cyberattack. Disaster recovery and switchover procedures involving this facility are regularly exercised. The replica is kept current through snapshot-based replication several times per day and currently covers all dataset metadata — including individual landing pages and all harvesting endpoints — full representations of tabular data publications, the Elasticsearch search index, and the relevant web frontend. Extension of the replica to include binary data files is under active development, with formal commitment to be finalized through the ongoing MARUM hosting agreement negotiations.&lt;br /&gt;
&lt;br /&gt;
The off-site installation operates in a fully isolated environment with Layer 2 network separation. Access is restricted to VPN connections from a single designated gateway host at AWI, with access tokens and keys limited to the corresponding gateway and PANGAEA DevOps staff.&lt;br /&gt;
== Middleware: Processing and Transformation ==&lt;br /&gt;
The middleware layer manages the flow of data and metadata between the backend storage systems and the frontend interfaces. Its core functions are marshaling, indexing, transformation, and dissemination.&lt;br /&gt;
&lt;br /&gt;
Metadata are dynamically marshaled from the relational database to a PANGAEA-specific internal XML format, from which they are transformed via &#039;&#039;&#039;XSLT and XML-to-JSON pipelines&#039;&#039;&#039; into a range of content standards for delivery to users and harvesting services. Currently supported output standards include JSON-LD (Schema.org with Croissant extension), DataCite XML, Dublin Core XML, ISO 19115/ISO 19139, and DIF/FGDC. Dissemination occurs via OAI-PMH, HTTP content negotiation following the Signposting standard (&amp;lt;nowiki&amp;gt;https://signposting.org/&amp;lt;/nowiki&amp;gt;), and other protocols. Metadata marshaling from the PostgreSQL database to the Elasticsearch search index is handled asynchronously by dedicated middleware components, which also manage automated DOI minting at DataCite upon dataset publication.&lt;br /&gt;
&lt;br /&gt;
The marshaled metadata are stored and indexed in &#039;&#039;&#039;Elasticsearch&#039;&#039;&#039;, which serves as the primary index for all public search and metadata access interfaces. This architecture separates the authoritative relational record from the search-optimized representation, allowing the search index to be rebuilt or updated without modifying archived data.&lt;br /&gt;
&lt;br /&gt;
The flexible metadata framework &#039;&#039;&#039;PanFMP&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://www.panfmp.org/&amp;lt;/nowiki&amp;gt;) underpins the mapping between the PANGAEA internal schema and the various external output standards, ensuring that new or updated standards can be accommodated without structural changes to the underlying data.&lt;br /&gt;
&lt;br /&gt;
Data submissions, user requests, and bug reports are managed through a &#039;&#039;&#039;JIRA&#039;&#039;&#039; (Atlassian) issue tracking system, which serves as the primary communication channel between data providers and the PANGAEA editorial team throughout the submission and publication process.&lt;br /&gt;
== Frontend: Editorial System ==&lt;br /&gt;
The PANGAEA editorial system is the primary tool through which data editors review, curate, harmonize, and publish submitted data and metadata. It is a &#039;&#039;&#039;web-based client/server application&#039;&#039;&#039; developed entirely in-house, operating directly on the PostgreSQL databases.&lt;br /&gt;
&lt;br /&gt;
The backend of the editorial system is built on &#039;&#039;&#039;Java 17&#039;&#039;&#039; using the &#039;&#039;&#039;Dropwizard&#039;&#039;&#039; framework, exposing a REST API. The frontend is implemented in &#039;&#039;&#039;React&#039;&#039;&#039; with the &#039;&#039;&#039;Ant Design&#039;&#039;&#039; component library. Source code for both components, together with shared libraries and test suites, is maintained in an institutional &#039;&#039;&#039;GitLab&#039;&#039;&#039; instance hosted at AWI, structured as an &#039;&#039;&#039;Nx monorepo&#039;&#039;&#039;. The repository encompasses automated unit tests and &#039;&#039;&#039;Cypress&#039;&#039;&#039; end-to-end tests. All new versions are deployed exclusively through a &#039;&#039;&#039;GitLab CI/CD pipeline&#039;&#039;&#039;, with releases gated on successful completion of the full automated test suite.&lt;br /&gt;
&lt;br /&gt;
In production, the editorial system runs on Ubuntu virtual machines hosted on VMware infrastructure. &#039;&#039;&#039;Apache&#039;&#039;&#039; serves the React frontend and acts as a reverse proxy to the backend services. Four parallel instances are operated simultaneously, providing dedicated test and production environments for two editorial system versions at any given time. This setup supports controlled feature validation alongside uninterrupted service operation. Infrastructure components — including VMware, file systems, monitoring, GitLab, and the database platform — are managed by AWI, while the PANGAEA group is responsible for the Apache web server, Java services, and React frontend.&lt;br /&gt;
&lt;br /&gt;
The relational database model underlying the editorial system enforces structural and semantic consistency between submitted metadata and the existing data inventory. The editorial system also integrates a &#039;&#039;&#039;Terminology Catalogue (TC)&#039;&#039;&#039;, which manages controlled vocabularies and ontologies used to harmonize data and metadata during ingest. Supported terminologies include WoRMS, ITIS, ChEBI, EnvO, PATO, QUDT, and the NERC vocabulary server, among others. A detailed description of the Terminology Catalogue and its role in data archiving and publication is given in Diepenbroek et al. (2017).&lt;br /&gt;
== Frontend: Public Web Interface and Search ==&lt;br /&gt;
&lt;br /&gt;
=== Web Delivery Stack ===&lt;br /&gt;
The public web infrastructure is fronted by an &#039;&#039;&#039;NGINX&#039;&#039;&#039; reverse proxy that manages incoming traffic and supports &#039;&#039;&#039;HTTP/3 (QUIC)&#039;&#039;&#039; and &#039;&#039;&#039;HTTP/2&#039;&#039;&#039; alongside HTTP/1.1 for broad client compatibility. Behind this entry point, a dual-backend architecture separates concerns by function: a &#039;&#039;&#039;PHP 8&#039;&#039;&#039; environment handles general information pages and the PANGAEA Wiki, while a &#039;&#039;&#039;Java 17&#039;&#039;&#039; application layer running on &#039;&#039;&#039;Eclipse Jetty 12&#039;&#039;&#039; manages core repository services — including dataset landing pages, DOI and handle resolution, and content negotiation for both human- and machine-readable data access. The Jetty-based layer also manages delivery of high-volume and binary files from the tape archive by staging requested files to a local hard-disk cache before serving them to users.&lt;br /&gt;
&lt;br /&gt;
=== Search ===&lt;br /&gt;
The PANGAEA search engine supports full-text and faceted search across all published metadata. Faceted navigation is enabled by semantic metadata enrichment performed during the marshaling process: terminology-based annotations are added to metadata records, allowing consistent filtering by topic, device type, geographic region, and other dimensions. PANGAEA&#039;s approach to data discovery complies with the recent recommendations of the Research Data Alliance (RDA) Interest Group &amp;quot;Data Discovery Paradigms&amp;quot; (Wu, M. et al., 2026, &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;). Search documentation is available in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/PANGAEA_search&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Map Search ===&lt;br /&gt;
Geographic search and visualization are implemented using &#039;&#039;&#039;leaflet.js&#039;&#039;&#039;, serving map tiles from four configurable sources: the AWI basemap, OpenStreetMap, Google (hybrid), and ESRI.&lt;br /&gt;
&lt;br /&gt;
=== Dataset Landing Pages ===&lt;br /&gt;
Each published dataset is represented by a landing page resolved through its DOI. Landing pages present the full dataset metadata in human-readable form and are enriched with structured &#039;&#039;&#039;schema.org&#039;&#039;&#039; markup in JSON-LD extended with Croissant properties, ensuring machine-actionability and compatibility with generic web search engines and data registries. The schema.org metadata is also accessible via HTTP content negotiation following the Signposting standard. Minor editorial corrections to datasets that do not affect scientific content are documented in a &amp;quot;Change history&amp;quot; section of the landing page, recording the date and a summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
=== Programmatic Access ===&lt;br /&gt;
PANGAEA offers programmatic access to data and metadata through a range of web services (REST and SOAP). The OAI-PMH endpoint supports metadata harvesting in all supported standards. Client libraries for &#039;&#039;&#039;Python&#039;&#039;&#039; (pangaeapy, developed by PANGAEA) and &#039;&#039;&#039;R&#039;&#039;&#039; (pangaear, developed by the community) allow researchers to load and transform PANGAEA data directly into native data structures for analysis in environments such as Jupyter notebooks.&lt;br /&gt;
&lt;br /&gt;
== Monitoring and Analytics ==&lt;br /&gt;
Service health across the PANGAEA infrastructure is monitored through the &#039;&#039;&#039;AWI Grafana/Telegraf&#039;&#039;&#039; stack, covering service availability, application logs, and resource saturation for all production systems. This is supplemented by external availability checks via &#039;&#039;&#039;UptimeRobot&#039;&#039;&#039;, which provides independent verification of public-facing service endpoints from outside the AWI network. All external links in PANGAEA metadata records — references to related literature, other dataset versions, and external resources — are automatically checked for broken (HTTP 404) or permanently redirected (HTTP 301) responses on a weekly basis.&lt;br /&gt;
&lt;br /&gt;
User engagement and download metrics are captured by an integrated **Matomo** analytics instance configured to produce usage statistics compliant with the **COUNTER** (Counting Online Usage of Networked Electronic Resources) standard, ensuring that data impact is measured according to international scholarly norms. Usage statistics are publicly visible on each dataset landing page; details are documented in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/Data_Usage_Statistics&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Security ==&lt;br /&gt;
Backend and middleware systems are protected behind a firewall; frontend systems operate in a demilitarized zone (DMZ) accessible from outside with restricted but still firewalled access. Frontend systems have no write access and only limited read access to the backend database and tape archives. Public access from the website and REST APIs is served from data replicas hosted in Elasticsearch or through read-only remote filesystem access. Data curators access production systems via virtual private networks (VPN), which enforce a basic check that client operating systems are up to date.&lt;br /&gt;
&lt;br /&gt;
Physical access to the AWI computing center is managed through an electronic access control system for all relevant entrances; key distribution is documented, and guest policies are formally established. The AWI maintains uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation. Fire alerts are automatically forwarded to the fire department with a contractual response time under 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Security of the technical infrastructure is further maintained through the use of asymmetric key infrastructures, mandatory minimum-length passwords for all user classes, short-cycle security patching for all hardware and software components, professional monitoring tools for hardware, firewall, software, services, performance, and attacks, and regular security training for all technical and non-technical staff. Security risks are assessed on an ongoing basis by AWI&#039;s institutional IT security team, which coordinates incident response and monitors threat landscapes relevant to research data infrastructure. PANGAEA technical staff participates in regular security reviews.&lt;br /&gt;
&lt;br /&gt;
Access controls for restricted or moratorium datasets are enforced at the application layer. Dataset metadata remains publicly accessible in all cases; access to the data itself is restricted to authorized users at the individual level for the duration of the moratorium, typically a maximum of two years from submission.&lt;br /&gt;
&lt;br /&gt;
The off-site replica at MARUM operates in a fully isolated network environment (Layer 2 separation) with access restricted to a firewalled VPN gateway at AWI. The Green IT Housing Center is monitored 24/7 by Bremen University staff located in very close proximity to the datacenter (15 m, adjacent building), with automated fire alarms, redundant power supply with battery backup, and multilevel physical access control.&lt;br /&gt;
&lt;br /&gt;
== Software Inventory ==&lt;br /&gt;
The following table summarizes the principal software components of the PANGAEA infrastructure and their development model.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Component !! Software !! Development Model&lt;br /&gt;
|-&lt;br /&gt;
| Primary database || PostgreSQL || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Data warehouse || Clickhouse || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Search and metadata index || Elasticsearch || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system backend || Java 17 / Dropwizard || In-house development&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system frontend || React / Ant Design || In-house development (open source framework)&lt;br /&gt;
|-&lt;br /&gt;
| Source control || GitLab || Open source (community-supported, self-hosted)&lt;br /&gt;
|-&lt;br /&gt;
| CI/CD pipeline || GitLab CI/CD || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web server / reverse proxy (editorial) || Apache || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web traffic management || NGINX || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| General web and Wiki frontend || PHP 8 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Repository services backend || Java 17 / Eclipse Jetty 12 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Map visualization || leaflet.js || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Metadata framework || PanFMP || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| Issue tracking || JIRA (Atlassian) || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Monitoring stack || Grafana / Telegraf || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| External uptime monitoring || UptimeRobot || Commercial SaaS&lt;br /&gt;
|-&lt;br /&gt;
| Analytics || Matomo || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Infrastructure virtualization || VMware || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Tape archive hardware || SpectraLogic TFinity ExaScale || Commercial hardware&lt;br /&gt;
|-&lt;br /&gt;
| Python data client || pangaeapy || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| R data client || pangaear || Community development (open source)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Documentation and Change Management ==&lt;br /&gt;
General documentation of PANGAEA systems and services is maintained in the public &#039;&#039;&#039;PANGAEA Wiki&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/&amp;lt;/nowiki&amp;gt;). A separate internal &#039;&#039;&#039;Confluence&#039;&#039;&#039; Wiki, kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes, contains detailed documentation of server configurations, installation and maintenance procedures, relationships between system components, VM snapshot procedures, backup routines, and service restart priorities for incident response.&lt;br /&gt;
&lt;br /&gt;
All changes to published data and metadata are recorded in the editorial system&#039;s version history. Substantive revisions to published datasets result in a new dataset version with a new DOI; all prior versions remain accessible and cross-referenced. Minor editorial corrections that do not affect scientific content are applied without creating a new identifier and are transparently documented in the &amp;quot;Change history&amp;quot; section of the dataset landing page, including the date and a brief summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science. &#039;&#039;Scientific Data&#039;&#039;, 10, 347. &amp;lt;nowiki&amp;gt;https://doi.org/10.1038/s41597-023-02269-x&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Diepenbroek, M., Schindler, U., Huber, R., Pesant, S., Stocker, M., Felden, J., Buss, M. &amp;amp; Weinrebe, M. (2017). Terminology supported archiving and publication of environmental science data in PANGAEA. &#039;&#039;Journal of Biotechnology&#039;&#039;, 261, 177–186. &amp;lt;nowiki&amp;gt;https://doi.org/10.1016/j.jbiotec.2017.07.016&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Wu, M. et al. (2026). RDA Interest Group on Data Discovery Paradigms. &#039;&#039;Data Science Journal&#039;&#039;. &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
*[[PANGAEA search]]&lt;br /&gt;
*[[Data submission]]&lt;br /&gt;
*[[Authors Guides]]&lt;br /&gt;
*[[Curation levels]]&lt;br /&gt;
*[[Processing levels]]&lt;br /&gt;
*[[Data Usage Statistics]]&lt;br /&gt;
*[[Format]]&lt;br /&gt;
*[[PANGAEA XML schema]]&lt;br /&gt;
*[[Best practice manuals and templates]]&lt;br /&gt;
*[[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
*[[Preservation Plan]] — https://www.pangaea.de/about/preservation.php&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16803</id>
		<title>Technology</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16803"/>
		<updated>2026-04-10T16:27:48Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Dataset Landing Pages */ added Croissant extension&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{construction}}&lt;br /&gt;
&lt;br /&gt;
PANGAEA is built on a three-tiered client/server architecture that has evolved continuously since the system&#039;s founding in the early 1990s. The architecture separates backend storage and database systems, middleware processing and transformation components, and frontend delivery and user interfaces into distinct layers that are individually maintainable and replaceable. This design philosophy has enabled PANGAEA to migrate core components — including its primary database engine, its editorial system, and its search infrastructure — without disruption to archived data or to users, and underpins the system&#039;s long-term stability as a trustworthy data repository.&lt;br /&gt;
&lt;br /&gt;
A detailed description of the PANGAEA information system and its workflows is provided in Felden et al. (2023) and Diepenbroek et al. (2017); this article provides an up-to-date overview oriented toward the general structure and key components.&lt;br /&gt;
&lt;br /&gt;
== System Architecture ==&lt;br /&gt;
&lt;br /&gt;
The technical architecture of PANGAEA follows a three-tiered model comprising a backend, a middleware layer, and a frontend. All hard- and software services are hosted and operated by the data and computing center of the Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research (AWI) in Bremerhaven. Most backend and middleware systems, as well as all frontend web servers and search engines, run on virtual machines (VMware) operated with Ubuntu Linux. Virtualization provides sufficient capacity and performance while enabling high availability and transparent hardware renewal on a typical cycle of three to four years.&lt;br /&gt;
[[File:PANGAEA overview architecture.png|alt=Schema illustrating the system architecture according to Felden et al., 2023|thumb|Schema illustrating the system architecture according to Felden et al., 2023|none|600x600px]]&lt;br /&gt;
&lt;br /&gt;
PANGAEA currently operates nine virtual machines at the AWI computing center, using a total of 53 CPUs, 162 GB RAM, and 28 TB of disk space. The relational database holdings currently occupy approximately 5 TB, with approximately 1 PB of data stored on tape.&lt;br /&gt;
&lt;br /&gt;
== Backend: Storage and Databases ==&lt;br /&gt;
&lt;br /&gt;
=== Relational Database ===&lt;br /&gt;
The primary store for all structured data and metadata in PANGAEA is a &#039;&#039;&#039;PostgreSQL&#039;&#039;&#039; relational database management system (RDBMS). The data model is highly normalized, reflecting the full observational context of each measurement: events (when and where data were collected), campaigns (cruises or field expeditions), methods and devices, parameters, references, and institutional provenance information are all stored in defined relational structures. This normalized model allows data descriptions to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived data.&lt;br /&gt;
&lt;br /&gt;
Database integrity is continuously maintained through &#039;&#039;&#039;PostgreSQL streaming replication&#039;&#039;&#039; to a dedicated backup system, which enables point-in-time recovery to any moment prior to a failure event. In addition, full database backup copies (`pg_base_backup`) are created each weekend and retained for three weeks, providing an independent recovery layer supplementary to the streaming replica.&lt;br /&gt;
&lt;br /&gt;
=== Data Warehouse ===&lt;br /&gt;
Fast access to large, aggregated compilations of data across the full PANGAEA holdings is provided by a &#039;&#039;&#039;Clickhouse&#039;&#039;&#039; data warehouse, which mirrors the numerical data inventory from the relational database. The data warehouse supports spatially and chronologically constrained queries at the parameter level across potentially hundreds of thousands of individual datasets, and is accessible both through the PANGAEA website and programmatically via a REST API. All data warehouse exports include the DOI for each contributing dataset, ensuring that provenance and citation remain intact in compiled data products.&lt;br /&gt;
&lt;br /&gt;
=== Archival Storage ===&lt;br /&gt;
For high-volume and binary data — including geophysical datasets, images, video, and community-specific formats such as NetCDF — data are stored in consistent formats on hard-disk arrays and robotic tape archives. The central archival storage system consists of two &#039;&#039;&#039;SpectraLogic TFinity ExaScale&#039;&#039;&#039; robotic tape libraries housed in separate buildings at AWI, with a combined capacity of up to 60 PB. Backup operations use high-capacity LTO-tape drives within this environment.&lt;br /&gt;
&lt;br /&gt;
All data are stored redundantly using &#039;&#039;&#039;erasure coding&#039;&#039;&#039; across disk and tape. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots. Write caches are battery-backed to ensure data integrity at the point of write.&lt;br /&gt;
&lt;br /&gt;
=== Off-Site Replica ===&lt;br /&gt;
Since 2025, PANGAEA operates a **minimal viable repository service** as an off-site replica at MARUM/University of Bremen, hosted in the Green IT Housing Center of Bremen University. This facility provides geographic and infrastructural separation from the primary AWI systems, enabling recovery of data delivery services within 24 hours following a technical breakdown or cyberattack. Disaster recovery and switchover procedures involving this facility are regularly exercised. The replica is kept current through snapshot-based replication several times per day and currently covers all dataset metadata — including individual landing pages and all harvesting endpoints — full representations of tabular data publications, the Elasticsearch search index, and the relevant web frontend. Extension of the replica to include binary data files is under active development, with formal commitment to be finalized through the ongoing MARUM hosting agreement negotiations.&lt;br /&gt;
&lt;br /&gt;
The off-site installation operates in a fully isolated environment with Layer 2 network separation. Access is restricted to VPN connections from a single designated gateway host at AWI, with access tokens and keys limited to the corresponding gateway and PANGAEA DevOps staff.&lt;br /&gt;
== Middleware: Processing and Transformation ==&lt;br /&gt;
The middleware layer manages the flow of data and metadata between the backend storage systems and the frontend interfaces. Its core functions are marshaling, indexing, transformation, and dissemination.&lt;br /&gt;
&lt;br /&gt;
Metadata are dynamically marshaled from the relational database to a PANGAEA-specific internal XML format, from which they are transformed via &#039;&#039;&#039;XSLT and XML-to-JSON pipelines&#039;&#039;&#039; into a range of content standards for delivery to users and harvesting services. Currently supported output standards include JSON-LD (Schema.org with Croissant extension), DataCite XML, Dublin Core XML, ISO 19115/ISO 19139, and DIF/FGDC. Dissemination occurs via OAI-PMH, HTTP content negotiation following the Signposting standard (&amp;lt;nowiki&amp;gt;https://signposting.org/&amp;lt;/nowiki&amp;gt;), and other protocols. Metadata marshaling from the PostgreSQL database to the Elasticsearch search index is handled asynchronously by dedicated middleware components, which also manage automated DOI minting at DataCite upon dataset publication.&lt;br /&gt;
&lt;br /&gt;
The marshaled metadata are stored and indexed in &#039;&#039;&#039;Elasticsearch&#039;&#039;&#039;, which serves as the primary index for all public search and metadata access interfaces. This architecture separates the authoritative relational record from the search-optimized representation, allowing the search index to be rebuilt or updated without modifying archived data.&lt;br /&gt;
&lt;br /&gt;
The flexible metadata framework &#039;&#039;&#039;PanFMP&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://www.panfmp.org/&amp;lt;/nowiki&amp;gt;) underpins the mapping between the PANGAEA internal schema and the various external output standards, ensuring that new or updated standards can be accommodated without structural changes to the underlying data.&lt;br /&gt;
&lt;br /&gt;
Data submissions, user requests, and bug reports are managed through a &#039;&#039;&#039;JIRA&#039;&#039;&#039; (Atlassian) issue tracking system, which serves as the primary communication channel between data providers and the PANGAEA editorial team throughout the submission and publication process.&lt;br /&gt;
== Frontend: Editorial System ==&lt;br /&gt;
The PANGAEA editorial system is the primary tool through which data editors review, curate, harmonize, and publish submitted data and metadata. It is a &#039;&#039;&#039;web-based client/server application&#039;&#039;&#039; developed entirely in-house, operating directly on the PostgreSQL databases.&lt;br /&gt;
&lt;br /&gt;
The backend of the editorial system is built on &#039;&#039;&#039;Java 17&#039;&#039;&#039; using the &#039;&#039;&#039;Dropwizard&#039;&#039;&#039; framework, exposing a REST API. The frontend is implemented in &#039;&#039;&#039;React&#039;&#039;&#039; with the &#039;&#039;&#039;Ant Design&#039;&#039;&#039; component library. Source code for both components, together with shared libraries and test suites, is maintained in an institutional &#039;&#039;&#039;GitLab&#039;&#039;&#039; instance hosted at AWI, structured as an &#039;&#039;&#039;Nx monorepo&#039;&#039;&#039;. The repository encompasses automated unit tests and &#039;&#039;&#039;Cypress&#039;&#039;&#039; end-to-end tests. All new versions are deployed exclusively through a &#039;&#039;&#039;GitLab CI/CD pipeline&#039;&#039;&#039;, with releases gated on successful completion of the full automated test suite.&lt;br /&gt;
&lt;br /&gt;
In production, the editorial system runs on Ubuntu virtual machines hosted on VMware infrastructure. &#039;&#039;&#039;Apache&#039;&#039;&#039; serves the React frontend and acts as a reverse proxy to the backend services. Four parallel instances are operated simultaneously, providing dedicated test and production environments for two editorial system versions at any given time. This setup supports controlled feature validation alongside uninterrupted service operation. Infrastructure components — including VMware, file systems, monitoring, GitLab, and the database platform — are managed by AWI, while the PANGAEA group is responsible for the Apache web server, Java services, and React frontend.&lt;br /&gt;
&lt;br /&gt;
The relational database model underlying the editorial system enforces structural and semantic consistency between submitted metadata and the existing data inventory. The editorial system also integrates a &#039;&#039;&#039;Terminology Catalogue (TC)&#039;&#039;&#039;, which manages controlled vocabularies and ontologies used to harmonize data and metadata during ingest. Supported terminologies include WoRMS, ITIS, ChEBI, EnvO, PATO, QUDT, and the NERC vocabulary server, among others. A detailed description of the Terminology Catalogue and its role in data archiving and publication is given in Diepenbroek et al. (2017).&lt;br /&gt;
== Frontend: Public Web Interface and Search ==&lt;br /&gt;
&lt;br /&gt;
=== Web Delivery Stack ===&lt;br /&gt;
The public web infrastructure is fronted by an &#039;&#039;&#039;NGINX&#039;&#039;&#039; reverse proxy that manages incoming traffic and supports &#039;&#039;&#039;HTTP/3 (QUIC)&#039;&#039;&#039; and &#039;&#039;&#039;HTTP/2&#039;&#039;&#039; alongside HTTP/1.1 for broad client compatibility. Behind this entry point, a dual-backend architecture separates concerns by function: a &#039;&#039;&#039;PHP 8&#039;&#039;&#039; environment handles general information pages and the PANGAEA Wiki, while a &#039;&#039;&#039;Java 17&#039;&#039;&#039; application layer running on &#039;&#039;&#039;Eclipse Jetty 12&#039;&#039;&#039; manages core repository services — including dataset landing pages, DOI and handle resolution, and content negotiation for both human- and machine-readable data access. The Jetty-based layer also manages delivery of high-volume and binary files from the tape archive by staging requested files to a local hard-disk cache before serving them to users.&lt;br /&gt;
&lt;br /&gt;
=== Search ===&lt;br /&gt;
The PANGAEA search engine supports full-text and faceted search across all published metadata. Faceted navigation is enabled by semantic metadata enrichment performed during the marshaling process: terminology-based annotations are added to metadata records, allowing consistent filtering by topic, device type, geographic region, and other dimensions. PANGAEA&#039;s approach to data discovery complies with the recent recommendations of the Research Data Alliance (RDA) Interest Group &amp;quot;Data Discovery Paradigms&amp;quot; (Wu, M. et al., 2026, &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;). Search documentation is available in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/PANGAEA_search&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Map Search ===&lt;br /&gt;
Geographic search and visualization are implemented using &#039;&#039;&#039;leaflet.js&#039;&#039;&#039;, serving map tiles from four configurable sources: the AWI basemap, OpenStreetMap, Google (hybrid), and ESRI.&lt;br /&gt;
&lt;br /&gt;
=== Dataset Landing Pages ===&lt;br /&gt;
Each published dataset is represented by a landing page resolved through its DOI. Landing pages present the full dataset metadata in human-readable form and are enriched with structured &#039;&#039;&#039;schema.org&#039;&#039;&#039; markup in JSON-LD extended with Croissant properties, ensuring machine-actionability and compatibility with generic web search engines and data registries. The schema.org metadata is also accessible via HTTP content negotiation following the Signposting standard. Minor editorial corrections to datasets that do not affect scientific content are documented in a &amp;quot;Change history&amp;quot; section of the landing page, recording the date and a summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
=== Programmatic Access ===&lt;br /&gt;
PANGAEA offers programmatic access to data and metadata through a range of web services (REST and SOAP). The OAI-PMH endpoint supports metadata harvesting in all supported standards. Client libraries for &#039;&#039;&#039;Python&#039;&#039;&#039; (pangaeapy, developed by PANGAEA) and &#039;&#039;&#039;R&#039;&#039;&#039; (pangaear, developed by the community) allow researchers to load and transform PANGAEA data directly into native data structures for analysis in environments such as Jupyter notebooks.&lt;br /&gt;
&lt;br /&gt;
== Monitoring and Analytics ==&lt;br /&gt;
Service health across the PANGAEA infrastructure is monitored through the &#039;&#039;&#039;AWI Grafana/Telegraf&#039;&#039;&#039; stack, covering service availability, application logs, and resource saturation for all production systems. This is supplemented by external availability checks via &#039;&#039;&#039;UptimeRobot&#039;&#039;&#039;, which provides independent verification of public-facing service endpoints from outside the AWI network. All external links in PANGAEA metadata records — references to related literature, other dataset versions, and external resources — are automatically checked for broken (HTTP 404) or permanently redirected (HTTP 301) responses on a weekly basis.&lt;br /&gt;
&lt;br /&gt;
User engagement and download metrics are captured by an integrated **Matomo** analytics instance configured to produce usage statistics compliant with the **COUNTER** (Counting Online Usage of Networked Electronic Resources) standard, ensuring that data impact is measured according to international scholarly norms. Usage statistics are publicly visible on each dataset landing page; details are documented in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/Data_Usage_Statistics&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Security ==&lt;br /&gt;
Backend and middleware systems are protected behind a firewall; frontend systems operate in a demilitarized zone (DMZ) accessible from outside with restricted but still firewalled access. Frontend systems have no write access and only limited read access to the backend database and tape archives. Public access from the website and REST APIs is served from data replicas hosted in Elasticsearch or through read-only remote filesystem access. Data curators access production systems via virtual private networks (VPN), which enforce a basic check that client operating systems are up to date.&lt;br /&gt;
&lt;br /&gt;
Physical access to the AWI computing center is managed through an electronic access control system for all relevant entrances; key distribution is documented, and guest policies are formally established. The AWI maintains uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation. Fire alerts are automatically forwarded to the fire department with a contractual response time under 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Security of the technical infrastructure is further maintained through the use of asymmetric key infrastructures, mandatory minimum-length passwords for all user classes, short-cycle security patching for all hardware and software components, professional monitoring tools for hardware, firewall, software, services, performance, and attacks, and regular security training for all technical and non-technical staff. Security risks are assessed on an ongoing basis by AWI&#039;s institutional IT security team, which coordinates incident response and monitors threat landscapes relevant to research data infrastructure. PANGAEA technical staff participates in regular security reviews.&lt;br /&gt;
&lt;br /&gt;
Access controls for restricted or moratorium datasets are enforced at the application layer. Dataset metadata remains publicly accessible in all cases; access to the data itself is restricted to authorized users at the individual level for the duration of the moratorium, typically a maximum of two years from submission.&lt;br /&gt;
&lt;br /&gt;
The off-site replica at MARUM operates in a fully isolated network environment (Layer 2 separation) with access restricted to a firewalled VPN gateway at AWI. The Green IT Housing Center is monitored 24/7 by Bremen University staff located in very close proximity to the datacenter (15 m, adjacent building), with automated fire alarms, redundant power supply with battery backup, and multilevel physical access control.&lt;br /&gt;
&lt;br /&gt;
== Software Inventory ==&lt;br /&gt;
The following table summarizes the principal software components of the PANGAEA infrastructure and their development model.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Component !! Software !! Development Model&lt;br /&gt;
|-&lt;br /&gt;
| Primary database || PostgreSQL || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Data warehouse || Clickhouse || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Search and metadata index || Elasticsearch || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system backend || Java 17 / Dropwizard || In-house development&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system frontend || React / Ant Design || In-house development (open source framework)&lt;br /&gt;
|-&lt;br /&gt;
| Source control || GitLab || Open source (community-supported, self-hosted)&lt;br /&gt;
|-&lt;br /&gt;
| CI/CD pipeline || GitLab CI/CD || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web server / reverse proxy (editorial) || Apache || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web traffic management || NGINX || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| General web and Wiki frontend || PHP 8 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Repository services backend || Java 17 / Eclipse Jetty 12 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Map visualization || leaflet.js || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Metadata framework || PanFMP || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| Issue tracking || JIRA (Atlassian) || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Monitoring stack || Grafana / Telegraf || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| External uptime monitoring || UptimeRobot || Commercial SaaS&lt;br /&gt;
|-&lt;br /&gt;
| Analytics || Matomo || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Infrastructure virtualization || VMware || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Tape archive hardware || SpectraLogic TFinity ExaScale || Commercial hardware&lt;br /&gt;
|-&lt;br /&gt;
| Python data client || pangaeapy || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| R data client || pangaear || Community development (open source)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Documentation and Change Management ==&lt;br /&gt;
General documentation of PANGAEA systems and services is maintained in the public &#039;&#039;&#039;PANGAEA Wiki&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/&amp;lt;/nowiki&amp;gt;). A separate internal &#039;&#039;&#039;Confluence&#039;&#039;&#039; Wiki, kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes, contains detailed documentation of server configurations, installation and maintenance procedures, relationships between system components, VM snapshot procedures, backup routines, and service restart priorities for incident response.&lt;br /&gt;
&lt;br /&gt;
All changes to published data and metadata are recorded in the editorial system&#039;s version history. Substantive revisions to published datasets result in a new dataset version with a new DOI; all prior versions remain accessible and cross-referenced. Minor editorial corrections that do not affect scientific content are applied without creating a new identifier and are transparently documented in the &amp;quot;Change history&amp;quot; section of the dataset landing page, including the date and a brief summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science. &#039;&#039;Scientific Data&#039;&#039;, 10, 347. &amp;lt;nowiki&amp;gt;https://doi.org/10.1038/s41597-023-02269-x&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Diepenbroek, M., Schindler, U., Huber, R., Pesant, S., Stocker, M., Felden, J., Buss, M. &amp;amp; Weinrebe, M. (2017). Terminology supported archiving and publication of environmental science data in PANGAEA. &#039;&#039;Journal of Biotechnology&#039;&#039;, 261, 177–186. &amp;lt;nowiki&amp;gt;https://doi.org/10.1016/j.jbiotec.2017.07.016&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Wu, M. et al. (2026). RDA Interest Group on Data Discovery Paradigms. &#039;&#039;Data Science Journal&#039;&#039;. &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
*[[PANGAEA search]]&lt;br /&gt;
*[[Data submission]]&lt;br /&gt;
*[[Authors Guides]]&lt;br /&gt;
*[[Curation levels]]&lt;br /&gt;
*[[Processing levels]]&lt;br /&gt;
*[[Data Usage Statistics]]&lt;br /&gt;
*[[Format]]&lt;br /&gt;
*[[PANGAEA XML schema]]&lt;br /&gt;
*[[Best practice manuals and templates]]&lt;br /&gt;
*[[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
*[[Preservation Plan]] — https://www.pangaea.de/about/preservation.php&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16802</id>
		<title>Technology</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16802"/>
		<updated>2026-04-10T16:25:19Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Middleware: Processing and Transformation */ added Croissant extension&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{construction}}&lt;br /&gt;
&lt;br /&gt;
PANGAEA is built on a three-tiered client/server architecture that has evolved continuously since the system&#039;s founding in the early 1990s. The architecture separates backend storage and database systems, middleware processing and transformation components, and frontend delivery and user interfaces into distinct layers that are individually maintainable and replaceable. This design philosophy has enabled PANGAEA to migrate core components — including its primary database engine, its editorial system, and its search infrastructure — without disruption to archived data or to users, and underpins the system&#039;s long-term stability as a trustworthy data repository.&lt;br /&gt;
&lt;br /&gt;
A detailed description of the PANGAEA information system and its workflows is provided in Felden et al. (2023) and Diepenbroek et al. (2017); this article provides an up-to-date overview oriented toward the general structure and key components.&lt;br /&gt;
&lt;br /&gt;
== System Architecture ==&lt;br /&gt;
&lt;br /&gt;
The technical architecture of PANGAEA follows a three-tiered model comprising a backend, a middleware layer, and a frontend. All hard- and software services are hosted and operated by the data and computing center of the Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research (AWI) in Bremerhaven. Most backend and middleware systems, as well as all frontend web servers and search engines, run on virtual machines (VMware) operated with Ubuntu Linux. Virtualization provides sufficient capacity and performance while enabling high availability and transparent hardware renewal on a typical cycle of three to four years.&lt;br /&gt;
[[File:PANGAEA overview architecture.png|alt=Schema illustrating the system architecture according to Felden et al., 2023|thumb|Schema illustrating the system architecture according to Felden et al., 2023|none|600x600px]]&lt;br /&gt;
&lt;br /&gt;
PANGAEA currently operates nine virtual machines at the AWI computing center, using a total of 53 CPUs, 162 GB RAM, and 28 TB of disk space. The relational database holdings currently occupy approximately 5 TB, with approximately 1 PB of data stored on tape.&lt;br /&gt;
&lt;br /&gt;
== Backend: Storage and Databases ==&lt;br /&gt;
&lt;br /&gt;
=== Relational Database ===&lt;br /&gt;
The primary store for all structured data and metadata in PANGAEA is a &#039;&#039;&#039;PostgreSQL&#039;&#039;&#039; relational database management system (RDBMS). The data model is highly normalized, reflecting the full observational context of each measurement: events (when and where data were collected), campaigns (cruises or field expeditions), methods and devices, parameters, references, and institutional provenance information are all stored in defined relational structures. This normalized model allows data descriptions to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived data.&lt;br /&gt;
&lt;br /&gt;
Database integrity is continuously maintained through &#039;&#039;&#039;PostgreSQL streaming replication&#039;&#039;&#039; to a dedicated backup system, which enables point-in-time recovery to any moment prior to a failure event. In addition, full database backup copies (`pg_base_backup`) are created each weekend and retained for three weeks, providing an independent recovery layer supplementary to the streaming replica.&lt;br /&gt;
&lt;br /&gt;
=== Data Warehouse ===&lt;br /&gt;
Fast access to large, aggregated compilations of data across the full PANGAEA holdings is provided by a &#039;&#039;&#039;Clickhouse&#039;&#039;&#039; data warehouse, which mirrors the numerical data inventory from the relational database. The data warehouse supports spatially and chronologically constrained queries at the parameter level across potentially hundreds of thousands of individual datasets, and is accessible both through the PANGAEA website and programmatically via a REST API. All data warehouse exports include the DOI for each contributing dataset, ensuring that provenance and citation remain intact in compiled data products.&lt;br /&gt;
&lt;br /&gt;
=== Archival Storage ===&lt;br /&gt;
For high-volume and binary data — including geophysical datasets, images, video, and community-specific formats such as NetCDF — data are stored in consistent formats on hard-disk arrays and robotic tape archives. The central archival storage system consists of two &#039;&#039;&#039;SpectraLogic TFinity ExaScale&#039;&#039;&#039; robotic tape libraries housed in separate buildings at AWI, with a combined capacity of up to 60 PB. Backup operations use high-capacity LTO-tape drives within this environment.&lt;br /&gt;
&lt;br /&gt;
All data are stored redundantly using &#039;&#039;&#039;erasure coding&#039;&#039;&#039; across disk and tape. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots. Write caches are battery-backed to ensure data integrity at the point of write.&lt;br /&gt;
&lt;br /&gt;
=== Off-Site Replica ===&lt;br /&gt;
Since 2025, PANGAEA operates a **minimal viable repository service** as an off-site replica at MARUM/University of Bremen, hosted in the Green IT Housing Center of Bremen University. This facility provides geographic and infrastructural separation from the primary AWI systems, enabling recovery of data delivery services within 24 hours following a technical breakdown or cyberattack. Disaster recovery and switchover procedures involving this facility are regularly exercised. The replica is kept current through snapshot-based replication several times per day and currently covers all dataset metadata — including individual landing pages and all harvesting endpoints — full representations of tabular data publications, the Elasticsearch search index, and the relevant web frontend. Extension of the replica to include binary data files is under active development, with formal commitment to be finalized through the ongoing MARUM hosting agreement negotiations.&lt;br /&gt;
&lt;br /&gt;
The off-site installation operates in a fully isolated environment with Layer 2 network separation. Access is restricted to VPN connections from a single designated gateway host at AWI, with access tokens and keys limited to the corresponding gateway and PANGAEA DevOps staff.&lt;br /&gt;
== Middleware: Processing and Transformation ==&lt;br /&gt;
The middleware layer manages the flow of data and metadata between the backend storage systems and the frontend interfaces. Its core functions are marshaling, indexing, transformation, and dissemination.&lt;br /&gt;
&lt;br /&gt;
Metadata are dynamically marshaled from the relational database to a PANGAEA-specific internal XML format, from which they are transformed via &#039;&#039;&#039;XSLT and XML-to-JSON pipelines&#039;&#039;&#039; into a range of content standards for delivery to users and harvesting services. Currently supported output standards include JSON-LD (Schema.org with Croissant extension), DataCite XML, Dublin Core XML, ISO 19115/ISO 19139, and DIF/FGDC. Dissemination occurs via OAI-PMH, HTTP content negotiation following the Signposting standard (&amp;lt;nowiki&amp;gt;https://signposting.org/&amp;lt;/nowiki&amp;gt;), and other protocols. Metadata marshaling from the PostgreSQL database to the Elasticsearch search index is handled asynchronously by dedicated middleware components, which also manage automated DOI minting at DataCite upon dataset publication.&lt;br /&gt;
&lt;br /&gt;
The marshaled metadata are stored and indexed in &#039;&#039;&#039;Elasticsearch&#039;&#039;&#039;, which serves as the primary index for all public search and metadata access interfaces. This architecture separates the authoritative relational record from the search-optimized representation, allowing the search index to be rebuilt or updated without modifying archived data.&lt;br /&gt;
&lt;br /&gt;
The flexible metadata framework &#039;&#039;&#039;PanFMP&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://www.panfmp.org/&amp;lt;/nowiki&amp;gt;) underpins the mapping between the PANGAEA internal schema and the various external output standards, ensuring that new or updated standards can be accommodated without structural changes to the underlying data.&lt;br /&gt;
&lt;br /&gt;
Data submissions, user requests, and bug reports are managed through a &#039;&#039;&#039;JIRA&#039;&#039;&#039; (Atlassian) issue tracking system, which serves as the primary communication channel between data providers and the PANGAEA editorial team throughout the submission and publication process.&lt;br /&gt;
== Frontend: Editorial System ==&lt;br /&gt;
The PANGAEA editorial system is the primary tool through which data editors review, curate, harmonize, and publish submitted data and metadata. It is a &#039;&#039;&#039;web-based client/server application&#039;&#039;&#039; developed entirely in-house, operating directly on the PostgreSQL databases.&lt;br /&gt;
&lt;br /&gt;
The backend of the editorial system is built on &#039;&#039;&#039;Java 17&#039;&#039;&#039; using the &#039;&#039;&#039;Dropwizard&#039;&#039;&#039; framework, exposing a REST API. The frontend is implemented in &#039;&#039;&#039;React&#039;&#039;&#039; with the &#039;&#039;&#039;Ant Design&#039;&#039;&#039; component library. Source code for both components, together with shared libraries and test suites, is maintained in an institutional &#039;&#039;&#039;GitLab&#039;&#039;&#039; instance hosted at AWI, structured as an &#039;&#039;&#039;Nx monorepo&#039;&#039;&#039;. The repository encompasses automated unit tests and &#039;&#039;&#039;Cypress&#039;&#039;&#039; end-to-end tests. All new versions are deployed exclusively through a &#039;&#039;&#039;GitLab CI/CD pipeline&#039;&#039;&#039;, with releases gated on successful completion of the full automated test suite.&lt;br /&gt;
&lt;br /&gt;
In production, the editorial system runs on Ubuntu virtual machines hosted on VMware infrastructure. &#039;&#039;&#039;Apache&#039;&#039;&#039; serves the React frontend and acts as a reverse proxy to the backend services. Four parallel instances are operated simultaneously, providing dedicated test and production environments for two editorial system versions at any given time. This setup supports controlled feature validation alongside uninterrupted service operation. Infrastructure components — including VMware, file systems, monitoring, GitLab, and the database platform — are managed by AWI, while the PANGAEA group is responsible for the Apache web server, Java services, and React frontend.&lt;br /&gt;
&lt;br /&gt;
The relational database model underlying the editorial system enforces structural and semantic consistency between submitted metadata and the existing data inventory. The editorial system also integrates a &#039;&#039;&#039;Terminology Catalogue (TC)&#039;&#039;&#039;, which manages controlled vocabularies and ontologies used to harmonize data and metadata during ingest. Supported terminologies include WoRMS, ITIS, ChEBI, EnvO, PATO, QUDT, and the NERC vocabulary server, among others. A detailed description of the Terminology Catalogue and its role in data archiving and publication is given in Diepenbroek et al. (2017).&lt;br /&gt;
== Frontend: Public Web Interface and Search ==&lt;br /&gt;
&lt;br /&gt;
=== Web Delivery Stack ===&lt;br /&gt;
The public web infrastructure is fronted by an &#039;&#039;&#039;NGINX&#039;&#039;&#039; reverse proxy that manages incoming traffic and supports &#039;&#039;&#039;HTTP/3 (QUIC)&#039;&#039;&#039; and &#039;&#039;&#039;HTTP/2&#039;&#039;&#039; alongside HTTP/1.1 for broad client compatibility. Behind this entry point, a dual-backend architecture separates concerns by function: a &#039;&#039;&#039;PHP 8&#039;&#039;&#039; environment handles general information pages and the PANGAEA Wiki, while a &#039;&#039;&#039;Java 17&#039;&#039;&#039; application layer running on &#039;&#039;&#039;Eclipse Jetty 12&#039;&#039;&#039; manages core repository services — including dataset landing pages, DOI and handle resolution, and content negotiation for both human- and machine-readable data access. The Jetty-based layer also manages delivery of high-volume and binary files from the tape archive by staging requested files to a local hard-disk cache before serving them to users.&lt;br /&gt;
&lt;br /&gt;
=== Search ===&lt;br /&gt;
The PANGAEA search engine supports full-text and faceted search across all published metadata. Faceted navigation is enabled by semantic metadata enrichment performed during the marshaling process: terminology-based annotations are added to metadata records, allowing consistent filtering by topic, device type, geographic region, and other dimensions. PANGAEA&#039;s approach to data discovery complies with the recent recommendations of the Research Data Alliance (RDA) Interest Group &amp;quot;Data Discovery Paradigms&amp;quot; (Wu, M. et al., 2026, &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;). Search documentation is available in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/PANGAEA_search&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Map Search ===&lt;br /&gt;
Geographic search and visualization are implemented using &#039;&#039;&#039;leaflet.js&#039;&#039;&#039;, serving map tiles from four configurable sources: the AWI basemap, OpenStreetMap, Google (hybrid), and ESRI.&lt;br /&gt;
&lt;br /&gt;
=== Dataset Landing Pages ===&lt;br /&gt;
Each published dataset is represented by a landing page resolved through its DOI. Landing pages present the full dataset metadata in human-readable form and are enriched with structured &#039;&#039;&#039;schema.org&#039;&#039;&#039; markup in JSON-LD, ensuring machine-actionability and compatibility with generic web search engines and data registries. The schema.org metadata is also accessible via HTTP content negotiation following the Signposting standard. Minor editorial corrections to datasets that do not affect scientific content are documented in a &amp;quot;Change history&amp;quot; section of the landing page, recording the date and a summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
=== Programmatic Access ===&lt;br /&gt;
PANGAEA offers programmatic access to data and metadata through a range of web services (REST and SOAP). The OAI-PMH endpoint supports metadata harvesting in all supported standards. Client libraries for &#039;&#039;&#039;Python&#039;&#039;&#039; (pangaeapy, developed by PANGAEA) and &#039;&#039;&#039;R&#039;&#039;&#039; (pangaear, developed by the community) allow researchers to load and transform PANGAEA data directly into native data structures for analysis in environments such as Jupyter notebooks.&lt;br /&gt;
&lt;br /&gt;
== Monitoring and Analytics ==&lt;br /&gt;
Service health across the PANGAEA infrastructure is monitored through the &#039;&#039;&#039;AWI Grafana/Telegraf&#039;&#039;&#039; stack, covering service availability, application logs, and resource saturation for all production systems. This is supplemented by external availability checks via &#039;&#039;&#039;UptimeRobot&#039;&#039;&#039;, which provides independent verification of public-facing service endpoints from outside the AWI network. All external links in PANGAEA metadata records — references to related literature, other dataset versions, and external resources — are automatically checked for broken (HTTP 404) or permanently redirected (HTTP 301) responses on a weekly basis.&lt;br /&gt;
&lt;br /&gt;
User engagement and download metrics are captured by an integrated **Matomo** analytics instance configured to produce usage statistics compliant with the **COUNTER** (Counting Online Usage of Networked Electronic Resources) standard, ensuring that data impact is measured according to international scholarly norms. Usage statistics are publicly visible on each dataset landing page; details are documented in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/Data_Usage_Statistics&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Security ==&lt;br /&gt;
Backend and middleware systems are protected behind a firewall; frontend systems operate in a demilitarized zone (DMZ) accessible from outside with restricted but still firewalled access. Frontend systems have no write access and only limited read access to the backend database and tape archives. Public access from the website and REST APIs is served from data replicas hosted in Elasticsearch or through read-only remote filesystem access. Data curators access production systems via virtual private networks (VPN), which enforce a basic check that client operating systems are up to date.&lt;br /&gt;
&lt;br /&gt;
Physical access to the AWI computing center is managed through an electronic access control system for all relevant entrances; key distribution is documented, and guest policies are formally established. The AWI maintains uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation. Fire alerts are automatically forwarded to the fire department with a contractual response time under 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Security of the technical infrastructure is further maintained through the use of asymmetric key infrastructures, mandatory minimum-length passwords for all user classes, short-cycle security patching for all hardware and software components, professional monitoring tools for hardware, firewall, software, services, performance, and attacks, and regular security training for all technical and non-technical staff. Security risks are assessed on an ongoing basis by AWI&#039;s institutional IT security team, which coordinates incident response and monitors threat landscapes relevant to research data infrastructure. PANGAEA technical staff participates in regular security reviews.&lt;br /&gt;
&lt;br /&gt;
Access controls for restricted or moratorium datasets are enforced at the application layer. Dataset metadata remains publicly accessible in all cases; access to the data itself is restricted to authorized users at the individual level for the duration of the moratorium, typically a maximum of two years from submission.&lt;br /&gt;
&lt;br /&gt;
The off-site replica at MARUM operates in a fully isolated network environment (Layer 2 separation) with access restricted to a firewalled VPN gateway at AWI. The Green IT Housing Center is monitored 24/7 by Bremen University staff located in very close proximity to the datacenter (15 m, adjacent building), with automated fire alarms, redundant power supply with battery backup, and multilevel physical access control.&lt;br /&gt;
&lt;br /&gt;
== Software Inventory ==&lt;br /&gt;
The following table summarizes the principal software components of the PANGAEA infrastructure and their development model.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Component !! Software !! Development Model&lt;br /&gt;
|-&lt;br /&gt;
| Primary database || PostgreSQL || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Data warehouse || Clickhouse || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Search and metadata index || Elasticsearch || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system backend || Java 17 / Dropwizard || In-house development&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system frontend || React / Ant Design || In-house development (open source framework)&lt;br /&gt;
|-&lt;br /&gt;
| Source control || GitLab || Open source (community-supported, self-hosted)&lt;br /&gt;
|-&lt;br /&gt;
| CI/CD pipeline || GitLab CI/CD || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web server / reverse proxy (editorial) || Apache || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web traffic management || NGINX || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| General web and Wiki frontend || PHP 8 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Repository services backend || Java 17 / Eclipse Jetty 12 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Map visualization || leaflet.js || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Metadata framework || PanFMP || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| Issue tracking || JIRA (Atlassian) || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Monitoring stack || Grafana / Telegraf || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| External uptime monitoring || UptimeRobot || Commercial SaaS&lt;br /&gt;
|-&lt;br /&gt;
| Analytics || Matomo || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Infrastructure virtualization || VMware || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Tape archive hardware || SpectraLogic TFinity ExaScale || Commercial hardware&lt;br /&gt;
|-&lt;br /&gt;
| Python data client || pangaeapy || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| R data client || pangaear || Community development (open source)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Documentation and Change Management ==&lt;br /&gt;
General documentation of PANGAEA systems and services is maintained in the public &#039;&#039;&#039;PANGAEA Wiki&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/&amp;lt;/nowiki&amp;gt;). A separate internal &#039;&#039;&#039;Confluence&#039;&#039;&#039; Wiki, kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes, contains detailed documentation of server configurations, installation and maintenance procedures, relationships between system components, VM snapshot procedures, backup routines, and service restart priorities for incident response.&lt;br /&gt;
&lt;br /&gt;
All changes to published data and metadata are recorded in the editorial system&#039;s version history. Substantive revisions to published datasets result in a new dataset version with a new DOI; all prior versions remain accessible and cross-referenced. Minor editorial corrections that do not affect scientific content are applied without creating a new identifier and are transparently documented in the &amp;quot;Change history&amp;quot; section of the dataset landing page, including the date and a brief summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science. &#039;&#039;Scientific Data&#039;&#039;, 10, 347. &amp;lt;nowiki&amp;gt;https://doi.org/10.1038/s41597-023-02269-x&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Diepenbroek, M., Schindler, U., Huber, R., Pesant, S., Stocker, M., Felden, J., Buss, M. &amp;amp; Weinrebe, M. (2017). Terminology supported archiving and publication of environmental science data in PANGAEA. &#039;&#039;Journal of Biotechnology&#039;&#039;, 261, 177–186. &amp;lt;nowiki&amp;gt;https://doi.org/10.1016/j.jbiotec.2017.07.016&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Wu, M. et al. (2026). RDA Interest Group on Data Discovery Paradigms. &#039;&#039;Data Science Journal&#039;&#039;. &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
*[[PANGAEA search]]&lt;br /&gt;
*[[Data submission]]&lt;br /&gt;
*[[Authors Guides]]&lt;br /&gt;
*[[Curation levels]]&lt;br /&gt;
*[[Processing levels]]&lt;br /&gt;
*[[Data Usage Statistics]]&lt;br /&gt;
*[[Format]]&lt;br /&gt;
*[[PANGAEA XML schema]]&lt;br /&gt;
*[[Best practice manuals and templates]]&lt;br /&gt;
*[[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
*[[Preservation Plan]] — https://www.pangaea.de/about/preservation.php&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16801</id>
		<title>Technology</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16801"/>
		<updated>2026-04-10T15:25:43Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* System Architecture */ corrected Relational database storage figures&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{construction}}&lt;br /&gt;
&lt;br /&gt;
PANGAEA is built on a three-tiered client/server architecture that has evolved continuously since the system&#039;s founding in the early 1990s. The architecture separates backend storage and database systems, middleware processing and transformation components, and frontend delivery and user interfaces into distinct layers that are individually maintainable and replaceable. This design philosophy has enabled PANGAEA to migrate core components — including its primary database engine, its editorial system, and its search infrastructure — without disruption to archived data or to users, and underpins the system&#039;s long-term stability as a trustworthy data repository.&lt;br /&gt;
&lt;br /&gt;
A detailed description of the PANGAEA information system and its workflows is provided in Felden et al. (2023) and Diepenbroek et al. (2017); this article provides an up-to-date overview oriented toward the general structure and key components.&lt;br /&gt;
&lt;br /&gt;
== System Architecture ==&lt;br /&gt;
&lt;br /&gt;
The technical architecture of PANGAEA follows a three-tiered model comprising a backend, a middleware layer, and a frontend. All hard- and software services are hosted and operated by the data and computing center of the Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research (AWI) in Bremerhaven. Most backend and middleware systems, as well as all frontend web servers and search engines, run on virtual machines (VMware) operated with Ubuntu Linux. Virtualization provides sufficient capacity and performance while enabling high availability and transparent hardware renewal on a typical cycle of three to four years.&lt;br /&gt;
[[File:PANGAEA overview architecture.png|alt=Schema illustrating the system architecture according to Felden et al., 2023|thumb|Schema illustrating the system architecture according to Felden et al., 2023|none|600x600px]]&lt;br /&gt;
&lt;br /&gt;
PANGAEA currently operates nine virtual machines at the AWI computing center, using a total of 53 CPUs, 162 GB RAM, and 28 TB of disk space. The relational database holdings currently occupy approximately 5 TB, with approximately 1 PB of data stored on tape.&lt;br /&gt;
&lt;br /&gt;
== Backend: Storage and Databases ==&lt;br /&gt;
&lt;br /&gt;
=== Relational Database ===&lt;br /&gt;
The primary store for all structured data and metadata in PANGAEA is a &#039;&#039;&#039;PostgreSQL&#039;&#039;&#039; relational database management system (RDBMS). The data model is highly normalized, reflecting the full observational context of each measurement: events (when and where data were collected), campaigns (cruises or field expeditions), methods and devices, parameters, references, and institutional provenance information are all stored in defined relational structures. This normalized model allows data descriptions to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived data.&lt;br /&gt;
&lt;br /&gt;
Database integrity is continuously maintained through &#039;&#039;&#039;PostgreSQL streaming replication&#039;&#039;&#039; to a dedicated backup system, which enables point-in-time recovery to any moment prior to a failure event. In addition, full database backup copies (`pg_base_backup`) are created each weekend and retained for three weeks, providing an independent recovery layer supplementary to the streaming replica.&lt;br /&gt;
&lt;br /&gt;
=== Data Warehouse ===&lt;br /&gt;
Fast access to large, aggregated compilations of data across the full PANGAEA holdings is provided by a &#039;&#039;&#039;Clickhouse&#039;&#039;&#039; data warehouse, which mirrors the numerical data inventory from the relational database. The data warehouse supports spatially and chronologically constrained queries at the parameter level across potentially hundreds of thousands of individual datasets, and is accessible both through the PANGAEA website and programmatically via a REST API. All data warehouse exports include the DOI for each contributing dataset, ensuring that provenance and citation remain intact in compiled data products.&lt;br /&gt;
&lt;br /&gt;
=== Archival Storage ===&lt;br /&gt;
For high-volume and binary data — including geophysical datasets, images, video, and community-specific formats such as NetCDF — data are stored in consistent formats on hard-disk arrays and robotic tape archives. The central archival storage system consists of two &#039;&#039;&#039;SpectraLogic TFinity ExaScale&#039;&#039;&#039; robotic tape libraries housed in separate buildings at AWI, with a combined capacity of up to 60 PB. Backup operations use high-capacity LTO-tape drives within this environment.&lt;br /&gt;
&lt;br /&gt;
All data are stored redundantly using &#039;&#039;&#039;erasure coding&#039;&#039;&#039; across disk and tape. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots. Write caches are battery-backed to ensure data integrity at the point of write.&lt;br /&gt;
&lt;br /&gt;
=== Off-Site Replica ===&lt;br /&gt;
Since 2025, PANGAEA operates a **minimal viable repository service** as an off-site replica at MARUM/University of Bremen, hosted in the Green IT Housing Center of Bremen University. This facility provides geographic and infrastructural separation from the primary AWI systems, enabling recovery of data delivery services within 24 hours following a technical breakdown or cyberattack. Disaster recovery and switchover procedures involving this facility are regularly exercised. The replica is kept current through snapshot-based replication several times per day and currently covers all dataset metadata — including individual landing pages and all harvesting endpoints — full representations of tabular data publications, the Elasticsearch search index, and the relevant web frontend. Extension of the replica to include binary data files is under active development, with formal commitment to be finalized through the ongoing MARUM hosting agreement negotiations.&lt;br /&gt;
&lt;br /&gt;
The off-site installation operates in a fully isolated environment with Layer 2 network separation. Access is restricted to VPN connections from a single designated gateway host at AWI, with access tokens and keys limited to the corresponding gateway and PANGAEA DevOps staff.&lt;br /&gt;
== Middleware: Processing and Transformation ==&lt;br /&gt;
The middleware layer manages the flow of data and metadata between the backend storage systems and the frontend interfaces. Its core functions are marshaling, indexing, transformation, and dissemination.&lt;br /&gt;
&lt;br /&gt;
Metadata are dynamically marshaled from the relational database to a PANGAEA-specific internal XML format, from which they are transformed via &#039;&#039;&#039;XSLT and XML-to-JSON pipelines&#039;&#039;&#039; into a range of content standards for delivery to users and harvesting services. Currently supported output standards include JSON-LD (Schema.org), DataCite XML, Dublin Core XML, ISO 19115/ISO 19139, and DIF/FGDC. Dissemination occurs via OAI-PMH, HTTP content negotiation following the Signposting standard (&amp;lt;nowiki&amp;gt;https://signposting.org/&amp;lt;/nowiki&amp;gt;), and other protocols. Metadata marshaling from the PostgreSQL database to the Elasticsearch search index is handled asynchronously by dedicated middleware components, which also manage automated DOI minting at DataCite upon dataset publication.&lt;br /&gt;
&lt;br /&gt;
The marshaled metadata are stored and indexed in &#039;&#039;&#039;Elasticsearch&#039;&#039;&#039;, which serves as the primary index for all public search and metadata access interfaces. This architecture separates the authoritative relational record from the search-optimized representation, allowing the search index to be rebuilt or updated without modifying archived data.&lt;br /&gt;
&lt;br /&gt;
The flexible metadata framework &#039;&#039;&#039;PanFMP&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://www.panfmp.org/&amp;lt;/nowiki&amp;gt;) underpins the mapping between the PANGAEA internal schema and the various external output standards, ensuring that new or updated standards can be accommodated without structural changes to the underlying data.&lt;br /&gt;
&lt;br /&gt;
Data submissions, user requests, and bug reports are managed through a &#039;&#039;&#039;JIRA&#039;&#039;&#039; (Atlassian) issue tracking system, which serves as the primary communication channel between data providers and the PANGAEA editorial team throughout the submission and publication process.&lt;br /&gt;
== Frontend: Editorial System ==&lt;br /&gt;
The PANGAEA editorial system is the primary tool through which data editors review, curate, harmonize, and publish submitted data and metadata. It is a &#039;&#039;&#039;web-based client/server application&#039;&#039;&#039; developed entirely in-house, operating directly on the PostgreSQL databases.&lt;br /&gt;
&lt;br /&gt;
The backend of the editorial system is built on &#039;&#039;&#039;Java 17&#039;&#039;&#039; using the &#039;&#039;&#039;Dropwizard&#039;&#039;&#039; framework, exposing a REST API. The frontend is implemented in &#039;&#039;&#039;React&#039;&#039;&#039; with the &#039;&#039;&#039;Ant Design&#039;&#039;&#039; component library. Source code for both components, together with shared libraries and test suites, is maintained in an institutional &#039;&#039;&#039;GitLab&#039;&#039;&#039; instance hosted at AWI, structured as an &#039;&#039;&#039;Nx monorepo&#039;&#039;&#039;. The repository encompasses automated unit tests and &#039;&#039;&#039;Cypress&#039;&#039;&#039; end-to-end tests. All new versions are deployed exclusively through a &#039;&#039;&#039;GitLab CI/CD pipeline&#039;&#039;&#039;, with releases gated on successful completion of the full automated test suite.&lt;br /&gt;
&lt;br /&gt;
In production, the editorial system runs on Ubuntu virtual machines hosted on VMware infrastructure. &#039;&#039;&#039;Apache&#039;&#039;&#039; serves the React frontend and acts as a reverse proxy to the backend services. Four parallel instances are operated simultaneously, providing dedicated test and production environments for two editorial system versions at any given time. This setup supports controlled feature validation alongside uninterrupted service operation. Infrastructure components — including VMware, file systems, monitoring, GitLab, and the database platform — are managed by AWI, while the PANGAEA group is responsible for the Apache web server, Java services, and React frontend.&lt;br /&gt;
&lt;br /&gt;
The relational database model underlying the editorial system enforces structural and semantic consistency between submitted metadata and the existing data inventory. The editorial system also integrates a &#039;&#039;&#039;Terminology Catalogue (TC)&#039;&#039;&#039;, which manages controlled vocabularies and ontologies used to harmonize data and metadata during ingest. Supported terminologies include WoRMS, ITIS, ChEBI, EnvO, PATO, QUDT, and the NERC vocabulary server, among others. A detailed description of the Terminology Catalogue and its role in data archiving and publication is given in Diepenbroek et al. (2017).&lt;br /&gt;
== Frontend: Public Web Interface and Search ==&lt;br /&gt;
&lt;br /&gt;
=== Web Delivery Stack ===&lt;br /&gt;
The public web infrastructure is fronted by an &#039;&#039;&#039;NGINX&#039;&#039;&#039; reverse proxy that manages incoming traffic and supports &#039;&#039;&#039;HTTP/3 (QUIC)&#039;&#039;&#039; and &#039;&#039;&#039;HTTP/2&#039;&#039;&#039; alongside HTTP/1.1 for broad client compatibility. Behind this entry point, a dual-backend architecture separates concerns by function: a &#039;&#039;&#039;PHP 8&#039;&#039;&#039; environment handles general information pages and the PANGAEA Wiki, while a &#039;&#039;&#039;Java 17&#039;&#039;&#039; application layer running on &#039;&#039;&#039;Eclipse Jetty 12&#039;&#039;&#039; manages core repository services — including dataset landing pages, DOI and handle resolution, and content negotiation for both human- and machine-readable data access. The Jetty-based layer also manages delivery of high-volume and binary files from the tape archive by staging requested files to a local hard-disk cache before serving them to users.&lt;br /&gt;
&lt;br /&gt;
=== Search ===&lt;br /&gt;
The PANGAEA search engine supports full-text and faceted search across all published metadata. Faceted navigation is enabled by semantic metadata enrichment performed during the marshaling process: terminology-based annotations are added to metadata records, allowing consistent filtering by topic, device type, geographic region, and other dimensions. PANGAEA&#039;s approach to data discovery complies with the recent recommendations of the Research Data Alliance (RDA) Interest Group &amp;quot;Data Discovery Paradigms&amp;quot; (Wu, M. et al., 2026, &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;). Search documentation is available in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/PANGAEA_search&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Map Search ===&lt;br /&gt;
Geographic search and visualization are implemented using &#039;&#039;&#039;leaflet.js&#039;&#039;&#039;, serving map tiles from four configurable sources: the AWI basemap, OpenStreetMap, Google (hybrid), and ESRI.&lt;br /&gt;
&lt;br /&gt;
=== Dataset Landing Pages ===&lt;br /&gt;
Each published dataset is represented by a landing page resolved through its DOI. Landing pages present the full dataset metadata in human-readable form and are enriched with structured &#039;&#039;&#039;schema.org&#039;&#039;&#039; markup in JSON-LD, ensuring machine-actionability and compatibility with generic web search engines and data registries. The schema.org metadata is also accessible via HTTP content negotiation following the Signposting standard. Minor editorial corrections to datasets that do not affect scientific content are documented in a &amp;quot;Change history&amp;quot; section of the landing page, recording the date and a summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
=== Programmatic Access ===&lt;br /&gt;
PANGAEA offers programmatic access to data and metadata through a range of web services (REST and SOAP). The OAI-PMH endpoint supports metadata harvesting in all supported standards. Client libraries for &#039;&#039;&#039;Python&#039;&#039;&#039; (pangaeapy, developed by PANGAEA) and &#039;&#039;&#039;R&#039;&#039;&#039; (pangaear, developed by the community) allow researchers to load and transform PANGAEA data directly into native data structures for analysis in environments such as Jupyter notebooks.&lt;br /&gt;
&lt;br /&gt;
== Monitoring and Analytics ==&lt;br /&gt;
Service health across the PANGAEA infrastructure is monitored through the &#039;&#039;&#039;AWI Grafana/Telegraf&#039;&#039;&#039; stack, covering service availability, application logs, and resource saturation for all production systems. This is supplemented by external availability checks via &#039;&#039;&#039;UptimeRobot&#039;&#039;&#039;, which provides independent verification of public-facing service endpoints from outside the AWI network. All external links in PANGAEA metadata records — references to related literature, other dataset versions, and external resources — are automatically checked for broken (HTTP 404) or permanently redirected (HTTP 301) responses on a weekly basis.&lt;br /&gt;
&lt;br /&gt;
User engagement and download metrics are captured by an integrated **Matomo** analytics instance configured to produce usage statistics compliant with the **COUNTER** (Counting Online Usage of Networked Electronic Resources) standard, ensuring that data impact is measured according to international scholarly norms. Usage statistics are publicly visible on each dataset landing page; details are documented in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/Data_Usage_Statistics&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Security ==&lt;br /&gt;
Backend and middleware systems are protected behind a firewall; frontend systems operate in a demilitarized zone (DMZ) accessible from outside with restricted but still firewalled access. Frontend systems have no write access and only limited read access to the backend database and tape archives. Public access from the website and REST APIs is served from data replicas hosted in Elasticsearch or through read-only remote filesystem access. Data curators access production systems via virtual private networks (VPN), which enforce a basic check that client operating systems are up to date.&lt;br /&gt;
&lt;br /&gt;
Physical access to the AWI computing center is managed through an electronic access control system for all relevant entrances; key distribution is documented, and guest policies are formally established. The AWI maintains uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation. Fire alerts are automatically forwarded to the fire department with a contractual response time under 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Security of the technical infrastructure is further maintained through the use of asymmetric key infrastructures, mandatory minimum-length passwords for all user classes, short-cycle security patching for all hardware and software components, professional monitoring tools for hardware, firewall, software, services, performance, and attacks, and regular security training for all technical and non-technical staff. Security risks are assessed on an ongoing basis by AWI&#039;s institutional IT security team, which coordinates incident response and monitors threat landscapes relevant to research data infrastructure. PANGAEA technical staff participates in regular security reviews.&lt;br /&gt;
&lt;br /&gt;
Access controls for restricted or moratorium datasets are enforced at the application layer. Dataset metadata remains publicly accessible in all cases; access to the data itself is restricted to authorized users at the individual level for the duration of the moratorium, typically a maximum of two years from submission.&lt;br /&gt;
&lt;br /&gt;
The off-site replica at MARUM operates in a fully isolated network environment (Layer 2 separation) with access restricted to a firewalled VPN gateway at AWI. The Green IT Housing Center is monitored 24/7 by Bremen University staff located in very close proximity to the datacenter (15 m, adjacent building), with automated fire alarms, redundant power supply with battery backup, and multilevel physical access control.&lt;br /&gt;
&lt;br /&gt;
== Software Inventory ==&lt;br /&gt;
The following table summarizes the principal software components of the PANGAEA infrastructure and their development model.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Component !! Software !! Development Model&lt;br /&gt;
|-&lt;br /&gt;
| Primary database || PostgreSQL || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Data warehouse || Clickhouse || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Search and metadata index || Elasticsearch || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system backend || Java 17 / Dropwizard || In-house development&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system frontend || React / Ant Design || In-house development (open source framework)&lt;br /&gt;
|-&lt;br /&gt;
| Source control || GitLab || Open source (community-supported, self-hosted)&lt;br /&gt;
|-&lt;br /&gt;
| CI/CD pipeline || GitLab CI/CD || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web server / reverse proxy (editorial) || Apache || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web traffic management || NGINX || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| General web and Wiki frontend || PHP 8 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Repository services backend || Java 17 / Eclipse Jetty 12 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Map visualization || leaflet.js || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Metadata framework || PanFMP || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| Issue tracking || JIRA (Atlassian) || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Monitoring stack || Grafana / Telegraf || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| External uptime monitoring || UptimeRobot || Commercial SaaS&lt;br /&gt;
|-&lt;br /&gt;
| Analytics || Matomo || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Infrastructure virtualization || VMware || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Tape archive hardware || SpectraLogic TFinity ExaScale || Commercial hardware&lt;br /&gt;
|-&lt;br /&gt;
| Python data client || pangaeapy || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| R data client || pangaear || Community development (open source)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Documentation and Change Management ==&lt;br /&gt;
General documentation of PANGAEA systems and services is maintained in the public &#039;&#039;&#039;PANGAEA Wiki&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/&amp;lt;/nowiki&amp;gt;). A separate internal &#039;&#039;&#039;Confluence&#039;&#039;&#039; Wiki, kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes, contains detailed documentation of server configurations, installation and maintenance procedures, relationships between system components, VM snapshot procedures, backup routines, and service restart priorities for incident response.&lt;br /&gt;
&lt;br /&gt;
All changes to published data and metadata are recorded in the editorial system&#039;s version history. Substantive revisions to published datasets result in a new dataset version with a new DOI; all prior versions remain accessible and cross-referenced. Minor editorial corrections that do not affect scientific content are applied without creating a new identifier and are transparently documented in the &amp;quot;Change history&amp;quot; section of the dataset landing page, including the date and a brief summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science. &#039;&#039;Scientific Data&#039;&#039;, 10, 347. &amp;lt;nowiki&amp;gt;https://doi.org/10.1038/s41597-023-02269-x&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Diepenbroek, M., Schindler, U., Huber, R., Pesant, S., Stocker, M., Felden, J., Buss, M. &amp;amp; Weinrebe, M. (2017). Terminology supported archiving and publication of environmental science data in PANGAEA. &#039;&#039;Journal of Biotechnology&#039;&#039;, 261, 177–186. &amp;lt;nowiki&amp;gt;https://doi.org/10.1016/j.jbiotec.2017.07.016&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Wu, M. et al. (2026). RDA Interest Group on Data Discovery Paradigms. &#039;&#039;Data Science Journal&#039;&#039;. &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
*[[PANGAEA search]]&lt;br /&gt;
*[[Data submission]]&lt;br /&gt;
*[[Authors Guides]]&lt;br /&gt;
*[[Curation levels]]&lt;br /&gt;
*[[Processing levels]]&lt;br /&gt;
*[[Data Usage Statistics]]&lt;br /&gt;
*[[Format]]&lt;br /&gt;
*[[PANGAEA XML schema]]&lt;br /&gt;
*[[Best practice manuals and templates]]&lt;br /&gt;
*[[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
*[[Preservation Plan]] — https://www.pangaea.de/about/preservation.php&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=PANGAEA_Community_Workshops&amp;diff=16782</id>
		<title>PANGAEA Community Workshops</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=PANGAEA_Community_Workshops&amp;diff=16782"/>
		<updated>2026-04-07T13:47:58Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Next Editions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[File:PANGAEA Banner CWS&amp;amp;Partner sans.png|thumb|773x773px]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== About ==&lt;br /&gt;
The &#039;&#039;&#039;PANGAEA Community Workshop Series (CWS)&#039;&#039;&#039; is an established training initiative launched in 2021. It was developed to strengthen research data management (RDM) competences among users of PANGAEA and to promote good scientific practice in the preparation, publication, and reuse of research data. Combining theoretical instruction with practical, hands-on components, the workshop series provides participants with a comprehensive understanding of how to work effectively with the PANGAEA data repository. It is primarily designed for early career scientists but is open for participation by anyone interested.&lt;br /&gt;
&lt;br /&gt;
The overarching goals of the series are to raise awareness of [https://www.go-fair.org/fair-principles/ FAIR data principles] and high-quality data management, empower researchers to make best use of PANGAEA services, and improve the overall efficiency of the data publication process. By educating authors on how to prepare their data appropriately, the workshops not only enhance participants’ individual scientific visibility but also contribute to reducing editorial workload and improving productivity within PANGAEA. &lt;br /&gt;
&lt;br /&gt;
The CWS is a contribution to the [https://www.denbi.de/training de.NBI training network] and covers the two major topics relevant to the use of a data repository.&lt;br /&gt;
&lt;br /&gt;
=== FAIR Data Publications with PANGAEA ===&lt;br /&gt;
The course &#039;&#039;&#039;“FAIR Data Publications with PANGAEA”&#039;&#039;&#039;, usually held in November, focuses on the submission and publication aspects of data management. Participants gain detailed insights into how to prepare datasets and metadata for submission, understand editorial procedures, and learn how to ensure their outputs meet FAIR principles. The benefits for researchers are manifold: publishing FAIR data increases citation rates, visibility, and scientific impact, while promoting transparency and encouraging collaboration. Data authors also gain recognition independent of traditional paper authorship, ensuring proper credit for data generation and processing.&lt;br /&gt;
&lt;br /&gt;
=== Finding and Retrieving Data from PANGAEA ===&lt;br /&gt;
The second course, &#039;&#039;&#039;“Finding and Retrieving Data from PANGAEA”&#039;&#039;&#039;, is devoted to the discovery and reuse of published data and is usually offered in May. It introduces participants to the repository’s powerful search capabilities, standardized metadata structures, and integration with major data portals such as [https://explore.openaire.eu/ OpenAIRE], [https://commons.datacite.org/ DataCite], and [https://datasetsearch.research.google.com/ Google Dataset Search]. Through practical exercises, attendees learn to efficiently locate, combine, and analyze datasets using standard web protocols and scripting tools based on [https://pypi.org/project/pangaeapy/ Python] and [https://github.com/ropensci/pangaear R]. High-quality, harmonized data products curated by domain experts ensure that users can confidently employ PANGAEA data in new studies and analytical workflows, thereby fostering reproducibility, innovation, and cross-disciplinary research.&lt;br /&gt;
&lt;br /&gt;
== Format Details ==&lt;br /&gt;
The workshops are conducted &#039;&#039;&#039;online via web conference&#039;&#039;&#039;, allowing participants from around the world to join easily. Each workshop spans &#039;&#039;&#039;two days&#039;&#039;&#039;, typically held on &#039;&#039;&#039;Thursday and Friday&#039;&#039;&#039;, with &#039;&#039;&#039;two-hour sessions per day&#039;&#039;&#039;. The format is designed to be compact yet interactive, combining presentations with live demonstrations and short exercises. &#039;&#039;&#039;No prior experience or prerequisites&#039;&#039;&#039; are required—participants of all backgrounds are welcome to attend.&lt;br /&gt;
&lt;br /&gt;
== Upcoming and Previous Editions and staying in the loop ==&lt;br /&gt;
New editions of the &#039;&#039;&#039;PANGAEA Community Workshop Series&#039;&#039;&#039; are held regularly, typically twice a year, featuring updated content that reflects emerging developments in open science, FAIR data practices, and digital research infrastructures. Researchers interested in participating or staying informed about future events are encouraged to subscribe to the &#039;&#039;&#039;PANGAEA Trainings mailing list&#039;&#039;&#039;. The list provides timely updates on upcoming workshops and other training activities and materials. You can subscribe to the list [https://lists.pangaea.de/listinfo/training &#039;&#039;&#039;here&#039;&#039;&#039;]. &lt;br /&gt;
&lt;br /&gt;
=== Next Editions ===&lt;br /&gt;
The next editions are scheduled for:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Date&lt;br /&gt;
!Topic&lt;br /&gt;
!Registration&lt;br /&gt;
|-&lt;br /&gt;
|07.-08.05.2026&lt;br /&gt;
|Finding and Retrieving Data from PANGAEA&lt;br /&gt;
|[https://events.hifis.net/e/CWS2605 Link] (open until May 03 2026)&lt;br /&gt;
|-&lt;br /&gt;
|12.-13.11.2026&lt;br /&gt;
|FAIR Data Publications with PANGAEA&lt;br /&gt;
|[https://events.hifis.net/e/CWS2611 Link] (reg. opens Oct. 2026)&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Slides and Materials for previous editions ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Date&lt;br /&gt;
!Title&lt;br /&gt;
!Slides&lt;br /&gt;
!Recordings&lt;br /&gt;
|-&lt;br /&gt;
|Nov. 2025&lt;br /&gt;
|FAIR Data Publications with PANGAEA&lt;br /&gt;
|[https://events.hifis.net/event/3172/attachments/4911/10323/202511_PANGAEA_CommunityWorkshop_DataIn_slides.pdf Link] &lt;br /&gt;
|[https://youtube.com/playlist?list=PLJpsMTFHswiaPYWCfKSxTTk1V-Hu8hDW5&amp;amp;si=iTUFnA_b_-eWEHGD @Youtube]&lt;br /&gt;
|-&lt;br /&gt;
|May 2025&lt;br /&gt;
|Finding and Retrieving Data from PANGAEA&lt;br /&gt;
|[https://events.hifis.net/event/2356/attachments/3920/8228/202505_SlideDeck_PANGAEA_CommunityWorkshop_Find&amp;amp;RetrieveData.pdf Link] &lt;br /&gt;
|n/a&lt;br /&gt;
|-&lt;br /&gt;
|Nov. 2024&lt;br /&gt;
|FAIR Data Publications with PANGAEA&lt;br /&gt;
|[https://nextcloud.awi.de/s/DRcFGpd6DMdjLja Link]&lt;br /&gt;
|[https://www.youtube.com/playlist?list=PLJpsMTFHswia0jH3XXHC-L2ZpXCYgv0xX @Youtube]&lt;br /&gt;
|-&lt;br /&gt;
|May 2024&lt;br /&gt;
|Finding and Retrieving Data from PANGAEA&lt;br /&gt;
|[https://nextcloud.awi.de/s/cfxZDarjzBj9mjr Link]&lt;br /&gt;
|[https://www.youtube.com/playlist?list=PLJpsMTFHswibjvzdC1yns2FFMVnxdiu5S @Youtube]&lt;br /&gt;
|}&lt;br /&gt;
You should also checkout our [https://github.com/pangaea-data-publisher/community-workshop-material CWS Github repository] for training material, Jupyter notebooks and scripts for interactions with PANGAEA content via R and Python.&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=PANGAEA_Community_Workshops&amp;diff=16772</id>
		<title>PANGAEA Community Workshops</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=PANGAEA_Community_Workshops&amp;diff=16772"/>
		<updated>2026-04-07T13:29:50Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Next Editions */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[File:PANGAEA Banner CWS&amp;amp;Partner sans.png|thumb|773x773px]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== About ==&lt;br /&gt;
The &#039;&#039;&#039;PANGAEA Community Workshop Series (CWS)&#039;&#039;&#039; is an established training initiative launched in 2021. It was developed to strengthen research data management (RDM) competences among users of PANGAEA and to promote good scientific practice in the preparation, publication, and reuse of research data. Combining theoretical instruction with practical, hands-on components, the workshop series provides participants with a comprehensive understanding of how to work effectively with the PANGAEA data repository. It is primarily designed for early career scientists but is open for participation by anyone interested.&lt;br /&gt;
&lt;br /&gt;
The overarching goals of the series are to raise awareness of [https://www.go-fair.org/fair-principles/ FAIR data principles] and high-quality data management, empower researchers to make best use of PANGAEA services, and improve the overall efficiency of the data publication process. By educating authors on how to prepare their data appropriately, the workshops not only enhance participants’ individual scientific visibility but also contribute to reducing editorial workload and improving productivity within PANGAEA. &lt;br /&gt;
&lt;br /&gt;
The CWS is a contribution to the [https://www.denbi.de/training de.NBI training network] and covers the two major topics relevant to the use of a data repository.&lt;br /&gt;
&lt;br /&gt;
=== FAIR Data Publications with PANGAEA ===&lt;br /&gt;
The course &#039;&#039;&#039;“FAIR Data Publications with PANGAEA”&#039;&#039;&#039;, usually held in November, focuses on the submission and publication aspects of data management. Participants gain detailed insights into how to prepare datasets and metadata for submission, understand editorial procedures, and learn how to ensure their outputs meet FAIR principles. The benefits for researchers are manifold: publishing FAIR data increases citation rates, visibility, and scientific impact, while promoting transparency and encouraging collaboration. Data authors also gain recognition independent of traditional paper authorship, ensuring proper credit for data generation and processing.&lt;br /&gt;
&lt;br /&gt;
=== Finding and Retrieving Data from PANGAEA ===&lt;br /&gt;
The second course, &#039;&#039;&#039;“Finding and Retrieving Data from PANGAEA”&#039;&#039;&#039;, is devoted to the discovery and reuse of published data and is usually offered in May. It introduces participants to the repository’s powerful search capabilities, standardized metadata structures, and integration with major data portals such as [https://explore.openaire.eu/ OpenAIRE], [https://commons.datacite.org/ DataCite], and [https://datasetsearch.research.google.com/ Google Dataset Search]. Through practical exercises, attendees learn to efficiently locate, combine, and analyze datasets using standard web protocols and scripting tools based on [https://pypi.org/project/pangaeapy/ Python] and [https://github.com/ropensci/pangaear R]. High-quality, harmonized data products curated by domain experts ensure that users can confidently employ PANGAEA data in new studies and analytical workflows, thereby fostering reproducibility, innovation, and cross-disciplinary research.&lt;br /&gt;
&lt;br /&gt;
== Format Details ==&lt;br /&gt;
The workshops are conducted &#039;&#039;&#039;online via web conference&#039;&#039;&#039;, allowing participants from around the world to join easily. Each workshop spans &#039;&#039;&#039;two days&#039;&#039;&#039;, typically held on &#039;&#039;&#039;Thursday and Friday&#039;&#039;&#039;, with &#039;&#039;&#039;two-hour sessions per day&#039;&#039;&#039;. The format is designed to be compact yet interactive, combining presentations with live demonstrations and short exercises. &#039;&#039;&#039;No prior experience or prerequisites&#039;&#039;&#039; are required—participants of all backgrounds are welcome to attend.&lt;br /&gt;
&lt;br /&gt;
== Upcoming and Previous Editions and staying in the loop ==&lt;br /&gt;
New editions of the &#039;&#039;&#039;PANGAEA Community Workshop Series&#039;&#039;&#039; are held regularly, typically twice a year, featuring updated content that reflects emerging developments in open science, FAIR data practices, and digital research infrastructures. Researchers interested in participating or staying informed about future events are encouraged to subscribe to the &#039;&#039;&#039;PANGAEA Trainings mailing list&#039;&#039;&#039;. The list provides timely updates on upcoming workshops and other training activities and materials. You can subscribe to the list [https://lists.pangaea.de/listinfo/training &#039;&#039;&#039;here&#039;&#039;&#039;]. &lt;br /&gt;
&lt;br /&gt;
=== Next Editions ===&lt;br /&gt;
The next editions are scheduled for:&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|+&lt;br /&gt;
!Date&lt;br /&gt;
!Topic&lt;br /&gt;
!Registration&lt;br /&gt;
|-&lt;br /&gt;
|07.-08.05.2026&lt;br /&gt;
|Finding and Retrieving Data from PANGAEA&lt;br /&gt;
|[https://events.hifis.net/e/CWS2605 Link] (reg. open until May 03 2026)&lt;br /&gt;
|-&lt;br /&gt;
|12.-13.11.2026&lt;br /&gt;
|FAIR Data Publications with PANGAEA&lt;br /&gt;
|[https://events.hifis.net/e/CWS2611 Link] (reg. opens Oct. 2026)&lt;br /&gt;
|-&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
=== Slides and Materials for previous editions ===&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Date&lt;br /&gt;
!Title&lt;br /&gt;
!Slides&lt;br /&gt;
!Recordings&lt;br /&gt;
|-&lt;br /&gt;
|Nov. 2025&lt;br /&gt;
|FAIR Data Publications with PANGAEA&lt;br /&gt;
|[https://events.hifis.net/event/3172/attachments/4911/10323/202511_PANGAEA_CommunityWorkshop_DataIn_slides.pdf Link] &lt;br /&gt;
|[https://youtube.com/playlist?list=PLJpsMTFHswiaPYWCfKSxTTk1V-Hu8hDW5&amp;amp;si=iTUFnA_b_-eWEHGD @Youtube]&lt;br /&gt;
|-&lt;br /&gt;
|May 2025&lt;br /&gt;
|Finding and Retrieving Data from PANGAEA&lt;br /&gt;
|[https://events.hifis.net/event/2356/attachments/3920/8228/202505_SlideDeck_PANGAEA_CommunityWorkshop_Find&amp;amp;RetrieveData.pdf Link] &lt;br /&gt;
|n/a&lt;br /&gt;
|-&lt;br /&gt;
|Nov. 2024&lt;br /&gt;
|FAIR Data Publications with PANGAEA&lt;br /&gt;
|[https://nextcloud.awi.de/s/DRcFGpd6DMdjLja Link]&lt;br /&gt;
|[https://www.youtube.com/playlist?list=PLJpsMTFHswia0jH3XXHC-L2ZpXCYgv0xX @Youtube]&lt;br /&gt;
|-&lt;br /&gt;
|May 2024&lt;br /&gt;
|Finding and Retrieving Data from PANGAEA&lt;br /&gt;
|[https://nextcloud.awi.de/s/cfxZDarjzBj9mjr Link]&lt;br /&gt;
|[https://www.youtube.com/playlist?list=PLJpsMTFHswibjvzdC1yns2FFMVnxdiu5S @Youtube]&lt;br /&gt;
|}&lt;br /&gt;
You should also checkout our [https://github.com/pangaea-data-publisher/community-workshop-material CWS Github repository] for training material, Jupyter notebooks and scripts for interactions with PANGAEA content via R and Python.&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16771</id>
		<title>Technology</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16771"/>
		<updated>2026-04-03T16:06:29Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Software Inventory */ Added revised Table&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{construction}}&lt;br /&gt;
&lt;br /&gt;
PANGAEA is built on a three-tiered client/server architecture that has evolved continuously since the system&#039;s founding in the early 1990s. The architecture separates backend storage and database systems, middleware processing and transformation components, and frontend delivery and user interfaces into distinct layers that are individually maintainable and replaceable. This design philosophy has enabled PANGAEA to migrate core components — including its primary database engine, its editorial system, and its search infrastructure — without disruption to archived data or to users, and underpins the system&#039;s long-term stability as a trustworthy data repository.&lt;br /&gt;
&lt;br /&gt;
A detailed description of the PANGAEA information system and its workflows is provided in Felden et al. (2023) and Diepenbroek et al. (2017); this article provides an up-to-date overview oriented toward the general structure and key components.&lt;br /&gt;
&lt;br /&gt;
== System Architecture ==&lt;br /&gt;
&lt;br /&gt;
The technical architecture of PANGAEA follows a three-tiered model comprising a backend, a middleware layer, and a frontend. All hard- and software services are hosted and operated by the data and computing center of the Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research (AWI) in Bremerhaven. Most backend and middleware systems, as well as all frontend web servers and search engines, run on virtual machines (VMware) operated with Ubuntu Linux. Virtualization provides sufficient capacity and performance while enabling high availability and transparent hardware renewal on a typical cycle of three to four years.&lt;br /&gt;
[[File:PANGAEA overview architecture.png|alt=Schema illustrating the system architecture according to Felden et al., 2023|thumb|Schema illustrating the system architecture according to Felden et al., 2023|none|600x600px]]&lt;br /&gt;
&lt;br /&gt;
PANGAEA currently operates nine virtual machines at the AWI computing center, using a total of 53 CPUs, 162 GB RAM, and 28 TB of disk space. The relational database holdings currently occupy approximately 130 TB, with approximately 1 PB of data stored on tape.&lt;br /&gt;
&lt;br /&gt;
== Backend: Storage and Databases ==&lt;br /&gt;
&lt;br /&gt;
=== Relational Database ===&lt;br /&gt;
The primary store for all structured data and metadata in PANGAEA is a &#039;&#039;&#039;PostgreSQL&#039;&#039;&#039; relational database management system (RDBMS). The data model is highly normalized, reflecting the full observational context of each measurement: events (when and where data were collected), campaigns (cruises or field expeditions), methods and devices, parameters, references, and institutional provenance information are all stored in defined relational structures. This normalized model allows data descriptions to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived data.&lt;br /&gt;
&lt;br /&gt;
Database integrity is continuously maintained through &#039;&#039;&#039;PostgreSQL streaming replication&#039;&#039;&#039; to a dedicated backup system, which enables point-in-time recovery to any moment prior to a failure event. In addition, full database backup copies (`pg_base_backup`) are created each weekend and retained for three weeks, providing an independent recovery layer supplementary to the streaming replica.&lt;br /&gt;
&lt;br /&gt;
=== Data Warehouse ===&lt;br /&gt;
Fast access to large, aggregated compilations of data across the full PANGAEA holdings is provided by a &#039;&#039;&#039;Clickhouse&#039;&#039;&#039; data warehouse, which mirrors the numerical data inventory from the relational database. The data warehouse supports spatially and chronologically constrained queries at the parameter level across potentially hundreds of thousands of individual datasets, and is accessible both through the PANGAEA website and programmatically via a REST API. All data warehouse exports include the DOI for each contributing dataset, ensuring that provenance and citation remain intact in compiled data products.&lt;br /&gt;
&lt;br /&gt;
=== Archival Storage ===&lt;br /&gt;
For high-volume and binary data — including geophysical datasets, images, video, and community-specific formats such as NetCDF — data are stored in consistent formats on hard-disk arrays and robotic tape archives. The central archival storage system consists of two &#039;&#039;&#039;SpectraLogic TFinity ExaScale&#039;&#039;&#039; robotic tape libraries housed in separate buildings at AWI, with a combined capacity of up to 60 PB. Backup operations use high-capacity LTO-tape drives within this environment.&lt;br /&gt;
&lt;br /&gt;
All data are stored redundantly using &#039;&#039;&#039;erasure coding&#039;&#039;&#039; across disk and tape. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots. Write caches are battery-backed to ensure data integrity at the point of write.&lt;br /&gt;
&lt;br /&gt;
=== Off-Site Replica ===&lt;br /&gt;
Since 2025, PANGAEA operates a **minimal viable repository service** as an off-site replica at MARUM/University of Bremen, hosted in the Green IT Housing Center of Bremen University. This facility provides geographic and infrastructural separation from the primary AWI systems, enabling recovery of data delivery services within 24 hours following a technical breakdown or cyberattack. Disaster recovery and switchover procedures involving this facility are regularly exercised. The replica is kept current through snapshot-based replication several times per day and currently covers all dataset metadata — including individual landing pages and all harvesting endpoints — full representations of tabular data publications, the Elasticsearch search index, and the relevant web frontend. Extension of the replica to include binary data files is under active development, with formal commitment to be finalized through the ongoing MARUM hosting agreement negotiations.&lt;br /&gt;
&lt;br /&gt;
The off-site installation operates in a fully isolated environment with Layer 2 network separation. Access is restricted to VPN connections from a single designated gateway host at AWI, with access tokens and keys limited to the corresponding gateway and PANGAEA DevOps staff.&lt;br /&gt;
== Middleware: Processing and Transformation ==&lt;br /&gt;
The middleware layer manages the flow of data and metadata between the backend storage systems and the frontend interfaces. Its core functions are marshaling, indexing, transformation, and dissemination.&lt;br /&gt;
&lt;br /&gt;
Metadata are dynamically marshaled from the relational database to a PANGAEA-specific internal XML format, from which they are transformed via &#039;&#039;&#039;XSLT and XML-to-JSON pipelines&#039;&#039;&#039; into a range of content standards for delivery to users and harvesting services. Currently supported output standards include JSON-LD (Schema.org), DataCite XML, Dublin Core XML, ISO 19115/ISO 19139, and DIF/FGDC. Dissemination occurs via OAI-PMH, HTTP content negotiation following the Signposting standard (&amp;lt;nowiki&amp;gt;https://signposting.org/&amp;lt;/nowiki&amp;gt;), and other protocols. Metadata marshaling from the PostgreSQL database to the Elasticsearch search index is handled asynchronously by dedicated middleware components, which also manage automated DOI minting at DataCite upon dataset publication.&lt;br /&gt;
&lt;br /&gt;
The marshaled metadata are stored and indexed in &#039;&#039;&#039;Elasticsearch&#039;&#039;&#039;, which serves as the primary index for all public search and metadata access interfaces. This architecture separates the authoritative relational record from the search-optimized representation, allowing the search index to be rebuilt or updated without modifying archived data.&lt;br /&gt;
&lt;br /&gt;
The flexible metadata framework &#039;&#039;&#039;PanFMP&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://www.panfmp.org/&amp;lt;/nowiki&amp;gt;) underpins the mapping between the PANGAEA internal schema and the various external output standards, ensuring that new or updated standards can be accommodated without structural changes to the underlying data.&lt;br /&gt;
&lt;br /&gt;
Data submissions, user requests, and bug reports are managed through a &#039;&#039;&#039;JIRA&#039;&#039;&#039; (Atlassian) issue tracking system, which serves as the primary communication channel between data providers and the PANGAEA editorial team throughout the submission and publication process.&lt;br /&gt;
== Frontend: Editorial System ==&lt;br /&gt;
The PANGAEA editorial system is the primary tool through which data editors review, curate, harmonize, and publish submitted data and metadata. It is a &#039;&#039;&#039;web-based client/server application&#039;&#039;&#039; developed entirely in-house, operating directly on the PostgreSQL databases.&lt;br /&gt;
&lt;br /&gt;
The backend of the editorial system is built on &#039;&#039;&#039;Java 17&#039;&#039;&#039; using the &#039;&#039;&#039;Dropwizard&#039;&#039;&#039; framework, exposing a REST API. The frontend is implemented in &#039;&#039;&#039;React&#039;&#039;&#039; with the &#039;&#039;&#039;Ant Design&#039;&#039;&#039; component library. Source code for both components, together with shared libraries and test suites, is maintained in an institutional &#039;&#039;&#039;GitLab&#039;&#039;&#039; instance hosted at AWI, structured as an &#039;&#039;&#039;Nx monorepo&#039;&#039;&#039;. The repository encompasses automated unit tests and &#039;&#039;&#039;Cypress&#039;&#039;&#039; end-to-end tests. All new versions are deployed exclusively through a &#039;&#039;&#039;GitLab CI/CD pipeline&#039;&#039;&#039;, with releases gated on successful completion of the full automated test suite.&lt;br /&gt;
&lt;br /&gt;
In production, the editorial system runs on Ubuntu virtual machines hosted on VMware infrastructure. &#039;&#039;&#039;Apache&#039;&#039;&#039; serves the React frontend and acts as a reverse proxy to the backend services. Four parallel instances are operated simultaneously, providing dedicated test and production environments for two editorial system versions at any given time. This setup supports controlled feature validation alongside uninterrupted service operation. Infrastructure components — including VMware, file systems, monitoring, GitLab, and the database platform — are managed by AWI, while the PANGAEA group is responsible for the Apache web server, Java services, and React frontend.&lt;br /&gt;
&lt;br /&gt;
The relational database model underlying the editorial system enforces structural and semantic consistency between submitted metadata and the existing data inventory. The editorial system also integrates a &#039;&#039;&#039;Terminology Catalogue (TC)&#039;&#039;&#039;, which manages controlled vocabularies and ontologies used to harmonize data and metadata during ingest. Supported terminologies include WoRMS, ITIS, ChEBI, EnvO, PATO, QUDT, and the NERC vocabulary server, among others. A detailed description of the Terminology Catalogue and its role in data archiving and publication is given in Diepenbroek et al. (2017).&lt;br /&gt;
== Frontend: Public Web Interface and Search ==&lt;br /&gt;
&lt;br /&gt;
=== Web Delivery Stack ===&lt;br /&gt;
The public web infrastructure is fronted by an &#039;&#039;&#039;NGINX&#039;&#039;&#039; reverse proxy that manages incoming traffic and supports &#039;&#039;&#039;HTTP/3 (QUIC)&#039;&#039;&#039; and &#039;&#039;&#039;HTTP/2&#039;&#039;&#039; alongside HTTP/1.1 for broad client compatibility. Behind this entry point, a dual-backend architecture separates concerns by function: a &#039;&#039;&#039;PHP 8&#039;&#039;&#039; environment handles general information pages and the PANGAEA Wiki, while a &#039;&#039;&#039;Java 17&#039;&#039;&#039; application layer running on &#039;&#039;&#039;Eclipse Jetty 12&#039;&#039;&#039; manages core repository services — including dataset landing pages, DOI and handle resolution, and content negotiation for both human- and machine-readable data access. The Jetty-based layer also manages delivery of high-volume and binary files from the tape archive by staging requested files to a local hard-disk cache before serving them to users.&lt;br /&gt;
&lt;br /&gt;
=== Search ===&lt;br /&gt;
The PANGAEA search engine supports full-text and faceted search across all published metadata. Faceted navigation is enabled by semantic metadata enrichment performed during the marshaling process: terminology-based annotations are added to metadata records, allowing consistent filtering by topic, device type, geographic region, and other dimensions. PANGAEA&#039;s approach to data discovery complies with the recent recommendations of the Research Data Alliance (RDA) Interest Group &amp;quot;Data Discovery Paradigms&amp;quot; (Wu, M. et al., 2026, &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;). Search documentation is available in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/PANGAEA_search&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Map Search ===&lt;br /&gt;
Geographic search and visualization are implemented using &#039;&#039;&#039;leaflet.js&#039;&#039;&#039;, serving map tiles from four configurable sources: the AWI basemap, OpenStreetMap, Google (hybrid), and ESRI.&lt;br /&gt;
&lt;br /&gt;
=== Dataset Landing Pages ===&lt;br /&gt;
Each published dataset is represented by a landing page resolved through its DOI. Landing pages present the full dataset metadata in human-readable form and are enriched with structured &#039;&#039;&#039;schema.org&#039;&#039;&#039; markup in JSON-LD, ensuring machine-actionability and compatibility with generic web search engines and data registries. The schema.org metadata is also accessible via HTTP content negotiation following the Signposting standard. Minor editorial corrections to datasets that do not affect scientific content are documented in a &amp;quot;Change history&amp;quot; section of the landing page, recording the date and a summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
=== Programmatic Access ===&lt;br /&gt;
PANGAEA offers programmatic access to data and metadata through a range of web services (REST and SOAP). The OAI-PMH endpoint supports metadata harvesting in all supported standards. Client libraries for &#039;&#039;&#039;Python&#039;&#039;&#039; (pangaeapy, developed by PANGAEA) and &#039;&#039;&#039;R&#039;&#039;&#039; (pangaear, developed by the community) allow researchers to load and transform PANGAEA data directly into native data structures for analysis in environments such as Jupyter notebooks.&lt;br /&gt;
&lt;br /&gt;
== Monitoring and Analytics ==&lt;br /&gt;
Service health across the PANGAEA infrastructure is monitored through the &#039;&#039;&#039;AWI Grafana/Telegraf&#039;&#039;&#039; stack, covering service availability, application logs, and resource saturation for all production systems. This is supplemented by external availability checks via &#039;&#039;&#039;UptimeRobot&#039;&#039;&#039;, which provides independent verification of public-facing service endpoints from outside the AWI network. All external links in PANGAEA metadata records — references to related literature, other dataset versions, and external resources — are automatically checked for broken (HTTP 404) or permanently redirected (HTTP 301) responses on a weekly basis.&lt;br /&gt;
&lt;br /&gt;
User engagement and download metrics are captured by an integrated **Matomo** analytics instance configured to produce usage statistics compliant with the **COUNTER** (Counting Online Usage of Networked Electronic Resources) standard, ensuring that data impact is measured according to international scholarly norms. Usage statistics are publicly visible on each dataset landing page; details are documented in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/Data_Usage_Statistics&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Security ==&lt;br /&gt;
Backend and middleware systems are protected behind a firewall; frontend systems operate in a demilitarized zone (DMZ) accessible from outside with restricted but still firewalled access. Frontend systems have no write access and only limited read access to the backend database and tape archives. Public access from the website and REST APIs is served from data replicas hosted in Elasticsearch or through read-only remote filesystem access. Data curators access production systems via virtual private networks (VPN), which enforce a basic check that client operating systems are up to date.&lt;br /&gt;
&lt;br /&gt;
Physical access to the AWI computing center is managed through an electronic access control system for all relevant entrances; key distribution is documented, and guest policies are formally established. The AWI maintains uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation. Fire alerts are automatically forwarded to the fire department with a contractual response time under 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Security of the technical infrastructure is further maintained through the use of asymmetric key infrastructures, mandatory minimum-length passwords for all user classes, short-cycle security patching for all hardware and software components, professional monitoring tools for hardware, firewall, software, services, performance, and attacks, and regular security training for all technical and non-technical staff. Security risks are assessed on an ongoing basis by AWI&#039;s institutional IT security team, which coordinates incident response and monitors threat landscapes relevant to research data infrastructure. PANGAEA technical staff participates in regular security reviews.&lt;br /&gt;
&lt;br /&gt;
Access controls for restricted or moratorium datasets are enforced at the application layer. Dataset metadata remains publicly accessible in all cases; access to the data itself is restricted to authorized users at the individual level for the duration of the moratorium, typically a maximum of two years from submission.&lt;br /&gt;
&lt;br /&gt;
The off-site replica at MARUM operates in a fully isolated network environment (Layer 2 separation) with access restricted to a firewalled VPN gateway at AWI. The Green IT Housing Center is monitored 24/7 by Bremen University staff located in very close proximity to the datacenter (15 m, adjacent building), with automated fire alarms, redundant power supply with battery backup, and multilevel physical access control.&lt;br /&gt;
&lt;br /&gt;
== Software Inventory ==&lt;br /&gt;
The following table summarizes the principal software components of the PANGAEA infrastructure and their development model.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
! Component !! Software !! Development Model&lt;br /&gt;
|-&lt;br /&gt;
| Primary database || PostgreSQL || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Data warehouse || Clickhouse || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Search and metadata index || Elasticsearch || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system backend || Java 17 / Dropwizard || In-house development&lt;br /&gt;
|-&lt;br /&gt;
| Editorial system frontend || React / Ant Design || In-house development (open source framework)&lt;br /&gt;
|-&lt;br /&gt;
| Source control || GitLab || Open source (community-supported, self-hosted)&lt;br /&gt;
|-&lt;br /&gt;
| CI/CD pipeline || GitLab CI/CD || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web server / reverse proxy (editorial) || Apache || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Web traffic management || NGINX || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| General web and Wiki frontend || PHP 8 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Repository services backend || Java 17 / Eclipse Jetty 12 || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Map visualization || leaflet.js || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Metadata framework || PanFMP || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| Issue tracking || JIRA (Atlassian) || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Monitoring stack || Grafana / Telegraf || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| External uptime monitoring || UptimeRobot || Commercial SaaS&lt;br /&gt;
|-&lt;br /&gt;
| Analytics || Matomo || Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
| Infrastructure virtualization || VMware || Commercial&lt;br /&gt;
|-&lt;br /&gt;
| Tape archive hardware || SpectraLogic TFinity ExaScale || Commercial hardware&lt;br /&gt;
|-&lt;br /&gt;
| Python data client || pangaeapy || In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
| R data client || pangaear || Community development (open source)&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Documentation and Change Management ==&lt;br /&gt;
General documentation of PANGAEA systems and services is maintained in the public &#039;&#039;&#039;PANGAEA Wiki&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/&amp;lt;/nowiki&amp;gt;). A separate internal &#039;&#039;&#039;Confluence&#039;&#039;&#039; Wiki, kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes, contains detailed documentation of server configurations, installation and maintenance procedures, relationships between system components, VM snapshot procedures, backup routines, and service restart priorities for incident response.&lt;br /&gt;
&lt;br /&gt;
All changes to published data and metadata are recorded in the editorial system&#039;s version history. Substantive revisions to published datasets result in a new dataset version with a new DOI; all prior versions remain accessible and cross-referenced. Minor editorial corrections that do not affect scientific content are applied without creating a new identifier and are transparently documented in the &amp;quot;Change history&amp;quot; section of the dataset landing page, including the date and a brief summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science. &#039;&#039;Scientific Data&#039;&#039;, 10, 347. &amp;lt;nowiki&amp;gt;https://doi.org/10.1038/s41597-023-02269-x&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Diepenbroek, M., Schindler, U., Huber, R., Pesant, S., Stocker, M., Felden, J., Buss, M. &amp;amp; Weinrebe, M. (2017). Terminology supported archiving and publication of environmental science data in PANGAEA. &#039;&#039;Journal of Biotechnology&#039;&#039;, 261, 177–186. &amp;lt;nowiki&amp;gt;https://doi.org/10.1016/j.jbiotec.2017.07.016&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Wu, M. et al. (2026). RDA Interest Group on Data Discovery Paradigms. &#039;&#039;Data Science Journal&#039;&#039;. &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
*[[PANGAEA search]]&lt;br /&gt;
*[[Data submission]]&lt;br /&gt;
*[[Authors Guides]]&lt;br /&gt;
*[[Curation levels]]&lt;br /&gt;
*[[Processing levels]]&lt;br /&gt;
*[[Data Usage Statistics]]&lt;br /&gt;
*[[Format]]&lt;br /&gt;
*[[PANGAEA XML schema]]&lt;br /&gt;
*[[Best practice manuals and templates]]&lt;br /&gt;
*[[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
*[[Preservation Plan]] — https://www.pangaea.de/about/preservation.php&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16770</id>
		<title>Technology</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16770"/>
		<updated>2026-04-03T16:03:22Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Updated according to new technical details introduced to the CTS re-certification 2026&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{construction}}&lt;br /&gt;
&lt;br /&gt;
PANGAEA is built on a three-tiered client/server architecture that has evolved continuously since the system&#039;s founding in the early 1990s. The architecture separates backend storage and database systems, middleware processing and transformation components, and frontend delivery and user interfaces into distinct layers that are individually maintainable and replaceable. This design philosophy has enabled PANGAEA to migrate core components — including its primary database engine, its editorial system, and its search infrastructure — without disruption to archived data or to users, and underpins the system&#039;s long-term stability as a trustworthy data repository.&lt;br /&gt;
&lt;br /&gt;
A detailed description of the PANGAEA information system and its workflows is provided in Felden et al. (2023) and Diepenbroek et al. (2017); this article provides an up-to-date overview oriented toward the general structure and key components.&lt;br /&gt;
&lt;br /&gt;
== System Architecture ==&lt;br /&gt;
&lt;br /&gt;
The technical architecture of PANGAEA follows a three-tiered model comprising a backend, a middleware layer, and a frontend. All hard- and software services are hosted and operated by the data and computing center of the Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research (AWI) in Bremerhaven. Most backend and middleware systems, as well as all frontend web servers and search engines, run on virtual machines (VMware) operated with Ubuntu Linux. Virtualization provides sufficient capacity and performance while enabling high availability and transparent hardware renewal on a typical cycle of three to four years.&lt;br /&gt;
[[File:PANGAEA overview architecture.png|alt=Schema illustrating the system architecture according to Felden et al., 2023|thumb|Schema illustrating the system architecture according to Felden et al., 2023|none|600x600px]]&lt;br /&gt;
&lt;br /&gt;
PANGAEA currently operates nine virtual machines at the AWI computing center, using a total of 53 CPUs, 162 GB RAM, and 28 TB of disk space. The relational database holdings currently occupy approximately 130 TB, with approximately 1 PB of data stored on tape.&lt;br /&gt;
&lt;br /&gt;
== Backend: Storage and Databases ==&lt;br /&gt;
&lt;br /&gt;
=== Relational Database ===&lt;br /&gt;
The primary store for all structured data and metadata in PANGAEA is a &#039;&#039;&#039;PostgreSQL&#039;&#039;&#039; relational database management system (RDBMS). The data model is highly normalized, reflecting the full observational context of each measurement: events (when and where data were collected), campaigns (cruises or field expeditions), methods and devices, parameters, references, and institutional provenance information are all stored in defined relational structures. This normalized model allows data descriptions to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived data.&lt;br /&gt;
&lt;br /&gt;
Database integrity is continuously maintained through &#039;&#039;&#039;PostgreSQL streaming replication&#039;&#039;&#039; to a dedicated backup system, which enables point-in-time recovery to any moment prior to a failure event. In addition, full database backup copies (`pg_base_backup`) are created each weekend and retained for three weeks, providing an independent recovery layer supplementary to the streaming replica.&lt;br /&gt;
&lt;br /&gt;
=== Data Warehouse ===&lt;br /&gt;
Fast access to large, aggregated compilations of data across the full PANGAEA holdings is provided by a &#039;&#039;&#039;Clickhouse&#039;&#039;&#039; data warehouse, which mirrors the numerical data inventory from the relational database. The data warehouse supports spatially and chronologically constrained queries at the parameter level across potentially hundreds of thousands of individual datasets, and is accessible both through the PANGAEA website and programmatically via a REST API. All data warehouse exports include the DOI for each contributing dataset, ensuring that provenance and citation remain intact in compiled data products.&lt;br /&gt;
&lt;br /&gt;
=== Archival Storage ===&lt;br /&gt;
For high-volume and binary data — including geophysical datasets, images, video, and community-specific formats such as NetCDF — data are stored in consistent formats on hard-disk arrays and robotic tape archives. The central archival storage system consists of two &#039;&#039;&#039;SpectraLogic TFinity ExaScale&#039;&#039;&#039; robotic tape libraries housed in separate buildings at AWI, with a combined capacity of up to 60 PB. Backup operations use high-capacity LTO-tape drives within this environment.&lt;br /&gt;
&lt;br /&gt;
All data are stored redundantly using &#039;&#039;&#039;erasure coding&#039;&#039;&#039; across disk and tape. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots. Write caches are battery-backed to ensure data integrity at the point of write.&lt;br /&gt;
&lt;br /&gt;
=== Off-Site Replica ===&lt;br /&gt;
Since 2025, PANGAEA operates a **minimal viable repository service** as an off-site replica at MARUM/University of Bremen, hosted in the Green IT Housing Center of Bremen University. This facility provides geographic and infrastructural separation from the primary AWI systems, enabling recovery of data delivery services within 24 hours following a technical breakdown or cyberattack. Disaster recovery and switchover procedures involving this facility are regularly exercised. The replica is kept current through snapshot-based replication several times per day and currently covers all dataset metadata — including individual landing pages and all harvesting endpoints — full representations of tabular data publications, the Elasticsearch search index, and the relevant web frontend. Extension of the replica to include binary data files is under active development, with formal commitment to be finalized through the ongoing MARUM hosting agreement negotiations.&lt;br /&gt;
&lt;br /&gt;
The off-site installation operates in a fully isolated environment with Layer 2 network separation. Access is restricted to VPN connections from a single designated gateway host at AWI, with access tokens and keys limited to the corresponding gateway and PANGAEA DevOps staff.&lt;br /&gt;
== Middleware: Processing and Transformation ==&lt;br /&gt;
The middleware layer manages the flow of data and metadata between the backend storage systems and the frontend interfaces. Its core functions are marshaling, indexing, transformation, and dissemination.&lt;br /&gt;
&lt;br /&gt;
Metadata are dynamically marshaled from the relational database to a PANGAEA-specific internal XML format, from which they are transformed via &#039;&#039;&#039;XSLT and XML-to-JSON pipelines&#039;&#039;&#039; into a range of content standards for delivery to users and harvesting services. Currently supported output standards include JSON-LD (Schema.org), DataCite XML, Dublin Core XML, ISO 19115/ISO 19139, and DIF/FGDC. Dissemination occurs via OAI-PMH, HTTP content negotiation following the Signposting standard (&amp;lt;nowiki&amp;gt;https://signposting.org/&amp;lt;/nowiki&amp;gt;), and other protocols. Metadata marshaling from the PostgreSQL database to the Elasticsearch search index is handled asynchronously by dedicated middleware components, which also manage automated DOI minting at DataCite upon dataset publication.&lt;br /&gt;
&lt;br /&gt;
The marshaled metadata are stored and indexed in &#039;&#039;&#039;Elasticsearch&#039;&#039;&#039;, which serves as the primary index for all public search and metadata access interfaces. This architecture separates the authoritative relational record from the search-optimized representation, allowing the search index to be rebuilt or updated without modifying archived data.&lt;br /&gt;
&lt;br /&gt;
The flexible metadata framework &#039;&#039;&#039;PanFMP&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://www.panfmp.org/&amp;lt;/nowiki&amp;gt;) underpins the mapping between the PANGAEA internal schema and the various external output standards, ensuring that new or updated standards can be accommodated without structural changes to the underlying data.&lt;br /&gt;
&lt;br /&gt;
Data submissions, user requests, and bug reports are managed through a &#039;&#039;&#039;JIRA&#039;&#039;&#039; (Atlassian) issue tracking system, which serves as the primary communication channel between data providers and the PANGAEA editorial team throughout the submission and publication process.&lt;br /&gt;
== Frontend: Editorial System ==&lt;br /&gt;
The PANGAEA editorial system is the primary tool through which data editors review, curate, harmonize, and publish submitted data and metadata. It is a &#039;&#039;&#039;web-based client/server application&#039;&#039;&#039; developed entirely in-house, operating directly on the PostgreSQL databases.&lt;br /&gt;
&lt;br /&gt;
The backend of the editorial system is built on &#039;&#039;&#039;Java 17&#039;&#039;&#039; using the &#039;&#039;&#039;Dropwizard&#039;&#039;&#039; framework, exposing a REST API. The frontend is implemented in &#039;&#039;&#039;React&#039;&#039;&#039; with the &#039;&#039;&#039;Ant Design&#039;&#039;&#039; component library. Source code for both components, together with shared libraries and test suites, is maintained in an institutional &#039;&#039;&#039;GitLab&#039;&#039;&#039; instance hosted at AWI, structured as an &#039;&#039;&#039;Nx monorepo&#039;&#039;&#039;. The repository encompasses automated unit tests and &#039;&#039;&#039;Cypress&#039;&#039;&#039; end-to-end tests. All new versions are deployed exclusively through a &#039;&#039;&#039;GitLab CI/CD pipeline&#039;&#039;&#039;, with releases gated on successful completion of the full automated test suite.&lt;br /&gt;
&lt;br /&gt;
In production, the editorial system runs on Ubuntu virtual machines hosted on VMware infrastructure. &#039;&#039;&#039;Apache&#039;&#039;&#039; serves the React frontend and acts as a reverse proxy to the backend services. Four parallel instances are operated simultaneously, providing dedicated test and production environments for two editorial system versions at any given time. This setup supports controlled feature validation alongside uninterrupted service operation. Infrastructure components — including VMware, file systems, monitoring, GitLab, and the database platform — are managed by AWI, while the PANGAEA group is responsible for the Apache web server, Java services, and React frontend.&lt;br /&gt;
&lt;br /&gt;
The relational database model underlying the editorial system enforces structural and semantic consistency between submitted metadata and the existing data inventory. The editorial system also integrates a &#039;&#039;&#039;Terminology Catalogue (TC)&#039;&#039;&#039;, which manages controlled vocabularies and ontologies used to harmonize data and metadata during ingest. Supported terminologies include WoRMS, ITIS, ChEBI, EnvO, PATO, QUDT, and the NERC vocabulary server, among others. A detailed description of the Terminology Catalogue and its role in data archiving and publication is given in Diepenbroek et al. (2017).&lt;br /&gt;
== Frontend: Public Web Interface and Search ==&lt;br /&gt;
&lt;br /&gt;
=== Web Delivery Stack ===&lt;br /&gt;
The public web infrastructure is fronted by an &#039;&#039;&#039;NGINX&#039;&#039;&#039; reverse proxy that manages incoming traffic and supports &#039;&#039;&#039;HTTP/3 (QUIC)&#039;&#039;&#039; and &#039;&#039;&#039;HTTP/2&#039;&#039;&#039; alongside HTTP/1.1 for broad client compatibility. Behind this entry point, a dual-backend architecture separates concerns by function: a &#039;&#039;&#039;PHP 8&#039;&#039;&#039; environment handles general information pages and the PANGAEA Wiki, while a &#039;&#039;&#039;Java 17&#039;&#039;&#039; application layer running on &#039;&#039;&#039;Eclipse Jetty 12&#039;&#039;&#039; manages core repository services — including dataset landing pages, DOI and handle resolution, and content negotiation for both human- and machine-readable data access. The Jetty-based layer also manages delivery of high-volume and binary files from the tape archive by staging requested files to a local hard-disk cache before serving them to users.&lt;br /&gt;
&lt;br /&gt;
=== Search ===&lt;br /&gt;
The PANGAEA search engine supports full-text and faceted search across all published metadata. Faceted navigation is enabled by semantic metadata enrichment performed during the marshaling process: terminology-based annotations are added to metadata records, allowing consistent filtering by topic, device type, geographic region, and other dimensions. PANGAEA&#039;s approach to data discovery complies with the recent recommendations of the Research Data Alliance (RDA) Interest Group &amp;quot;Data Discovery Paradigms&amp;quot; (Wu, M. et al., 2026, &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;). Search documentation is available in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/PANGAEA_search&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Map Search ===&lt;br /&gt;
Geographic search and visualization are implemented using &#039;&#039;&#039;leaflet.js&#039;&#039;&#039;, serving map tiles from four configurable sources: the AWI basemap, OpenStreetMap, Google (hybrid), and ESRI.&lt;br /&gt;
&lt;br /&gt;
=== Dataset Landing Pages ===&lt;br /&gt;
Each published dataset is represented by a landing page resolved through its DOI. Landing pages present the full dataset metadata in human-readable form and are enriched with structured &#039;&#039;&#039;schema.org&#039;&#039;&#039; markup in JSON-LD, ensuring machine-actionability and compatibility with generic web search engines and data registries. The schema.org metadata is also accessible via HTTP content negotiation following the Signposting standard. Minor editorial corrections to datasets that do not affect scientific content are documented in a &amp;quot;Change history&amp;quot; section of the landing page, recording the date and a summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
=== Programmatic Access ===&lt;br /&gt;
PANGAEA offers programmatic access to data and metadata through a range of web services (REST and SOAP). The OAI-PMH endpoint supports metadata harvesting in all supported standards. Client libraries for &#039;&#039;&#039;Python&#039;&#039;&#039; (pangaeapy, developed by PANGAEA) and &#039;&#039;&#039;R&#039;&#039;&#039; (pangaear, developed by the community) allow researchers to load and transform PANGAEA data directly into native data structures for analysis in environments such as Jupyter notebooks.&lt;br /&gt;
&lt;br /&gt;
== Monitoring and Analytics ==&lt;br /&gt;
Service health across the PANGAEA infrastructure is monitored through the &#039;&#039;&#039;AWI Grafana/Telegraf&#039;&#039;&#039; stack, covering service availability, application logs, and resource saturation for all production systems. This is supplemented by external availability checks via &#039;&#039;&#039;UptimeRobot&#039;&#039;&#039;, which provides independent verification of public-facing service endpoints from outside the AWI network. All external links in PANGAEA metadata records — references to related literature, other dataset versions, and external resources — are automatically checked for broken (HTTP 404) or permanently redirected (HTTP 301) responses on a weekly basis.&lt;br /&gt;
&lt;br /&gt;
User engagement and download metrics are captured by an integrated **Matomo** analytics instance configured to produce usage statistics compliant with the **COUNTER** (Counting Online Usage of Networked Electronic Resources) standard, ensuring that data impact is measured according to international scholarly norms. Usage statistics are publicly visible on each dataset landing page; details are documented in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/Data_Usage_Statistics&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
== Security ==&lt;br /&gt;
Backend and middleware systems are protected behind a firewall; frontend systems operate in a demilitarized zone (DMZ) accessible from outside with restricted but still firewalled access. Frontend systems have no write access and only limited read access to the backend database and tape archives. Public access from the website and REST APIs is served from data replicas hosted in Elasticsearch or through read-only remote filesystem access. Data curators access production systems via virtual private networks (VPN), which enforce a basic check that client operating systems are up to date.&lt;br /&gt;
&lt;br /&gt;
Physical access to the AWI computing center is managed through an electronic access control system for all relevant entrances; key distribution is documented, and guest policies are formally established. The AWI maintains uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation. Fire alerts are automatically forwarded to the fire department with a contractual response time under 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Security of the technical infrastructure is further maintained through the use of asymmetric key infrastructures, mandatory minimum-length passwords for all user classes, short-cycle security patching for all hardware and software components, professional monitoring tools for hardware, firewall, software, services, performance, and attacks, and regular security training for all technical and non-technical staff. Security risks are assessed on an ongoing basis by AWI&#039;s institutional IT security team, which coordinates incident response and monitors threat landscapes relevant to research data infrastructure. PANGAEA technical staff participates in regular security reviews.&lt;br /&gt;
&lt;br /&gt;
Access controls for restricted or moratorium datasets are enforced at the application layer. Dataset metadata remains publicly accessible in all cases; access to the data itself is restricted to authorized users at the individual level for the duration of the moratorium, typically a maximum of two years from submission.&lt;br /&gt;
&lt;br /&gt;
The off-site replica at MARUM operates in a fully isolated network environment (Layer 2 separation) with access restricted to a firewalled VPN gateway at AWI. The Green IT Housing Center is monitored 24/7 by Bremen University staff located in very close proximity to the datacenter (15 m, adjacent building), with automated fire alarms, redundant power supply with battery backup, and multilevel physical access control.&lt;br /&gt;
&lt;br /&gt;
== Software Inventory ==&lt;br /&gt;
The following table summarizes the principal software components of the PANGAEA infrastructure and their development model.&lt;br /&gt;
&lt;br /&gt;
== Documentation and Change Management ==&lt;br /&gt;
General documentation of PANGAEA systems and services is maintained in the public &#039;&#039;&#039;PANGAEA Wiki&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/&amp;lt;/nowiki&amp;gt;). A separate internal &#039;&#039;&#039;Confluence&#039;&#039;&#039; Wiki, kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes, contains detailed documentation of server configurations, installation and maintenance procedures, relationships between system components, VM snapshot procedures, backup routines, and service restart priorities for incident response.&lt;br /&gt;
&lt;br /&gt;
All changes to published data and metadata are recorded in the editorial system&#039;s version history. Substantive revisions to published datasets result in a new dataset version with a new DOI; all prior versions remain accessible and cross-referenced. Minor editorial corrections that do not affect scientific content are applied without creating a new identifier and are transparently documented in the &amp;quot;Change history&amp;quot; section of the dataset landing page, including the date and a brief summary of each change applied.&lt;br /&gt;
&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science. &#039;&#039;Scientific Data&#039;&#039;, 10, 347. &amp;lt;nowiki&amp;gt;https://doi.org/10.1038/s41597-023-02269-x&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Diepenbroek, M., Schindler, U., Huber, R., Pesant, S., Stocker, M., Felden, J., Buss, M. &amp;amp; Weinrebe, M. (2017). Terminology supported archiving and publication of environmental science data in PANGAEA. &#039;&#039;Journal of Biotechnology&#039;&#039;, 261, 177–186. &amp;lt;nowiki&amp;gt;https://doi.org/10.1016/j.jbiotec.2017.07.016&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Wu, M. et al. (2026). RDA Interest Group on Data Discovery Paradigms. &#039;&#039;Data Science Journal&#039;&#039;. &amp;lt;nowiki&amp;gt;https://doi.org/10.5334/dsj-2026-006&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
*[[PANGAEA search]]&lt;br /&gt;
*[[Data submission]]&lt;br /&gt;
*[[Authors Guides]]&lt;br /&gt;
*[[Curation levels]]&lt;br /&gt;
*[[Processing levels]]&lt;br /&gt;
*[[Data Usage Statistics]]&lt;br /&gt;
*[[Format]]&lt;br /&gt;
*[[PANGAEA XML schema]]&lt;br /&gt;
*[[Best practice manuals and templates]]&lt;br /&gt;
*[[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
*[[Preservation Plan]] — https://www.pangaea.de/about/preservation.php&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Preservation_Plan&amp;diff=16769</id>
		<title>Preservation Plan</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Preservation_Plan&amp;diff=16769"/>
		<updated>2026-04-03T11:25:55Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Finalized updating&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;PANGAEA is committed to the long-term preservation of all data and metadata entrusted to it by the research community. This commitment extends beyond bit-level storage integrity to encompass active management of format usability, semantic consistency over time, and formally documented procedures for all stages of the archival lifecycle. The following article describes the technical and organizational measures that together constitute PANGAEA&#039;s preservation strategy. It supplements the information provided in Felden et al. (2023) and is one of the reference documents for PANGAEA&#039;s CoreTrustSeal certification.&lt;br /&gt;
== Principles ==&lt;br /&gt;
PANGAEA&#039;s preservation approach is grounded in three guiding principles. First, data are not merely stored as files but are ingested into a structured, normalized relational database that preserves the full semantic context of each measurement — ensuring that data remain interpretable independently of any external documentation. Second, preservation is an active process: PANGAEA monitors the long-term usability of archived formats and takes preventive action against obsolescence, including the creation of format-migrated copies when required. Third, institutional commitment is formally secured: the AWI/MARUM cooperation agreement (AMAR) guarantees that all archived data and metadata will remain accessible for a minimum of ten years following any formal decommissioning of PANGAEA, and that the host institutions will maintain the necessary infrastructure and expertise to honor this commitment.&lt;br /&gt;
&lt;br /&gt;
PANGAEA&#039;s ingest and archiving workflow is compliant with the Open Archival Information System (OAIS) standard (ISO 14721).&lt;br /&gt;
== Metadata Preservation ==&lt;br /&gt;
PANGAEA treats metadata as essential for the long-term reusability of data. Metadata are stored in a highly normalized PostgreSQL relational database, whose schema is modeled to be compatible with international standards including Schema.org and ISO 19115. This normalized structure allows dataset representations to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived records.&lt;br /&gt;
&lt;br /&gt;
The following metadata categories are collected and preserved for every published dataset:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Citation metadata:&#039;&#039;&#039; author and contributor names with ORCID iDs; institutional affiliations with ROR identifiers; dataset title; publication year; publisher;  DOI name;  [resource type] according to Stall et al., 2023&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Funding information:&#039;&#039;&#039; project names, grant numbers, and funder identifiers (Crossref Funder IDs or ROR identifiers).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Event information:&#039;&#039;&#039; detailed spatial and temporal coverage of sampling or measurement events, including methods, devices, and campaign context.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Related documentation:&#039;&#039;&#039; links (using DOIs or other persistent identifiers) to related scientific articles, reports, and supplementary materials. Where related documentation is not held in an external repository with a persistent identifier, PANGAEA stores a local copy in PDF/A format. PDF/A is preferred for its long-term stability; copies will be migrated to successor standards if PDF/A itself becomes obsolete.&lt;br /&gt;
&lt;br /&gt;
PANGAEA&#039;s database schema is continuously adapted to accommodate new and evolving metadata standards. When schema extensions are introduced, the metadata of existing datasets are reviewed and updated accordingly. All such changes are managed carefully to avoid incompatible modifications to existing records.&lt;br /&gt;
== Data Object Preservation ==&lt;br /&gt;
&lt;br /&gt;
=== Preservation Responsibility by Data Type ===&lt;br /&gt;
PANGAEA differentiates its preservation responsibility according to data type. For structured tabular data stored in the relational database, PANGAEA assumes full responsibility for structural and semantic integrity over time. For binary objects, PANGAEA commits to monitoring format viability and taking remedial action — including format migration — when long-term accessibility is threatened; the original submitted files are always retained regardless of any migration.&lt;br /&gt;
&lt;br /&gt;
=== Tabular Data ===&lt;br /&gt;
Except for binary objects, all submitted data values are imported into the PANGAEA relational database as structured data series. Each data entry in a data series carries metadata about its type (numeric, date/time, string), the responsible scientist (PI), the methodology applied, and, for numerical values, format information including significant digits. This structured representation decouples the data from any particular file format, ensuring long-term interpretability regardless of changes in software environments.&lt;br /&gt;
&lt;br /&gt;
At the time of archival, a copy of each dataset — together with a checksum and timestamp — is additionally marshaled to disk as a tab-delimited text file. These reference copies serve as the basis for fixity verification: in the event of suspected data loss or corruption, the relational database contents can be compared against these checksummed disk copies. The data submission system independently retains the original files uploaded by data providers, providing a further reference point for integrity checking.&lt;br /&gt;
&lt;br /&gt;
=== Binary Data ===&lt;br /&gt;
Not all data held in PANGAEA is available in tabular form. Some datasets are archived in compact, community-specific binary formats — including NetCDF files, images, video recordings, and geophysical data products. High-volume and binary files are stored on hard-disk arrays and robotic tape archives; when a user requests access to a file held on tape, it is staged to a local disk cache before delivery, ensuring that tape-based archival does not compromise accessibility for the Designated Community.&lt;br /&gt;
&lt;br /&gt;
For binary objects, long-term usability is an active responsibility: the PANGAEA team monitors software dependencies, version changes, and backward compatibility issues for all archived binary formats. Where continued readability requires it, new format-migrated copies are created and archived; the original submitted file is always retained alongside any migrated copy. Data deposited in non-preferred formats are either transformed to PANGAEA-compatible formats during editorial processing or, if transformation is not feasible, stored as static binary files with appropriate metadata documentation.&lt;br /&gt;
&lt;br /&gt;
PANGAEA applies format rules before accepting binary data for archival. Where possible, uncompressed or widely supported open formats are preferred. Currently accepted formats are:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Images:&#039;&#039;&#039; JPEG, PNG, TIFF&lt;br /&gt;
* &#039;&#039;&#039;Documents:&#039;&#039;&#039; PDF/A (preferred), ODF, OOXML&lt;br /&gt;
* &#039;&#039;&#039;Media containers:&#039;&#039;&#039; MP4, MPG, OGG, Matroska; audio and video content within containers must comply with the following codecs:&lt;br /&gt;
** &#039;&#039;Video:&#039;&#039; uncompressed, MPEG-1, MPEG-4 Part 2, AVC/H.264, H.265&lt;br /&gt;
** &#039;&#039;Audio:&#039;&#039; uncompressed, MPEG Layer III (MP3), MPEG-4 Part 3/AAC&lt;br /&gt;
* &#039;&#039;&#039;Scientific data:&#039;&#039;&#039; NetCDF, preferably using the Climate and Forecast (CF) Metadata Conventions; detailed documentation is required in all other cases&lt;br /&gt;
&lt;br /&gt;
This list is not exhaustive and is updated as community standards evolve. If any accepted format becomes deprecated or is superseded, PANGAEA will migrate archived copies to modern equivalents while retaining the originals.&lt;br /&gt;
&lt;br /&gt;
Raw data — defined as level-0 data without any accompanying metadata — are not accepted for archival. Data at processing level 1 (raw data with a minimum set of metadata) may be accepted if adequate contextual information is provided. No guarantees are given for the long-term usability of level-0 or level-1 datasets. Processing levels are documented here:  [[Processing levels]].&lt;br /&gt;
== Storage and Physical Infrastructure ==&lt;br /&gt;
PANGAEA&#039;s storage infrastructure is operated by the computing center of the Alfred Wegener Institute (AWI) in Bremerhaven, in accordance with the AMAR cooperation agreement. PANGAEA currently uses approximately 5 TB for its relational databases and approximately 1 PB for data on tape. A comprehensive set of technical and organizational measures (TOM) is in place:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Redundant storage:&#039;&#039;&#039; all data are stored using erasure coding across disk and tape, with write caches battery-backed to ensure integrity at the point of write. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Tape archive:&#039;&#039;&#039; the central archival storage system consists of two SpectraLogic TFinity ExaScale robotic tape libraries, housed in separate buildings at AWI, with a combined capacity of up to 60 PB and using high-capacity LTO-tape drives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Database integrity:&#039;&#039;&#039; PostgreSQL streaming replication to a dedicated backup system enables point-in-time recovery to any moment prior to a failure event. In addition, full database backup copies (`pg_base_backup`) are created each weekend and retained for three weeks, providing an additional recovery layer independent of the streaming replica.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Facility measures:&#039;&#039;&#039; fire and smoke detection systems; server room monitoring of temperature and humidity; server room air-conditioning; uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation; RAID and hard disk mirroring in the virtualization environment; user permission management; network firewall and intrusion detection systems; anti-virus email filtering.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Documentation:&#039;&#039;&#039; all systems are documented in an internal Confluence Wiki kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes. A ticket system is used to track and manage incidents.&lt;br /&gt;
&lt;br /&gt;
Hardware is typically renewed every three to four years through the AWI computing center&#039;s lifecycle management program, implemented transparently via virtualization. AWI holds support contracts for all relevant hardware and non-open-source software, with five business-day response times and one-day part replacement guarantees, ensuring that hardware failures can be remedied without extended service interruption.&lt;br /&gt;
&lt;br /&gt;
== Off-Site Replica at MARUM ==&lt;br /&gt;
Since 2025, PANGAEA operates a minimal viable repository service as an off-site replica at MARUM/University of Bremen, hosted in the Green IT Housing Center (Rechenzentrum) of Bremen University. This facility provides geographic and institutional separation from the primary AWI infrastructure and is a key component of PANGAEA&#039;s resilience strategy against both technical failure and cyberattack scenarios. The replica enables recovery of data delivery services within 24 hours following a catastrophic failure at AWI. Disaster recovery and switchover procedures involving this facility are regularly exercised.&lt;br /&gt;
&lt;br /&gt;
The off-site installation currently covers:&lt;br /&gt;
&lt;br /&gt;
* All dataset metadata, including individual DOI landing pages and all metadata serializations available via harvesting endpoints (OAI-PMH - Schema.org, DataCite, Dublin Core, DIF, ISO 19139)&lt;br /&gt;
* Full representations of tabular data publications&lt;br /&gt;
* The Elasticsearch search index&lt;br /&gt;
* The relevant web frontend&lt;br /&gt;
&lt;br /&gt;
Extension of the replica to include binary data files is under active development, with formal commitment to be finalized through the ongoing MARUM hosting agreement negotiations.&lt;br /&gt;
&lt;br /&gt;
The Green IT Housing Center maintains the following physical and technical safeguards: two separate fire sections with servers distributed across both for site redundancy; 24/7 on-site monitoring by Bremen University staff located in very close proximity to the datacenter (15 m, adjacent building); automated fire alarms with a fire brigade station less than 1 km away; redundant power supply with battery backup; and multilevel physical access control ensuring that only authorized personnel can access the physical hardware. The off-site installation operates in a fully isolated network environment (Layer 2 separation), accessible only via a firewalled VPN gateway at AWI. The number of access tokens and keys is strictly limited to the corresponding gateway host and PANGAEA DevOps staff. Replication is unidirectional from AWI to MARUM, using snapshot-based transfers executed multiple times per day.&lt;br /&gt;
&lt;br /&gt;
== Versioning and Persistent Identifiers ==&lt;br /&gt;
Every published dataset is assigned a universally unique Digital Object Identifier (DOI) minted at DataCite. DOI minting is automated as part of the archival workflow. DOI resolution is actively maintained: PANGAEA keeps its authoritative metadata records at DataCite synchronized with dataset landing pages, and all external links in metadata records are checked automatically on a weekly basis for broken (HTTP 404) or permanently redirected (HTTP 301) responses.&lt;br /&gt;
&lt;br /&gt;
New DOI names are created under clearly defined conditions:&lt;br /&gt;
&lt;br /&gt;
* A new identifier is issued upon the initial publication of each dataset.&lt;br /&gt;
* A new identifier is issued when a published dataset undergoes a substantive revision of data or metadata that would affect reproducibility or scientific interpretation. The prior version remains accessible and is cross-referenced in the metadata record of the new version.&lt;br /&gt;
* Minor editorial corrections that do not affect scientific content are applied without creating a new identifier. Such corrections are transparently documented in the “Change history” section of the dataset landing pages, including the date and a brief summary of the changes applied.&lt;br /&gt;
&lt;br /&gt;
All versions are linked in the metadata record, ensuring full traceability of the publication history for data users and citing authors.&lt;br /&gt;
== Deletion and Tombstone Records ==&lt;br /&gt;
PANGAEA does not routinely delete or remove published datasets. In exceptional cases where retraction is required — for example, due to confirmed scientific error, misconduct, copyright infringement, or data privacy obligations under applicable law (e.g., GDPR Art. 17) — the following procedure applies: the data itself is made inaccessible, but the DOI and the dataset landing page are retained as a tombstone record. The tombstone record clearly indicates the dataset&#039;s status and the documented reason for its withdrawal, in accordance with DataCite&#039;s metadata practices for retracted records. All such actions are logged in the editorial system&#039;s change history. &lt;br /&gt;
&lt;br /&gt;
== Custody Transfer and Decommissioning ==&lt;br /&gt;
Should PANGAEA cease operations, the host institutions guarantee that all data and metadata will remain accessible for a minimum of ten years following any formal decommissioning. In such a scenario, only the submission and editorial system would be terminated; the database and data delivery services would remain operational. A full transition of custody to another repository could be supported by the off-site replica at MARUM as a concrete technical starting point, providing immediate access to all metadata and tabular data holdings. As a further fallback, PANGAEA can be reduced to a file-based repository: a complete file-based copy of all datasets, including binary objects, can be assembled and made available independently by AWI and/or the University of Bremen. The legal and institutional basis for these guarantees is the AWI/MARUM cooperation agreement (AMAR); a summary of its key commitments is available in the [[Continuity Plan]].&lt;br /&gt;
&lt;br /&gt;
== Community-Specific Preservation Documentation ==&lt;br /&gt;
In coordination with scientific communities, PANGAEA has developed detailed documentation on the harmonization and preservation of specific data types. These include guidance on CTD data, Thermosalinograph (TSG) underway data, and bathymetric data, among others. Where applicable, these documents contain information on format choices and long-term preservation handling specific to the relevant data type. See  [[Best practice manuals and templates]] for the full collection.&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science.  &#039;&#039;Scientific Data&#039;&#039;, 10, 347. https://doi.org/10.1038/s41597-023-02269-x&lt;br /&gt;
* Stall, S., Bilder, G., Cannon, M.  &#039;&#039;et al.&#039;&#039; Journal Production Guidance for Software and Data Citations.  &#039;&#039;Sci Data&#039;&#039;  &#039;&#039;&#039;10&#039;&#039;&#039;, 656 (2023). https://doi.org/10.1038/s41597-023-02491-7&lt;br /&gt;
* Consultative Committee for Space Data Systems (2012). Reference Model for an Open Archival Information System (OAIS). Recommended Practice CCSDS 650.0-M-2. https://public.ccsds.org/Pubs/650x0m2.pdf&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
* [[Technology]]&lt;br /&gt;
* [[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
* [[Authors Guides]]&lt;br /&gt;
* [[Format]]&lt;br /&gt;
* [[Processing levels]]&lt;br /&gt;
* [[Curation levels]]&lt;br /&gt;
* [[Best practice manuals and templates]]&lt;br /&gt;
* [[PANGAEA XML schema]]&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Preservation_Plan&amp;diff=16768</id>
		<title>Preservation Plan</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Preservation_Plan&amp;diff=16768"/>
		<updated>2026-04-03T11:01:40Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Start updating the article with recent reviewed adaptations to the CTS  certification documentation&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;PANGAEA is committed to the long-term preservation of all data and metadata entrusted to it by the research community. This commitment extends beyond bit-level storage integrity to encompass active management of format usability, semantic consistency over time, and formally documented procedures for all stages of the archival lifecycle. The following article describes the technical and organizational measures that together constitute PANGAEA&#039;s preservation strategy. It supplements the information provided in Felden et al. (2023) and is one of the reference documents for PANGAEA&#039;s CoreTrustSeal certification.&lt;br /&gt;
== Principles ==&lt;br /&gt;
PANGAEA&#039;s preservation approach is grounded in three guiding principles. First, data are not merely stored as files but are ingested into a structured, normalized relational database that preserves the full semantic context of each measurement — ensuring that data remain interpretable independently of any external documentation. Second, preservation is an active process: PANGAEA monitors the long-term usability of archived formats and takes preventive action against obsolescence, including the creation of format-migrated copies when required. Third, institutional commitment is formally secured: the AWI/MARUM cooperation agreement (AMAR) guarantees that all archived data and metadata will remain accessible for a minimum of ten years following any formal decommissioning of PANGAEA, and that the host institutions will maintain the necessary infrastructure and expertise to honor this commitment.&lt;br /&gt;
&lt;br /&gt;
PANGAEA&#039;s ingest and archiving workflow is compliant with the Open Archival Information System (OAIS) standard (ISO 14721).&lt;br /&gt;
== Metadata Preservation ==&lt;br /&gt;
PANGAEA treats metadata as essential for the long-term reusability of data. Metadata are stored in a highly normalized PostgreSQL relational database, whose schema is modeled to be compatible with international standards including Schema.org and ISO 19115. This normalized structure allows dataset representations to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived records.&lt;br /&gt;
&lt;br /&gt;
The following metadata categories are collected and preserved for every published dataset:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Citation metadata:&#039;&#039;&#039; author and contributor names with ORCID iDs; institutional affiliations with ROR identifiers; dataset title; publication year; publisher;  DOI name;  [resource type] according to Stall et al., 2023&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Funding information:&#039;&#039;&#039; project names, grant numbers, and funder identifiers (Crossref Funder IDs or ROR identifiers).&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Event information:&#039;&#039;&#039; detailed spatial and temporal coverage of sampling or measurement events, including methods, devices, and campaign context.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Related documentation:&#039;&#039;&#039; links (using DOIs or other persistent identifiers) to related scientific articles, reports, and supplementary materials. Where related documentation is not held in an external repository with a persistent identifier, PANGAEA stores a local copy in PDF/A format. PDF/A is preferred for its long-term stability; copies will be migrated to successor standards if PDF/A itself becomes obsolete.&lt;br /&gt;
&lt;br /&gt;
PANGAEA&#039;s database schema is continuously adapted to accommodate new and evolving metadata standards. When schema extensions are introduced, the metadata of existing datasets are reviewed and updated accordingly. All such changes are managed carefully to avoid incompatible modifications to existing records.&lt;br /&gt;
== Data Object Preservation ==&lt;br /&gt;
&lt;br /&gt;
=== Tabular Data ===&lt;br /&gt;
Except for binary objects, all submitted data values are imported into the PANGAEA relational database as structured data series. Each data entry carries metadata about its type (numeric, date/time, string), the responsible scientist (PI), the methodology applied, and, for numerical values, format information including significant digits. This structured representation decouples the data from any particular file format, ensuring long-term interpretability regardless of changes in software environments.&lt;br /&gt;
&lt;br /&gt;
At the time of archival, a copy of each dataset (with checksum and timestamp) is additionally marshaled to disk as a tab-delimited text file. These copies serve as reference files for integrity verification, and the tab-delimited format ensures readability without specialized software.&lt;br /&gt;
&lt;br /&gt;
=== Binary Data ===&lt;br /&gt;
Not all data held in PANGAEA is available in tabular form. Some datasets are archived in compact, community-specific binary formats — including NetCDF files, images, video recordings, and geophysical data products. For these, long-term usability is an active responsibility: the PANGAEA team monitors software dependencies, version changes, and backward compatibility issues for all archived binary formats. Where continued readability requires it, new format-migrated copies are created and archived; the original submitted file is always retained alongside any migrated copy.&lt;br /&gt;
&lt;br /&gt;
PANGAEA applies format rules before accepting binary data for archival. Where possible, uncompressed or widely supported open formats are preferred. Currently accepted formats are:&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;Images:&#039;&#039;&#039; JPEG, PNG, TIFF&lt;br /&gt;
* &#039;&#039;&#039;Documents:&#039;&#039;&#039; PDF/A (preferred), ODF, OOXML&lt;br /&gt;
* &#039;&#039;&#039;Media containers:&#039;&#039;&#039; MP4, MPG, OGG, Matroska; audio and video content within containers must comply with the following codecs:&lt;br /&gt;
** &#039;&#039;Video:&#039;&#039; uncompressed, MPEG-1, MPEG-4 Part 2, AVC/H.264, H.265&lt;br /&gt;
** &#039;&#039;Audio:&#039;&#039; uncompressed, MPEG Layer III (MP3), MPEG-4 Part 3/AAC&lt;br /&gt;
* &#039;&#039;&#039;Scientific data:&#039;&#039;&#039; NetCDF, preferably using the Climate and Forecast (CF) Metadata Conventions; detailed documentation is required in all other cases&lt;br /&gt;
&lt;br /&gt;
This list is not exhaustive and is updated as community standards evolve. If any accepted format becomes deprecated or is superseded, PANGAEA will migrate archived copies to modern equivalents while retaining the originals.&lt;br /&gt;
&lt;br /&gt;
Raw data — defined as level-0 data without any accompanying metadata — are not accepted for archival. Data at processing level 1 (raw data with a minimum set of metadata) may be accepted if adequate contextual information is provided. No guarantees are given for the long-term usability of level-0 or level-1 datasets. Processing levels are documented here:  [[Processing levels]].&lt;br /&gt;
== Storage and Physical Infrastructure ==&lt;br /&gt;
PANGAEA&#039;s storage infrastructure is operated by the computing center of the Alfred Wegener Institute (AWI) in Bremerhaven, in accordance with the AMAR cooperation agreement. A comprehensive set of technical and organizational measures (TOM) is in place:&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Redundant storage:&#039;&#039;&#039; all data are stored using erasure coding across disk and tape, with write caches battery-backed to ensure integrity at the point of write. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Tape archive:&#039;&#039;&#039; the central archival storage system consists of two SpectraLogic TFinity ExaScale robotic tape libraries, housed in separate buildings at AWI, with a combined capacity of up to 60 PB and using high-capacity LTO-tape drives.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Database integrity:&#039;&#039;&#039; PostgreSQL streaming replication to a dedicated backup system enables point-in-time recovery to any moment prior to a failure event.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Facility measures:&#039;&#039;&#039; fire and smoke detection systems; server room monitoring of temperature and humidity; server room air-conditioning; uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation; RAID and hard disk mirroring in the virtualization environment; user permission management; network firewall and intrusion detection systems; anti-virus email filtering.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Documentation:&#039;&#039;&#039; all systems are documented in an internal Confluence Wiki kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes. A ticket system is used to track and manage incidents.&lt;br /&gt;
&lt;br /&gt;
Hardware is typically renewed every three to four years through the AWI computing center&#039;s lifecycle management program, implemented transparently via virtualization.&lt;br /&gt;
== Off-Site Replica at MARUM ==&lt;br /&gt;
Since 2025, PANGAEA operates an off-site replica of the relational database and web frontend at MARUM/University of Bremen, hosted in the Green IT Housing Center (Rechenzentrum) of Bremen University. This facility provides geographic and institutional separation from the primary AWI infrastructure and is a key component of PANGAEA&#039;s resilience strategy against both technical failure and cyberattack scenarios.&lt;br /&gt;
&lt;br /&gt;
The off-site installation currently covers:&lt;br /&gt;
&lt;br /&gt;
* All dataset metadata, including individual DOI landing pages and all metadata serializations available via harvesting endpoints (OAI-PMH, schema.org, DataCite, Dublin Core, DIF, ISO 19139)&lt;br /&gt;
* Full representations of tabular data publications&lt;br /&gt;
&lt;br /&gt;
Extension of the replica to include binary data files is planned as part of the ongoing development of this facility. The replica enables recovery of data delivery services within 24 hours following a catastrophic failure at AWI.&lt;br /&gt;
&lt;br /&gt;
The MARUM Green IT Housing Center maintains the following physical and technical safeguards: two separate fire sections with servers distributed across both for site redundancy; 24/7 on-site monitoring by Bremen University staff; automated fire alarms with a fire brigade station less than 1 km away; redundant power supply with battery backup; and multilevel physical access control. The off-site installation operates in a fully isolated network environment (Layer 2 separation), accessible only via a firewalled VPN gateway at AWI. The number of access tokens and keys is strictly limited to the corresponding gateway host and PANGAEA DevOps staff. Replication is unidirectional from AWI to MARUM, using snapshot-based transfers executed multiple times per day.&lt;br /&gt;
== Versioning and Persistent Identifiers ==&lt;br /&gt;
Every published dataset is assigned a universally unique Digital Object Identifier (DOI) minted at DataCite. DOI resolution is actively maintained: PANGAEA keeps its authoritative metadata records at DataCite synchronized with dataset landing pages, and all external links in metadata records are checked automatically on a weekly basis for broken (HTTP 404) or permanently redirected (HTTP 301) responses.&lt;br /&gt;
&lt;br /&gt;
New DOI names are created under clearly defined conditions:&lt;br /&gt;
&lt;br /&gt;
* A new identifier is issued upon the initial publication of each dataset.&lt;br /&gt;
* A new identifier is issued when a published dataset undergoes a substantive revision of data or metadata that would affect reproducibility or scientific interpretation. The prior version remains accessible and is cross-referenced in the metadata record of the new version.&lt;br /&gt;
* Minor editorial corrections that do not affect scientific content are applied without creating a new identifier. Instead, the corrections are transparently documented in the “Change history” section of the dataset landing pages, including the date and a short summary of the changes applied.&lt;br /&gt;
&lt;br /&gt;
All versions are linked in the metadata record, ensuring full traceability of the publication history for data users and citing authors.&lt;br /&gt;
== Deletion and Tombstone Records ==&lt;br /&gt;
PANGAEA does not routinely delete or remove published datasets. In exceptional cases where retraction is required — for example, due to demonstrated scientific error, misconduct, copyright infringement, or data privacy obligations under applicable law (e.g., GDPR Art. 17) — the following procedure applies: the data itself is made inaccessible, but the DOI and the dataset landing page are retained as a tombstone record. The tombstone record clearly indicates the dataset&#039;s status and the documented reason for its withdrawal, in accordance with DataCite&#039;s tombstone policy. All such actions are logged in the editorial system&#039;s change history. &lt;br /&gt;
== Custody Transfer and Decommissioning ==&lt;br /&gt;
Should PANGAEA cease operations, the host institutions guarantee that all data and metadata will remain accessible for a minimum of ten years following any formal decommissioning. In such a scenario, only the submission and editorial system would be terminated; the database and data delivery services would remain operational. A full transition of custody to another repository could be supported by the off-site replica at MARUM as a concrete technical starting point. As a further fallback, PANGAEA can be reduced to a file-based repository: a complete file-based copy of all datasets, including binary objects, can be assembled and made available independently by AWI and/or the University of Bremen. The legal and institutional basis for these guarantees is the AWI/MARUM cooperation agreement (AMAR); a summary of its key commitments is available in the [[Continuity Plan]].&lt;br /&gt;
== Community-Specific Preservation Documentation ==&lt;br /&gt;
In coordination with scientific communities, PANGAEA has developed detailed documentation on the harmonization and preservation of specific data types. These include guidance on CTD data, Thermosalinograph (TSG) underway data, and bathymetric data, among others. Where applicable, these documents contain information on format choices and long-term preservation handling specific to the relevant data type. See  [[Best practice manuals and templates]] for the full collection.&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science.  &#039;&#039;Scientific Data&#039;&#039;, 10, 347. https://doi.org/10.1038/s41597-023-02269-x&lt;br /&gt;
* Stall, S., Bilder, G., Cannon, M.  &#039;&#039;et al.&#039;&#039; Journal Production Guidance for Software and Data Citations.  &#039;&#039;Sci Data&#039;&#039;  &#039;&#039;&#039;10&#039;&#039;&#039;, 656 (2023). https://doi.org/10.1038/s41597-023-02491-7&lt;br /&gt;
* Consultative Committee for Space Data Systems (2012). Reference Model for an Open Archival Information System (OAIS). Recommended Practice CCSDS 650.0-M-2. https://public.ccsds.org/Pubs/650x0m2.pdf&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
* [[Technology]]&lt;br /&gt;
* [[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
* [[Authors Guides]]&lt;br /&gt;
* [[Format]]&lt;br /&gt;
* [[Processing levels]]&lt;br /&gt;
* [[Curation levels]]&lt;br /&gt;
* [[Best practice manuals and templates]]&lt;br /&gt;
* [[PANGAEA XML schema]]&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Format&amp;diff=16757</id>
		<title>Format</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Format&amp;diff=16757"/>
		<updated>2026-03-31T15:09:01Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Proprietary formats */ changed wording a bit&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;PANGAEA accepts and publishes a wide spectrum of data formats. Thereby, we classify these file formats into three categories – [[Format#Documentation|Documentation formats]], and [[Format#Tabular data|Tabular]] and [[Format#Binary data|Binary data]] formats – based on how these formats are treated and processed during our editorial work, and how they are represented in our data publications. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Important:&#039;&#039;&#039; If you consider submitting data and supplementary documentation to PANGAEA, please make sure to provide open (ideally non-proprietary) formats widely accepted and endorsed in your scientific communities in order to support the accessibility and (re-)usability on long time-scales and by openly available tools.  &lt;br /&gt;
&lt;br /&gt;
== Documentation ==&lt;br /&gt;
As documentation we consider all files provided in typical text data formats meant to supplement or further describe data submissions (whether in tabular or binary formats), such as processing reports, instrument calibration protocols, standard operating procedures.&lt;br /&gt;
&lt;br /&gt;
Accepted formats for documentation are:&lt;br /&gt;
* &#039;&#039;&#039;PDF&#039;&#039;&#039;/A (ISO19005) - http://en.wikipedia.org/wiki/PDF/A&lt;br /&gt;
* &#039;&#039;&#039;RTF&#039;&#039;&#039; or &#039;&#039;&#039;ODF&#039;&#039;&#039; (ISO26300) - http://en.wikipedia.org/wiki/OpenDocument&lt;br /&gt;
* Microsoft &#039;&#039;&#039;Office files&#039;&#039;&#039; such as &#039;&#039;&#039;*.docx&#039;&#039;&#039; for MS Word, and &#039;&#039;&#039;*.xlsx&#039;&#039;&#039; for Excel spreadsheet documents – compliant to standard OOXML (ISO/IEC 29500:2008 since Office 2013 – https://de.wikipedia.org/wiki/Microsoft_Office),&lt;br /&gt;
* or (our favourite and recommended) &#039;&#039;&#039;plain UTF-8 encoded text files&#039;&#039;&#039; - https://en.wikipedia.org/wiki/UTF-8&lt;br /&gt;
&lt;br /&gt;
== Tabular data ==&lt;br /&gt;
PANGAEA is specifically good at (or, to put it differently, is able to make the most of) tabular field observation data. Owing to the many efforts we put in harmonizing parameter/variable names, methods, dimensions and units during our editorial processing, and due to the fact that this kind of data is stored in a relational database (PostgreSQL) at PANGAEA, our users are able to easily compile specific parameters/variables of interest from many similar studies (i.e. related individual publications) into meta-studies targeting new research questions and contexts using our [[PANGAEA search#Data warehouse|Data Warehouse]]. Or apply similar functionality with the help of the Python module pangaeapy or the R pendant pangaear (see our [https://www.pangaea.de/tools/ Tools site] for details). &lt;br /&gt;
&lt;br /&gt;
The &#039;&#039;&#039;preferred&#039;&#039;&#039; &#039;&#039;&#039;formats&#039;&#039;&#039; for data tables are:  &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;TAB-delimited TEXT-files&#039;&#039;&#039; (UTF-8 encoded), or &lt;br /&gt;
* &#039;&#039;&#039;(open) spreadsheet file formats&#039;&#039;&#039; (MS Excel .xlsx, OpenOffice &amp;amp; LibreOffice Calc .ods etc)&#039;&#039;&#039;.&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Please note that data tables are &#039;&#039;&#039;not&#039;&#039;&#039; accepted as encapsulated objects (e.g., in .mat files).&lt;br /&gt;
&lt;br /&gt;
Example for a tabular dataset: https://doi.org/10.1594/PANGAEA.934148&lt;br /&gt;
&lt;br /&gt;
== Binary data ==&lt;br /&gt;
Binary objects and documentations are usually stored in a combination of hard-drive arrays (for immediate and performant access) and tape archives. File formats should follow ISO standards or at least &#039;&#039;de facto&#039;&#039; standards. Online preview is available for raster graphics and videos (e.g. .tif, .png, .jpeg, .mp4).&lt;br /&gt;
&lt;br /&gt;
Example: https://doi.org/10.1594/PANGAEA.936185&lt;br /&gt;
&lt;br /&gt;
=== Images ===&lt;br /&gt;
* tiff&lt;br /&gt;
** http://en.wikipedia.org/wiki/Tagged_Image_File_Format&lt;br /&gt;
** [http://www.iso.org/iso/iso_catalogue/catalogue_ics/catalogue_detail_ics.htm?csnumber=2181 ISO12639:1998]&lt;br /&gt;
* jpeg&lt;br /&gt;
** http://en.wikipedia.org/wiki/Jpeg&lt;br /&gt;
** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=18902 ISO10918-1:1994]&lt;br /&gt;
*png&lt;br /&gt;
**https://en.wikipedia.org/wiki/Portable_Network_Graphics&lt;br /&gt;
**[https://www.iso.org/standard/29581.html ISO/IEC 15948:2004]&lt;br /&gt;
&lt;br /&gt;
=== Video ===&lt;br /&gt;
see:  http://de.wikipedia.org/wiki/Digital_Video&lt;br /&gt;
* &#039;&#039;&#039;MPG Container&#039;&#039;&#039;&lt;br /&gt;
** MP3&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MP3&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=45524 ISO/IEC 14496-4:2004]?&lt;br /&gt;
** MPEG2 (for PAL)&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MPEG_2&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=37680 ISO/IEC 13818-11:2004]&lt;br /&gt;
**** Software [http://www.videolan.org/ VLC media player]&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;MP4 Container&#039;&#039;&#039;&lt;br /&gt;
** AAC&lt;br /&gt;
*** http://en.wikipedia.org/wiki/Advanced_Audio_Coding&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=43345 ISO/IEC 13818-7:2006]&lt;br /&gt;
** MPEG-4 (for HDTV)&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MPEG-4&lt;br /&gt;
*** [http://www.iso.org/iso/search.htm?qt=14496&amp;amp;published=on&amp;amp;active_tab=standards ISO/IEC 14496]&lt;br /&gt;
&lt;br /&gt;
=== Audio ===&lt;br /&gt;
* MP3&lt;br /&gt;
* WAVE (WAV)&lt;br /&gt;
** description http://en.wikipedia.org/wiki/WAV&lt;br /&gt;
** example {{doi|10.1594/PANGAEA.339110}}&lt;br /&gt;
&lt;br /&gt;
=== Seismic data ===&lt;br /&gt;
* segy&lt;br /&gt;
&lt;br /&gt;
=== ADCP ===&lt;br /&gt;
* proprietary binary &#039;&#039;ping&#039;&#039;-format, archived on hs, linked to metadescription in PANGAEA&lt;br /&gt;
** ping: http://currents.soest.hawaii.edu/docs/doc/codas_doc/CODAS_pingdemo.html&lt;br /&gt;
* final processed data in UTF-8, archived in &#039;&#039;data numeric&#039;&#039; of PANGAEA (file size 100-500 MB!)&lt;br /&gt;
** Example {{doi|10.1594/PANGAEA.701279}}&lt;br /&gt;
&lt;br /&gt;
=== Large array-oriented, scientific data (no models - please see our corresponding Wiki article &amp;quot;[[Model data and PANGAEA]]&amp;quot;!) ===&lt;br /&gt;
* Network Common Data Form (NetCDF), &lt;br /&gt;
** description http://en.wikipedia.org/wiki/NetCDF&lt;br /&gt;
** Unidata/NSF http://www.unidata.ucar.edu/software/netcdf/&lt;br /&gt;
** example https://doi.org/10.1594/PANGAEA.940846&lt;br /&gt;
** viewer &#039;&#039;panoply&#039;&#039; http://www.giss.nasa.gov/tools/panoply/&lt;br /&gt;
&lt;br /&gt;
=== Compression ===&lt;br /&gt;
* zip is ISO-standard and supported - *.tar, *.rar and *.7z are NOT standard and (at least the latter two) not supported by PANGAEA&lt;br /&gt;
&lt;br /&gt;
=== Proprietary formats ===&lt;br /&gt;
If proprietary data formats cannot be avoided, please include a reference to open source software, preferably with a DOI, that can be used to open such files (e.g. at GitHub, pypi.org).&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Format&amp;diff=16756</id>
		<title>Format</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Format&amp;diff=16756"/>
		<updated>2026-03-31T15:06:08Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Tabular data */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;PANGAEA accepts and publishes a wide spectrum of data formats. Thereby, we classify these file formats into three categories – [[Format#Documentation|Documentation formats]], and [[Format#Tabular data|Tabular]] and [[Format#Binary data|Binary data]] formats – based on how these formats are treated and processed during our editorial work, and how they are represented in our data publications. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Important:&#039;&#039;&#039; If you consider submitting data and supplementary documentation to PANGAEA, please make sure to provide open (ideally non-proprietary) formats widely accepted and endorsed in your scientific communities in order to support the accessibility and (re-)usability on long time-scales and by openly available tools.  &lt;br /&gt;
&lt;br /&gt;
== Documentation ==&lt;br /&gt;
As documentation we consider all files provided in typical text data formats meant to supplement or further describe data submissions (whether in tabular or binary formats), such as processing reports, instrument calibration protocols, standard operating procedures.&lt;br /&gt;
&lt;br /&gt;
Accepted formats for documentation are:&lt;br /&gt;
* &#039;&#039;&#039;PDF&#039;&#039;&#039;/A (ISO19005) - http://en.wikipedia.org/wiki/PDF/A&lt;br /&gt;
* &#039;&#039;&#039;RTF&#039;&#039;&#039; or &#039;&#039;&#039;ODF&#039;&#039;&#039; (ISO26300) - http://en.wikipedia.org/wiki/OpenDocument&lt;br /&gt;
* Microsoft &#039;&#039;&#039;Office files&#039;&#039;&#039; such as &#039;&#039;&#039;*.docx&#039;&#039;&#039; for MS Word, and &#039;&#039;&#039;*.xlsx&#039;&#039;&#039; for Excel spreadsheet documents – compliant to standard OOXML (ISO/IEC 29500:2008 since Office 2013 – https://de.wikipedia.org/wiki/Microsoft_Office),&lt;br /&gt;
* or (our favourite and recommended) &#039;&#039;&#039;plain UTF-8 encoded text files&#039;&#039;&#039; - https://en.wikipedia.org/wiki/UTF-8&lt;br /&gt;
&lt;br /&gt;
== Tabular data ==&lt;br /&gt;
PANGAEA is specifically good at (or, to put it differently, is able to make the most of) tabular field observation data. Owing to the many efforts we put in harmonizing parameter/variable names, methods, dimensions and units during our editorial processing, and due to the fact that this kind of data is stored in a relational database (PostgreSQL) at PANGAEA, our users are able to easily compile specific parameters/variables of interest from many similar studies (i.e. related individual publications) into meta-studies targeting new research questions and contexts using our [[PANGAEA search#Data warehouse|Data Warehouse]]. Or apply similar functionality with the help of the Python module pangaeapy or the R pendant pangaear (see our [https://www.pangaea.de/tools/ Tools site] for details). &lt;br /&gt;
&lt;br /&gt;
The &#039;&#039;&#039;preferred&#039;&#039;&#039; &#039;&#039;&#039;formats&#039;&#039;&#039; for data tables are:  &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;TAB-delimited TEXT-files&#039;&#039;&#039; (UTF-8 encoded), or &lt;br /&gt;
* &#039;&#039;&#039;(open) spreadsheet file formats&#039;&#039;&#039; (MS Excel .xlsx, OpenOffice &amp;amp; LibreOffice Calc .ods etc)&#039;&#039;&#039;.&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Please note that data tables are &#039;&#039;&#039;not&#039;&#039;&#039; accepted as encapsulated objects (e.g., in .mat files).&lt;br /&gt;
&lt;br /&gt;
Example for a tabular dataset: https://doi.org/10.1594/PANGAEA.934148&lt;br /&gt;
&lt;br /&gt;
== Binary data ==&lt;br /&gt;
Binary objects and documentations are usually stored in a combination of hard-drive arrays (for immediate and performant access) and tape archives. File formats should follow ISO standards or at least &#039;&#039;de facto&#039;&#039; standards. Online preview is available for raster graphics and videos (e.g. .tif, .png, .jpeg, .mp4).&lt;br /&gt;
&lt;br /&gt;
Example: https://doi.org/10.1594/PANGAEA.936185&lt;br /&gt;
&lt;br /&gt;
=== Images ===&lt;br /&gt;
* tiff&lt;br /&gt;
** http://en.wikipedia.org/wiki/Tagged_Image_File_Format&lt;br /&gt;
** [http://www.iso.org/iso/iso_catalogue/catalogue_ics/catalogue_detail_ics.htm?csnumber=2181 ISO12639:1998]&lt;br /&gt;
* jpeg&lt;br /&gt;
** http://en.wikipedia.org/wiki/Jpeg&lt;br /&gt;
** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=18902 ISO10918-1:1994]&lt;br /&gt;
*png&lt;br /&gt;
**https://en.wikipedia.org/wiki/Portable_Network_Graphics&lt;br /&gt;
**[https://www.iso.org/standard/29581.html ISO/IEC 15948:2004]&lt;br /&gt;
&lt;br /&gt;
=== Video ===&lt;br /&gt;
see:  http://de.wikipedia.org/wiki/Digital_Video&lt;br /&gt;
* &#039;&#039;&#039;MPG Container&#039;&#039;&#039;&lt;br /&gt;
** MP3&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MP3&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=45524 ISO/IEC 14496-4:2004]?&lt;br /&gt;
** MPEG2 (for PAL)&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MPEG_2&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=37680 ISO/IEC 13818-11:2004]&lt;br /&gt;
**** Software [http://www.videolan.org/ VLC media player]&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;MP4 Container&#039;&#039;&#039;&lt;br /&gt;
** AAC&lt;br /&gt;
*** http://en.wikipedia.org/wiki/Advanced_Audio_Coding&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=43345 ISO/IEC 13818-7:2006]&lt;br /&gt;
** MPEG-4 (for HDTV)&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MPEG-4&lt;br /&gt;
*** [http://www.iso.org/iso/search.htm?qt=14496&amp;amp;published=on&amp;amp;active_tab=standards ISO/IEC 14496]&lt;br /&gt;
&lt;br /&gt;
=== Audio ===&lt;br /&gt;
* MP3&lt;br /&gt;
* WAVE (WAV)&lt;br /&gt;
** description http://en.wikipedia.org/wiki/WAV&lt;br /&gt;
** example {{doi|10.1594/PANGAEA.339110}}&lt;br /&gt;
&lt;br /&gt;
=== Seismic data ===&lt;br /&gt;
* segy&lt;br /&gt;
&lt;br /&gt;
=== ADCP ===&lt;br /&gt;
* proprietary binary &#039;&#039;ping&#039;&#039;-format, archived on hs, linked to metadescription in PANGAEA&lt;br /&gt;
** ping: http://currents.soest.hawaii.edu/docs/doc/codas_doc/CODAS_pingdemo.html&lt;br /&gt;
* final processed data in UTF-8, archived in &#039;&#039;data numeric&#039;&#039; of PANGAEA (file size 100-500 MB!)&lt;br /&gt;
** Example {{doi|10.1594/PANGAEA.701279}}&lt;br /&gt;
&lt;br /&gt;
=== Large array-oriented, scientific data (no models - please see our corresponding Wiki article &amp;quot;[[Model data and PANGAEA]]&amp;quot;!) ===&lt;br /&gt;
* Network Common Data Form (NetCDF), &lt;br /&gt;
** description http://en.wikipedia.org/wiki/NetCDF&lt;br /&gt;
** Unidata/NSF http://www.unidata.ucar.edu/software/netcdf/&lt;br /&gt;
** example https://doi.org/10.1594/PANGAEA.940846&lt;br /&gt;
** viewer &#039;&#039;panoply&#039;&#039; http://www.giss.nasa.gov/tools/panoply/&lt;br /&gt;
&lt;br /&gt;
=== Compression ===&lt;br /&gt;
* zip is ISO-standard and supported - *.tar, *.rar and *.7z are NOT standard and (at least the latter two) not supported by PANGAEA&lt;br /&gt;
&lt;br /&gt;
=== Proprietary formats ===&lt;br /&gt;
Include a reference, preferably with a DOI, to open source software (e.g. at GitHub, pypi.org) that can be used to open such files.&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Format&amp;diff=16755</id>
		<title>Format</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Format&amp;diff=16755"/>
		<updated>2026-03-31T15:04:11Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: minor changes to wording&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;PANGAEA accepts and publishes a wide spectrum of data formats. Thereby, we classify these file formats into three categories – [[Format#Documentation|Documentation formats]], and [[Format#Tabular data|Tabular]] and [[Format#Binary data|Binary data]] formats – based on how these formats are treated and processed during our editorial work, and how they are represented in our data publications. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Important:&#039;&#039;&#039; If you consider submitting data and supplementary documentation to PANGAEA, please make sure to provide open (ideally non-proprietary) formats widely accepted and endorsed in your scientific communities in order to support the accessibility and (re-)usability on long time-scales and by openly available tools.  &lt;br /&gt;
&lt;br /&gt;
== Documentation ==&lt;br /&gt;
As documentation we consider all files provided in typical text data formats meant to supplement or further describe data submissions (whether in tabular or binary formats), such as processing reports, instrument calibration protocols, standard operating procedures.&lt;br /&gt;
&lt;br /&gt;
Accepted formats for documentation are:&lt;br /&gt;
* &#039;&#039;&#039;PDF&#039;&#039;&#039;/A (ISO19005) - http://en.wikipedia.org/wiki/PDF/A&lt;br /&gt;
* &#039;&#039;&#039;RTF&#039;&#039;&#039; or &#039;&#039;&#039;ODF&#039;&#039;&#039; (ISO26300) - http://en.wikipedia.org/wiki/OpenDocument&lt;br /&gt;
* Microsoft &#039;&#039;&#039;Office files&#039;&#039;&#039; such as &#039;&#039;&#039;*.docx&#039;&#039;&#039; for MS Word, and &#039;&#039;&#039;*.xlsx&#039;&#039;&#039; for Excel spreadsheet documents – compliant to standard OOXML (ISO/IEC 29500:2008 since Office 2013 – https://de.wikipedia.org/wiki/Microsoft_Office),&lt;br /&gt;
* or (our favourite and recommended) &#039;&#039;&#039;plain UTF-8 encoded text files&#039;&#039;&#039; - https://en.wikipedia.org/wiki/UTF-8&lt;br /&gt;
&lt;br /&gt;
== Tabular data ==&lt;br /&gt;
PANGAEA is specifically good at (or, to put it differently, is able to make the most of) tabular field observation data. Owing to the many efforts we put in harmonizing parameter/variable names, methods, dimensions and units during our editorial processing, and due to the fact that this kind of data is stored in a relational database (PostgreSQL) at PANGAEA, our users are able to easily compile specific parameters/variables of interest from many similar studies (i.e. related individual publications) into meta-studies targeting new research questions and contexts using our [[PANGAEA search#Data warehouse|Data Warehouse]]. Or apply similar functionality with the help of the Python module pangaeapy or the R pendant pangaear (see our [https://www.pangaea.de/tools/ Tools site] for details). &lt;br /&gt;
&lt;br /&gt;
The preferred &#039;&#039;&#039;Format&#039;&#039;&#039; for data tables is  &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;TAB-delimited TEXT-files&#039;&#039;&#039; (UTF-8 encoded), or &lt;br /&gt;
* &#039;&#039;&#039;(open) spreadsheet file formats&#039;&#039;&#039; (MS Excel .xlsx, OpenOffice &amp;amp; LibreOffice Calc .ods etc)&#039;&#039;&#039;.&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Please note that tables are &#039;&#039;&#039;not&#039;&#039;&#039; accepted as encapsulated objects (e.g., in .mat files).&lt;br /&gt;
&lt;br /&gt;
Example for a tabular dataset: https://doi.org/10.1594/PANGAEA.934148&lt;br /&gt;
&lt;br /&gt;
== Binary data ==&lt;br /&gt;
Binary objects and documentations are usually stored in a combination of hard-drive arrays (for immediate and performant access) and tape archives. File formats should follow ISO standards or at least &#039;&#039;de facto&#039;&#039; standards. Online preview is available for raster graphics and videos (e.g. .tif, .png, .jpeg, .mp4).&lt;br /&gt;
&lt;br /&gt;
Example: https://doi.org/10.1594/PANGAEA.936185&lt;br /&gt;
&lt;br /&gt;
=== Images ===&lt;br /&gt;
* tiff&lt;br /&gt;
** http://en.wikipedia.org/wiki/Tagged_Image_File_Format&lt;br /&gt;
** [http://www.iso.org/iso/iso_catalogue/catalogue_ics/catalogue_detail_ics.htm?csnumber=2181 ISO12639:1998]&lt;br /&gt;
* jpeg&lt;br /&gt;
** http://en.wikipedia.org/wiki/Jpeg&lt;br /&gt;
** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=18902 ISO10918-1:1994]&lt;br /&gt;
*png&lt;br /&gt;
**https://en.wikipedia.org/wiki/Portable_Network_Graphics&lt;br /&gt;
**[https://www.iso.org/standard/29581.html ISO/IEC 15948:2004]&lt;br /&gt;
&lt;br /&gt;
=== Video ===&lt;br /&gt;
see:  http://de.wikipedia.org/wiki/Digital_Video&lt;br /&gt;
* &#039;&#039;&#039;MPG Container&#039;&#039;&#039;&lt;br /&gt;
** MP3&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MP3&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=45524 ISO/IEC 14496-4:2004]?&lt;br /&gt;
** MPEG2 (for PAL)&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MPEG_2&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=37680 ISO/IEC 13818-11:2004]&lt;br /&gt;
**** Software [http://www.videolan.org/ VLC media player]&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;MP4 Container&#039;&#039;&#039;&lt;br /&gt;
** AAC&lt;br /&gt;
*** http://en.wikipedia.org/wiki/Advanced_Audio_Coding&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=43345 ISO/IEC 13818-7:2006]&lt;br /&gt;
** MPEG-4 (for HDTV)&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MPEG-4&lt;br /&gt;
*** [http://www.iso.org/iso/search.htm?qt=14496&amp;amp;published=on&amp;amp;active_tab=standards ISO/IEC 14496]&lt;br /&gt;
&lt;br /&gt;
=== Audio ===&lt;br /&gt;
* MP3&lt;br /&gt;
* WAVE (WAV)&lt;br /&gt;
** description http://en.wikipedia.org/wiki/WAV&lt;br /&gt;
** example {{doi|10.1594/PANGAEA.339110}}&lt;br /&gt;
&lt;br /&gt;
=== Seismic data ===&lt;br /&gt;
* segy&lt;br /&gt;
&lt;br /&gt;
=== ADCP ===&lt;br /&gt;
* proprietary binary &#039;&#039;ping&#039;&#039;-format, archived on hs, linked to metadescription in PANGAEA&lt;br /&gt;
** ping: http://currents.soest.hawaii.edu/docs/doc/codas_doc/CODAS_pingdemo.html&lt;br /&gt;
* final processed data in UTF-8, archived in &#039;&#039;data numeric&#039;&#039; of PANGAEA (file size 100-500 MB!)&lt;br /&gt;
** Example {{doi|10.1594/PANGAEA.701279}}&lt;br /&gt;
&lt;br /&gt;
=== Large array-oriented, scientific data (no models - please see our corresponding Wiki article &amp;quot;[[Model data and PANGAEA]]&amp;quot;!) ===&lt;br /&gt;
* Network Common Data Form (NetCDF), &lt;br /&gt;
** description http://en.wikipedia.org/wiki/NetCDF&lt;br /&gt;
** Unidata/NSF http://www.unidata.ucar.edu/software/netcdf/&lt;br /&gt;
** example https://doi.org/10.1594/PANGAEA.940846&lt;br /&gt;
** viewer &#039;&#039;panoply&#039;&#039; http://www.giss.nasa.gov/tools/panoply/&lt;br /&gt;
&lt;br /&gt;
=== Compression ===&lt;br /&gt;
* zip is ISO-standard and supported - *.tar, *.rar and *.7z are NOT standard and (at least the latter two) not supported by PANGAEA&lt;br /&gt;
&lt;br /&gt;
=== Proprietary formats ===&lt;br /&gt;
Include a reference, preferably with a DOI, to open source software (e.g. at GitHub, pypi.org) that can be used to open such files.&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Format&amp;diff=16754</id>
		<title>Format</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Format&amp;diff=16754"/>
		<updated>2026-03-31T14:55:59Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: /* Tabular data */ Den Bandwurmsatz rund ums DataWareHouse gekürzt&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;PANGAEA accepts and publishes a wide spectrum of data formats. Thereby, we classify these file formats into three categories – [[Format#Documentation|Documentation formats]], and [[Format#Tabular data|Tabular]] and [[Format#Binary data|Binary data]] formats – based on how these formats are treated and processed during the editorial processing and how they are represented in our data publications. &lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;Important:&#039;&#039;&#039; If you consider submitting data and supplementary documentation to PANGAEA, please make sure to provide open (ideally non-proprietary) formats widely accepted and endorsed in your scientific communities in order to support the accessibility and (re-)usability on long time-scales and by openly available tools.  &lt;br /&gt;
&lt;br /&gt;
== Documentation ==&lt;br /&gt;
As documentation we consider all files provided in typical text data formats meant to supplement or further describe data submissions (whether in tabular or binary formats), such as processing reports, instrument calibration protocols, standard operating procedures.&lt;br /&gt;
&lt;br /&gt;
Accepted formats for documentation are:&lt;br /&gt;
* &#039;&#039;&#039;PDF&#039;&#039;&#039;/A (ISO19005) - http://en.wikipedia.org/wiki/PDF/A&lt;br /&gt;
* &#039;&#039;&#039;RTF&#039;&#039;&#039; or &#039;&#039;&#039;ODF&#039;&#039;&#039; (ISO26300) - http://en.wikipedia.org/wiki/OpenDocument&lt;br /&gt;
* MS Office files - standard OOXML (ISO/IEC 29500:2008 since Office 2013 - https://de.wikipedia.org/wiki/Microsoft_Office), e.g. &#039;&#039;&#039;*.docx&#039;&#039;&#039; for MS Office document and *.xlsx for Excel spreadsheet files&lt;br /&gt;
* or (our favourite and recommended) &#039;&#039;&#039;plain UTF-8 encoded text files&#039;&#039;&#039; - https://en.wikipedia.org/wiki/UTF-8&lt;br /&gt;
&lt;br /&gt;
== Tabular data ==&lt;br /&gt;
PANGAEA is specifically good at (or, to put it differently, is able to make the most of) tabular field observation data. Owing to the many efforts we put in harmonizing parameter/variable names, methods, dimensions and units during our editorial processing, and due to the fact that this kind of data is stored in a relational database (PostgreSQL) at PANGAEA, our users are able to easily compile specific parameters/variables of interest from many similar studies (i.e. related individual publications) into meta-studies targeting new research questions and contexts using our [[PANGAEA search#Data warehouse|Data Warehouse]]. Or apply similar functionality with the help of the Python module pangaeapy or the R pendant pangaear (see our [https://www.pangaea.de/tools/ Tools site] for details). &lt;br /&gt;
&lt;br /&gt;
The preferred &#039;&#039;&#039;Format&#039;&#039;&#039; for data tables is  &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;TAB-delimited TEXT-files&#039;&#039;&#039; (UTF-8 encoded), or &lt;br /&gt;
* &#039;&#039;&#039;(open) spreadsheet file formats&#039;&#039;&#039; (MS Excel .xlsx, OpenOffice &amp;amp; LibreOffice Calc .ods etc)&#039;&#039;&#039;.&#039;&#039;&#039; &lt;br /&gt;
&lt;br /&gt;
Please note that tables are &#039;&#039;&#039;not&#039;&#039;&#039; accepted as encapsulated objects (e.g., in .mat files).&lt;br /&gt;
&lt;br /&gt;
Example for a tabular dataset: https://doi.org/10.1594/PANGAEA.934148&lt;br /&gt;
&lt;br /&gt;
== Binary data ==&lt;br /&gt;
Binary objects and documentations are usually stored in a combination of hard-drive arrays (for immediate and performant access) and tape archives. File formats should follow ISO standards or at least &#039;&#039;de facto&#039;&#039; standards. Online preview is available for raster graphics and videos (e.g. .tif, .png, .jpeg, .mp4).&lt;br /&gt;
&lt;br /&gt;
Example: https://doi.org/10.1594/PANGAEA.936185&lt;br /&gt;
&lt;br /&gt;
=== Images ===&lt;br /&gt;
* tiff&lt;br /&gt;
** http://en.wikipedia.org/wiki/Tagged_Image_File_Format&lt;br /&gt;
** [http://www.iso.org/iso/iso_catalogue/catalogue_ics/catalogue_detail_ics.htm?csnumber=2181 ISO12639:1998]&lt;br /&gt;
* jpeg&lt;br /&gt;
** http://en.wikipedia.org/wiki/Jpeg&lt;br /&gt;
** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=18902 ISO10918-1:1994]&lt;br /&gt;
*png&lt;br /&gt;
**https://en.wikipedia.org/wiki/Portable_Network_Graphics&lt;br /&gt;
**[https://www.iso.org/standard/29581.html ISO/IEC 15948:2004]&lt;br /&gt;
&lt;br /&gt;
=== Video ===&lt;br /&gt;
see:  http://de.wikipedia.org/wiki/Digital_Video&lt;br /&gt;
* &#039;&#039;&#039;MPG Container&#039;&#039;&#039;&lt;br /&gt;
** MP3&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MP3&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=45524 ISO/IEC 14496-4:2004]?&lt;br /&gt;
** MPEG2 (for PAL)&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MPEG_2&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=37680 ISO/IEC 13818-11:2004]&lt;br /&gt;
**** Software [http://www.videolan.org/ VLC media player]&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;MP4 Container&#039;&#039;&#039;&lt;br /&gt;
** AAC&lt;br /&gt;
*** http://en.wikipedia.org/wiki/Advanced_Audio_Coding&lt;br /&gt;
*** [http://www.iso.org/iso/iso_catalogue/catalogue_tc/catalogue_detail.htm?csnumber=43345 ISO/IEC 13818-7:2006]&lt;br /&gt;
** MPEG-4 (for HDTV)&lt;br /&gt;
*** http://en.wikipedia.org/wiki/MPEG-4&lt;br /&gt;
*** [http://www.iso.org/iso/search.htm?qt=14496&amp;amp;published=on&amp;amp;active_tab=standards ISO/IEC 14496]&lt;br /&gt;
&lt;br /&gt;
=== Audio ===&lt;br /&gt;
* MP3&lt;br /&gt;
* WAVE (WAV)&lt;br /&gt;
** description http://en.wikipedia.org/wiki/WAV&lt;br /&gt;
** example {{doi|10.1594/PANGAEA.339110}}&lt;br /&gt;
&lt;br /&gt;
=== Seismic data ===&lt;br /&gt;
* segy&lt;br /&gt;
&lt;br /&gt;
=== ADCP ===&lt;br /&gt;
* proprietary binary &#039;&#039;ping&#039;&#039;-format, archived on hs, linked to metadescription in PANGAEA&lt;br /&gt;
** ping: http://currents.soest.hawaii.edu/docs/doc/codas_doc/CODAS_pingdemo.html&lt;br /&gt;
* final processed data in UTF-8, archived in &#039;&#039;data numeric&#039;&#039; of PANGAEA (file size 100-500 MB!)&lt;br /&gt;
** Example {{doi|10.1594/PANGAEA.701279}}&lt;br /&gt;
&lt;br /&gt;
=== Large array-oriented, scientific data (no models - please see our corresponding Wiki article &amp;quot;[[Model data and PANGAEA]]&amp;quot;!) ===&lt;br /&gt;
* Network Common Data Form (NetCDF), &lt;br /&gt;
** description http://en.wikipedia.org/wiki/NetCDF&lt;br /&gt;
** Unidata/NSF http://www.unidata.ucar.edu/software/netcdf/&lt;br /&gt;
** example https://doi.org/10.1594/PANGAEA.940846&lt;br /&gt;
** viewer &#039;&#039;panoply&#039;&#039; http://www.giss.nasa.gov/tools/panoply/&lt;br /&gt;
&lt;br /&gt;
=== Compression ===&lt;br /&gt;
* zip is ISO-standard and supported - *.tar, *.rar and *.7z are NOT standard and (at least the latter two) not supported by PANGAEA&lt;br /&gt;
&lt;br /&gt;
=== Proprietary formats ===&lt;br /&gt;
Include a reference, preferably with a DOI, to open source software (e.g. at GitHub, pypi.org) that can be used to open such files.&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16715</id>
		<title>Technology</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Technology&amp;diff=16715"/>
		<updated>2026-03-26T17:10:33Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Edited label&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;{{construction}}&lt;br /&gt;
&lt;br /&gt;
PANGAEA is built on a three-tiered client/server architecture that has evolved continuously since the system&#039;s founding in the early 1990s. The architecture separates backend storage and database systems, middleware processing and transformation components, and frontend delivery and user interfaces into distinct layers that are individually maintainable and replaceable. This design philosophy has enabled PANGAEA to migrate core components — including its primary database engine, its editorial system, and its search infrastructure — without disruption to archived data or to users, and underpins the system&#039;s long-term stability as a trustworthy data repository.&lt;br /&gt;
&lt;br /&gt;
A detailed description of the PANGAEA information system and its workflows is provided in Felden et al. (2023) and Diepenbroek et al. (2017); this article provides an up-to-date overview oriented toward the general structure and key components.&lt;br /&gt;
&lt;br /&gt;
== System Architecture ==&lt;br /&gt;
&lt;br /&gt;
The technical architecture of PANGAEA follows a three-tiered model comprising a backend, a middleware layer, and a frontend. All hard- and software services are hosted and operated by the data and computing center of the Alfred Wegener Institute, Helmholtz Centre for Polar and Marine Research (AWI) in Bremerhaven. Most backend and middleware systems, as well as all frontend web servers and search engines, run on virtual machines (VMware) operated with Ubuntu Linux. Virtualization provides sufficient capacity and performance while enabling high availability and transparent hardware renewal on a typical cycle of three to four years.&lt;br /&gt;
[[File:PANGAEA overview architecture.png|alt=Schema illustrating the system architecture according to Felden et al., 2023|thumb|Schema illustrating the system architecture according to Felden et al., 2023|none|600x600px]]&lt;br /&gt;
&lt;br /&gt;
PANGAEA currently operates nine virtual machines at the AWI computing center, using a total of 53 CPUs, 162 GB RAM, and 28 TB of disk space. The relational database holdings currently occupy approximately 130 TB, with approximately 1 PB of data stored on tape.&lt;br /&gt;
&lt;br /&gt;
== Backend: Storage and Databases ==&lt;br /&gt;
&lt;br /&gt;
=== Relational Database ===&lt;br /&gt;
The primary store for all structured data and metadata in PANGAEA is a &#039;&#039;&#039;PostgreSQL&#039;&#039;&#039; relational database management system (RDBMS). The data model is highly normalized, reflecting the full observational context of each measurement: events (when and where data were collected), campaigns (cruises or field expeditions), methods and devices, parameters, references, and institutional provenance information are all stored in defined relational structures. This normalized model allows data descriptions to be compiled dynamically and serialized into a wide range of output formats on demand, without modifying the underlying archived data.&lt;br /&gt;
&lt;br /&gt;
Database integrity is continuously maintained through &#039;&#039;&#039;PostgreSQL streaming replication&#039;&#039;&#039; to a dedicated backup system, which enables point-in-time recovery to any moment prior to a failure event.&lt;br /&gt;
&lt;br /&gt;
=== Data Warehouse ===&lt;br /&gt;
Fast access to large, aggregated compilations of data across the full PANGAEA holdings is provided by a &#039;&#039;&#039;Clickhouse&#039;&#039;&#039; data warehouse, which mirrors the numerical data inventory from the relational database. The data warehouse supports spatially and chronologically constrained queries at the parameter level across potentially hundreds of thousands of individual datasets, and is accessible both through the PANGAEA website and programmatically via a REST API. All data warehouse exports include the DOI for each contributing dataset, ensuring that provenance and citation remain intact in compiled data products.&lt;br /&gt;
&lt;br /&gt;
=== Archival Storage ===&lt;br /&gt;
For high-volume and binary data — including geophysical datasets, images, video, and community-specific formats such as NetCDF — data are stored in consistent formats on hard-disk arrays and robotic tape archives. The central archival storage system consists of two &#039;&#039;&#039;SpectraLogic TFinity ExaScale&#039;&#039;&#039; robotic tape libraries housed in separate buildings at AWI, with a combined capacity of up to 60 PB. Backup operations use high-capacity LTO-tape drives within this environment.&lt;br /&gt;
&lt;br /&gt;
All data are stored redundantly using &#039;&#039;&#039;erasure coding&#039;&#039;&#039; across disk and tape. Data on disk is replicated to tape nightly and saved to snapshots retained for six months. Tape copies are replicated to a physically separate building within two hours of creation; decommissioned tapes are retained for one year before reuse. Virtual machine working data is captured in nightly machine snapshots. Write caches are battery-backed to ensure data integrity at the point of write.&lt;br /&gt;
&lt;br /&gt;
=== Off-Site Replica ===&lt;br /&gt;
Since 2025, PANGAEA operates an &#039;&#039;&#039;off-site replica&#039;&#039;&#039; of the relational database and web frontend at MARUM/University of Bremen, hosted in the Green IT Housing Center of Bremen University. This facility provides geographic and infrastructural separation from the primary AWI systems, enabling recovery of data delivery services within 24 hours following a technical breakdown or cyberattack. The replica is kept current through snapshot-based replication several times per day and currently covers all dataset metadata — including individual landing pages and all harvesting endpoints — as well as full representations of tabular data publications. Extension of the replica to include binary data files is planned as a further development of this facility.&lt;br /&gt;
&lt;br /&gt;
The off-site installation operates in a fully isolated environment with Layer 2 network separation. Access is restricted to VPN connections from a single designated gateway host at AWI, with access tokens and keys limited to the corresponding gateway and PANGAEA DevOps staff.&lt;br /&gt;
== Middleware: Processing and Transformation ==&lt;br /&gt;
The middleware layer manages the flow of data and metadata between the backend storage systems and the frontend interfaces. Its core functions are marshaling, indexing, transformation, and dissemination.&lt;br /&gt;
&lt;br /&gt;
Metadata are dynamically marshaled from the relational database to a PANGAEA-specific internal XML format, from which they are transformed via &#039;&#039;&#039;XSLT and XML-to-JSON pipelines&#039;&#039;&#039; into a range of content standards for delivery to users and harvesting services. Currently supported output standards include JSON-LD according to schema.org, DataCite XML, Dublin Core XML, ISO 19115/ISO 19139, DIF (Directory Interchange Format), and Darwin Core. Dissemination occurs via OAI-PMH, HTTP content negotiation based on HTTP standards and technical FAIR recommendations, and other protocols.&lt;br /&gt;
&lt;br /&gt;
The marshaled metadata are stored and indexed in &#039;&#039;&#039;Elasticsearch&#039;&#039;&#039;, which serves as the primary index for all public search and metadata access interfaces. This architecture separates the authoritative relational record from the search-optimized representation, allowing the search index to be rebuilt or updated without modifying archived data.&lt;br /&gt;
&lt;br /&gt;
The flexible metadata framework &#039;&#039;&#039;PanFMP&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://www.panfmp.org/&amp;lt;/nowiki&amp;gt;) underpins the mapping between the PANGAEA internal schema and the various external output standards, ensuring that new or updated standards can be accommodated without structural changes to the underlying data.&lt;br /&gt;
&lt;br /&gt;
Data submissions, user requests, and bug reports are managed through a &#039;&#039;&#039;JIRA&#039;&#039;&#039; (Atlassian) issue tracking system, which serves as the primary communication channel between data providers and the PANGAEA editorial team throughout the submission and publication process.&lt;br /&gt;
== Frontend: Editorial System ==&lt;br /&gt;
The PANGAEA editorial system is the primary tool through which data editors review, curate, harmonize, and publish submitted data and metadata. It is a &#039;&#039;&#039;web-based client/server application&#039;&#039;&#039; developed entirely in-house, operating directly on the PostgreSQL databases.&lt;br /&gt;
&lt;br /&gt;
The backend of the editorial system is built on &#039;&#039;&#039;Java 17&#039;&#039;&#039; using the &#039;&#039;&#039;Dropwizard&#039;&#039;&#039; framework, exposing a REST API. The frontend is implemented in &#039;&#039;&#039;React&#039;&#039;&#039; with the &#039;&#039;&#039;Ant Design&#039;&#039;&#039; component library. Source code for both components, together with shared libraries and test suites, is maintained in an institutional &#039;&#039;&#039;GitLab&#039;&#039;&#039; instance hosted at AWI, structured as an &#039;&#039;&#039;Nx monorepo&#039;&#039;&#039;. The repository encompasses automated unit tests and &#039;&#039;&#039;Cypress&#039;&#039;&#039; end-to-end tests. All new versions are deployed exclusively through a &#039;&#039;&#039;GitLab CI/CD pipeline&#039;&#039;&#039;, with releases gated on successful completion of the full automated test suite.&lt;br /&gt;
&lt;br /&gt;
In production, the editorial system runs on Ubuntu virtual machines hosted on VMware infrastructure. &#039;&#039;&#039;Apache&#039;&#039;&#039; serves the React frontend and acts as a reverse proxy to the backend services. Four parallel instances are operated simultaneously, providing dedicated test and production environments for two editorial system versions at any given time. This setup supports controlled feature validation alongside uninterrupted service operation. Infrastructure components — including VMware, file systems, monitoring, GitLab, and the database platform — are managed by AWI, while the PANGAEA group is responsible for the Apache web server, Java services, and React frontend.&lt;br /&gt;
&lt;br /&gt;
The relational database model underlying the editorial system enforces structural and semantic consistency between submitted metadata and the existing data inventory. The editorial system also integrates a &#039;&#039;&#039;Terminology Catalogue (TC)&#039;&#039;&#039;, which manages controlled vocabularies and ontologies used to harmonize data and metadata during ingest. Supported terminologies include WoRMS, ITIS, ChEBI, EnvO, PATO, QUDT, and the NERC vocabulary server, among others. A detailed description of the Terminology Catalogue and its role in data archiving and publication is given in Diepenbroek et al. (2017).&lt;br /&gt;
== Frontend: Public Web Interface and Search ==&lt;br /&gt;
The public PANGAEA website and search interface is served through Apache web servers in the frontend tier, backed by the Elasticsearch index for fast metadata retrieval.&lt;br /&gt;
&lt;br /&gt;
=== Search ===&lt;br /&gt;
The PANGAEA search engine supports full-text and faceted search across all published metadata. Faceted navigation is enabled by semantic metadata enrichment performed during the marshaling process: terminology-based annotations are added to metadata records, allowing consistent filtering by topic, device type, geographic region, and other dimensions. Search documentation is available in the Wiki at &amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/wiki/PANGAEA_search&amp;lt;/nowiki&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
=== Map Search ===&lt;br /&gt;
Geographic search and visualization are implemented using &#039;&#039;&#039;leaflet.js&#039;&#039;&#039;, serving map tiles from four configurable sources: the AWI basemap, OpenStreetMap, Google (hybrid), and ESRI.&lt;br /&gt;
&lt;br /&gt;
=== Dataset Landing Pages ===&lt;br /&gt;
Each published dataset is represented by a landing page resolved through its DOI. Landing pages present the full dataset metadata in human-readable form and are enriched with structured &#039;&#039;&#039;schema.org&#039;&#039;&#039; markup in JSON-LD, ensuring machine-actionability and compatibility with generic web search engines and data registries. The schema.org metadata is also accessible via HTTP content negotiation.&lt;br /&gt;
&lt;br /&gt;
=== Programmatic Access ===&lt;br /&gt;
PANGAEA offers programmatic access to data and metadata through a range of web services (SOAP and REST). The OAI-PMH endpoint supports metadata harvesting in all supported standards. Client libraries for &#039;&#039;&#039;Python&#039;&#039;&#039; (pangaeapy, developed by PANGAEA) and &#039;&#039;&#039;R&#039;&#039;&#039; (pangaear, developed by the community) allow researchers to load and transform PANGAEA data directly into native data structures for analysis in environments such as Jupyter notebooks.&lt;br /&gt;
== Monitoring ==&lt;br /&gt;
Service health across the PANGAEA infrastructure is monitored through the &#039;&#039;&#039;AWI Grafana/Telegraf&#039;&#039;&#039; stack, covering service availability, application logs, and resource saturation for all production systems. This is supplemented by external availability checks via &#039;&#039;&#039;UptimeRobot&#039;&#039;&#039;, which provides independent verification of public-facing service endpoints from outside the AWI network. All external links in PANGAEA metadata records — references to related literature, other dataset versions, and external resources — are automatically checked for broken (HTTP 404) or permanently redirected (HTTP 301) responses on a weekly basis.&lt;br /&gt;
== Security ==&lt;br /&gt;
Backend and middleware systems are protected behind a firewall; frontend systems operate in a demilitarized zone (DMZ) accessible from outside with restricted but still firewalled access. Frontend systems have no write access and only limited read access to the backend database and tape archives. Public access from the website and REST APIs is served from data replicas hosted in Elasticsearch or through read-only remote filesystem access. Data curators access production systems via virtual private networks (VPN), which enforce a basic check that client operating systems are up to date.&lt;br /&gt;
&lt;br /&gt;
Physical access to the AWI computing center is managed through an electronic access control system for all relevant entrances; key distribution is documented, and guest policies are formally established. The AWI maintains uninterruptible power supplies (UPS) capable of sustaining all PANGAEA-relevant hardware for up to 60 minutes, backed by a diesel-powered emergency generator providing a further 23 hours of operation. Fire alerts are automatically forwarded to the fire department with a contractual response time under 15 minutes.&lt;br /&gt;
&lt;br /&gt;
Security of the technical infrastructure is further maintained through the use of asymmetric key infrastructures, mandatory minimum-length passwords for all user classes, short-cycle security patching for all hardware and software components, professional monitoring tools for hardware, firewall, software, services, performance, and attacks, and regular security training for all technical and non-technical staff. Security risks are assessed on an ongoing basis by AWI&#039;s institutional IT security team, which coordinates incident response and monitors threat landscapes relevant to research data infrastructure. PANGAEA technical staff participates in regular security reviews.&lt;br /&gt;
&lt;br /&gt;
Access controls for restricted or moratorium datasets are enforced at the application layer. Dataset metadata remains publicly accessible in all cases; access to the data itself is restricted to authorized users at the individual level for the duration of the moratorium, typically a maximum of two years from submission.&lt;br /&gt;
&lt;br /&gt;
The off-site replica at MARUM operates in a fully isolated network environment (Layer 2 separation) with access restricted to a firewalled VPN gateway at AWI. The Green IT Housing Center is monitored 24/7 by Bremen University staff, with automated fire alarms, redundant power supply with battery backup, and multilevel physical access control.&lt;br /&gt;
== Software Inventory ==&lt;br /&gt;
The following table summarizes the principal software components of the PANGAEA infrastructure and their development model.&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
!Component&lt;br /&gt;
!Software&lt;br /&gt;
!Development Model&lt;br /&gt;
|-&lt;br /&gt;
|Primary database&lt;br /&gt;
|PostgreSQL&lt;br /&gt;
|Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
|Data warehouse&lt;br /&gt;
|Clickhouse&lt;br /&gt;
|Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
|Search and metadata index&lt;br /&gt;
|Elasticsearch&lt;br /&gt;
|Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
|Editorial system backend&lt;br /&gt;
|Java 17 / Dropwizard&lt;br /&gt;
|In-house development&lt;br /&gt;
|-&lt;br /&gt;
|Editorial system frontend&lt;br /&gt;
|React / Ant Design&lt;br /&gt;
|In-house development (open source framework)&lt;br /&gt;
|-&lt;br /&gt;
|Source control&lt;br /&gt;
|GitLab&lt;br /&gt;
|Open source (community-supported, self-hosted)&lt;br /&gt;
|-&lt;br /&gt;
|CI/CD pipeline&lt;br /&gt;
|GitLab CI/CD&lt;br /&gt;
|Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
|Web server / reverse proxy&lt;br /&gt;
|Apache&lt;br /&gt;
|Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
|Map visualization&lt;br /&gt;
|leaflet.js&lt;br /&gt;
|Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
|Metadata framework&lt;br /&gt;
|PanFMP&lt;br /&gt;
|In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
|Issue tracking&lt;br /&gt;
|JIRA (Atlassian)&lt;br /&gt;
|Commercial&lt;br /&gt;
|-&lt;br /&gt;
|Monitoring stack&lt;br /&gt;
|Grafana / Telegraf&lt;br /&gt;
|Open source (community-supported)&lt;br /&gt;
|-&lt;br /&gt;
|External uptime monitoring&lt;br /&gt;
|UptimeRobot&lt;br /&gt;
|Commercial SaaS&lt;br /&gt;
|-&lt;br /&gt;
|Infrastructure virtualization&lt;br /&gt;
|VMware&lt;br /&gt;
|Commercial&lt;br /&gt;
|-&lt;br /&gt;
|Tape archive hardware&lt;br /&gt;
|SpectraLogic TFinity ExaScale&lt;br /&gt;
|Commercial hardware&lt;br /&gt;
|-&lt;br /&gt;
|Python data client&lt;br /&gt;
|pangaeapy&lt;br /&gt;
|In-house development (open source)&lt;br /&gt;
|-&lt;br /&gt;
|R data client&lt;br /&gt;
|pangaear&lt;br /&gt;
|Community development (open source)&lt;br /&gt;
|}&lt;br /&gt;
== Documentation and Change Management ==&lt;br /&gt;
General documentation of PANGAEA systems and services is maintained in the public &#039;&#039;&#039;PANGAEA Wiki&#039;&#039;&#039; (&amp;lt;nowiki&amp;gt;https://wiki.pangaea.de/&amp;lt;/nowiki&amp;gt;). A separate internal &#039;&#039;&#039;Confluence&#039;&#039;&#039; Wiki, kept operationally isolated from the main PANGAEA infrastructure for disaster-safety purposes, contains detailed documentation of server configurations, installation and maintenance procedures, relationships between system components, VM snapshot procedures, backup routines, and service restart priorities for incident response.&lt;br /&gt;
&lt;br /&gt;
All changes to published data and metadata are recorded in the editorial system&#039;s version history. Substantive revisions to published datasets result in a new dataset version with a new DOI; all prior versions remain accessible and cross-referenced. Minor editorial corrections that do not affect scientific content are applied without creating a new identifier.&lt;br /&gt;
== References ==&lt;br /&gt;
&lt;br /&gt;
* Felden, J., Möller, L., Schindler, U., Huber, R., Schumacher, S., Koppe, R., Diepenbroek, M. &amp;amp; Glöckner, F.O. (2023). PANGAEA — Data Publisher for Earth &amp;amp; Environmental Science. &#039;&#039;Scientific Data&#039;&#039;, 10, 347. &amp;lt;nowiki&amp;gt;https://doi.org/10.1038/s41597-023-02269-x&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
* Diepenbroek, M., Schindler, U., Huber, R., Pesant, S., Stocker, M., Felden, J., Buss, M. &amp;amp; Weinrebe, M. (2017). Terminology supported archiving and publication of environmental science data in PANGAEA. &#039;&#039;Journal of Biotechnology&#039;&#039;, 261, 177–186. &amp;lt;nowiki&amp;gt;https://doi.org/10.1016/j.jbiotec.2017.07.016&amp;lt;/nowiki&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== See Also ==&lt;br /&gt;
&lt;br /&gt;
*[[PANGAEA search]]&lt;br /&gt;
*[[Data submission]]&lt;br /&gt;
*[[Authors Guides]]&lt;br /&gt;
*[[Curation levels]]&lt;br /&gt;
*[[Processing levels]]&lt;br /&gt;
*[[Data Usage Statistics]]&lt;br /&gt;
*[[Format]]&lt;br /&gt;
*[[PANGAEA XML schema]]&lt;br /&gt;
*[[Best practice manuals and templates]]&lt;br /&gt;
*[[Continuity Plan]] — https://www.pangaea.de/about/continuity.php&lt;br /&gt;
*[[Preservation Plan]] — https://www.pangaea.de/about/preservation.php&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Version&amp;diff=16711</id>
		<title>Version</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Version&amp;diff=16711"/>
		<updated>2026-03-26T17:07:05Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Lmoeller moved page Version to Intern:Version&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Intern:Version]]&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Data_Seal_of_Approval&amp;diff=16708</id>
		<title>Data Seal of Approval</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Data_Seal_of_Approval&amp;diff=16708"/>
		<updated>2026-03-26T17:05:55Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Lmoeller moved page Data Seal of Approval to Intern:Data Seal of Approval&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Intern:Data Seal of Approval]]&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=Portal&amp;diff=16703</id>
		<title>Portal</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=Portal&amp;diff=16703"/>
		<updated>2026-03-26T16:41:43Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Lmoeller moved page Portal to Intern:Portal: outdated and content better explained elsewhere&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Intern:Portal]]&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=DOI&amp;diff=16698</id>
		<title>DOI</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=DOI&amp;diff=16698"/>
		<updated>2026-03-26T16:13:02Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: deleted link to 3 sentence wiki Backup article&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Digital Object Identifier (DOI) ==&lt;br /&gt;
&lt;br /&gt;
DOIs provide [[persistent identifier|persistent links]] to scholarly content, helping users get to the authoritative, published version of the content they are searching for, even when the content changes location or ownership. With about 35 Million registered DOI for publications (2009), the system is established and consequently used by scientific publishers and organisations.&lt;br /&gt;
&lt;br /&gt;
Through the project &#039;&#039;&#039;[[STD-DOI]]&#039;&#039;&#039;, the TIB (German National Library of Science and Technology, Hannover) was established as an agency for &#039;&#039;&#039;data DOI&#039;&#039;&#039;. PANGAEA among four data providers was the first system using DOI for automated persistent identification of data sets. A data DOI has the prefix &#039;&#039;&#039;10.1594&#039;&#039;&#039; which is assigned to the publication of primary data through the TIB. The suffix, separated by a slash, is composed of the data system or center acronym and a system specific part. In a Pangaea DOI, this part is equivalent to the &#039;&#039;&#039;internal ID&#039;&#039;&#039;, automaticaly assigned to a data set by the relational database management system during import; thus the uniqueness of each DOI is assured.&lt;br /&gt;
&lt;br /&gt;
A valid Pangaea-DOI has the syntax &#039;&#039;&#039;10.1594/PANGAEA.738357&#039;&#039;&#039;&lt;br /&gt;
* spelled {{doi|10.1594/PANGAEA.738357}}&lt;br /&gt;
* and resolved as https://doi.org/10.1594/PANGAEA.738357&lt;br /&gt;
----&lt;br /&gt;
Data citation and DOI are defined in three steps during the publication process:&lt;br /&gt;
&lt;br /&gt;
* registry status &#039;&#039;will be registered&#039;&#039;, then &lt;br /&gt;
* registry status &#039;&#039;registration is in the lead time&#039;&#039; with DOI registration in progress for &#039;&#039;&#039;30 days&#039;&#039;&#039; followed by&lt;br /&gt;
* registry status &#039;&#039;registered&#039;&#039; after transfer of the DOI to the DOI-registry &amp;gt; DOI can be resolved globally, e.g. at https://doi.org/ &lt;br /&gt;
#If a data set is imported and the status is set to &#039;&#039;validated&#039;&#039;, its internal ID can only be resolved as a &#039;&#039;&#039;preliminary DOI&#039;&#039;&#039; through &#039;&#039;doi.pangaea.de&#039;&#039; (PANGAEAs own DOI resolver). In the citation, the data set is identified as &#039;&#039;&#039;Dataset #738509&#039;&#039;&#039;&lt;br /&gt;
#If a data set status is set to &#039;&#039;published&#039;&#039;, the internal ID is changed to a global resolvable &#039;&#039;&#039;technical DOI&#039;&#039;&#039; 4 weeks after the last edit and the data set gets the status [[Citation|citable]]. In the citation, the data set is identified as &#039;&#039;&#039;Dataset #738509 (DOI registration in progress)&#039;&#039;&#039;, changing after 4 weeks to &#039;&#039;&#039;doi:10.1594/PANGAEA.82361&#039;&#039;&#039; which can be resolved globally.&lt;br /&gt;
#On request the dataset can be defined as an offical &#039;&#039;&#039;data publication&#039;&#039;&#039; and is thus added to the library catalog of the TIB, see [[citation]].&lt;br /&gt;
&lt;br /&gt;
[[PangaVista]] and the [http://doi.pangaea.de/ DOI resolver of PANGAEA] can be used for any registered DOI, including preliminary DOIs of Pangaea.&lt;br /&gt;
&lt;br /&gt;
If a data set to be archived in Pangaea already has a DOI from an other repository, this citation will be indicated with its citation as &amp;quot;other version&amp;quot; in the metadata header.&lt;br /&gt;
&lt;br /&gt;
In case a registered data set has to be deleted, in the field &#039;&#039;other version&#039;&#039; the link/DOI to the substitute must be given &#039;&#039;&#039;before&#039;&#039;&#039; deletion.&lt;br /&gt;
* Example: [http://doi.pangaea.de/10.1594/PANGAEA.58757 doi:10.1594/PANGAEA.58757]&lt;br /&gt;
&lt;br /&gt;
== Prerequisites to become an agent for the registration of scientific primary data ==&lt;br /&gt;
Any data provider, interested in assigning DOI for data may use one of the agents listed below or become a new agent of the &#039;&#039;data-DOI&#039;&#039; agency TIB. When establishing a data system/center new agents need to assure the following points defined through a &#039;&#039;&#039;concept&#039;&#039;&#039; and a &#039;&#039;&#039;[[data policy]]&#039;&#039;&#039;:&lt;br /&gt;
* &#039;&#039;&#039;Metadata&#039;&#039;&#039;&lt;br /&gt;
** metadata are mandatory and should follow standards of the specific scientific field the data are covering (e.g. ISO19115 for geo-data)&lt;br /&gt;
** data sets must be accompanied by a citation, consisting of bibliographic fields according to the STD-DOI application profile&lt;br /&gt;
* &#039;&#039;&#039;Access and availibility&#039;&#039;&#039;&lt;br /&gt;
** long-term availability must be assured, stable linking is provided by means of a DOI&lt;br /&gt;
** data must be available online, assuring Open Access to metadata; Open Access to data is highly recommended (access restrictions may appear for a moratorium period); data should be provided under a CC-[[license]]&lt;br /&gt;
** it is highly recommended, that data are &#039;&#039;machine readable&#039;&#039;, giving data in the repository an &#039;&#039;added values&#039;&#039;. This means, that&lt;br /&gt;
*** (1) data are provided in a standard technical format (best is ascii and ISO formats)&lt;br /&gt;
*** (2) data are organized in a way, that further processing of any part of the repository can easily be performed (data model, relational database)&lt;br /&gt;
** a full backup of the data repository must be assured&lt;br /&gt;
* &#039;&#039;&#039;Data review and integrity&#039;&#039;&#039;&lt;br /&gt;
** once registered, data sets are static&lt;br /&gt;
** versioning is allowed, different versions should be linked to each other&lt;br /&gt;
** data curation must include an editorial process with proofread by the author/principle investigator (the author is responsible for the scientific quality of the data!)&lt;br /&gt;
** an external peer-review of data publications is recommended&lt;br /&gt;
&lt;br /&gt;
== Links to agents for archiving geoscientific primary data with DOI ==&lt;br /&gt;
&lt;br /&gt;
*[http://www.mad.zmaw.de/wdc-for-climate/ World Data Center for Climate (&#039;&#039;&#039;WDCC&#039;&#039;&#039;)] for climate models&lt;br /&gt;
** &#039;&#039;example {{doi|10.1594/WDCC/EH5-T63L31_OM_1CO2_2_MM}}&#039;&#039;&lt;br /&gt;
** contact [mailto:lautenschlager@dkrz.de &#039;&#039;&#039;Michael Lautenschlager&#039;&#039;&#039;]&lt;br /&gt;
----&lt;br /&gt;
*[http://www.pangaea.de &#039;&#039;&#039;PANGAEA&#039;&#039;&#039; data library] for georeferenced observational data (including [http://www.wdc-mare.org WDC-MARE])&lt;br /&gt;
** &#039;&#039;example: {{doi|10.1594/PANGAEA.484677}}&#039;&#039;&lt;br /&gt;
** [https://www.pangaea.de/contact/ Contact]&lt;br /&gt;
----&lt;br /&gt;
*[http://wdc.dlr.de/ World Data Center for Remote Sensing of the Atmosphere (&#039;&#039;&#039;WDC-RSAT&#039;&#039;&#039;)] for remote sensing&lt;br /&gt;
** &#039;&#039;example {{doi|10.1594/WDCRSAT.5Q6Q9Q9B}}&#039;&#039;&lt;br /&gt;
** contact [mailto:michael.bittner@dlr.de &#039;&#039;&#039;Michael Bittner&#039;&#039;&#039;]&lt;br /&gt;
----&lt;br /&gt;
*[http://www.gfz-potsdam.de/ GeoForschungszentrum Potsdam (&#039;&#039;&#039;GFZ&#039;&#039;&#039;) with &#039;&#039;&#039;ICDP&#039;&#039;&#039;]&lt;br /&gt;
** &#039;&#039;example {{doi|10.1594/GFZ/ICDP/KTB/ktb-geoch-gaschr-p}}&#039;&#039;&lt;br /&gt;
** contact [mailto:jklump@gfz-potsdam.de &#039;&#039;&#039;Jens Klump&#039;&#039;&#039;]&lt;br /&gt;
&lt;br /&gt;
== DOI provision service for reports and grey literature by PANGAEA/TIB ==&lt;br /&gt;
For reports and grey literature such as Master or PhD theses a DOI can be assigned by the TIB. The DOI prefix for these kinds of documents is 10.2312. &#039;&#039;&#039;Update 2020-12&#039;&#039;&#039;: a new prefix was provided by the TIB to account for inconsistencies with the DataCite policy: please use 10.48433 from now on.&lt;br /&gt;
&lt;br /&gt;
For submitting documents as PDF to TIB:&amp;lt;br /&amp;gt;&lt;br /&gt;
1) Put all PDF files into one directory. The file names of the PDF files have to be the suffix of the DOI (case sensitive!).&amp;lt;br /&amp;gt;&lt;br /&gt;
2) Create a control file for [[Intern:PanXML]]. See also [[File:Metadata_grey_literature_v3.pdf|Metadata for grey literature]]&amp;lt;br /&amp;gt;&lt;br /&gt;
3) Execute the control file with [[Intern:PanXML]].&amp;lt;br /&amp;gt;&lt;br /&gt;
&lt;br /&gt;
PanXML creates an XML file for each DOI. Send all PDF files together with XML files to TIB (Frauke Ziedorn or Britta Dreyer) in one zip archive.&lt;br /&gt;
&lt;br /&gt;
== Links ==&lt;br /&gt;
*[http://www.tib-hannover.de/de/die-tib/doi-registrierungsagentur/ DOI-Registrieragentur @ TIB]&lt;br /&gt;
*The International DOI Foundation (IDF) https://doi.org/&lt;br /&gt;
*shortDOI service http://shortdoi.org&lt;br /&gt;
*DOI handbook: {{Doi|10.1000/182}}&lt;br /&gt;
*DOI Project for scientific primary data http://www.std-doi.de&lt;br /&gt;
*DOI of the DOI system: {{Doi|10.1000/1}}&lt;br /&gt;
*DOI of the Pangaea data library: {{Doi|10.1594/PANGAEA}}&lt;br /&gt;
*[http://doi.pangaea.de DOI resolver of Pangaea]&lt;br /&gt;
*[http://de.wikipedia.org/wiki/Uniform_Resource_Identifier What is a resource identifier?]&lt;br /&gt;
*Publication using a &#039;&#039;child-&#039;&#039;DOI for an image: {{doi|10.1371/journal.pbio.0020449}}, see &#039;&#039;Reconstruction of Neanderthal woman&#039;&#039;&lt;br /&gt;
*[http://hdl.handle.net Handle resolver]&lt;br /&gt;
*[http://handle.net/ Handle system]&lt;br /&gt;
**to check handle values, type in a DOI and check &#039;&#039;Don&#039;t Redirect to URLs&#039;&#039;&lt;br /&gt;
----&lt;br /&gt;
* provider für DOI im &#039;&#039;deutschsprachigen Raum&#039;&#039; http://www.mvb-online.de&lt;br /&gt;
* provider für DOI in Europa http://www.medra.org/&lt;br /&gt;
----&lt;br /&gt;
* [http://kopal.langzeitarchivierung.de &#039;&#039;&#039;kopal&#039;&#039;&#039; - Kooperativer Aufbau eines Langzeitarchivs digitaler Informationen]&lt;br /&gt;
* [http://www.langzeitarchivierung.de/ &#039;&#039;&#039;nestor&#039;&#039;&#039; - Kompetenznetzwerk Langzeitarchivierung]&lt;br /&gt;
** [http://nestor.sub.uni-goettingen.de/handbuch/index.php nestor-Handbuch]&lt;br /&gt;
* [http://www.tib.uni-hannover.de/ueber_uns/projekte/vascoda/ &#039;&#039;&#039;vascoda&#039;&#039;&#039; - Internet-Portal für wissenschaftliche Information]&lt;br /&gt;
** [http://www.vascoda.de/ vascoda search]&lt;br /&gt;
* [http://www.parse-insight.eu/ &#039;&#039;&#039;PARSE.Insight&#039;&#039;&#039; - Permanent access to the records of science in Europe]&lt;br /&gt;
&lt;br /&gt;
== Resolver ==&lt;br /&gt;
* The CNRI Handle Extension for Firefox is part of the official [http://addons.mozilla.org/addon/10820 Resolver Add-ons for Firefox]&lt;br /&gt;
* &#039;&#039;&#039;https://doi.org&#039;&#039;&#039; - priority in use!&lt;br /&gt;
** https://doi.org/10.1007/s00367-006-0049-8&lt;br /&gt;
*** spelling {{doi|10.1007/s00367-006-0049-8}}&lt;br /&gt;
* https://doi.pangaea.de - resolves doi and handle and unregistered &amp;quot;doi&amp;quot; of pangaea&lt;br /&gt;
** https://doi.pangaea.de/10.1594/PANGAEA.547989&lt;br /&gt;
*** spelling {{doi|10.1594/PANGAEA.547989}}&lt;br /&gt;
* http://hdl.handle.net&lt;br /&gt;
** http://hdl.handle.net/10013/epic.32128&lt;br /&gt;
*** spelling {{hdl|10013/epic.32128}}&lt;br /&gt;
* http://nbn-resolving.de&lt;br /&gt;
** http://nbn-resolving.de/urn:nbn:de:gbv:46-ep000103869&lt;br /&gt;
** http://nbn-resolving.de/urn/resolver.pl?urn=urn:nbn:de:tib-10.1594/PANGAEA.5479896 (special case with doi as part of the urn)&lt;br /&gt;
* http://www.sref.org (invented by publisher &#039;&#039;Copernicus&#039;&#039;, out of order, substituted by DOI in 2009)&lt;br /&gt;
** http://direct.sref.org/1814-9332/cp/2005-1-19&lt;br /&gt;
*** spelling {{sref|1814-9332/cp/2005-1-19}}&lt;br /&gt;
* http://lsid.tdwg.org/ (life science identifier)&lt;br /&gt;
** http://lsid.tdwg.org/urn:lsid:ubio.org:namebank:2659717&lt;br /&gt;
&lt;br /&gt;
for &#039;&#039;&#039;ISBN&#039;&#039;&#039; there is no online-redirect and thus no direct resolver, see&lt;br /&gt;
* http://en.wikipedia.org/wiki/Special:BookSources?isbn=9783000050282&lt;br /&gt;
* http://en.wikipedia.org/wiki/Wikipedia:ISBN&lt;br /&gt;
&lt;br /&gt;
IGSN (International Geological Sampling Number) &#039;&#039;development stage&#039;&#039;&lt;br /&gt;
* IGSN:	ODP010MEY&lt;br /&gt;
** http://app.geosamples.org/sample/igsn/ODP010MEY&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
	<entry>
		<id>https://wiki.pangaea.de/w/handler?title=NADC&amp;diff=16696</id>
		<title>NADC</title>
		<link rel="alternate" type="text/html" href="https://wiki.pangaea.de/w/handler?title=NADC&amp;diff=16696"/>
		<updated>2026-03-26T16:06:22Z</updated>

		<summary type="html">&lt;p&gt;Lmoeller: Lmoeller moved page NADC to Intern:NADC: last edit 2016 - relevance for PANGAEA docu questionable&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;#REDIRECT [[Intern:NADC]]&lt;/div&gt;</summary>
		<author><name>Lmoeller</name></author>
	</entry>
</feed>