Workflow

This chapter describes the workflow from the data source to its final publication. The workflow for a data publication is similar to the established submission>peer-review>publication process in scientific literature. Author(s) who want to will be integrated into an editorial process in communication with the projects or institutes data curator or an editor of the Pangaea system. Since 2011 the workflow is documented through a service ticket system.
 * archive a supplement related to a publication or
 * making results available to the scientific community, e.g. within a research project,
 * publish a data report,

Provided that the data publication workflow is an interaction beween (corresponding) author and curator and consists of 5 steps:
 * the granularity of the data sets is defined; depending on the scientific field and the topology of the data,
 * parameters are defined with unit; as close to international standard usage as possible,
 * metadata are on hand,
 * any information, necessary to understand the generation and content of the data set, is available,
 * 1) Submission - the author sends data sets with metadata description for archiving/publication following the submission guidelines and the policy of the project. Central address for submissions is info@pangaea.de or via the ticket system.
 * 2) Completeness check - the data package as submitted by the author is checked by the curator/editor for completeness of metadata and validity of the data. A request will be send to the author if mandatory information is missing.
 * 3) Archiving - the availability of the required metadata is checked via the editorial system (4D), missing information is defined by the curator. Data are converted to the Pangaea import format. For single files an editor or spreadsheet software is used, for multiple files with similar format, the use of PanTool and Split2Events is recommended. With the editorial system the relations are set between metadata and data prior to import. External documents should have PDF/A-format and must be linked via a persistent identifier. After import the availability and vailidity of the dataset(s) is checked in browser-view by the curator. This includes a validity check of all external links.
 * 4) Proof-read - the curator sends the DOI as link to the author(s) and requests for proofread. Through an iterative process between author and curator via email the dataset is edited until it is finaly approved by the author.
 * 5) Publication - the data set status is set to published, the DOI becomes valid 4 weeks after the final editing. On request of the author, a password may be set for a muratorium period as defined through a projects data policy or - for supplements - until a paper is public available. Between submission and publication of a paper, the reviewer may have data access via the DOI and thus data can easily be included in the peer-review process.

Data supplements with status supplementary data can automatically be linked to the publications splash page in the journals/publishers catalog through an automatic web service (e.g. doi:10.1016/0025-3227(89)90086-8 ).

The ressources required to establish this workflow need to become a mandatory part of each research project. The amount is estimated in the business model.