Documentation: HMC Dashboard on Open and FAIR Data in Helmholtz

Introduction

The HMC Dashboard on Open and FAIR Data in the Helmholtz Association is a data-driven approach to monitor and improve open- and FAIR data practices in the Helmholtz Association, Germany’s largest non-university research organization. It is based on a modular pipeline to identify and evaluate data published alongside journal publications and is directly transferable and reusable by other research organizations.

The project addresses a variety of target groups ranging from the management level, research data professionals, and research data management staff at the eighteen Helmholtz centers, to providers of research data infrastructure as well as researchers who are interested in learning more about open- and FAIR data practices in the Helmholtz Association.

Three key challenges addressed with this project are:

  1. Where is data in a cross-disciplinary, federated research organization published?
  2. How can we identify gaps towards FAIR data?
  3. How can we use this information to engage research data communities and -infrastructure?

Structure of the project

The project is divided into two parts, represented by two project repositories and two sections in this documentation, respectively:

structure

Disclaimer

Please note that the list of data publications obtained from data harvesting using the HMC Toolbox for Data Mining, as presented in the HMC FAIR Data Dashboard is affected by method-specific biases and is neither complete nor entirely free of falsely identified data. If you wish to reuse the data shown in this dashboard for sensitive topics such as funding mechanisms, we highly recommend a manual review of the data.

We also recommend careful interpretation of evaluation-results derived from automatized FAIR assessment. The FAIR principles are a set of high-level principles and applying them depends on the specific context such as discipline-specific aspects. There are various quantitative and qualitative methods to assess the FAIRness of data (see also FAIRassist.org (opens in a new tab) but no definitive methodology (see Wilkinson et al. (opens in a new tab)). For this reason, different FAIR assessment tools can provide different scores for the same dataset. We may include alternate, complementary methodologies in future versions of this project. To illustrate the potentials of identifying systematic gaps with automated evaluation approaches, in this dashboard you can observe evaluation results obtained from F-UJI (opens in a new tab) as one selected approach. Both, the F-UJI framework and the underlying metrics (opens in a new tab) are subject of continuous development. The evaluation results can be useful in providing guidance for improving the FAIRness of data and repository infrastructure, respectively, but focus on machine-actionable aspects and are limited with respect to human-understandable and discipline-specific aspects of metadata. Evaluations results obtained from F-UJI can be useful in providing guidance for improving the FAIRness of data and repository infrastructure, respectively, but cannot truly assess how FAIR research data really is.

How to cite this work

Scientific papers

  • Kubin, M., Sedeqi, M.R., Schmidt, A., Gilein, A., Glodowski, T., Serve, V., Günther, G., Weisweiler, N.L., Preuß, G. and Mannix, O. (2024) "A Data-Driven Approach to Monitor and Improve Open and FAIR Research Data in a Federated Research Ecosystem", Data Science Journal, 23(1), p. 41. Available at: https://doi.org/10.5334/dsj-2024-041 (opens in a new tab).

Software publications

Data availability

  • A selective export of the dataset included in the dashboard is published on Zenodo (opens in a new tab) and will be updated in regular intervals.