ALL Metrics
-
Views
-
Downloads
Get PDF
Get XML
Cite
Export
Track
Method Article

Towards a quality indicator for research data publications and research software publications – a vision from the Helmholtz Association

[version 1; peer review: awaiting peer review]
PUBLISHED 10 May 2024
Author details Author details
OPEN PEER REVIEW
REVIEWER STATUS AWAITING PEER REVIEW

This article is included in the Research on Research, Policy & Culture gateway.

Abstract

Research data and software are widely accepted as an outcome of scientific work. However, in comparison to text-based publications, there is not yet an established process to assess and evaluate quality of research data and research software publications. This paper presents an attempt to fill this gap. Initiated by the Working Group Open Science of the Helmholtz Association, the Task Group Helmholtz Quality Indicators for Data and Software Publications currently develops a quality indicator for research data and research software publications to be used within the Association. This report summarizes the vision of the group of what all contributes to such an indicator. The proposed approach relies on generic well-established concepts for quality criteria, such as the FAIR Principles and the COBIT Maturity Model. It does – on purpose – not limit itself to technical implementation possibilities to avoid using an existing metric for a new purpose. The intention of this paper is to share the current state for further discussion with all stakeholders, particularly with other groups also working on similar metrics but also with entities that use the metrics.

Keywords

Data publication, Software publication, bibliometrics

1. Background and objective

Measuring the performance and progress of research based on means of output is a common task of research assessment. While it is commonly accepted to use quality-assured scientific publications as a proxy for quality assessment in research evaluation, other artifacts of scientific work such as research data and research software are often not yet considered. For text-based scientific work, there are well established processes of quality assurance during the process of publication. For publishing research data and research software no commonly accepted standards exist yet, although research data and software represent a significant output of the scientific endeavor especially given the rise of data- and software-intensive research.

The Helmholtz Association has already decided to integrate research data publications and research software publications in its internal key performance indicator system and is currently developing and implementing a quality indicator for those artifacts. It should become a driver within Helmholtz in a threefold way: a) to create awareness and appreciation for the diversity of research output, b) to align the assessment of research and research practice to the conditions of digitalization and openness of science, and c) as a stimulus and incentive for scientists to improve their data and software publications related to specific quality criteria, and thus, to improve the quality and trustworthiness of science as a whole. Helmholtz Centers from all research fields of the Association contributed in an iterative discussion process to cover a broad perspective originating from their respective research environment. To meet all aspects relevant for the quality indicator, a group of scientists, scientific software developers, data curators, and colleagues reporting on research output were involved in its development. This paper provides insight to the current status of the discussion.

2. Terminology

The following definitions should clarify what type of research data and research software we address in this paper.

2.1 Research data

In accordance with the EOSC-Glossary1 we define research data as data that has been collected, created and/or is being used in “scientific research activities and used as evidence in the research process, or commonly accepted in the research community as necessary to validate research findings and results”. It ranges from highly complex and individual long-tail-data from experiments and observations to automatically produced generic data collections.

2.2 Research software

We define research software as software that has been developed in the research process.2 We consider software of all levels of maturity: from personal use, reuse by others, long term support, and criticality as defined by3 in the form of software classes.

2.3 Published data and software

In both cases we concentrate on research output which is published. We define published data and software as digital objects that have been made as permanently available as possible on the internet.4 The different levels of “published” are addressed in the framework of quality dimensions and attributes (Section 4 and 5).

3. Methodical framework – quality criteria and quality quantification

Recalling the publishing process of scientific publications: A paper published e.g. in a scientific journal has been quality checked with respect to several criteria over the course of the publication process. When accepted, a manuscript is marked as being approved by the underlying quality assurance mechanisms. Aggregating the number of papers that went through this process of quality assurance is therefore used as a quantitative measure of the output performance of a scientific entity. Yet, this process of solely using simple and one-dimensional metrics to assess research quality and performance is broadly criticized. But the convenience of having one number easy at hand has so far hindered all attempts to replace this practice. Taking this into consideration a quality indicator for research data and software has to be developed with the goal to have one easily comparable value that still does more then only counting - by representing a diverse set of quality features in itself.

Therefore, any analog of this process for scientific data publications or research software publications has to technically map a multi-dimensional suite of quality criteria onto one value, the quality indicator. Thus, developing a quality indicator has to address the following questions: 1) What quality criteria should be considered? 2) How to measure each of the quality criteria? 3) How to condense individual quality measures into one number, the quality indicator? In addition, for practicability, 4) how to collect the quality measures for the most part in an automated fashion? Since the Helmholtz quality indicator has to match all types of research data publications and research software publications generated in the diverse scientific disciplines within the Helmholtz Association, generic approaches are required to provide answers to these questions. We decided to build on the FAIR Principles5 to specify relevant quality criteria. In order to provide a generic way to estimate quality for a wide range of criteria, we are using an approach based on the COBIT Maturity Model6 to quantify quality in terms of a maturity scale. Finally, a weighted summation and radar plot approach is used to derive one single value from the various quantitative measures while also providing a more detailed representation. The collection of quality measures should rely on automated procedures and tools as much as possible. However, in the end, any indicator must match the target set, technical issues must not lead to deviation from that target. Therefore, our basic approach for developing the quality indicator was to define the quality criteria that are essential for the intended objectives in a first step, and to then condense the list of criteria in a second step with respect to technical feasibility of automated collection. This approach discusses quality criteria of research data publications and software publications without immediately considering the technical implementation of measuring them. Thus, we selected metrics not simply for the ease of measurement. Rather, the measurement of some criteria deemed highly relevant might be deferred until a corresponding, automated approach of evaluating them becomes available.

The overall approach then consists of the following five steps.

  • (i) Define a suitable number of dimensions, according to which quality of research data and software can be assessed.

  • (ii) For each of these dimensions, collect a set of specific attributes capturing the respective dimension.

  • (iii) Using a generic maturity model, each attribute is mapped onto a numeric scale

  • (iv) Using weighted averages, an aggregated maturity level is obtained for each dimension.

  • (v) The summarizing quality assessment can then be visualized using radar plots.

These steps lead to the current state of a quality summary of the assessed research data and software. Using more than one dimension allows us to take different perspectives into account. For example, an organization aiming at building up a large, integrated software library might have other targets for quality assessment than a single software developer working on individual projects. As well as an international campaign gathering huge amounts of automatically generated sensor-data has other objectives, restrictions, and conditions than long-tail data generated in experiments in a laboratory setting or a medical scientist working with confidential patient data. In analogy to the aggregation of quality-assured scientific publications, a quantitative estimate of high quality data and software publications could then be obtained by aggregating the number of data/software showing a maturity above a given minimum in each of the dimensions. Afterwards, the maturity of each dimension also has to be aggregated to form the final indicator for a publication.

To determine an initial set of dimensions for quality assessment, we are building on the FAIR principles.5 The FAIR principles have been widely adopted as a minimum set of quality criteria for research data and research software thus, our approach generally also builds on the four principles findability, accessibility, interoperability and re-usability. Considering upcoming discussions within communities on the limitations of pure implementation of FAIR to grasp a comprehensive view on the quality of research data and research software publications, the partly interrelated aspects and minor inaccuracies of wording within the four core principles we adopted and extended the FAIR principles to best suit our needs.

For research software we amended the FAIR criteria with two additional criteria: scientific quality, addressing the question of research software being scientifically well grounded, and technical competence, capturing aspects of the software being technically well grounded (FAIR-ST).

For evaluating research data publications, we regrouped the FAIR criteria into four dimensions that are easier to grasp related to datasets but still cover the same aspects. Those being 1. Publishing, 2. Openness, 3. Curation, and 4. Metadata. To compensate for inadequacies between different disciplines when applying a general framework like the proposed, the 5. dimension of the External View has been added to the set (POCME).

In order to allow for a comparable and unbiased quality assessment for data and software publications based on our modified FAIR-framework, we need a broadly applicable scheme leading to a numerical value for each quality criterion. In certification processes, maturity models are commonly used to solve this task.

3.1 The COBIT maturity model

To achieve comparability among different attributes of quality, stages of product evolution can be assessed using a generic model of maturity. Being generic, the model can be applied to all categories of scientific output leading to quantitative results which can be compared. Our approach utilizes the COBIT Maturity Model.6 The COBIT Maturity Model quantifies maturity of a product in six levels by mainly evaluating the processes that lead to the product. Looking at processes for achieving a certain quality is a common approach in quality management, where quality management systems (QMS) are defined as “formalized systems that document processes, procedures, and responsibilities for achieving quality policies and objectives”.7 Typically, lower levels of maturity focus on individual outcomes of a process to be certified. The higher the maturity, the more the process is considered, accounting for the fact that a well-developed process naturally leads to products of a certain quality. COBIT starts with a default level 0 (non-existent), two levels of maturity acknowledge the successive completion of a set of quality criteria for the product to be targeted. Starting with level 3 maturity implies a process to be existent which guarantees the addressed quality to be kept. Two further levels finally focus on the improvement of the underlying process towards full optimization with maturity level 5. Clearly, the explicit meaning of the maturity levels for a given quality criterion has to be defined as part of the community process. Obviously, the maturation levels need to be interpreted within the context of a given aspect of quality assessment. The levels are cumulative in the sense that a higher level includes the requirements of all lower levels, i.e. marking a higher level as fulfilled means that the requirements of all lower levels are fulfilled plus the requirements of the highest marked level. Thus, a higher level can either refine/specialize characteristics that were already evaluated for compliance with a lower level or introduce a new and additional characteristic to assess. As a result, comparability across quality dimensions is given.

The specific application of maturity levels derived from the COBIT Maturity Model with regards to data and software publications are introduced in the respective subsection of this paper (Sections 4 and 5).

3.2 Weighted summation and radar plot

Guidelines for reviewers for academic journals typically provide a set of criteria being considered in preparing a review for a manuscript. Examples are scientific originality, proper use of methods, organization of the material/figures/tables, and comprehensibility. Such criteria can likewise be grouped into different dimensions of assessment such as scientific novelty, being scientifically grounded, quality of presentation etc. However, such dimensions are commonly not transparently weighted against each other. Nevertheless, there is a common understanding in each scientific community on what has to be considered more important. To make such an understanding open and transparent, we introduce radar plots to visualize the joint maturity of a data and software publication with respect to all quality dimensions. The relative importance of every single attribute within each dimension of assessment can be fine-tuned by agreeing on weights to be used in a weighted average. Doing so, minor criteria can also be included into the assessment without the risk of leading to a biased assessment.

For the overall assessment of the data and software publication, we deliberately do not suggest to use weighted averages as a default. The multi-dimensionality of each quality assessment should rather be kept transparent. This allows us to optimize towards different goals in using the quality indicator. Figure 1 exemplifies the result of the quality assessment of two fictitious research software or data publications.

3c09abb6-11f0-42de-b60c-97098e1d80e8_figure1.gif

Figure 1. Example of a fictitious assessment for two research software or data publications showing how the radar plot can visualize the fulfillment of some criteria at higher quality levels than others.

Maturity levels are ordered from center (0, non-existent).

4. The Quality Indicator for research data

4.1 Dimensions of quality for research data assessment

The quality of published research data and hence its potential value for future reuse in other contexts can be assessed based on the quality of the metadata describing the data formally, on metadata describing the data on the content-level, on the quality of the data itself and the conditions of Openness/Licensing. With regards to the diversity of research data over the different scientific domains represented at Helmholtz, the heterogeneous ways of collecting, creating, editing, and curating data, a single quality criterion for research data assessment would not be constructive. To allow for a multi-dimensional perspective on research data, the quality assessment being presented in this document is spanning five dimensions of data quality being partly based on the FAIR-principles and the FAIR Data Maturity Model of the Research Data Alliance.8

The derived quality dimensions of the POCME-framework for research data publications, each with indication of the covered FAIR-principles, are:

Publishing

This dimension covers especially the areas of Findability and Accessibility. It addresses the use of identifiers to enable unambiguous identification as well as the level of storage and hence the possibilities and information on the general accessibility to published data.

Openness

As publishing a dataset does not necessarily mean to make it openly available (Intelligent Openness), this dimension takes this circumstance into account and covers aspects of Accessibility, Interoperability and Reusability.

Curation

Curation can be crucial for the quality of published data itself as well as for the provided formal and content related information. Hence, it can significantly improve the value of a published dataset for Reusability as well as for Findability, Accessibility, and Interoperability.

Metadata

Qualitative metadata can help in many aspects of Findability, Accessibility, Interoperability, and Reusability of published data. By providing content related aspects, helpful information to retrieve a dataset as well as context around a dataset.

External view

This dimension is meant as a handle to compensate inadequacies between different disciplines which is a sensible approach when applying a general framework like the proposed.

These five dimensions of quality for research data are still too generic to be evaluated using the maturity model. Therefore, each dimension is split into a set of attributes. The latter will be assessed using the maturity model. To derive a single value for each of the five dimensions, an aggregation step has to be incorporated. This is done by computing a weighted average of the maturation values for all attributes constituting a quality dimension. Using weights provides the freedom not to be forced to work with attributes of equal importance. Minor aspects are just given smaller weight.

4.2 Quality assessment of the dimensions

To assess the quality of research data publications, the developed POCME assessment-framework measures/estimates the maturity of different quality dimensions and attributes. To be able to do this, a generic five-level scale of maturity derived from the COBIT Maturity Model is applied. The levels are always to be seen in the context of the respective dimension and attribute they are applied on.

The first level (0) usually implies that the respective attribute is non-existent or not applied. The following levels built upon each other, always following the assumption, that the previous levels are largely met, leading to the highest possible level of maturity and hence quality of the data itself and the accompanying meta-information maximizing the findability, accessibility, interoperability, and re-usability.

  • (0) Non-existent: no information available or not applied

  • (1) Most necessary information provided or measure taken

  • (2) Basic information provided or measure taken (sensible level of information/measures)

  • (3) Advanced information provided or measure taken, allowing to generally understand and (re) use the published data

  • (4) Complete and accurate information provided or measure taken, to an extend that allows maximal understanding and usage of data

The following sections list the attributes and their maturity levels of research data publications that are proposed for each dimension. It must again be noted that the COBIT model is cumulative, i.e. a higher level includes the fulfillment of all requirements of a lower level.

4.2.1 Dimension publishing

Published with identifier

  • (0) No identifier (resource may only be found via personal communication)

  • (1) Basic Uniform Resource Identifier

  • (2) Dataset is identifiable via internal handle (does not resolve globally, generally no metadata)

  • (3) Dataset is basically identifiable via formalized, standardized, persistent identifier (resolves globally, general metadata provided)

  • (4) Dataset is identifiable via globally unique, formalized, standardized, persistent identifier supported by general metadata (e.g. DOI).

Published via a repository or collection, that is indexed in a meta-repository (e.g. re3data)

  • (0) No information available, the data is not published via a repository/collection

  • (1) The data is published in a repository/collection which is not listed in an eligible meta-repository

  • (2) The repository/collection is listed in an eligible meta-repository, basic number of quality indicators assigned by the meta-repository are achieved

  • (3) The repository/collection is listed in an eligible meta-repository, medium number of quality indicators assigned by the meta-repository are achieved

  • (4) The repository/collection is listed in an eligible meta-repository, high number of quality indicators assigned by the meta-repository are achieved

Remark: Meta-repositories like re3data usually do not perform in depth checks on the quality of published research data in the indexed sources. Still, if sources indexed in meta-repositories like re3data fulfill certain quality-criteria respective indicators (e.g. icons) are assigned. The icons themselves represent different quality-levels on the level of the source, assuming that contained dataset likely meet those quality levels represented by the source.

Published with information on access to the data

  • (0) No metadata available

  • (1) Metadata available, but no data access-information available in the metadata

  • (2) Metadata available, data access-information available only in human-readable form

  • (3) Metadata available, data access-information available only in human readable form, including general license information

  • (4) Metadata available, data access-information available in human-readable and machine-readable form [1], including license information

4.2.2 Dimension openness

General degree of openness

  • (0) No information on open accessibility/availability of the data at all

  • (1) Information available: no open accessibility/availability of the data. No justification, no information on possible contact or restrictions

  • (2) Like (1) + information on possible contact, restrictions or potential use cases on request available

  • (3) Like (2) + with justification AND/OR date of moratorium

  • (4) Open accessibility with corresponding license (no login or contact needed or otherwise with justification)

Primary data formats

  • (0) No primary data available in digital form

  • (1) Primary data generally available

  • (2) Primary data stored in common proprietary data formats

  • (3) Primary data stored in open formats

  • (4) Primary data makes use of common, domain specific terminologies (e.g., codelists)

4.2.3 Dimension curation

Level of curation

  • (0) Data is published in raw form without any curation or documentation (e.g. raw long-tail data)

  • (1) Data is published in raw form without curation but according to standard with basic documentation like readme (e.g. automatic generated sensor data, long-tail data following a basic scheme)

  • (2) Data is published in cleaned form with some curation (e.g. brief checking, documentation according to standard)

  • (3) Data is published in cleaned form with enhanced curation and/or reprocessing (e.g. conversion to new formats, enhancement of documentation)

  • (4) Data is published after undergoing extensive curation and/or reprocessing according to discipline specific standards in order to enhance to max. quality (like (3) + additional editing of deposited data for accuracy)

4.2.4 Dimension metadata

Metadata to find/retrieve a resource/formal metadata

  • (0) No metadata available

  • (1) Metadata available for/with the data publication that is not structured according to a commonly accepted scheme (i.e. no scheme applied)

  • (2) Metadata provided with the data publication that is structured in a basic way according to a commonly accepted scheme (e.g. completed DataCite mandatory-properties/discovery; Dublin Core, etc.)

  • (3) Metadata provided with the data publication that is structured in an advanced way, according to a commonly accepted scheme (e.g., completed DataCite mandatory- and recommended-properties for discovery + discovery-supporting basic content metadata according to DataCite scheme)

  • (4) Full Metadata provided with the data publication (complete DataCite mandatory- and recommended- and optional-properties for discovery + comprehensive discovery-supporting content metadata according to DataCite scheme)

Content related metadata

  • (0) No content related metadata available

  • (1) Some content related metadata available, following a (generic) scheme (e.g. DataCite)

  • (2) Complete content related metadata available following a (generic) scheme (e.g. DataCite)

  • (3) Some content related metadata available following standardized form or domain specific scheme

  • (4) Complete and curated content related metadata available following a standardized form and domain specific scheme (see 3)

4.2.5 Dimension external view

Score from domain specific fair assessment tool

  • (0) 0-20% Score reached

  • (1) 21-40% Score reached

  • (2) 41-60% Score reached

  • (3) 61-80% Score reached

  • (4) 81-100% Score reached

Remark: By integrating this dimension, it would be possible to take a domain specific evaluation of a data publication in regard to community best practices and domain specific requirements on metadata and ideally the data itself into account. This is based on the assumption that in addition to more generalistic FAIR assessment tools currently available, further tools with domain-specific evaluation schemes will be developed. The rating scales of these tools can then presumably always be converted into a percentage scale. A rather low weighting of this dimension must be considered in order to minimize the stimulus for gamification.

5. The Quality Indicator for research software

5.1 Dimensions of quality for research software assessment

Quality of research software has various aspects to be considered. Depending on the type of software, the field of application and the scientific goals to be followed in using research software, different aspects of research software are being considered as important. Therefore, a single quality criterion for research software assessment would not be constructive. To allow a multi-dimensional perspective on research software, the quality assessment being presented in this document is spanning six dimensions of software quality being given by the proposed FAIR-ST framework.

Using the principles for quality assessment, the FAIR principles have to be spelled out explicitly with respect to research software. This is given by FAIR4RS.9 As mentioned above, the four FAIR principles have been augmented with two further quality dimensions. Thus, the six dimensions of research software quality being presented in the assessment scheme are given as follows.

  • (a) Findable: Researchers need to be able to find the software using typical search strategies. Having found a given software product, they have to be able to identify a given version/release and be provided with enough information to value the software for their specific research.

  • (b) Accessible: Software needs to be accessible in order to unfold its potential. Hereby, accessibility has both a legal side as well as an operational side.

  • (c) Interoperable: For software to be interoperable, it must be capable of being included into larger contexts or frameworks. Towards this aim, input/output formats need to be considered as well as interfaces to use the software in automated pipelines.

  • (d) Reusable: Using software for exactly the purpose it has been developed for is considered as accessibility within the quality assessment. In contrast, reusability considered the aspect of being able/allowed to use the code for one’s own purposes, to adapt and extend the software.

  • (e) Scientific basis: Research software is an integral part of the research process. It therefore has to also follow community/organization specified common standards in performing research. While the contribution of software to a certain scientific achievement must be evaluated by dedicated experts in the field, some aspects of scientific quality can also be considered in wider generality.

  • (f) Technical basis: Quality of software also reflects general aspects of software engineering. At the end, software has to guarantee to actually produce what it has been aimed to do. Professional software development aims at producing software following state-of-the-art software engineering concepts. Thus, aspects of professional software engineering must also be considered as part of a quality assessment of research software development.

The six dimensions of quality for research software are still too generic to be evaluated using the maturity model. Therefore similar to the approach for research data, each dimension is split into a set of attributes. The same principles as outlined previously for research data are applied.

5.2 Quality assessment of the dimensions

The following descriptions summarize the proposed five levels of maturity used for evaluating the quality of software publications. Level 0 is the default level, when no information for the evaluated attribute is provided.

  • (5) Non-existent: no information available

  • (6) Initial: initial information available being obtained in an ad-hoc, unorganized manner

  • (7) Repeatable: the information is complete, being produced in a repeatable, yet intuitive manner

  • (8) Defined: a process is established guaranteeing the complete compilation of the required information

  • (9) Managed: the process being established is managed, i.e. monitoring/measuring is included

  • (10) Optimized: practices are put in place optimizing the way the process is operated, leading to improved quality over time

For each of the quality dimensions, the assessment is done by rating every corresponding attribute according to its maturity. The explicit interpretation of each maturity level for a given attribute is given in terms of asking of compliance with statements. Since the maturity levels are building on each other in a cumulative way, the number of the last achieved statement defines the maturity of the attribute.

The following sections provide the assessment scheme for the six dimensions of the FAIR-ST framework.

5.2.1 Dimension findable

The following statements address the aspect of being able to find and uniquely identify the software.

Open publication repository

  • (0) There is no information available on where to find the software.

  • (1) The software is contained in an online repository.

  • (2) Some kind of description is available giving further information on the software in this repository (e.g. readme file).

  • (3) A structured meta data description (e.g. following DataCite) given for software is in this repository.

  • (4) The repository is listed in some overarching meta-repository (e.g. Helmholtz Research Software Directory (RSD)).

  • (5) The meta-repository performs quality checks (e.g. re3data) for the used publication repository.

Versioning

  • (0) No software versioning applied.

  • (1) There is some kind of versioning for the software.

  • (2) The software uses structured (e.g. semantic) versioning.

  • (3) A description of the versioning scheme is available.

  • (4) There is a documentation on release cycles for the software.

  • (5) The versioning scheme allows for automatic tagging by CI/CD processes.

Persistent Identifier (PID)

  • (0) No PIDs given.

  • (1) A handle/URL is provided to identify the software.

  • (2) The handle/URL is provided with a defined metadata scheme.

  • (3) A persistent identifier is provided.

  • (4) A PID allowing for automated harvesting of metadata information is provided.

  • (5) The PID is part of an established community standard.

Rich metadata

  • (0) No metadata given.

  • (1) Some metadata information is provided with the software.

  • (2) The metadata information is following a given metadata scheme completely.

  • (3) A metadata curation process reflects changes/updates.

  • (4) All metadata information following the given metadata scheme can be automatically harvested.

  • (5) An external quality assessment of the metadata exists.

5.2.2 Dimension accessible

The following statements address the aspect of being able to access research software. Accessing included the possibility to run the software, which might also be in terms of a web service. However, accessibility does not include the possibility to adjust the code which is rather being captured under the aspect of reusability.

Access conditions (organizational)

  • (0) Not specified.

  • (1) A contact is given which to inquire about the right to use the software.

  • (2) The software has a license describing rights of use.

  • (3) The license allows for open use of the software (e.g. OSI licenses).

  • (4) There is a way to also obtain some kind of support in using the software.

  • (5) There is a community, providing the opportunity of support and exchange concerning aspects of using the software.

Access options (process)

  • (0) There is only one specific form of accessing the software or no option at all.

  • (1) The software (source code or executable) is provided.

  • (2) The sources or executables being provided include some documentation on how to install/use the software.

  • (3) Provided test cases allow to determine whether installation/execution worked as expected.

  • (4) Provided checks make sure the software works correctly.

  • (5) A software service is provided, i.e. are reported bugs taken into the development cycle.

Technical accessibility (run/start)

  • (0) No information given.

  • (1) “How to install” information is provided.

  • (2) Installation scripts are provided.

  • (3) The software allows for (semi-)automated installation, e.g. a Makefile or manual package (like Python modules).

  • (4) Sources are provided such that a package manager or automated build tools, e.g. automake, can be used.

  • (5) A complete package that enables execution (e.g. container, app package) is available.

5.2.3 Dimension interoperable

The following statements address the aspect of being interoperable, i.e., the possibilities of being able to integrate the software into one’s own software framework or execution pipelines.

Input/Output formats

  • (0) Not specified.

  • (1) Some description of input and output formats is provided.

  • (2) The software builds on standard formats for input and output.

  • (3) Additional options for varying input/output formats are provided.

  • (4) The software builds on accepted community standards for input/output data.

  • (5) The software provides in addition further tools for processing input/output data.

Adaptability/Flexibility of use

  • (0) No information given.

  • (1) There is a way to use the software with one defined set of input data.

  • (2) There are parameters to adjust the way the software is working.

  • (3) There is some way of logging what is done during execution.

  • (4) Documented API(s) are provided to integrate the software into one’s own framework.

  • (5) There is a documented way to integrate the software into open workflows, e.g. via containers, web-services etc.

5.2.4 Dimension reusable

The following statements address the aspect of being reusable. In addition of being accessible, i.e. executable, reusability includes the possibility to actually change/adapt the code.

Reusability conditions

  • (0) Not clear.

  • (1) The software uses a custom license allowing reuse.

  • (2) The software uses a FOSS/OSI approved license including that license dependencies are at least being checked manually.

  • (3) The software uses an appropriate license for different file types (code, text, images etc.) following e.g. the REUSE specification.

  • (4) There is a process available for automatically checking e.g. the REUSE specification.

  • (5) There is a process available such that all license dependencies are automatically controlled.

5.2.5 Dimension scientific basis

The following statements address the aspect of the software being scientifically well grounded. While domain specific scientific requirements have to be assessed as part of a scientific peer-review process, certain generic aspect of good scientific practice can be assessed for all research software.

Community standards

  • (0) No information given.

  • (1) The connection to known (scientific) standards is drawn.

  • (2) The software follows the standards of the relevant scientific community.

  • (3) The software complies with relevant scientific standards of the field.

  • (4) There is an indication on how further evolution of community standards will be addressed.

  • (5) A closed feedback-loop is established, making sure that further evolutions of community standards are being adopted.

Team expertise

  • (0) No information given.

  • (1) Clear expertise from a single, relevant domain is part of the software development team.

  • (2) The software development team has access to expertise in several relevant domains.

  • (3) The software development team has access to expertise in all relevant domains.

  • (4) A fixed, established, interdisciplinary team works on the software.

  • (5) An established and coordinated community of software developers works on the software.

Scientific embedding

  • (0) No information given.

  • (1) At least one scientific use case is documented.

  • (2) A broader scientific context is documented including several examples.

  • (3) The software development is at least loosely connected to some scientific initiative.

  • (4) The software development is part of a larger scientific initiative.

  • (5) The software development is part of a larger scientific initiative with dedicated processes for software development.

5.2.6 Dimension technical basis

The following statements address aspects of professional software development leading to sustainable, high quality research software.

Project management

  • (0) No information on project management and code history being provided.

  • (1) Some kind of version control is used.

  • (2) A version control system is used.

  • (3) A version control system being part of a code project management platform (e.g. GitHub, GitLab) and an associated ticket system is in place.

  • (4) A transparent process for ticket resolving, code review by other developer, and merge requests is established.

  • (5) A release process with guaranteed changelog generation, testing, and product provisioning is established.

Repository structure

  • (0) No information given.

  • (1) All files are provided in some structured/unstructured way inside the repository.

  • (2) The repository is structured albeit maybe in a manner such that every contributor is free to follow their own way of organizing files.

  • (3) A contribution mechanism is documented, e.g. CONTRIBUTORS.md file, as well as a defined structure for the repository and a documented onboarding process.

  • (4) A common template for the repository structure is available, as well as some kind of identification of deviation.

  • (5) A repository structure is enforced following community standards.

Code structure

  • (0) No information given.

  • (1) Every developer is free to use his/her own style of coding.

  • (2) There are general recommendations for coding, albeit every developer being able to follow his/her own style.

  • (3) There is some harmonization of code style being enforced following common standards including meaningful naming of functions/variables etc.

  • (4) The code style is checked when accepting changes into the repository.

  • (5) The code style is enforced via a review process (e.g. failed pipelines or auto-formatting).

Reproducibility (Code)

  • (0) No tests, or duplicated code.

  • (1) The code follows a modular structure allowing for component reusability.

  • (2) Clear system requirements are documented with min/max versions, albeit version pinning, modularity etc. being enforced manually.

  • (3) A package manager is used for dependency pinning and testing enforced.

  • (4) Test coverage is measured, albeit tests may be written on a voluntary basis.

  • (5) Automated testing for different system environments, requirements for minimal test coverage, and provisioning of containerized packages is done.

Code change process

  • (0) No information

  • (1) Internal 4-eye principle for accepting changes

  • (2) Code changes via transparent processes, e.g. merge/pull request

  • (3) Approval of code changes via transparent processes and with a 4-eye principle

  • (4) Integration of code changes into main development branch/releases only allowed for specifically named/trained persons.

  • (5) Software releases involve an external review (by someone outside of the core developer team)

Security

  • (0) No security concepts given.

  • (1) There are at least sporadic updates and dependency checks.

  • (2) There is a systematic assessment of dependencies and documentation of the software stack.

  • (3) Deployment is provided within a CI/CD framework for different environments including tools for check for security leaks.

  • (4) There is some process for monitoring dependency updates including reporting.

  • (5) There are regular and automated security monitoring and an automated update process in place allowing merges only of security checks have been passed.

6. Results

The procedure described so far represents the result of the conceptualization phase of defining a quality indicator for data and software publications as initiated within the Helmholtz Association. For implementation of the procedure, we plan four phases: 1) proof of concept, 2) automatization supported by suitable tools, 3) piloting phase, 4) final implementation within the Helmholtz Association.

6.1 Proof of concept

Feasibility of the concept is currently evaluated in various research communities by running a couple of exemplifying pen and paper assessments. While the attributes for each quality dimension have been discussed and iterated in several rounds of community engagement, their actual application to assess the quality of given data or software will show whether the formulation of the attributes is clear enough to be applied in a wider community, where documentation of the assessment needs to be improved, or where even manual assessment is not easily possible at the moment.

6.2 Automatization supported by suitable tools

The quality assessment should only lead to a minimal increase in resources required for the assessment itself. Therefore, the assessment of all quality dimensions and attributes should be automatized as much as possible, the final assessment and visualization should be done in a digital manner. We currently evaluate existing tools (such as r3data, F-UJI, HERMES, the Helmholtz Research Software Directory (RSD) or the ESCAPE-OSSR onboarding procedures) to identify the quality attributes that can already be assessed automatically. An additional issue in the evaluation is to find out the potential and effort for enhancement of existing tools to assess further quality attributes.

6.3 Piloting phase

The piloting phase is considered as a test and improvement of the quality assessment procedure and tools in real world conditions. All Helmholtz Centers are asked to contribute to the piloting phase to get comprehensive feedback about strength and weakness of the quality assessment approach. Iterative improvement will take place in this phase to develop a suitable and robust approach and toolset for quality assessment.

6.4 Final implementation within the Helmholtz Association

The suitable and robust approach and toolset for quality assessment will be implemented in the Helmholtz Centers.

The value of our procedure for quality assessment can be summarized following: Result of the procedure are multi-dimensional quality indicators for data and software publications. A multi-dimensional assessment rather than single numbers will better reflect the complexity and diversity of research outcome and lead to a fairer assessment. Our approach for developing the quality indicators is primarily guided by quality criteria not by technical tools, it starts with defining essential quality criteria and to then condense the list of criteria with respect to technical tool support. Following this approach, we ensure that quality criteria are directed to the intended objectives of quality assessment and not technical feasibility.

7. Conclusion

The described approach to derive multi-dimensional quality indicators for data and software publications has been initiated within the Helmholtz Association. The following main principles were guiding the developing process:

  • The indicator must match the quality criteria that are essential for the intended objectives, technical issues must not lead to deviation from that target. The objectives defined in the Helmholtz Association are: a) to create awareness and appreciation to the diversity of research output, b) to align the assessment of research and research practice to the conditions of digitalization and openness of science, and c) as a stimulus and incentive for scientists to improve their data and software publications related to specific quality criteria, and thus, to improve the quality and trustworthiness of science as a whole. Therefore, our basic approach for developing the quality indicator was to define the quality criteria that are essential for the intended objectives in a first step, and to then condense the list of criteria in a second step with respect to technical feasibility of automated collection.

  • The indicator should map a multi-dimensional suite of quality criteria onto one value, the quality indicator. The approach of solely using simple and one-dimensional metrics to assess research quality and performance is broadly criticized. Taking this into consideration the quality indicator for research data and research software publications should provide one easily comparable value that still does more then only counting - by representing a diverse set of quality features in itself.

  • The quality indicator has to match all types of research data publications and research software publications generated in the diverse scientific disciplines within the Helmholtz Association, generic approaches are required to provide answers to these questions.

  • The indicator should rely on generic well-established concepts for quality criteria and quality assessment to make use of already existing preliminary work.

The result of the conceptual development process is a generic multi-dimensional quality indicator based on a defined set of goal-oriented quality dimensions and attributes as well as well-established, maturity-based concepts for quality measurement.

The described approach to derive multi-dimensional quality indicators can be used not only to assess quality of data and software but also – in an analog way – the research endeavor as a whole. In the end, the adoption of a precise definition of these indicators, the fine-tuned weighting scheme of quality dimensions and attributes against each other, as well as the definition of a minimal level of maturity to be expected for research data and research software, will make the process of research assessment transparent and open. Moving towards multi-dimensional assessments rather than single numbers will better reflect the complexity and diversity of modern scientific work and, at the end, lead to a fairer assessment of the contributions of the various parts of the scientific system towards the common goal of contributing to increase collective human knowledge.

With this paper, we share the current state of the process within the Helmholtz Association. We invite all interested parties to further expand, refine, or change the dimensions and attributes being proposed here. It is our belief that any list of criteria has to be a living document to account for the dynamic nature of research. Only through constant evolution, the proposed framework is able to keep up with new thoughts, ideas, and findings on how to improve the scientific process.

Ethics and consent

Ethical approval and consent were not required.

CRediT author statement

Wolfgang zu Castell Conceptualization, Methodology, Writing – Original Draft

Doris Dransch Conceptualization, Investigation, Writing – Review and Editing, Supervision, Project administration

Guido Juckeland Conceptualization, Writing – Review and Editing, Visualization, Supervision

Marcel Meistring Conceptualization, Writing – Review and Editing, Supervision, Project administration

Bernadette Fritzsch Conceptualization, Writing – Review and Editing

Ronny Gey Conceptualization, Writing – Review and Editing

Britta Höpfner Conceptualization, Writing – Review and Editing

Martin Köhler Conceptualization, Writing – Review and Editing

Christian Meeßen Conceptualization, Investigation, Writing – Original Draft

Hela Mehrtens Conceptualization, Writing – Review and Editing

Felix Mühlbauer Conceptualization, Investigation, Writing – Original Draft

Sirko Schindler Conceptualization, Writing – Review and Editing

Thomas Schnicke Conceptualization, Writing – Review and Editing

Roland Bertelmann Supervision, Project administration

Comments on this article Comments (0)

Version 1
VERSION 1 PUBLISHED 10 May 2024
Comment
Author details Author details
Competing interests
Grant information
Copyright
Download
 
Export To
metrics
Views Downloads
F1000Research - -
PubMed Central
Data from PMC are received and updated monthly.
- -
Citations
CITE
how to cite this article
zu Castell W, Dransch D, Juckeland G et al. Towards a quality indicator for research data publications and research software publications – a vision from the Helmholtz Association [version 1; peer review: awaiting peer review] F1000Research 2024, 13:471 (https://doi.org/10.12688/f1000research.147663.1)
NOTE: it is important to ensure the information in square brackets after the title is included in all citations of this article.
track
receive updates on this article
Track an article to receive email alerts on any updates to this article.

Open Peer Review

Current Reviewer Status:
AWAITING PEER REVIEW
AWAITING PEER REVIEW
?
Key to Reviewer Statuses VIEW
ApprovedThe paper is scientifically sound in its current form and only minor, if any, improvements are suggested
Approved with reservations A number of small changes, sometimes more significant revisions are required to address specific details and improve the papers academic merit.
Not approvedFundamental flaws in the paper seriously undermine the findings and conclusions

Comments on this article Comments (0)

Version 1
VERSION 1 PUBLISHED 10 May 2024
Comment
Alongside their report, reviewers assign a status to the article:
Approved - the paper is scientifically sound in its current form and only minor, if any, improvements are suggested
Approved with reservations - A number of small changes, sometimes more significant revisions are required to address specific details and improve the papers academic merit.
Not approved - fundamental flaws in the paper seriously undermine the findings and conclusions
Sign In
If you've forgotten your password, please enter your email address below and we'll send you instructions on how to reset your password.

The email address should be the one you originally registered with F1000.

Email address not valid, please try again

You registered with F1000 via Google, so we cannot reset your password.

To sign in, please click here.

If you still need help with your Google account password, please click here.

You registered with F1000 via Facebook, so we cannot reset your password.

To sign in, please click here.

If you still need help with your Facebook account password, please click here.

Code not correct, please try again
Email us for further assistance.
Server error, please try again.