Data archiving system implementation in ITER's CODAC Core System

https://doi.org/10.1016/j.fusengdes.2015.06.076Get rights and content

Highlights

  • Implementation of ITER's data archiving solution.

  • Integration of the solution into CODAC Core System.

  • Data archiving structure.

  • High efficient data transmission into fast plant system controllers.

  • Fast control and data acquisition in Linux.

Abstract

The aim of this work is to present the implementation of data archiving in ITER's CODAC Core System software. This first approach provides a client side API and server side software allowing the creation of a simplified version of ITERDB data archiving software, and implements all required elements to complete data archiving flow from data acquisition until its persistent storage technology. The client side includes all necessary components that run on devices that acquire or produce data, distributing and streaming to configure remote archiving servers. The server side comprises an archiving service that stores into HDF5 files all received data. The archiving solution aims at storing data coming for the data acquisition system, the conventional control and also processed/simulated data.

Introduction

CODAC Core System (CCS) [1] is a software kit distribution for plant systems I&C. It is based on the widely used open source software EPICS (Experimental Physics and Industrial Control System) and CSS (Control System Studio). During the last year, the consortium formed by INDRA, CIEMAT, SGENIA and UPM, has been developing data archiving technologies (commonly referenced as DAN) to be included into ITER CODAC Core System. From the CCS version 4.3, a simplified implementation of data archiving [2] has been included.

This new development has been driven by the ITER operation and diagnostics requirements:

  • Steady state data archiving: ITER pulse length is over thirty minutes and all the data gathered by acquisition systems or data processor systems must be archived while they are being generated in the experiment.

  • Concurrent reading while writing: because of ITER experiments’ length, it is not possible to wait until each experiment finishes to read archived data, and thus it is mandatory that any archived solution enables concurrency at least between one writer and multiple readers.

  • Archiving on remote: data archiving services will be located in the CODAC system and data generator systems will archive on remote.

The DAN system in CCS is formed by client and server subsystems. In this work both will be presented.

HDF-5 [3] is a scientific-oriented file format with a complete set of libraries and management tools (supported by a wide scientific community), and with the main objective of making data platforms independent and portable. There is a list of important advantages that support the use of HDF-5 files for data archiving. Here are the most important ones:

  • Self-description data management: any client can completely know and manage the format of the data stored in a file. This feature also helps to keep version compatibility.

  • Huge file size support: there are no size limits apart from those derived from the operating system or file system.

  • Optimized performance for data segment access: HDF5 perfectly fits the data access requirements for scientific data archiving in long pulse experiments.

  • Data appending: it is possible to append new data to a file without copies or redefinitions.

Although HDF5 does not support concurrent write and read access by default, which is one of the most important ITER requirements, this feature is currently supported in a parallel version of the library called HDF5 SWMR (Single-Writer Multiple-Reader), and will be officially supported by HDF5 in the short term.

Section snippets

DAN system in CCS

The general data archiving architecture is formed by two main elements, shown in Fig. 1. On the one hand, the client systems (DAN clients) which generate data to be archived from a set of data sources (either acquisition or processing sources), and on the other hand, data archiving servers (DAN servers) that receive data from DAN clients and are responsible to store them in HDF5 files using the DAN writer services. DAN clients can be distributed and transmit data to the DAN archiving servers

DAN API

The DAN API is a library that implements an internal data transmission engine between processing components. It is included into the CCS “dan-daq” package with the other necessary libraries and management tools that will help developers incorporate the DAN functionality into their systems. Its present design and implementation make it ready for multi-processing operation, which will be extended to be thread compatible in the next future.

The DAN API is designed with the aim of minimizing data

Data archiving server

The DAN archiving server is the software that responds to the requests of DAN clients, configured on the basis of the streaming data structure. It receives stream data from the clients and stores them into HDF5 files. A DAN archiving server listens to the dedicated port of the DAN network and provides connections with the various DAN clients in order to transmit data from the DAQ devices to the central data storage. A DAN stream is the elementary communication channel between a DAN client and

Conclusions

The DAN software for client systems supports data from acquisition systems and data from simulation/processing elements. Thanks to the DAN API library, these data can be efficiently accessed not only for data archiving purposes but also for additional uses such as data processing or monitoring. Additionally, within the DAN archiving services, data are stored in HDF-5 files with special single-writer multiple-reader extension. This type of file is widely used by the scientific community and

References (3)

  • ITER Organization, CODAC Core System....
There are more references available in the full text version of this article.

Cited by (18)

  • Design for the distributed data locator service for multi-site data repositories

    2021, Fusion Engineering and Design
    Citation Excerpt :

    Under the Broader Approach (BA) activities, the ITER Remote Experimentation Centre (REC) has been preparing to make a full replication of ITER data at Rokkasho, Japan [13]. ITER data system and its access methods are named as ITERDB and the Unified Data Access (UDA), respectively [14–16]. ITER UDA adopts the broker-type facilitator model [9], in which all the client/server communications will be carried through the UDA server.

  • Data model implementation in ITER data archiving system

    2019, Fusion Engineering and Design
    Citation Excerpt :

    ITER data archiving technology has to archive different types of data coming from their systems. ITER data can be grouped in three main groups: DAN, SDN and PON [1,2], according to the network used for their distribution. DAN (Data Archiving Network) data come from acquisition systems, fast control and diagnostics.

  • J-TEXT distributed data storage and management system

    2018, Fusion Engineering and Design
    Citation Excerpt :

    This means the applications based on MDSplus and HDF5 can seamlessly migrate onto these types of storage [9,10]. The experiment data analysis tooling and workflow are well established on MDSplus and HDF5, which are very popular among fusion community [3,11]. But MDSplus and HDF5 are not designed for distributed file systems, they are optimized for block devices like hard drives or RAIDs (Redundant Arrays of Independent Disks).

View all citing articles on Scopus
View full text