Next Article in Journal
UAV-Based Servicing of IoT Nodes: Assessment of Ecological Impact
Next Article in Special Issue
Using a Smartwatch to Record Precordial Electrocardiograms: A Validation Study
Previous Article in Journal
Sustainable Health-Related Quality of Life in Older Adults as Supported by the vINCI Technology
Previous Article in Special Issue
Improved Estimation of Exercise Intensity Thresholds by Combining Dual Non-Invasive Biomarker Concepts: Correlation Properties of Heart Rate Variability and Respiratory Frequency
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Architecture Using 3D Vectorcardiogram to Detect R-Peaks in ECG with Enhanced Precision

1
Epsidy, 54000 Nancy, France
2
Ecole Nationale d’Ingénieurs de Sousse, LATIS-Laboratory of Advanced Technology and Intelligent Systems, Université de Sousse, Sousse 4023, Tunisia
3
IADI-Imagerie Adaptative Diagnostique et Interventionnelle, Inserm U1254, Université de Lorraine, 54000 Nancy, France
4
CIC-IT 1433, Inserm, CHRU de Nancy, Université de Lorraine, 54000 Nancy, France
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 2288; https://doi.org/10.3390/s23042288
Submission received: 9 January 2023 / Revised: 14 February 2023 / Accepted: 15 February 2023 / Published: 18 February 2023

Abstract

:
Providing reliable detection of QRS complexes is key in automated analyses of electrocardiograms (ECG). Accurate and timely R-peak detections provide a basis for ECG-based diagnoses and to synchronize radiologic, electrophysiologic, or other medical devices. Compared with classical algorithms, deep learning (DL) architectures have demonstrated superior accuracy and high generalization capacity. Furthermore, they can be embedded on edge devices for real-time inference. 3D vectorcardiograms (VCG) provide a unifying framework for detecting R-peaks regardless of the acquisition strategy or number of ECG leads. In this article, a DL architecture was demonstrated to provide enhanced precision when trained and applied on 3D VCG, with no pre-processing nor post-processing steps. Experiments were conducted on four different public databases. Using the proposed approach, high F1-scores of 99.80% and 99.64% were achieved in leave-one-out cross-validation and cross-database validation protocols, respectively. False detections, measured by a precision of 99.88% or more, were significantly reduced compared with recent state-of-the-art methods tested on the same databases, without penalty in the number of missed peaks, measured by a recall of 99.39% or more. This approach can provide new applications for devices where precision, or positive predictive value, is essential, for instance cardiac magnetic resonance imaging.

1. Introduction

It is estimated that over 300 million resting ECG tests are performed every year [1], to which exercise ECG tests, and a number of procedures involving ECG synchronization must be added. Such procedures may be imaging the heart or ablating electrical pathways. Over the last few decades, ECG has extensively been studied, focusing on detecting QRS complexes, and classifying features [2]. In an ECG, P-waves indicate atrial depolarization, QRS complexes correspond to ventricular depolarization and T-wave to ventricular repolarization (Figure 1). Automated analyses of ECG signals have been accelerated with the public availability of electronic ECG databases [3].
In ECG, QRS complexes consist of short-time pulses that contain R-peaks with steep slopes and high amplitudes. Algorithms to detect R-peaks and to measure heart rates are ubiquitous in medicine, and automated analyses of ECG are increasingly common. Automatically detecting R-peaks provides essential information to clinicians about the heart activity status. Furthermore, R-peaks play a key role in the subsequent classification of cardiac cycles and identification of abnormalities (e.g., arrhythmia), eventually supporting computer-aided diagnoses [4,5,6].
In this article, our focus is the detection of R-peaks in multi-lead ECG recordings. Arguably, the timings of R-peaks are shifted between leads; however, we aim to define a single R-peak timing representative of the onset of ventricular systole. There has been extensive research focusing on different ECG key features. Besides the detection of R-peaks, the most common approach has been to segment QRS complexes from their onsets to their offsets [6].
In good quality ECG recordings, classical algorithms can detect an R-peak with recall and precision both over 99%. Teaching these algorithms new use cases featuring high levels of noise or artifacts requires adapting their parameters or developing new ones, whereas deep learning (DL) architectures can be tailored through the adaptation of their training databases [7]. Compared with classical algorithms, DL architectures require large amounts of annotated data and high computing resources, which are not available in all circumstances. In the last decade, DL architectures have successfully been explored across many sub-fields following early successes in computer vision and pattern recognition [8,9]. Many studies confirmed that DL architectures outperform classical algorithms for signal analysis, image recognition, and object/pattern detection [10]. Many medical applications have benefited from the development and progress in computing, data availability, and DL to address a large variety of complex challenges [11]. Hence, DL architectures have recently become interesting alternatives to classical ECG signal analysis methods [2,7,12]. Furthermore, DL architectures can automatically identify relationships in data independently from handcrafted features [13]. DL is among the best techniques to detect patterns or to classify objects degraded by different noise levels or types [14]. One of the advantages of DL architectures in the analysis of ECG is that they can easily deal with single or multiple lead ECG recordings [7].
While many R-peak detection techniques have had a strong focus on reducing the number of missed detections (FN), our focus and motivation have been to avoid false detections (FPs). Some diagnostic or therapeutic medical devices use R-peaks as systolic triggers. One instance is cardiac magnetic resonance imaging (CMR): R-peaks trigger the pulse sequence, leading to the acquisition of raw (K-space) data. FPs cause data corruption leading to image artifacts. In noninvasive cardiac radioablations [15], R-peaks’ timings may be used to trigger irradiations of an arrhythmogenic substrate. Reliable R-peak triggers may enhance the precision of radiotherapy beams. Both cases require a low latency, leaving little to no time for a decision-making step. In other imaging, interventional, or robotic fields, high precision cardiac triggers may also be relevant.

2. Related Work

In this section, we review the main R-peak detection techniques. Classical algorithms were initially designed with the advent of electronic recordings and computerized analysis. More recently, DL architectures demonstrated groundbreaking performance in detecting R-peaks. This was further accelerated by the public availability of annotated ECG databases, such as MIT-BIH arrhythmia [16,17], and organized challenges, such as CinC [18,19,20,21,22] and CPSC [23,24,25,26]. Finally, hybrid approaches merging DL architectures, signal processing, and structural analysis techniques (e.g., clustering algorithms) have emerged. Methods to detect R-peaks can be categorized into: classical, DL, and hybrid.

2.1. Classical Approaches

Classical approaches are based on using various classical signal processing and structural analysis techniques. They usually feature pre-processing and decision-making steps. Classical algorithms are often distinguished according to their pre-processing steps, while the majority of the decision-making steps are either heuristic or based on machine learning (ML), such as support vector machine (SVM) [27] or decision tree algorithms [28], and dependent on pre-processing results.
Pahlm and Sörnmo reviewed pre-1984 methods based on defining rules and signal processing techniques for detecting QRS complexes in one-channel ECG recordings [29]. In 1985, Pan and Tompkins proposed the Pan–Tompkins (PT) algorithm based on using differential thresholds and setting the slope, amplitude, and width of a moving window for locating R-peaks [30]. Noise elimination, signal smoothing, and enhancing width and QRS slope were introduced in a pre-processing step. A decision-making step was used to determine whether or not a detection fits an R -peak. In 1986, Hamilton and Tompkins measured the effects of several parameters (median, iterative, or mean peak estimators) used in the decision-making step and proposed an optimized decision rule to discriminate true R-peak events from false, noise-induced events [31]. In 2022, Khan and Imtiaz introduced Pan-Tompkins++, an improved PT algorithm [32], by adding a filter with a band-pass of 5–18 Hz followed by an N-point moving average filter and different rules to adjust the thresholds based on the signal pattern particularities.
Kohler et al. classified QRS detection methods into several categories based on signal derivatives and digital filters, wavelets, and additional approaches (e.g., hidden Markov models [33]) [34].
Derivative-based algorithms define threshold levels from differentiator filters and difference operation methods [35,36,37,38,39]. For instance, Gutiérrez-Rivas et al. computed a dynamic threshold defined by a finite state machine (FSM) which depends on the sampling frequency of ECG recordings [40]. Typical frequency components of QRS complexes range from ~10 Hz to ~25 Hz. Many sophisticated digital filters have been explored, such as MaMeMi [41] or quadratic [42]. Many linear and non-linear transform operators have been applied in the pre-processing step, such as U3 [43], S [44], Hilbert [45], fast Fourier [46,47], wavelet [48,49,50]. Traditional Fourier or wavelet transform based methods are not suitable for all QRS morphologies: they are sensitive to intra- and inter-subject variations. To address this issue, Zhou et al. proposed a sparse representation-based ECG signal enhancement method that models ECG signals by a combination of inner structures and additive random noise [51] and then used these structures during the training phase to extract original ECG signals and to remove artifacts (e.g., wandering baseline). QRS complexes were identified in enhanced ECG signals as structures having the larger kurtosis values. The Zhou et al. algorithm yielded a high rate of FP [51]. Hossain et al. presented a complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN) for QRS complexes and P-waves detection without using conventional filtering techniques [52]. After carrying a pre-processing step (e.g. baseline drift removal), two different signals are reconstructed by the CEEMDAN method, one for QRS complexes and one for P-waves; the one matching the signal dynamics is selected.
Many algorithms based on the Shannon energy envelope were proposed to detect R-peaks with low delay and high accuracy and speed [53,54], provided that two key thresholds are correctly set. A fast R-peak detection method requiring low resource consumption was presented by Elgendi [55], using two moving averages which are calibrated by means of an optimized knowledge base with two different parameters. For QRS detections, a weighted total variation (WTV) denoising technique was proposed in the pre-processing step [56]. A regularization parameter in WTV minimization and weights is required to adapt locally the amount of applied smoothing and to select QRS complexes over P- and T-waves.
Several classical methods use multiple modalities, combining ECG with other channels, such as blood pressure (BP), electroencephalography (EEG), electrooculography (EOG), electromyography (EMG) and others, to detect R-peaks. For instance, Johnson et al. proposed a method based on defining a signal quality index derived from two different modalities: R-peaks detected in the ECG using energy and R-peaks detected in the arterial BP waveform using length transform [57]. From noisy multi-modal recordings, Gierałtowski et al. used more than two channels (ECG, BP, EEG, EOG, and EMG) for QRS detection [58]. Combining multiple channels requires adjusting rules accordingly. The performance of this approach varies, especially in ECG recordings of short durations.
Liu et al. compared the performances of ten widely used QRS detection algorithms in different cases [59]: PT [30], Hamilton-mean [31], Hamilton-median [31], RS slope [58], sixth power algorithm [37], FSM [40], U3 transform [43], difference operation algorithm [38], window-based peak energy detector [57,60,61], and optimized knowledge base [55]. To guarantee high performance, pre-processing and post-processing steps were usually required, at an increased computing cost. Other reviews and benchmarks focused on pre-processing and decision-making steps [62,63,64]. Van and Podmasteryev reviewed in detail four ML-based algorithms [65], including artificial neural network (ANN) [66], k-nearest neighbor (k-NN) algorithm [67], k-means [68], and SVM [69]. Performances of investigated ML algorithms were shown to depend on the quality of their training datasets, which was an essential drawback.
Few approaches were designed to work with multiple lead ECG recordings. Śmigiel et al. used 6 classical detectors and a k-means clustering method after processing a 12-lead ECG signal [70]. Chen et al. used the PT algorithm to detect R-peaks in each lead of the ECG recordings of the LUDB database [71,72,73].
The main disadvantage of classical algorithms is that they depend on many empirical parameters (e.g., self-adaptive threshold values) and lack robustness, particularly with low-quality, noisy, or pathological signals, changes in sensors or different QRS morphologies. PT for instance, one of the most widely used algorithm to detect R-peaks in commercial cardiac monitoring devices, is sensitive to various sources of noise, including wandering baseline, power-line interferences, muscle artifacts, and electrode contact noise. Moreover, they are often unsuitable for ECG recordings of short durations.

2.2. DL-Based Approaches

Several DL architectures have been proposed for different tasks related to computer vision and pattern recognition [8,9], including ECG analysis [2,4]. A DL architecture is an ML algorithm based on a neural network (NN). It works by learning a correlation between input features using a large amount of data. The larger the database, the better the DL architecture can learn by optimizing its weights and perform the targeted task by taking the most suitable decision. Commonly used DL architectures to detect R-peaks are: convolutional neural network (CNN) [74], recurrent neural network (RNN) [75], long short-term memory (LSTM) [76], and gated recurrent unit (GRU) [77].
CNN-based solutions represent the most used techniques for ECG analysis, and particularly for feature extraction tasks [2]. Šarlija et al. reported a 1D CNN-based architecture exceeding 99% in recall and precision on unseen data from the MIT-BIH arrhythmia database [78]. Oudkerk-Pool et al. reported 92.6% recall, 91% precision, and 91.8% F1-score on 100 recordings from the PhysioNet/CinC challenge 2017 using a fully convolutional dilated NN [20,79]. Tison et al. used a modified CNN (U-Net) architecture for automating the classification of six types of ECG segments (P-wave, PR segment, QRS complex, ST segment, T-wave, and TP segment) [80]. Using the MIT-BIH arrhythmia database and a modified U-Net architecture, Oh et al. reported a high recall of 98.76% but a low precision of 29.55% [81]. Jimenez-Perez et al. used a U-Net architecture to analyze dual lead ECG recordings from the QTDB database and achieved a recall of 99.94% [82,83,84].
Some novel DL-based methods combine variant CNN architectures with other DL models. Liu et al. achieved high accuracy by using three variants of bidirectional long short-term memory (BLSTM) architectures: with attention model, combined with U-Net, and combined with U-Net++ [85]. Peimankar and Puthusserypady demonstrated a recall of 97.95% and a precision of 95.68% on 105 recordings from QTDB by combining a CNN architecture with a BLSTM model to detect onset, peak, and offset of different heartbeat waveforms (P-waves, QRS complexes, and T-waves) [86]. Vijayarangan et al. proposed a fused CNN-ResNet, called RPnet, an architecture combining the 1D U-Net model with inception and residual modules, to extract R-peaks from noisy ECG [87]. This approach contributed to alleviate the vanishing gradient issues of DL architectures. RPnet outperformed three classical algorithms (Hamilton and Tompkins [31], Christov [35], and stationary wavelet transform [50]), with an F 1-score of 98.37%, on the second CPSC database [24]. Duraj et al. showed high performance on LUDB by incorporating the residual and squeeze-excitation blocks into the 1D U-Net architecture for extracting segments, such as P-waves, QRS complexes, and T-waves, regardless of the lead [88]. Gabbouj et al. proposed a 1D self-organized operational neural network (ONN), evaluated on second CPSC, and achieved recall of 99.79%, precision of 98.42%, and F1-score of 99.10% [24,89]. However, during the training phase, it featured a high complexity in the number of multiply–accumulate operations and the number of parameters compared with 1D CNN architectures.
Many DL architectures are computationally expensive and hence not well-adapted to embedded devices. Furthermore, CNN and their variants, such as U-Net, are the most common DL architectures for R-peak detection. Yet, many researchers stated that the best accuracy was achieved by combining different DL architectures, at a cost of increasing complexity [90]. Single ECG lead DL architectures already outperform classical algorithms, but few use multiple leads. Increasingly available 12-lead ECG databases may increase the precision and robustness of DL-based approaches.

2.3. Hybrid Approaches

To refine the performances of DL-based approaches, signal processing and structural analysis techniques, such as clustering algorithms, ensemble learning methods, and heuristic approaches, can be added into a post-processing step.
Sereda et al. introduced a pre-processing step (wandering baseline removal) [91]. Then, they used a 12 CNN set on LUDB to correct errors occurring in a single CNN. They reported an F 1-score (over 95%) higher than when using a single CNN (94%). In multi-lead ECG recordings, Moskalenko et al. proposed to process the DL output of each lead independently and to average the resulting scores [92]. Their method demonstrated an F 1-score of 99.97%, outperforming other methods, such as the 12-lead set of Sereda et al. [91] or the wavelet-based method of Kalyakulina et al. [12].
To validate detected R-peaks in a single ECG lead from the MIT-BIH arrhythmia and third CPSC databases, Zahid et al. carried out a verification model based on a timing criterion, considering the following hypothesis: if the predicted R-peak locations of two beats fall within 300 ms, one is an F P [25,93].
In multi-lead ECG signals, Han et al. introduced a post-processing step based on an adaptive dynamic threshold strategy and electrophysiology knowledge, after applying a linear ensemble method averaging the outputs of two DL models [94]. They used CNN and LSTM models, both based on the U-Net architecture. A pre-processing step consisted of removing the wandering baseline and high frequency noise based on discrete wavelet transforms and a third order Butterworth band-pass filter. They reported recall and precision, ranging from 99.45% to 100%, on QTDB, LUDB, and CCDD databases [95,96,97]. Their strategy effectively reduced FN and FP; however, it did not correctly detect some arrhythmic QRS complexes, such as left bundle branch blocks.
Cascading several steps of variable complexities may require extensive computational power, and make the application of hybrid approaches difficult in embedded devices. Furthermore, introducing thresholds, heuristics, and rules in post-processing steps may reduce their generalization capacity to other data and reduce their robustness, particularly in the case of pathological QRS morphologies.

3. Proposed Approach

The originality of the proposed R-peak detection approach lies in adapting, training, and testing a DL architecture using a 3D VCG instead of a single or multiple lead ECG, without pre-processing or post-processing steps.

3.1. VCG Transformation

There is an inherent complexity in ECG signal analysis due to the plethora of configurations of ECG recording electrodes. R-peaks can have positive or negative polarities depending on the lead position, and often, the underlying cardiac pathology. In the 1950s, researchers introduced 3D VCG as a simplified representation of the electrical activity of the heart along three vectors (X, Y, and Z) [98].
Several mathematical transformations allow reconstructing a 3D VCG from a 12-lead ECG [99] based on a matrix product:
V = E × M
where V, E, and M denote a 3D VCG vector, a vector representing individual ECG leads, and a transformation matrix, respectively.
Leads III, aVR, aVL, and aVF, obtained by trivial linear combinations of leads I and II, can be eliminated. To convert a 12-lead ECG into a 3D VCG, a transformation was performed according to Equation (1), where E represents the 8 independent ECG leads (I, II, and V1 to V6) and M denotes the matrix obtained by Kors regression transformation [100]:
M = 0.38 0.07 0.11 0.07 0.93 0.23 0.13 0.06 0.43 0.05 0.02 0.06 0.01 0.05 0.14 0.14 0.06 0.20 0.06 0.17 0.11 0.54 0.13 0.31
Comparing the Kors quasi-orthogonal transformation with three other VCG transformations (inverse Dower, Kors regression, and Frank’s orthogonal lead system), Kors et al. reported the best results with the Kors regression transformation [100].
Figure 2 illustrates the Kors regression transformation in a recording from the INCART database. In the figures below, all signals were scaled between −1 and 1. All annotated R-peaks were highlighted using two yellow bars delineating a five-sample segment.

3.2. DL Architecture

Normalized 3D VCG segments of durations ~4 s (2048 samples at 500 Hz) were fed as input to the DL architecture. A short sample size was decided after initial experiments showing no negative impact on precision. A power of two was required for the encoding/decoding cascade of the DL model. The proposed approach can be performed in any practical durations of ECG recordings, using shorter segments of ~1 s for instance. R-peak times may be shifted between leads, and in the proposed approach, the timings of lead II were used, together with an acceptance window of ± 75 ms [59,93,101]. Neither pre-processing nor post-processing steps were used. Hence, the proposed approach can be generalized to low-quality, noisy, or pathological signals acquired from any sensors providing three independent directions of ECG data.
R-peak detection was approached as a segmentation task using the U-Net model. U-Net, used first for 2D medical image segmentation in 2015 by Ronneberger et al., is a DL model following the encoder–decoder architecture and is reputed for its high accuracy and reduced computational complexity [102].
The proposed DL architecture consists of convolutional layers (Conv1D), distributed along two symmetric contracting and expanding paths, which focus on the encoding and decoding processes, respectively. The contracting path focuses on extracting high level abstraction features (context information involving local information) using the convolution and pooling layers to reduce the input dimensionality. The expanding path applies the opposite operations, often called deconvolution (Conv1DTranspose). It is leveraged for precise localization by combining simultaneously global and contextual information captured from the contracting path through skip connections. Indeed, it restores the characteristics of the high level abstraction without information loss using skip connection and up-sampling processes in order to achieve an accurate semantic segmentation task. The bottleneck is the intermediate part between the contracting and expanding paths that contains an encoded information of the input data representing the latent space (relevant information to be able to reconstruct the input data).
Figure 3 illustrates the proposed DL architecture. A detailed description of its layers is provided in Appendix A.
The U-Net model used in this work is composed of multiple hidden layers to guarantee an optimized learning of high-level feature representation of ECG data. It also features a small number of trainable parameters (79,409) compared to other DL models (e.g., transformer [103]). The contracting path down-samples inputs through six convolutional layers with a down-sampling factor of two. The convolutional layers had filters with the same padding. For every 2 consecutive layers, we set the kernel size to 9, 6, and 3 and the number of filters to 16, 32, and 64. Each convolution layer of the last five is followed by a batch normalization layer. A 1D MaxPooling layer was added after each batch normalization layer to down-sample the 1D feature map by a factor of 2. The expanding path decompresses the feature map back to its original size with a reverse configuration of the contracting path. Only the first convolution layer of the expanding path has a dropout layer with a drop rate of 25%. All convolution layers of the proposed DL architecture have LeakyReLU activation functions (with a negative slope coefficient of 25%), except the last one which has a Sigmoid activation to get 1D segmentation maps for R-peaks. We used a filter size of 1 in the convolution operation of the last layer of the expanding path to output a 1D segmentation map. The proposed DL architecture outputs the probability value for every sample point, from which the R-peaks in ECG were centered in a five-sample segment during the annotating process.
To overcome issues related to missed detections of R-peaks located on segments borders, a stride of ~2 s (1024 samples at 500 Hz) was used when extracting 3D VCG segments of ~4 s. Thus, we computed average predictions from overlapping predictions for every sample point of the original signal. Afterward, we filtered predictions by selecting those that were above the probability threshold of 50%. Selected predictions were subsequently adjusted to coincide with a local maximum. Finally, we retained R-peaks if at least five sample points had the same local maximum.
The parameter settings of our DL architecture were first determined based on recently published works [93], and empirically validated. Multiple experiments were carried out, and the best parameters for our setup were retained.

4. Experiments

To evaluate the performance of the proposed approach, experiments were conducted on 4 publicly available 12-lead ECG databases, using leave-one-out cross-validation and cross-database validation protocols.

4.1. Experimental Corpora

In the context of ECG analyses, not all available databases provide expert-validated annotations of R-peaks locations. Furthermore, public databases are often built using single or dual lead ECG recordings, such as MIT-BIH arrhythmia [16,17], QTDB [83,84], and third CPSC [25] databases.
PTB Diagnostic ECG [104], PTB-XL [105,106], INCART [21,107], first CPSC [23,108], SPH [109], LUDB [72,73], CSE [110], CCDD [95,96,97], and Georgia [111] feature 12-lead ECG recordings. However, the number of publicly available databases having both 12-lead ECG recordings and expert-validated annotations is limited. To the best of our knowledge, CCDD, LUDB, and INCART are the only publicly available databases with the 12 conventional leads (I, II, III, AVR, AVL, AVF, and V1 to V6) and expert-validated R-peak annotations. These 3 databases, more precisely, all recordings from LUDB and INCART, and the first 251 recordings from the 943 publicly available recordings from CCDD were used in our experiments. Another 103 recordings were pseudo-randomly extracted from the remaining CCDD recordings to create an additional database (CCDD-Extra) which was never seen in the training phase of our DL architecture.
1.
Chinese Cardiovascular Disease Database (CCDD): is composed of 193,690 12-lead ECG datasets collected and annotated by 2 cardiologists [95,96,97]. Only 943 complete annotated recordings are publicly available. The positions of the QRS complexes, onsets, peaks, offsets for P and T-waves were provided. Among the 943 recordings, the first 251 recordings are usually selected by researchers in their experiments, since the corresponding annotations were made following the same protocol [94]. Based on the set segment size and stride, 753 windows with a total of 3973 R-peaks were obtained and used in our experiments. To appropriately evaluate the proposed approach and conduct an objective comparison with other methods, few adaptations of the annotations of QRS complexes were required. In the first and/or last cycles of the recordings, a few QRS complexes were not annotated. We discarded them, considering only annotated cycles. All samples located inside a QRS interval were annotated, while the proposed approach required only one R-peak location. A k-means algorithm was hence applied to extract one precise R-peak location per QRS interval after setting the number of R-peaks. Figure 4 illustrates a 12-lead ECG recording example of the CCDD database.
2.
Lobachevsky University Electrocardiography Database (LUDB): is composed of 200 12-lead ECG recordings with a duration of 10 s, and a sampling rate of 500 Hz, recorded using a Schiller Cardiovit AT-101 cardiograph (Schiller AG, Baar, Switzerland) and released on the PhysioNet Website in 2020. It features a large variety of QRS morphologies, from healthy volunteers to patients with different cardiovascular diseases, some of them with pacemakers. All boundaries and peaks of P-waves, QRS complexes and T-waves were manually annotated by cardiologists for each of 12 leads independently. We considered the annotations of lead II in our experiments. Based on the set segment size and stride, 400 windows with a total of 1831 R-peaks were used. In the first and last cycles of the recordings, some R-peaks were not annotated. We discarded them, considering annotated cycles only during the evaluation of the proposed approach. Figure 5 illustrates a 12-lead ECG recording example of LUDB.
3.
St. Petersburg Institute of Cardiological Technics 12-lead Arrhythmia Database (INCART): is a public 12-lead ECG database sampled at 257 Hz and released on the PhysioNet Website in 2008. It comprises 75 half-hour recordings extracted from 32 Holter recordings which were collected from patients undergoing tests for ischemia, coronary artery disease, conduction abnormalities, and arrhythmia. Based on the set analysis window size and stride, 32,925 windows with a total of 175,907 R-peaks were used in our experiments. Annotations were automatically produced by using an algorithm that detected beat annotations from all 12 leads in the middle of the QRS complexes, and then a few automatic annotations were corrected manually. Hence, there are misaligned annotations in INCART. Figure 6 illustrates a 12-lead ECG recording from INCART.
4.
CCDD-Extra: is a set of 103 recordings pseudo-randomly selected from the 692 ( 943 251 ) remaining recordings of the CCDD database. Annotations of these recordings followed the same protocol as CCDD (i.e. extracting the positions of R-peaks with the k-means algorithm). Based on the set analysis window size and stride, 309 windows with a total of 1616 R-peaks were used. Figure 7 illustrates a 12-lead ECG recording example of the CCDD-Extra database. CCDD-Extra was used to evaluate the performance of our DL architecture in a cross-database validation protocol.
Table 1 summarizes the features of the four databases used in our experiments.

4.2. Experimental Protocol

After converting the ECG recordings into 3D VCG, a set of thorough experiments were conducted to evaluate and validate the proposed approach. They followed the experimental protocols described below:
1.
Leave-one-out cross-validation protocol: was applied to the CCDB, LUDB, and INCART databases independently. One recording was extracted for the evaluation phase, while the remaining recordings of the same database were used for the training phase. For instance, in the case of the INCART database, 74 recordings were used to train our DL architecture, and the 75th remaining one was used for evaluating its performance. Then, an average of the 75 different evaluation experiments was computed.
2.
Cross-database validation protocol: was carried out by evaluating the proposed approach on a database different from those used in the training phase to demonstrate its generalization capacity [112,113]. In our experiments, all 526 recordings ( 251 + 200 + 75 ) of CCDB, LUDB, and INCART databases were used to train the DL architecture, while the 103 recordings of CCDD-Extra were used for its evaluation. A total of 34,078 windows and 181,711 R-peaks were used during the training phase.
For both protocols, the same hyperparameters, summarized in Table 2, were selected. For stochastic optimization, an Adam optimizer was used with an initial learning rate of 0.001 and a batch size of 64. Weights were randomly initialized with the Xavier uniform distribution. Training was stopped early only when no more improvement in the cross-entropy loss value was recorded for at least 10 epochs (i.e. patience of 10).
Python version 3.10 was used to implement the proposed approach, and Keras version 2.9 was employed for training the DL architecture. Our experiments were carried out on a physical computer equipped with 2.4 GHz Intel® Core™ i9-12900F having 32 GB of RAM and a single NVIDIA® GeForce RTX™ 3060 with 12 GB of RAM. Google Colaboratory servers were also used (Tesla T4 GPU with 16 GB of RAM). The training phase was processed using CUDA kernels, while the testing phase was implemented with a single CPU. Training on all data from CCDD, LUDB, and INCART took 7274 s on the computer (resp. 4297 s on Google Colaboratory servers). As an example, testing the proposed approach on a ~13 s recording (6484 samples at 500 Hz) of CCDD-Extra took 1.15 s on the computer (0.555 s on Google Colaboratory servers) to detect 13 R-peaks.
Figure 8a–c show the loss curves obtained using the leave-one-out cross-validation protocol. Figure 8d shows the loss curve obtained using the cross-database validation protocol.
All 4 learning curves in Figure 8 demonstrate a good fit as they decrease to a point of stability (500 epochs for both CCDD and LUDB, 78 epochs for INCART, and 479 epochs for the three databases). Moreover, the DL architecture trained on INCART shows superior loss convergence properties (i.e. speed) compared to using CCDD or LUDB. This can be explained by the fact that data in INCART were most representative to suitably train a DL architecture: its statistical characteristics were correctly captured. Indeed, the number of annotated R-peaks in INCART is ~44× (resp. ~96×) larger than in CCDD (resp. LUDB). Furthermore, due to the data heterogeneity yielded by combining the three different databases during training, the minimum loss value (0.00781) is higher in the cross-database validation protocol compared to the loss values obtained in the leave-one-out cross-validation protocol (Figure 8).
To validate the 3D VCG approach, the performances of individual lead-by-lead trainings of the DL architecture were assessed, first using representative ECG leads (I, II, and V1 to V6), and second, using individual 3D VCG vectors (X, Y, and Z). Multi-lead trainings were performed using all 8 ECG leads (I, II, and V1 to V6) as input and the 3 best leads from the first experiment.

4.3. Performance Evaluation Metrics

To quantify the performance of the proposed approach and compare it with other methods, three standard performance evaluation metrics, defined below, were adopted [6].
1.
Recall (or sensitivity): evaluates the rate of defined R-peak annotations that are correctly predicted.
Recall ( % ) = TP TP + FN × 100
2.
Precision (or positive predictive value): evaluates the rate of predicted R-peaks that are correctly matched to defined annotations.
Precision ( % ) = TP TP + FP × 100
3.
F1-score: corresponds to a harmonic mean of recall and precision.
F 1 - score ( % ) = 2 × Precision × Recall Precision + Recall × 100
where TP, FN, and FP denote the number of true positives, false negatives, and false positives, respectively. Since the maximum duration of QRS complexes is 150 ms, a tolerance of ± 75 ms of the annotated R-peak location was considered when counting TP, FN, and FP [59,93,101].

5. Results

The performance of the proposed approach was evaluated on 4 different public 12-lead ECG databases.

5.1. Qualitative Results

Figure 9, Figure 10, Figure 11 and Figure 12 illustrate the R-peaks detected using the proposed approach on recordings from CCDD, LUDB, INCART, and CCDD-Extra, respectively. Detected R-peaks are marked with red triangles.
Visual inspection of the qualitative results revealed that the proposed approach for detecting R-peaks is reliable and robust. First and/or last cardiac cycles were sometimes not annotated in some ECG recordings and were discarded from the analysis, as can be seen in Figure 9 and Figure 10. Figure 11 demonstrates that the proposed approach succeeded in detecting the R-peak located between 8 s and 9 s, despite a different QRS morphology compared with its predecessors and successors. In Figure 12, the proposed approach detected all R-peaks from an ECG recording of the CCDD-Extra database, never seen during the training phase. This reproducible result, along with quantitative results described below, confirmed that the proposed approach has the capacity to be generalized.
To illustrate some failed detections, a recording from the INCART database shows noice-induced FP in Figure 13. Another source of FP is pacemaker patients (LUDB). However, these errors were limited as shown in the quantitative analysis. Missed detections (FN) are mostly related to pathological QRS morphologies that are less often represented in the training datasets. They were mostly seen in CCDD and INCART databases.

5.2. Quantitative Results

Since a visual assessment of the effectiveness and robustness of a method is inherently subjective, we evaluated the accuracy of the locations of detected R-peaks by computing the three metrics described in Section 4.3. Numbers of TP, missed beats (FN), and false detections (FP) are reproduced when relevant.

5.2.1. Quantitative Results on CCDD, LUDB, INCART, and CCDD-Extra Databases

In the leave-one-out cross-validation protocol, the proposed approach produced results with recall ranging from 99.39% to 99.84% and precision ranging from 99.88% to 100.00% (Table 3). In the cross-database validation protocol, the proposed approach produced results with recall of 99.41% and precision of 99.89% (Table 4).

5.2.2. Training on 3D VCG vs. Single Lead, Single VCG Vector, Three Best Leads, or Eight Leads

The DL architecture was trained on 3D VCG data and could also be trained on a different number of leads, extracted or reconstructed from the 12-lead ECG recordings. Four different experiments were carried out, using the leave-one-out cross-validation protocol and the INCART database. Results are summarized in Table 5.
In the first and second experiments, the DL architecture was trained on one lead/vector at a time. Independent leads (I, II, and V1 to V6) and 3D VCG vectors (X, Y, and Z) were used. The testing of the obtained models, following the leave-one-out cross-validation protocol, consisted in 75 × ( 8 + 3 ) trainings and as many experiments.
None of the models trained on independent ECG leads (I, II, and V1 to V6) outperformed recall or precision of the DL architecture trained on combined 3D VCG data. In single lead training, recall and precision were highest with leads II, V2, and V5.
When using individual X, Y, and Z vectors for training, results were good, with recall ranging from 99.39% to 99.82% and a precision of 99.98%. It was interesting to see that models trained on the X, Y, and Z vectors were on par with the most effective individual leads from the experiments above, with X and Y vectors providing slightly better results. The 3D VCG trained model outperformed any of the models trained on individual X, Y, and Z vectors.
A third experiment, called the 03-leads experiment, consisted of training the DL architecture with leads II, V2, and V5 combined. These leads were chosen because they provided the best results in the first experiment. The 03-leads experiment demonstrated a lower number of FP than the 3D VCG trained model (9 vs. 16). However, a larger number of FN were observed (393 vs. 276). Comparing the performances of the 03-leads experiment with those of the 3D VCG trained model, recall was 99.77% vs. 99.84%, and precision 100.00% vs. 99.99%.
A fourth experiment, called the 08-leads experiment, consisted of training the DL architecture with the combination of all 8 independent leads. The 08-leads experiment demonstrated a larger number of FP (25 vs. 16), and the number of FN were 6× higher (1633 vs. 276). Comparing the performances of the 08-leads experiment with those of the 3D VCG trained model, recall was 99.14% vs. 99.84%, and precision was 99.99% in both cases.

5.2.3. Comparison with the PT Algorithm

To compare our results with the PT algorithm, we selected an open-source implementation from [114]. Results from PT were compared with the results of the proposed approach in the leave-one-out cross-validation protocol (Table 6).
Regarding missed detections (FN) and recall, the PT algorithm performance was superior to the proposed approach in several leads (II, V2, V3, V5, and V6) for CCDD, in one lead (V3) for LUDB, and was inferior in all leads for INCART. Regarding false detections (or FP) and precision, the PT algorithm performance was inferior to the proposed approach, regardless of the chosen lead. Average PT values of recall were comparable or inferior to the proposed approach (97.73% to 99.50% vs. 99.39% to 99.84%). Average PT values of precision were always inferior to the proposed approach (92.31% to 99.14% vs. 99.88% to 100.00%).
Results from PT were also compared with the results of the proposed approach in the cross-database validation protocol (Table 7). This is more fitting than the previous comparison as in both cases, the algorithms used to analyze the data were not previously exposed to them. The PT algorithm outperformed the proposed approach in recall in all leads except V6 (99.84% vs. 99.41%). However, its precision was inferior, regardless of the chosen lead (98.42% vs. 99.89%). The two best leads for precision were leads II and V4 (98.96%), exhibiting 16 FP vs. 2 for the proposed approach.

5.2.4. Comparison with Other Methods

In the literature, few R-peak detection methods were evaluated using the same databases. We selected recent methods to compare their performances with ours [71,94,115].
In CCDD, the hybrid method from Han et al. [94] outperformed the proposed approach in recall (99.99% vs. 99.39%), but its precision was inferior (99.45% vs. 100.00%) (Table 8).
In LUDB, the classical algorithm from Chen et al. [71] outperformed the proposed approach in recall (100.00% vs. 99.75%) and was equivalent in precision (99.86% vs. 99.88%) (Table 9).
In INCART, the performance of the classical algorithm from Schmidt et al. [115] was inferior to the proposed approach both in recall (99.43% vs. 99.84%) and in precision (99.91% vs. 99.99%) (Table 10).

6. Discussion

This work proposes a DL-based approach to detect R-peaks. The novelty of the proposed approach consists in using reconstructed 3D vectorcardiogram (VCG) leads [100] as input instead of single or multiple ECG lead(s). A 3D VCG is obtained by a linear combination of multiple ECG leads. It reduces the computational complexity (i.e. memory and time requirements) both for training and inference compared to using all available leads. Using a 3D VCG as input during the training phase of a DL architecture has the advantage that both strong temporal correlation features and fine morphological features are extracted. Indeed, a 3D VCG provides a calibrated maximal QRS vector within a specific timing. In contrast, when using multiple ECG leads, each scalar lead has its own sensitivity and timing for the R-peak maximum, which could bias data and subsequently results [116].
The proposed approach has an enhanced precision compared to other methods. The number of false detections (FP) was low without the use of any post-processing step. The number of missed beats (FN) yielded a recall on par with or better than published classical algorithms, DL architectures, or hybrid approaches. Without any post-processing step, the proposed approach outperformed the PT algorithm in precision, dividing the number of FP by a factor eight or more. Reducing FP has an important implication when synchronizing diagnostic or therapeutic devices which rely on a systolic trigger, and assume the heart to be in a given phase to perform certain tasks. In CMR, triggering data acquisition by a false detection could lead to severe imaging artifacts impairing proper image interpretation [117,118].
The proposed approach could have applications in automated measurements of clinical features [119]. For instance, heart rates and heart rate variability measurements were computed based on R-peak detections, using the leave-one-out cross-validation protocol in the INCART database [120]. The obtained values are almost identical to values obtained from ground truth annotations (see Table 11). This is further evidence of the clinical validity of the proposed approach.
The proposed approach was thoroughly evaluated on multiple 12-lead ECG databases in leave-one-out cross-validation and cross-database validation protocols. The cross-database validation protocol, exposing the proposed approach to data never seen in the training phase, demonstrated a tiny drop in performance: recall dropped from 99.66% to 99.41% and precision from 99.96% to 99.89%. This shows the robustness of our approach, acquired from training on a large number of ECG cycles (181,711). However, the databases used for training were unbalanced. The INCART database features a larger number of ECG cycles (175,907) than LUDB (1831) or CCDD (3973); it also has less patients: 75 vs. 200 (LUDB) or 251 (CCDD). More balanced ECG databases may further improve the results and robustness of the trained model. It could also provide more insights to fine-tune the model hyperparameters. The CCDD-Extra database used in the cross-database validation protocol was not completely independent from the CCDD database used for training. It was acquired using the same technique as CCDD, yet consisted of different patients. Testing the proposed approach on ECG databases acquired with different parameters may provide additional insights on its accuracy.
The hybrid method from Han et al. [94] and the classical algorithm from Chen et al. [71] outperformed the proposed approach. They used pre-processing or post-processing steps. Han et al. applied a post-processing step based on electrophysiology knowledge, after combining CNN and LSTM to detect QRS complexes [94]. It reduced the number of missed and false detections at the cost of an increased complexity. Chen et al. applied several pre-processing steps, including an adaptive PT algorithm, to determine the ranges and locations of QRS complexes in each lead and to remove false QRS locations taking other leads into account [71]. The recall of the proposed approach was comparable but slightly inferior to these two methods; however, its precision exceeded theirs.
Qualitatively examining the results, we observed that missed beats (FN) were mostly attributed to extrasystoles. This can be explained by the insufficient statistical representations of such QRS morphologies in the training databases. Zahid et al. identified this behavior and used a data augmentation strategy, generating additional extrasystole segments to address issues related to limited training examples [93]. False detections (FP), although extremely rare, usually happened in situations of noise, artifacts, or pacemaker pulses. They could be dealt with in a post-processing step, which could also reduce the number of FN. However, avoiding any post-processing is an advantage, especially for applications requiring a low latency (e.g., CMR).
Our hypothesis was to use a 3D VCG reconstruction scheme to avoid the arbitrary decision of selecting which lead(s) to use. A 3D VCG can be reconstructed from many ECG acquisition schemes and provides increased robustness to a large variety of changes, such as errors in positioning electrodes, artifacts, acquisition protocols, and devices. Results show that a 3D VCG is an optimal strategy in our application. The experiments resulted in one unexpected outcome that 3D VCG is even better than using all eight independent leads from the full ECG signal. A possible, although partial, explanation is that the noise or artifacts from individual leads are smoothed out in a 3D VCG. Another one is the weighting of individual leads: in a 3D VCG, the precordial leads are captured in the Z vector with a weight of 2 / 3 , whereas in the 08-leads experiment, their weight is 6 / 8 . Scaling down on the number of dimensions is an advantage to reduce training and testing complexity. A 3D VCG seems to provide a good compromise and avoids the arbitrary decision to pick the best lead(s). It can potentially improve the automaticity of detecting R-peaks in a variety of use cases.
Another hypothesis was to define a single QRS label (or timing) representative of the onset of ventricular systole. We used the expert-validated annotations from lead II as the gold standard. This choice may have had a negative impact on the accuracy of the PT algorithm in other ECG leads. It could also have reduced the performance of the DL architecture after training on other ECG leads. Finally, it may explain why 3D VCG vectors X and Y have the highest accuracy. To enhance the performance of our DL model, a refined timing of R-peaks could be modeled on a lead-by-lead basis.
To demonstrate both robustness and generalization capacity of the proposed approach in detecting premature ventricular contraction or supra-ventricular premature beats, we evaluated its performances on the testing set of the MIT-BIH arrhythmia database (DS2) after training it using lead II data from INCART only and compared them with other state-of-the-art R-peak detection methods (Table 12).
The DL architecture used in this work, a modified U-Net, has the potential to run with low latency [93]. It could be implemented on a powerful computer or on a GPU-capable edge device, and run in quasi real-time. Since it does not require CPU intensive pre-processing nor post-processing steps, it could be applicable in a number of use cases. Furthermore, adapting the training database, and thus the trained model, could provide similar accuracy in various use cases. One example of use cases is CMR, where the ECG signal suffers from a magneto-hydrodynamic artifact [121,122]. This artifact could be added to the training datasets. Other use cases featuring high noise could also be addressed by adding similar noise features to the training datasets. A performance evaluation of the impact of adding synthetic noise using a signal to noise ratio metric could be developed to assess the robustness of the proposed approach [123].

7. Conclusions and Further Work

In this article, we propose a novel approach for detecting R-peaks based on training a DL architecture with 3D VCG data, without pre-processing nor post-processing steps. R-peak detection was approached as a segmentation task using the U-Net architecture which was fed with normalized 3D VCG segments of durations ~4 s (2048 samples at 500 Hz). The proposed approach can be performed in ECG recordings of any practical durations (e.g., from tens of minutes down to a few seconds) and is robust to low-quality, noisy, or pathological signals acquired from different sensors. It is also robust to various QRS morphologies.
Using the proposed approach, high performances were achieved in leave-one-out cross-validation and cross-database validation protocols. Furthermore, false detections (FP) and missed beats (FN) were significantly reduced compared to recent state-of-the-art methods or the PT algorithm. This approach could be adapted for several use cases and seems relevant to synchronize medical devices with R-peaks with a low latency.
In future work, we intend to embed the proposed approach on different hardware backends (e.g., Raspberry Pi, Jetson Nano, …), and to investigate some optimization tools (e.g., TensorRT, Triton inference server, …) for real-time inference on edge platforms. Furthermore, data augmentation strategies could enhance the performance of the proposed approach in some high noise use cases.

Author Contributions

Conceptualization and software, M.M.; methodology, data curation, investigation, formal analysis, visualization and writing—original draft preparation, M.M. and G.C.; formal analysis and writing—review and editing, F.O. and J.O.; supervision, project administration and funding acquisition, G.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded under the grant agreement number 0187793/00 by the Bpifrance program that is gratefully acknowledged.

Institutional Review Board Statement

The data used in this research was approved for publication in PhysioNet at https://physionet.org/about/database accessed on 13 December 2022, according to the open-access policy.

Informed Consent Statement

Not applicable.

Data Availability Statement

The different databases used in this research are publicly available in PhysioNet at https://physionet.org/about/database and http://www.ecgdb.com accessed on 13 December 2022.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANNArtificial neural network
BLSTMBidirectional long short-term memory
BPBlood pressure signal
CEEMDANComplete ensemble empirical mode decomposition with adaptive noise
CCDD       Chinese cardiovascular disease database
CMRCardiac magnetic resonance imaging
CNNConvolutional neural network
DLDeep learning
DNNDeep neural network
ECGElectrocardiogram
EEGElectroencephalography
EMGElectromyography
EOGElectrooculography
FNFalse negative
FPFalse positive
FSMFinite state machine
GRUGated recurrent unit
INCARTSt. Petersburg institute of cardiological technics 12-lead arrhythmia database
k-NNk-Nearest Neighbor
LSTMLong short-term memory
LUDBLobachevsky university electrocardiography database
MLMachine learning
NNNeural network
ONNOperational neural network
PLSVLinear regression-based transformation for deriving the P-waves
QLSVLinear regression-based transformation for deriving the QRS complexes
RNNRecurrent neural network
SVMSupport vector machine
TPTrue positive
VCGVectorcardiogram
WTVWeighted total variation

Appendix A

Figure A1. Detailed description of the DL architecture.
Figure A1. Detailed description of the DL architecture.
Sensors 23 02288 g0a1

References

  1. Chang, R.K. Resting 12-lead ECG tests performed by patients at home amid the COVID-19 pandemic—Results from the first 1000 patients. J. Electrocardiol. 2022, 73, 108–112. [Google Scholar] [CrossRef]
  2. Ebrahimi, Z.; Loni, M.; Daneshtalab, M.; Gharehbaghi, A. A review on deep learning methods for ECG arrhythmia classification. Expert Syst. Appl. 2020, 7, 100033. [Google Scholar] [CrossRef]
  3. Goldberger, A.L.; Amaral, L.A.N.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 2000, 101, 215–220. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Liu, X.; Wang, H.; Li, Z.; Qin, L. Deep learning in ECG diagnosis: A review. Knowl.-Based Syst. 2021, 227, 107187. [Google Scholar] [CrossRef]
  5. Husain, K.; Mohd Zahid, M.S.; Ul Hassan, S.; Hasbullah, S.; Mandala, S. Advances of ECG Sensors from Hardware, Software and Format Interoperability Perspectives. Electronics 2021, 10, 105. [Google Scholar] [CrossRef]
  6. Merdjanovska, E.; Rashkovska, A. Comprehensive survey of computational ECG analysis: Databases, methods and applications. Expert Syst. Appl. 2022, 203, 117206. [Google Scholar] [CrossRef]
  7. Somani, S.; Russak, A.J.; Richter, F.; Zhao, S.; Vaid, A.; Chaudhry, F.; De Freitas, J.K.; Naik, N.; Miotto, R.; Nadkarni, G.N.; et al. Deep learning and the electrocardiogram: Review of the current state-of-the-art. EP Eur. 2021, 23, 1179–1191. [Google Scholar] [CrossRef] [PubMed]
  8. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  9. Matsuo, Y.; LeCun, Y.; Sahani, M.; Precup, D.; Silver, D.; Sugiyama, M.; Uchibe, E.; Morimoto, J. Deep learning, reinforcement learning, and world models. Neural Netw. 2022, 152, 267–275. [Google Scholar] [CrossRef]
  10. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  11. Shehab, M.; Abualigah, L.; Shambour, Q.; Abu-Hashem, M.A.; Shambour, M.K.Y.; Alsalibi, A.I.; Gandomi, A.H. Machine learning in medical applications: A review of state-of-the-art methods. Comput. Biol. Med. 2022, 145, 105458. [Google Scholar] [CrossRef] [PubMed]
  12. Kalyakulina, A.; Yusipov, I.; Moskalenko, V.; Nikolskiy, A.; Kozlov, A.; Zolotykh, N.; Ivanchenko, M. Finding morphology points of electrocardiographic signal waves using wavelet analysis. Radiophys. Quantum Electron. 2019, 61, 689–703. [Google Scholar] [CrossRef] [Green Version]
  13. Siontis, K.C.; Noseworthy, P.A.; Attia, Z.I.; Friedman, P.A. Artificial intelligence-enhanced electrocardiography in cardiovascular disease management. Nat. Rev. Cardiol. 2021, 18, 465–478. [Google Scholar] [CrossRef] [PubMed]
  14. Steinfath, E.; Palacios-Muñoz, A.; Rottschäfer, J.R.; Yuezak, D.; Clemens, J. Fast and accurate annotation of acoustic signals with deep neural networks. eLife 2021, 10, e68837. [Google Scholar] [CrossRef]
  15. Cuculich, P.S.; Schill, M.R.; Kashani, R.; Mutic, S.; Lang, A.; Cooper, D.; Faddis, M.; Gleva, M.; Noheria, A.; Smith, T.W.; et al. Noninvasive Cardiac Radiation for Ablation of Ventricular Tachycardia. N. Engl. J. Med. 2017, 377, 2325–2336. [Google Scholar] [CrossRef] [PubMed]
  16. Moody, G.; Mark, R. The impact of the MIT-BIH Arrhythmia Database. IEEE Eng. Med. Biol. Mag. 2001, 20, 45–50. [Google Scholar] [CrossRef]
  17. MIT-BIH Arrhythmia Dataset. Available online: https://physionet.org/content/mitdb/1.0.0/ (accessed on 13 December 2022).
  18. Moody, G. RR interval time series modeling: The PhysioNet/Computers in Cardiology Challenge 2002. In Proceedings of the Computers in Cardiology, Memphis, TN, USA, 22–25 September 2002; pp. 125–128. [Google Scholar]
  19. Moody, G.; Moody, B.; Silva, I. Robust detection of heart beats in multimodal data: The PhysioNet/Computing in Cardiology Challenge 2014. In Proceedings of the Computing in Cardiology, Cambridge, MA, USA, 7–10 September 2014; pp. 549–552. [Google Scholar]
  20. Clifford, G.D.; Liu, C.; Moody, B.; Lehman, L.w.H.; Silva, I.; Li, Q.; Johnson, A.E.; Mark, R.G. AF classification from a short single lead ECG recording: The PhysioNet/Computing in Cardiology Challenge 2017. In Proceedings of the Computing in Cardiology, Rennes, France, 24–27 September 2017; pp. 1–4. [Google Scholar]
  21. Reyna, M.A.; Alday, E.A.P.; Gu, A.; Liu, C.; Seyedi, S.; Rad, A.B.; Elola, A.; Li, Q.; Sharma, A.; Clifford, G.D. Classification of 12-lead ECGs: The PhysioNet/Computing in Cardiology Challenge 2020. In Proceedings of the Computing in Cardiology, Rimini, Italy, 13–16 September 2020; pp. 1–4. [Google Scholar]
  22. Reyna, M.A.; Sadr, N.; Alday, E.A.P.; Gu, A.; Shah, A.J.; Robichaux, C.; Rad, A.B.; Elola, A.; Seyedi, S.; Ansari, S.; et al. Will Two Do? Varying Dimensions in Electrocardiography: The PhysioNet/Computing in Cardiology Challenge 2021. In Proceedings of the Computing in Cardiology, Brno, Czech Republic, 13–15 September 2021; pp. 1–4. [Google Scholar]
  23. Liu, F.; Liu, C.; Zhao, L.; Zhang, X.; Wu, X.; Xu, X.; Liu, Y.; Ma, C.; Wei, S.; He, Z.; et al. An Open Access Database for Evaluating the Algorithms of Electrocardiogram Rhythm and Morphology Abnormality Detection. J. Med. Imaging Health Inform. 2018, 8, 1368–1373. [Google Scholar] [CrossRef]
  24. Hongxiang, G.; Liu, C.; Wang, X.; Zhao, L.; Shen, Q.; Ng, E.; Li, J. An Open-Access ECG Database for Algorithm Evaluation of QRS Detection and Heart Rate Estimation. J. Med. Imaging Health Inform. 2019, 9, 1853–1858. [Google Scholar]
  25. Cai, Z.; Liu, C.; Hongxiang, G.; Wang, X.; Zhao, L.; Shen, Q.; Ng, E.; Li, J. An Open-Access Long-Term Wearable ECG Database for Premature Ventricular Contractions and Supraventricular Premature Beat Detection. J. Med. Imaging Health Inform. 2020, 10, 2663–2667. [Google Scholar] [CrossRef]
  26. Wang, X.; Ma, C.; Zhang, X.; Gao, H.; Clifford, G.D.; Liu, C. Paroxysmal Atrial Fibrillation Events Detection from Dynamic ECG Recordings: The 4th China Physiological Signal Challenge 2021. 2021. Available online: http://2021.icbeb.org/CPSC2021 (accessed on 13 December 2022).
  27. Mehta, S.; Lingayat, N. Identification of QRS complexes in 12-lead electrocardiogram. Expert Syst. Appl. 2009, 36, 820–828. [Google Scholar] [CrossRef]
  28. Abdullah Al, Z.M.; Thapa, K.; Yang, S.H. Improving R Peak Detection in ECG Signal Using Dynamic Mode Selected Energy and Adaptive Window Sizing Algorithm with Decision Tree Algorithm. Sensors 2021, 21, 6682. [Google Scholar] [CrossRef] [PubMed]
  29. Pahlm, O.; Sörnmo, L. Software QRS detection in ambulatory monitoring—A review. Med Biol. Eng. Comput. 1984, 22, 289–297. [Google Scholar] [CrossRef] [PubMed]
  30. Pan, J.; Tompkins, W.J. A real-time QRS detection algorithm. IEEE Trans. Biomed. Eng. 1985, 32, 230–236. [Google Scholar] [CrossRef] [PubMed]
  31. Hamilton, P.S.; Tompkins, W.J. Quantitative Investigation of QRS Detection Rules Using the MIT/BIH Arrhythmia Database. IEEE Trans. Biomed. Eng. 1986, BME-33, 1157–1165. [Google Scholar] [CrossRef] [PubMed]
  32. Khan, N.; Imtiaz, M.N. Pan-Tompkins++: A Robust Approach to Detect R-peaks in ECG Signals. In Proceedings of the International Conference on Bioinformatics and Biomedicine, Las Vegas, NV, USA, 6–8 December 2022; pp. 2905–2912. [Google Scholar]
  33. Coast, D.; Stern, R.; Cano, G.; Briller, S. An approach to cardiac arrhythmia analysis using hidden Markov models. IEEE Trans. Biomed. Eng. 1990, 37, 826–836. [Google Scholar] [CrossRef]
  34. Kohler, B.U.; Hennig, C.; Orglmeister, R. The principles of software QRS detection. IEEE Eng. Med. Biol. Mag. 2002, 21, 42–57. [Google Scholar] [CrossRef]
  35. Christov, I.I. Real time electrocardiogram QRS detection using combined adaptive threshold. BioMedical Eng. OnLine 2004, 3, 28. [Google Scholar] [CrossRef] [Green Version]
  36. Qin, Q.; Li, J.; Yue, Y.; Liu, C. An Adaptive and Time-Efficient ECG R-Peak Detection Algorithm. J. Healthc. Eng. 2017, 2017, 1157–1165. [Google Scholar] [CrossRef] [Green Version]
  37. Dohare, A.K.; Kumar, V.; Kumar, R. An efficient new method for the detection of QRS in electrocardiogram. Comput. Electr. Eng. 2014, 40, 1717–1730. [Google Scholar] [CrossRef]
  38. De Cooman, T.; Goovaerts, G.; Varon, C.; Widjaja, D.; Willemen, T.; Huffel, S. Heart beat detection in multimodal data using automatic relevant signal detection. Physiol. Meas. 2015, 36, 1691–1704. [Google Scholar] [CrossRef]
  39. Burguera, A. Fast QRS Detection and ECG Compression Based on Signal Structural Analysis. IEEE J. Biomed. Health Inform. 2019, 23, 123–131. [Google Scholar] [CrossRef] [PubMed]
  40. Gutiérrez-Rivas, R.; García, J.J.; Marnane, W.P.; Hernández, Á. Novel Real-Time Low-Complexity QRS Complex Detector Based on Adaptive Thresholding. IEEE Sens. J. 2015, 15, 6036–6043. [Google Scholar] [CrossRef]
  41. Castells-Rufas, D.; Carrabina, J. Simple real-time QRS detector with the MaMeMi filter. Biomed. Signal Process. Control 2015, 21, 137–145. [Google Scholar] [CrossRef] [Green Version]
  42. Phukpattaranont, P. QRS detection algorithm based on the quadratic filter. Expert Syst. Appl. 2015, 42, 4867–4877. [Google Scholar] [CrossRef]
  43. Paoletti, M.; Marchesi, C. Discovering Dangerous Patterns in Long-Term Ambulatory ECG Recordings Using a Fast QRS Detection Algorithm and Explorative Data Analysis. Comput. Methods Programs Biomed. 2006, 82, 20–30. [Google Scholar] [CrossRef] [PubMed]
  44. Zidelmal, Z.; Amirou, A.; Ould-Abdeslam, D.; Moukadem, A.; Dieterlen, A. QRS detection using S-Transform and Shannon energy. Comput. Methods Programs Biomed. 2014, 116, 1–9. [Google Scholar] [CrossRef] [Green Version]
  45. Mukhopadhyay, S.K.; Krishnan, S. Robust identification of QRS-complexes in electrocardiogram signals using a combination of interval and trigonometric threshold values. Biomed. Signal Process. Control 2020, 61, 102007. [Google Scholar] [CrossRef]
  46. Kumar, A.; Ranganatham, R.; Komaragiri, R.; Kumar, M. Efficient QRS complex detection algorithm based on Fast Fourier Transform. Biomed. Eng. Lett. 2019, 9, 145–151. [Google Scholar] [CrossRef]
  47. Fathail, I.; Bhagile, V.D. ECG Paper Digitization and R Peaks Detection Using FFT. Appl. Comput. Intell. Soft Comput. 2022, 2022, 1238864. [Google Scholar] [CrossRef]
  48. Li, C.; Zheng, C.; Tai, C. Detection of ECG characteristic points using wavelet transforms. IEEE Trans. Biomed. Eng. 1995, 42, 21–28. [Google Scholar]
  49. Zidelmal, Z.; Amirou, A.; Adnane, M.; Belouchrani, A. QRS detection based on wavelet coefficients. Comput. Methods Programs Biomed. 2012, 107, 490–496. [Google Scholar] [CrossRef] [PubMed]
  50. Merah, M.; Abdelmalik, T.; Larbi, B. R-peaks detection based on stationary wavelet transform. Comput. Methods Programs Biomed. 2015, 121, 149–160. [Google Scholar] [CrossRef] [PubMed]
  51. Zhou, Y.; Hu, X.; Tang, Z.; Ahn, A. Sparse representation-based ECG signal enhancement and QRS detection. Physiol. Meas. 2016, 37, 2093–2110. [Google Scholar] [CrossRef] [PubMed]
  52. Hossain, M.B.; Bashar, S.K.; Walkey, A.J.; McManus, D.D.; Chon, K.H. An Accurate QRS Complex and P Wave Detection in ECG Signals Using Complete Ensemble Empirical Mode Decomposition with Adaptive Noise Approach. IEEE Access 2019, 7, 128869–128880. [Google Scholar] [CrossRef]
  53. Lee, M.; Park, D.; Dong, S.Y.; Youn, I. A Novel R Peak Detection Method for Mobile Environments. IEEE Access 2018, 6, 51227–51237. [Google Scholar] [CrossRef]
  54. Zhu, H.; Dong, J. An R-peak detection method based on peaks of Shannon energy envelope. Biomed. Signal Process. Control 2013, 8, 466–474. [Google Scholar] [CrossRef]
  55. Elgendi, M. Fast QRS Detection with an Optimized Knowledge-Based Method: Evaluation on 11 Standard ECG Databases. PLoS ONE 2013, 8, e73557. [Google Scholar] [CrossRef] [Green Version]
  56. Sharma, T.; Sharma, K.K. QRS complex detection in ECG signals using locally adaptive weighted total variation denoising. Comput. Biol. Med. 2017, 87, 187–199. [Google Scholar] [CrossRef]
  57. Johnson, A.; Behar, J.; Andreotti, F.; Clifford, G.; Oster, J. Multimodal heart beat detection using signal quality indices. Physiol. Meas. 2015, 36, 1665–1677. [Google Scholar] [CrossRef]
  58. Gierałtowski, J.; Ciuchciński, K.; Grzegorczyk, I.; Stępień, K.; Soliński, M.; Podziemski, P. RS slope detection algorithm for extraction of heart rate from noisy, multimodal recordings. Physiol. Meas. 2015, 36, 1743–1761. [Google Scholar] [CrossRef]
  59. Liu, F.; Liu, C.; Jiang, X.; Zhang, Z.; Zhang, Y.; Li, J.; Wei, S. Performance analysis of ten common QRS detectors on different ECG application cases. J. Healthc. Eng. 2018, 2018, 9050812. [Google Scholar] [CrossRef] [Green Version]
  60. Behar, J.; Oster, J.; Li, Q.; Clifford, G.D. ECG Signal Quality During Arrhythmia and Its Application to False Alarm Reduction. IEEE Trans. Biomed. Eng. 2013, 60, 1660–1666. [Google Scholar] [CrossRef] [PubMed]
  61. Behar, J.; Oster, J.; Clifford, G.D. Combining and benchmarking methods of foetal ECG extraction without maternal or scalp electrode data. Physiol. Meas. 2014, 35, 1569–1589. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  62. Saraswat, S.; Srivastava, G.; Shukla, S. Review: Comparison of QRS detection algorithms. In Proceedings of the International Conference on Computing, Communication & Automation, Greater Noida, India, 15–16 May 2015; pp. 354–359. [Google Scholar]
  63. Raj, S.; Ray, K.C.; Shankar, O. Development of robust, fast and efficient QRS complex detector: A methodological review. Australas. Phys. Eng. Sci. Med. 2018, 41, 581–600. [Google Scholar] [CrossRef] [PubMed]
  64. Kirti.; Sohal, H.; Jain, S. Review and Comparison of QRS Detection Algorithms for Arrhythmia Diagnosis. In Proceedings of the International Conference on Signal Processing, Computing and Control, Solan, India, 10–12 October 2019; pp. 200–205. [Google Scholar]
  65. Van, G.V.; Podmasteryev, K.V. Review of Algorithms for Detection the QRS-complex Based on Machine Learning. International Scientific Symposium “Sense. Enable. SPITSE”. 2015, pp. 61–63. Available online: https://www.semanticscholar.org/paper/Review-of-Algorithms-for-Detection-the-QRS-complex-Mehta-Shete/7cc7642f7216e61624dab127c89114aed97e8431 (accessed on 5 January 2023).
  66. Arbateni, K.; Bennia, A. Sigmoidal radial basis function ANN for QRS complex detection. Neurocomputing 2014, 145, 438–450. [Google Scholar] [CrossRef]
  67. Saini, I.; Singh, D.; Khosla, A. QRS detection using K-Nearest Neighbor algorithm (KNN) and evaluation on standard ECG databases. J. Adv. Res. 2013, 4, 331–344. [Google Scholar] [CrossRef] [Green Version]
  68. Mehta, S.S.; Shete, D.A.; Lingayat, N.S.; Chouhan, V.S. K-means algorithm for the detection and delineation of QRS-complexes in Electrocardiogram. IRBM 2010, 31, 48–54. [Google Scholar] [CrossRef]
  69. Mehta, S.S.; Lingayat, N.S. Development of SVM based classification techniques for the delineation of wave components in 12-lead electrocardiogram. Biomed. Signal Process. Control 2008, 3, 341–349. [Google Scholar] [CrossRef]
  70. Śmigiel, S.; Pałczyński, K.; Ledziński, D. Deep Learning Techniques in the Classification of ECG Signals Using R-Peak Detection Based on the PTB-XL Dataset. Sensors 2021, 21, 8174. [Google Scholar] [CrossRef]
  71. Chen, G.; Chen, M.; Zhang, J.; Zhang, L.; Pang, C. A Crucial Wave Detection and Delineation Method for Twelve-Lead ECG Signals. IEEE Access 2020, 8, 10707–10717. [Google Scholar] [CrossRef]
  72. Kalyakulina, A.I.; Yusipov, I.I.; Moskalenko, V.A.; Nikolskiy, A.V.; Kosonogov, K.A.; Osipov, G.V.; Zolotykh, N.Y.; Ivanchenko, M.V. LUDB: A New Open-Access Validation Tool for Electrocardiogram Delineation Algorithms. IEEE Access 2020, 8, 186181–186190. [Google Scholar] [CrossRef]
  73. LUDB Dataset. Available online: https://physionet.org/content/ludb/1.0.1/ (accessed on 13 December 2022).
  74. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  75. Medsker, L.R.; Jain, L.C. Recurrent Neural Networks: Design and Applications; CRC Press, Inc.: Boca Raton, FL, USA, 1999. [Google Scholar]
  76. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  77. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  78. Šarlija, M.; Jurišić, F.; Popović, S. A convolutional neural network based approach to QRS detection. In Proceedings of the International Symposium on Image and Signal Processing and Analysis, Ljubljana, Slovenia, 18–20 September 2017; pp. 121–125. [Google Scholar]
  79. Oudkerk-Pool, M.D.; de Vos, B.D.; Winter, M.M.; Išgum, I. Deep Learning-Based Data-Point Precise R-Peak Detection in Single-Lead Electrocardiograms. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine & Biology Society, Mexico City, Mexico, 1–5 November 2021; pp. 718–721. [Google Scholar]
  80. Tison, G.; Zhang, J.; Delling, F.; Deo, R. Automated and Interpretable Patient ECG Profiles for Disease Detection, Tracking, and Discovery. Circ. Cardiovasc. Qual. Outcomes 2019, 12, e005289. [Google Scholar] [CrossRef] [PubMed]
  81. Oh, S.L.; Ng, E.Y.; Tan, R.S.; Acharya, U.R. Automated beat-wise arrhythmia diagnosis using modified U-Net on extended electrocardiographic recordings with heterogeneous arrhythmia types. Comput. Biol. Med. 2019, 105, 92–101. [Google Scholar] [CrossRef]
  82. Jimenez-Perez, G.; Alcaine, A.; Camara, O. Delineation of the electrocardiogram with a mixed-quality-annotations dataset using convolutional neural networks. Sci. Rep. 2021, 11, 1–11. [Google Scholar] [CrossRef]
  83. Laguna, P.; Mark, R.G.; Goldberg, A.; Moody, G.B. A database for evaluation of algorithms for measurement of QT and other waveform intervals in the ECG. In Proceedings of the Computers in Cardiology 1997, Lund, Sweden, 7–10 September 1997; pp. 673–676. [Google Scholar]
  84. QT Dataset. Available online: https://physionet.org/content/qtdb/1.0.0/ (accessed on 13 December 2022).
  85. Liu, L.; Bai, B.; Chen, X.; Xia, Q. Semantic Segmentation of QRS Complex in Single Channel ECG with Bidirectional LSTM Networks. J. Med. Imaging Health Inform. 2020, 10, 758–762. [Google Scholar] [CrossRef]
  86. Peimankar, A.; Puthusserypady, S. DENS-ECG: A deep learning approach for ECG signal delineation. Expert Syst. Appl. 2021, 165, 113911. [Google Scholar] [CrossRef]
  87. Vijayarangan, S.; Vignesh, R.; Murugesan, B.; Preejith, S.P.; Joseph, J.; Sivaprakasam, M. RPnet: A Deep Learning approach for robust R Peak detection in noisy ECG. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine & Biology Society, Montreal, QC, Canada, 20–24 July 2020; pp. 345–348. [Google Scholar]
  88. Duraj, K.; Piaseczna, N.; Kostka, P.; Tkacz, E. Semantic Segmentation of 12-Lead ECG Using 1D Residual U-Net with Squeeze-Excitation Blocks. Appl. Sci. 2022, 12, 3332. [Google Scholar] [CrossRef]
  89. Gabbouj, M.; Kiranyaz, S.; Malik, J.; Zahid, M.U.; Ince, T.; Chowdhury, M.E.H.; Khandakar, A.; Tahir, A. Robust Peak Detection for Holter ECGs by self-organized operational neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–12. [Google Scholar] [CrossRef] [PubMed]
  90. Petmezas, G.; Stefanopoulos, L.; Kilintzis, V.; Tzavelis, A.; Rogers, J.A.; Katsaggelos, A.K.; Maglaveras, N. State-of-the-Art Deep Learning Methods on Electrocardiogram Data: Systematic Review. JMIR Med. Inform. 2022, 10, 38454. [Google Scholar] [CrossRef] [PubMed]
  91. Sereda, I.; Alekseev, S.; Koneva, A.; Kataev, R.; Osipov, G. ECG Segmentation by Neural Networks: Errors and Correction. In Proceedings of the International Joint Conference on Neural Networks, Budapest, Hungary, 14–19 July 2019; pp. 1–7. [Google Scholar]
  92. Moskalenko, V.; Zolotykh, N.; Osipov, G. Deep Learning for ECG Segmentation. In Advances in Neural Computation, Machine Learning, and Cognitive Research III; Springer: Cham, Switzerland, 2020; pp. 246–254. [Google Scholar]
  93. Zahid, M.U.; Kiranyaz, S.; Ince, T.; Devecioglu, O.C.; Chowdhury, M.E.; Khandakar, A.; Tahir, A.; Gabbouj, M. Robust R-peak detection in low-quality Holter ECGs using 1D convolutional neural network. IEEE Trans. Biomed. Eng. 2021, 69, 119–128. [Google Scholar] [CrossRef] [PubMed]
  94. Han, C.; Que, W.; Wang, S.; Zhang, J.; Zhao, J.; Shi, L. QRS complexes and T waves localization in multi-lead ECG signals based on deep learning and electrophysiology knowledge. Expert Syst. Appl. 2022, 199, 117187. [Google Scholar] [CrossRef]
  95. Chinese Cardiovascular Disease Database—CCDD Dataset. Available online: http://www.ecgdb.com (accessed on 13 December 2022).
  96. Zhang, J.w.; Wang, L.p.; Liu, X.; Zhu, H.h.; Dong, J. Chinese Cardiovascular Disease Database (CCDD) and Its Management Tool. In Proceedings of the International Conference on BioInformatics and BioEngineering, Philadelphia, PA, USA, 31 May–3 June 2010; pp. 66–72. [Google Scholar]
  97. Zhang, J.w.; Liu, X.; Dong, J. CCDD: An Enhanced Standard ECG Database with its Management and Annotation Tools. Int. J. Artif. Intell. Tools 2012, 21, 1240020. [Google Scholar] [CrossRef]
  98. Frank, E. An Accurate, Clinically Practical System For Spatial Vectorcardiography. Circulation 1956, 13, 737–749. [Google Scholar] [CrossRef] [Green Version]
  99. Vozda, M.; Cerny, M. Methods for derivation of orthogonal leads from 12-lead electrocardiogram: A review. Biomed. Signal Process. Control 2015, 19, 23–34. [Google Scholar] [CrossRef]
  100. Kors, J.A.; van Herpen, G.; Sittig, A.C.; van Bemmel, J. Reconstruction of the Frank vectorcardiogram from standard electrocardiographic leads: Diagnostic comparison of different methods. Eur. Heart J. 1990, 11, 1083–1092. [Google Scholar] [CrossRef]
  101. Sadhukhan, D.; Mitra, M. R-Peak Detection Algorithm for ECG using Double Difference And RR Interval Processing. Procedia Technol. 2012, 4, 873–877. [Google Scholar] [CrossRef] [Green Version]
  102. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  103. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Annu. Conf. Neural Inf. Process. Syst.-Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  104. PTB Diagnostic ECG Dataset. Available online: https://physionet.org/content/ptbdb/1.0.0/ (accessed on 13 December 2022).
  105. PTB-XL Dataset. Available online: https://physionet.org/content/ptb-xl/1.0.3/ (accessed on 13 December 2022).
  106. Wagner, P.; Strodthoff, N.; Bousseljot, R.D.; Kreiseler, D.; Lunze, F.I.; Samek, W.; Schaeffter, T. PTB-XL, a large publicly available electrocardiography dataset. Sci. Data 2020, 7, 154. [Google Scholar] [CrossRef] [PubMed]
  107. St Petersburg INCART 12-Lead Arrhythmia Dataset. Available online: https://physionet.org/content/incartdb/1.0.0/ (accessed on 13 December 2022).
  108. 1st CPSC Dataset. Available online: http://2018.icbeb.org/Challenge.html (accessed on 13 December 2022).
  109. Shaoxing People’s Hospital—SPH Dataset. Available online: https://physionet.org/content/ecg-arrhythmia/1.0.0/ (accessed on 13 December 2022).
  110. CSE Dataset. Available online: https://www.medteq.net/ctscse-database-information (accessed on 13 December 2022).
  111. Georgia Dataset. Available online: https://physionet.org/content/challenge-2020/1.0.2/ (accessed on 13 December 2022).
  112. Habib, A.; Karmakar, C.; Yearwood, J. Impact of ECG Dataset Diversity on Generalization of CNN Model for Detecting QRS Complex. IEEE Access 2019, 7, 93275–93285. [Google Scholar] [CrossRef]
  113. Yun, D.; Lee, H.C.; Jung, C.W.; Kwon, S.; Lee, S.R.; Kim, K.; Kim, S.U. Robust R-peak detection in an electrocardiogram with stationary wavelet transformation and separable convolution. Sci. Rep. 2022, 12, 19638. [Google Scholar] [CrossRef]
  114. Sedghamiz, H. Matlab Implementation of Pan Tompkins ECG QRS Detector. 2014. Available online: https://hooman650.github.io/ECG-QRS.html (accessed on 13 December 2022).
  115. Schmidt, M.; Krug, J.W.; Gierstorfer, A.; Rose, G. A real-time QRS detector based on higher-order statistics for ECG gated cardiac MRI. In Proceedings of the Computing in Cardiology 2014, Cambridge, MA, USA, 7–10 September 2014; pp. 733–736. [Google Scholar]
  116. Man, S.; Maan, A.C.; Schalij, M.J.; Swenne, C.A. Vectorcardiographic diagnostic & prognostic information derived from the 12-lead electrocardiogram: Historical review and clinical perspective. J. Electrocardiol. 2015, 48, 463–475. [Google Scholar]
  117. Oster, J.; Clifford, G.D. Acquisition of electrocardiogram signals during magnetic resonance imaging. Physiol. Meas. 2017, 38, R119–R142. [Google Scholar] [CrossRef] [PubMed]
  118. Isaieva, K.; Fauvel, M.; Weber, N.; Vuissoz, P.; Felblinger, J.; Oster, J.; Odille, F. A hardware and software system for MRI applications requiring external device data. Magn. Reson. Med. 2022, 88, 1406–1418. [Google Scholar] [CrossRef]
  119. Malik, M. Heart Rate Variability - Standards of Measurement, Physiological Interpretation, and Clinical Use. Circulation 1996, 93, 1043–1065. [Google Scholar] [CrossRef] [Green Version]
  120. Hrv-Analysis. 2018. Available online: https://aura-healthcare.github.io/hrv-analysis/index.html (accessed on 13 December 2022).
  121. Krug, J.W.; Clifford, G.D.; Rose, G.H.; Oster, J. The limited applicability of Wiener filtering to ECG signals disturbed by the MHD effect. In Proceedings of the 20th European Signal Processing Conference (EUSIPCO), Bucharest, Romania, 27–31 August 2012; pp. 959–963. [Google Scholar]
  122. Krug, J.; Schmidt, M.; Rose, G.; Friebe, M. A Database of Electrocardiogram Signals Acquired in Different Magnetic Resonance Imaging Scanners. In Proceedings of the Computing in Cardiology, Rennes, France, 24–27 September 2017; pp. 1–4. [Google Scholar]
  123. Oster, J.; Clifford, G.D. Impact of the presence of noise on RR interval-based atrial fibrillation detection. J. Electrocardiol. 2015, 48, 947–951. [Google Scholar] [CrossRef]
Figure 1. Illustrative schema of the ECG waveform components. P: atrial depolarization; QRS: ventricular depolarization; T: ventricular repolarization.
Figure 1. Illustrative schema of the ECG waveform components. P: atrial depolarization; QRS: ventricular depolarization; T: ventricular repolarization.
Sensors 23 02288 g001
Figure 2. Kors regression transformation of a 12-lead ECG into a 3D VCG in a recording from the INCART database: (a) 12-lead ECG; (b) 3D VCG.
Figure 2. Kors regression transformation of a 12-lead ECG into a 3D VCG in a recording from the INCART database: (a) 12-lead ECG; (b) 3D VCG.
Sensors 23 02288 g002
Figure 3. Schema of the proposed DL architecture for R-peak detection.
Figure 3. Schema of the proposed DL architecture for R-peak detection.
Sensors 23 02288 g003
Figure 4. A 12-lead ECG recording example of the CCDD database.
Figure 4. A 12-lead ECG recording example of the CCDD database.
Sensors 23 02288 g004
Figure 5. A 12-lead ECG recording example of the LUDB database.
Figure 5. A 12-lead ECG recording example of the LUDB database.
Sensors 23 02288 g005
Figure 6. A 12-lead ECG recording from the INCART database.
Figure 6. A 12-lead ECG recording from the INCART database.
Sensors 23 02288 g006
Figure 7. A 12-lead ECG recording example of the CCDD-Extra database.
Figure 7. A 12-lead ECG recording example of the CCDD-Extra database.
Sensors 23 02288 g007
Figure 8. Loss curves obtained using the leave-one-out cross-validation protocol on the training data from CCDD (a); LUDB (b); and INCART (c); loss curve obtained in the cross-database validation protocol using all 526 recordings of the three databases (d).
Figure 8. Loss curves obtained using the leave-one-out cross-validation protocol on the training data from CCDD (a); LUDB (b); and INCART (c); loss curve obtained in the cross-database validation protocol using all 526 recordings of the three databases (d).
Sensors 23 02288 g008
Figure 9. Qualitative results: recording from the CCDD database showing the 3D VCG vectors (X, Y, and Z) and R-peak detected by the proposed approach in the leave-one-out cross-validation protocol. The first and last R-peaks were ignored since there were no annotations.
Figure 9. Qualitative results: recording from the CCDD database showing the 3D VCG vectors (X, Y, and Z) and R-peak detected by the proposed approach in the leave-one-out cross-validation protocol. The first and last R-peaks were ignored since there were no annotations.
Sensors 23 02288 g009
Figure 10. Qualitative results: recording from the LUDB database showing the 3D VCG vectors (X, Y, and Z) and R-peak detected by the proposed approach in the leave-one-out cross-validation protocol. The first and last R-peaks were ignored since there were no annotations.
Figure 10. Qualitative results: recording from the LUDB database showing the 3D VCG vectors (X, Y, and Z) and R-peak detected by the proposed approach in the leave-one-out cross-validation protocol. The first and last R-peaks were ignored since there were no annotations.
Sensors 23 02288 g010
Figure 11. Qualitative results: recording from the INCART database showing the 3D VCG vectors (X, Y, and Z) and R-peak detected by the proposed approach in the leave-one-out cross-validation protocol.
Figure 11. Qualitative results: recording from the INCART database showing the 3D VCG vectors (X, Y, and Z) and R-peak detected by the proposed approach in the leave-one-out cross-validation protocol.
Sensors 23 02288 g011
Figure 12. Qualitative results: recording from the CCDD-Extra database showing the 3D VCG vectors (X, Y, and Z) and R-peak detected by the proposed approach in the cross-database validation protocol.
Figure 12. Qualitative results: recording from the CCDD-Extra database showing the 3D VCG vectors (X, Y, and Z) and R-peak detected by the proposed approach in the cross-database validation protocol.
Sensors 23 02288 g012
Figure 13. Illustration of a false R-peak detection (FP) when applying the proposed approach on a noisy recording from the INCART database: (a) 12-lead ECG; (b) 3D VCG.
Figure 13. Illustration of a false R-peak detection (FP) when applying the proposed approach on a noisy recording from the INCART database: (a) 12-lead ECG; (b) 3D VCG.
Sensors 23 02288 g013
Table 1. Key features of the four databases used in our experiments.
Table 1. Key features of the four databases used in our experiments.
DatabaseSampling Frequency (Hz)Recording DurationNo. of RecordingsNo. of Analysis WindowsNo. of R-Peaks
CCDD500~10 s2517533973
LUDB50010 s2004001831
INCART25730 min7532,925175,907
CCDD-Extra500~10 s1033091616
Total62934,387183,327
Table 2. Selected hyperparameters for training the DL architecture.
Table 2. Selected hyperparameters for training the DL architecture.
HyperparameterValue
OptimizerAdam
Loss functionBinary cross entropy
Weight decayNone
Learning rate0.001
Number of epochs500
Patience10
Batch size64
Table 3. Quantitative results on CCDD, LUDB, and INCART databases using the leave-one-out cross-validation protocol.
Table 3. Quantitative results on CCDD, LUDB, and INCART databases using the leave-one-out cross-validation protocol.
DatabaseTPFNFPRecallPrecisionF1-Score
CCDD394627099.39100.0099.69
LUDB18274299.7599.8899.80
INCART175,6312761699.8499.9999.91
Average99.6699.9699.80
Table 4. Quantitative results on the CCDD-Extra database using the cross-database validation protocol.
Table 4. Quantitative results on the CCDD-Extra database using the cross-database validation protocol.
DatabaseTPFNFPRecallPrecisionF1-Score
CCDD-Extra160610299.4199.8999.64
Table 5. Quantitative results on the INCART database using the leave-one-out cross-validation protocol and different inputs (single lead, single VCG vector, or multiple leads); 03-leads: II, V2, and V5; 08-leads: I, II, and V1 to V6.
Table 5. Quantitative results on the INCART database using the leave-one-out cross-validation protocol and different inputs (single lead, single VCG vector, or multiple leads); 03-leads: II, V2, and V5; 08-leads: I, II, and V1 to V6.
DatabaseInput TPFNFPRecallPrecisionF1-Score
INCARTSingle leadI174,227168060699.1299.7299.40
II175,5163914699.7799.9799.87
V1171,07139903097.4299.9898.57
V2174,04618613398.9599.9899.44
V3175,025882460499.4899.0999.10
V4174,8451062474399.4098.7798.93
V5175,2416662699.6299.9999.79
V6174,6211286251399.3499.1799.23
Single VCG vectorX175,6023054099.8299.9899.90
Y175,4574503699.7499.9899.86
Z174,77011373899.3999.9899.68
Multiple leads03-leads175,514393999.77100.0099.88
08-leads174,27416332599.1499.9999.55
3D VCGX, Y, and Z175,6312761699.8499.9999.91
Table 6. Comparison of the performances of the proposed approach with the PT algorithm (single lead) with the leave-one-out cross-validation protocol.
Table 6. Comparison of the performances of the proposed approach with the PT algorithm (single lead) with the leave-one-out cross-validation protocol.
DatabaseMethodLeadTPFNFPRecallPrecisionF1-Score
CCDDPT algorithmI3930436098.9998.7298.58
II3956174099.6199.2099.33
V13926475299.0198.9598.72
V2396764099.8599.0999.41
V3396676599.8698.7199.16
V43944295299.2198.9898.94
V53954191599.5399.6199.53
V639712599.9599.8699.91
Average 99.5099.1499.20
Proposed approach394627099.39100.0099.69
LUDBPT algorithmI18121916899.1691.7095.00
II18171415399.3492.3695.42
V118161516099.0992.2395.28
V218171414899.2192.4595.42
V31829219299.8991.5495.11
V41827413899.7892.9896.04
V518191213299.3392.8895.84
V618062514498.8192.3795.31
Average 99.3392.3195.43
Proposed approach18274299.7599.8899.80
INCARTPT algorithmI167,264864311,58195.7793.8394.65
II171,9203987266097.9298.4098.02
V1173,8632044133398.8999.3499.04
V2174,373153495499.1599.4699.27
V3171,8174090111797.8598.4897.91
V4169,8306077121096.8098.5697.26
V5172,818308977698.3899.4898.82
V6170,312559584497.1098.6297.58
Average 97.7398.2797.82
Proposed approach175,6312761699.8499.9999.91
Table 7. Comparison of the performances of the proposed approach with the PT algorithm (single lead) with the cross-database validation protocol (i.e. after training on CCDD, LUDB and INCART databases).
Table 7. Comparison of the performances of the proposed approach with the PT algorithm (single lead) with the cross-database validation protocol (i.e. after training on CCDD, LUDB and INCART databases).
DatabaseMethodLeadTPFNFPRecallPrecisionF1-Score
CCDD-ExtraPT algorithmI1616068100.0097.0898.30
II161511699.9698.9699.44
V1161511999.9698.7799.34
V2161511899.9698,8499.38
V3161424199.9198.0698.83
V4161511699.9698.9699.44
V51616017100.0098.9099.43
V61599172098.9997.7598.35
Average 99.8498.4299.06
Proposed approach160610299.4199.8999.64
Table 8. Comparison of the performances of the proposed approach with a state-of-the-art method in the CCDD database.
Table 8. Comparison of the performances of the proposed approach with a state-of-the-art method in the CCDD database.
DatabaseMethodRecallPrecisionF1-Score
CCDDHan et al. [94]99.9999.4599.71
Proposed approach99.39100.0099.69
Table 9. Comparison of the performances of the proposed approach with a state-of-the-art method in the LUDB database.
Table 9. Comparison of the performances of the proposed approach with a state-of-the-art method in the LUDB database.
DatabaseMethodRecallPrecisionF1-Score
LUDBChen et al. [71]100.0099.8699.92
Proposed approach99.7599.8899.80
Table 10. Comparison of the performances of the proposed approach with a state-of-the-art method in the INCART database.
Table 10. Comparison of the performances of the proposed approach with a state-of-the-art method in the INCART database.
DatabaseMethodRecallPrecisionF1-Score
INCARTSchmidt et al. [115]99.4399.9199.66
Proposed approach99.8499.9999.91
Table 11. Heart rates and heart rate variability measurements derived from R-peak detections using the leave-one-out cross-validation protocol in INCART: AVG-HR: average heart rate; SD-HR: standard deviation of heart rates; RMSSD (ms): square root of the mean sum of squares of differences between adjacent RR intervals.
Table 11. Heart rates and heart rate variability measurements derived from R-peak detections using the leave-one-out cross-validation protocol in INCART: AVG-HR: average heart rate; SD-HR: standard deviation of heart rates; RMSSD (ms): square root of the mean sum of squares of differences between adjacent RR intervals.
AVG-HR (Beats/min)SD-HR (Beats/min)RMSSD (ms)
Ground truth82.1817.45200.84
Proposed approach82.0718.35201.63
Table 12. Performances of some state-of-the-art R-peak detection methods using the MIT-BIH arrhythmia database compared with our DL architecture. Top three lines: classical algorithms; bottom three lines: DL-based architectures; Proposed approach*: our DL architecture trained using only lead II data from INCART.
Table 12. Performances of some state-of-the-art R-peak detection methods using the MIT-BIH arrhythmia database compared with our DL architecture. Top three lines: classical algorithms; bottom three lines: DL-based architectures; Proposed approach*: our DL architecture trained using only lead II data from INCART.
DatabaseMethodNo. of R-PeaksRecallPrecisionF1-Score
MIT-BIH arrhythmiaPan-Tompkins [30]109,98590.9599.5695.06
Pan-Tompkins++ [32]N/R99.4799.6099.54
Hamilton and Tompkins [31]109,26799.6999.7799.72
Habib et al. [112]N/R97.6191.9394.68
Zahid et al. [93]109,47599.8599.8299.83
Šarlija et al. [78]49,71299.8199.9399.86
Proposed approach*49,71299.7499.1599.44
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mehri, M.; Calmon, G.; Odille, F.; Oster, J. A Deep Learning Architecture Using 3D Vectorcardiogram to Detect R-Peaks in ECG with Enhanced Precision. Sensors 2023, 23, 2288. https://doi.org/10.3390/s23042288

AMA Style

Mehri M, Calmon G, Odille F, Oster J. A Deep Learning Architecture Using 3D Vectorcardiogram to Detect R-Peaks in ECG with Enhanced Precision. Sensors. 2023; 23(4):2288. https://doi.org/10.3390/s23042288

Chicago/Turabian Style

Mehri, Maroua, Guillaume Calmon, Freddy Odille, and Julien Oster. 2023. "A Deep Learning Architecture Using 3D Vectorcardiogram to Detect R-Peaks in ECG with Enhanced Precision" Sensors 23, no. 4: 2288. https://doi.org/10.3390/s23042288

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop