Next Article in Journal
Optoelectronic Pressure Sensor Based on the Bending Loss of Plastic Optical Fibers Embedded in Stretchable Polydimethylsiloxane
Previous Article in Journal
A Channel Allocation Method Considering the Asymmetry of Available Channels for Centralized Underwater Cognitive Acoustic Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Left Ventricle Detection from Cardiac Magnetic Resonance Relaxometry Images Using Visual Transformer

by
Lisa Anita De Santi
1,2,
Antonella Meloni
2,3,
Maria Filomena Santarelli
4,
Laura Pistoia
3,
Anna Spasiano
5,
Tommaso Casini
6,
Maria Caterina Putti
7,
Liana Cuccia
8,
Filippo Cademartiri
3 and
Vincenzo Positano
2,3,*
1
Department of Information Engineering, University of Pisa, 56122 Pisa, Italy
2
U.O.C. Bioingegneria, Fondazione G. Monasterio CNR-Regione Toscana, 56124 Pisa, Italy
3
Department of Radiology, Fondazione G. Monasterio CNR-Regione Toscana, 56124 Pisa, Italy
4
CNR Institute of Clinical Physiology, 56124 Pisa, Italy
5
Unità Operativa Semplice Dipartimentale Malattie Rare del Globulo Rosso, Azienda Ospedaliera di Rilievo Nazionale “A. Cardarelli”, 80131 Napoli, Italy
6
Centro Talassemie ed Emoglobinopatie, Ospedale “Meyer”, 50139 Firenze, Italy
7
Clinica di Emato-Oncologia Pediatrica, Dipartimento di Salute della Donna e del Bambino, Azienda Ospedale Università, 35128 Padova, Italy
8
Unità Operativa Complessa Ematologia con Talassemia, ARNAS Civico “Benfratelli-Di Cristina”, 90127 Palermo, Italy
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 3321; https://doi.org/10.3390/s23063321
Submission received: 2 March 2023 / Revised: 16 March 2023 / Accepted: 20 March 2023 / Published: 21 March 2023
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)

Abstract

:
Left Ventricle (LV) detection from Cardiac Magnetic Resonance (CMR) imaging is a fundamental step, preliminary to myocardium segmentation and characterization. This paper focuses on the application of a Visual Transformer (ViT), a novel neural network architecture, to automatically detect LV from CMR relaxometry sequences. We implemented an object detector based on the ViT model to identify LV from CMR multi-echo T2* sequences. We evaluated performances differentiated by slice location according to the American Heart Association model using 5-fold cross-validation and on an independent dataset of CMR T2*, T2, and T1 acquisitions. To the best of our knowledge, this is the first attempt to localize LV from relaxometry sequences and the first application of ViT for LV detection. We collected an Intersection over Union ( I o U ) index of 0.68 and a Correct Identification Rate ( C I R ) of blood pool centroid of 0.99, comparable with other state-of-the-art methods.  I o U  and  C I R  values were significantly lower in apical slices. No significant differences in performances were assessed on independent T2* dataset ( I o U  = 0.68, p = 0.405;  C I R  = 0.94, p = 0.066). Performances were significantly worse on the T2 and T1 independent datasets (T2:  I o U  = 0.62,  C I R  = 0.95; T1:  I o U  = 0.67,  C I R  = 0.98), but still encouraging considering the different types of acquisition. This study confirms the feasibility of the application of ViT architectures in LV detection and defines a benchmark for relaxometry imaging.

1. Introduction

Cardiac Magnetic Resonance (CMR) represents the gold standard non-invasive imaging tool for quantifying cardiac anatomy and function, as well as performing myocardial tissue characterization [1]. The Left Ventricle (LV) contraction is the source for the distribution of oxygenated blood to the entire body; hence, CMR image acquisition and analysis are focused on LV. In particular, the analysis of cardiac Short Axis Views (SAV) covering the Left Ventricle (LV) is mandatory in CMR to assess global and segmental cardiac function, myocardial perfusion, late gadolinium enhancement, and to characterize the myocardial tissue by T1, T2, and T2* relaxometry [2]. SAV planes are perpendicular to the long axis of LV that intercepts the apex and the center of the mitral valve, so that SAV planes preserve the integrity of the cardiac chambers, allowing us to perform standardized segmentation of the myocardium according to the segmental model defined by the American Heart Association (AHA) guidelines [3]. Image analysis of SAV series can be performed manually or by semi-automatic or fully-automatic algorithms. The use of automatic algorithms is strongly desirable to reduce processing time and avoid intra- and inter-subject variability [4]. SAV series are acquired over a large field of view, including the whole patient’s body, to avoid aliasing artifacts (Figure 1). Consequently, a typical pre-processing step in SAV analysis is a cropping operation centered on the object of interest (i.e., LV). Accurate identification and location of the LV in the SAV represent a pre-requisite for automatic evaluation of cardiac function indices, as this step provides good initialization for segmentation methods by extracting an ROI around the myocardium and cropping the initial volume ignoring all the other organs [5,6,7,8]. LV region extraction was also demonstrated to be useful in ischemic scar detection [9] and myocardial perfusion quantization [10]. In MR relaxometry, limiting the MR signal fitting to the LV region would allow a significant reduction in the required processing time [11,12].
In Computer Vision (CV) applications, the identification of LV represents an Object Detection (OD) problem, where the object detected will be marked using a Region Of Interest (ROI) such as a bounding box. The identification of the LV blood pool cavity from SAV images is a nontrivial task for different reasons [4]: (i) misalignment of SAV slices; (ii) inter-patient variability of LV cavity location; (iii) inter and within slices variability of the blood signal intensity (due to coil sensitivity variations and cardiac flow dynamics, respectively); (iv) presence of similarly shaped and isointense structures.
Object detection methods can be divided into two main classes: handcrafted-based and Deep Learning (DL)-based approaches [13].
Handcrafted approaches refer to the class of conventional algorithms based on mathematical models designed by field experts using domain knowledge [13,14]. Some examples in LV detection on fast-cine sequences include the usage of geometrical-based techniques such as a combination of scout-view geometry, blood-to-myocardial tissue contrast, and geometrical continuity constraint [4,5]; motion information obtained using Fourier transform [6,15,16]; or the usage of harmonic images to produce edge maps followed by anisotropic weighted circular Hough transform [17]. All the works mentioned have been applied to fast-cine MR images using internal or publicly available datasets.
At present, Artificial Intelligence (AI) is increasingly applied in the field of medical imaging analysis to help physicians in solving diagnostic problems and improve efficiency. DL models are a subset of AI-based algorithms that can be trained to learn discriminant features to perform a certain task automatically from data. There are several studies that deal with LV identification based on deep learning algorithms from CMR images. Abdeltawab et al. proposed a novel DL approach for the automated segmentation and quantification of the LV from single frames of fast-cine CMR images [18]. The authors developed and validated their model using both a locally acquired and a public dataset collected from the Automated Cardiac Diagnosis Challenge (ACDC-2017). Their framework started with a localization of the LV blood pool center point using a Fully Convolutional Neural Network (FCN) architecture to extract an ROI that contains the LV from all heart sections. Niu et al. developed a model to detect myocardium from the CMR image series from Cardiac Atlas Project (CAP) dataset. The model consists of a region proposal followed by a deep-Stacked Sparse Auto-Encoder (SSAE), and finally, a candidate regions classification and regression to refine the location of myocardium [19].
Region-based Convolutional Neural Networks (R-CNN) are a particular class of DL models commonly employed in OD tasks. They can be distinguished into (i) two-stage detection framework, e.g., R-CNN, Fast R-CNN, Faster R-CNN, Feature Pyramid Network (FPN); (ii) and one-stage detection framework, e.g., You Only Look Once (YOLO), Single Shot Detector (SSD), CenterNet and EfficientDet [20]. Wang et al. proposed a CNN model for LV detection in cardiac MRI slices from the CAP dataset by combining discriminative dictionary learning and sequence tracking. They compared their proposed model with different R-CNN models (Faster R-CNN, SSD, and Yolov3) to investigate the model’s performance [21]. Shaaf et al. [8] also used an R-CNN, in particular Faster R-CNN, to detect LV from SAV images using the Sunnybrook Cardiac Dataset (SCD). The network takes as input a single slice of a selected frame of a Steady-State Free Precession (SSFP) sequence.
Despite the incredible success achieved by CNNs, interesting performances in several CV tasks have been performed for another architecture: the so-called Vision Transformers (ViTs) [22,23,24]. Transformers were originally introduced in the field of Natural Language Processing (NLP), where they achieved state-of-the-art performances. Inspired by the incredible success achieved in NLP, researchers started to apply transformers directly to images, generating the ViT models [22]. ViTs can overcome some of the major shortcomings of CNN architectures, such as the inability to model long-distance spatial relationships and the overreliance of CNNs on textures with a weakness in modeling shapes. Due to their local receptive field, CNNs have a limited ability to capture long-range spatial dependencies in an input image while ViTs use an alternative architectural design able to capture the global context of an input image and can easily model long-distance visual relationships. In addition, the ability to function based on shapes rather than on textures is also highly attractive and promises better generalizability and robustness [25]. ViTs’ architecture includes a new attention-driven building block, which is a neural network layer that aggregates information from the entire input sequence, exploiting the so-called “self-attention (SA) mechanism” [23]. We can identify several ViT architectures for OD, some based on “pure” transformer models (only ViT layers), and others on a hybrid framework (both convolutional and transformer layers), e.g., DEtection TRansformer (DETR), COnvolutional and TRansformer layers (COTR) [23].
It should be emphasized that at this time it is unclear if the level of generalization and performance improvement that is seen in NLP applications with large networks and large datasets will also emerge in medical imaging applications. Many questions regarding the optimal architecture are unresolved, and we are truly just at the beginning of this exploration [25]. Finally, to the best of our knowledge, only a few attempts have been made in applying transformer models in medical image detection [23] and no one applied them to CMR relaxometry sequences.
In this work, we implemented an object detector architecture based on ViT model [22] to automatically detect LV from CMR multi-echo T2* sequences collected from the Italian Myocardial Iron Overload in Thalassemia (MIOT) Network [26]. We also tested the detector’s performances using as input an independent test set of T2*, T2, and T1 sequences to check its generalizability power also in different relaxometry acquisitions.

2. Materials and Methods

2.1. Study Population

Baseline images from 530 patients were retrospectively studied. All patients were consecutively enrolled from 2009 to 2020 in the core center of the MIOT/eMIOT (Myocardial Iron Overload in Thalassemia) network. This network is constituted by thalassemia and MRI centers where MRI exams are performed using homogeneous, standardized, and validated procedures and where patients’ clinical-instrumental data are collected in a centralized, web-based database [26,27]. The inclusion criterion was the availability of a CMR T2* multiecho sequence that is mandatory in the study protocol and was available in all patients. Twenty patients enrolled in an ancillary study with T2*, T1, and T2 mapping sequences available were also added to the study to serve as the external test set. The study complied with the Declaration of Helsinki. All subjects gave written informed consent to the protocol. The project was approved by the institutional ethics committee.

2.2. MR Imaging

MR exams were performed on 1.5T scanners (GE Signa/Excite HD, Milwaukee, WI, USA) using a cardiac phased-array receiver surface coil for signal reception. Three parallel short-axis views (basal, middle, and apical) of the LV were obtained by a T2* gradient–echo multiecho sequence (matrix = 256 × 256, FOV = 35 × 35 cm, thickness = 8.0 mm) with electrocardiogram triggering. Each single SAV was acquired at ten echo times (TEs 2.0–22 ms with an echo spacing of 2.26 ms) in a single end-expiratory breath-hold [28]. For T2 mapping, three parallel short-axis slices (basal, mid-ventricular, and apical) of the LV were acquired in end diastole by a Multi-Echo Fast-Spin-Echo (MEFSE) sequence (matrix = 256 × 256, the field of view—FOV = 35 × 35 cm, thickness = 8.0 mm) with 4 echo times (TEs): 9.78, 34.22, 58.66, and 83.10 ms. A black blood-prepulse (double-inversion-recovery) was applied [29]. For T1 mapping, three parallel short-axis slices (basal, middle, and apical) of the LV were acquired in end-diastole by a Modified Look-Locker Inversion Recovery (MOLLI) sequence (matrix = 256 × 256, FOV = 35 × 35 cm, thickness = 8.0 mm) with a 3 (3 s) 3 (3 s) 5 scheme In-line motion correction was applied to MOLLI images [30]. Figure 1 reports typical T2*, T2, and T1 image sequences.

2.3. Ground Truth

T2* image analysis was performed by trained MRI operators using a custom-written, previously validated software (HIPPO MIOT®) [28]. Endocardial and epicardial contours were manually traced on the three slices at the TE value providing a better contrast between myocardium and surrounding tissues and propagated among the other frames. The myocardium defined in the previous step was automatically segmented into equiangular segments. Six segments were used in the basal and middle slices, and four were used in the apical slice. For each segment, the mean value of the signal intensity along all TE values was calculated and fitted with a single exponential decay model, obtaining the corresponding T2* segmental value. The global heart T2* value was obtained by averaging all 16 segmental T2* values, according to the standard AHA model [3]. The assessed global T2* values ranged from 2 to 57 ms (mean 37.8 ± 8.97 ms), covering the entire spectrum of clinically relevant T2* values. Ground Truth (GT) bounding boxes were generated by creating the least rectangular region of interest that contains the entire epicardial mask. Each bounding box was defined by the  ( x , y )  location of the top left and bottom right corners, normalized for the image size. As the correct detection of the Blood Pool (BP) region is important in several applications, the ground truth BP was defined as the center of the bounding box containing the endocardial contour. A modified version of the HIPPO MIOT® software was used to define endocardium and epicardium in T2 and T1 sequences following a similar approach.

2.4. ViT Architecture

We selected the ViT architecture developed by Dosovitskiy et al. [22], substituting the final classification layer with a regression layer to predict bounding box coordinates. This section focuses on the description of the basic ViT components [23].
A ViT splits the input image into patches, then flattens and projects them into a positionally-embedded feature space. An encoder subsequently processes the embedded input to produce the final output. The transformer encoder implements the SA mechanism to determine the relative importance of a single embedded patch with respect to all others in the input. ViT uses a channel-wise concatenation of multiple SA blocks to model complex dependencies between the different elements of the input [23].
In Figure 2, we reported an overall scheme of the operation performed by the model implemented.
Given a certain 2D image sequence:
x R R × C × N f
ViT splits it into patches of fixed sizes and vectorizes them using a flattening operation. We obtain a sequence of flattened 2D patches:
x P R N × ( P 2 N f )
where
  • R × C × N f = R o w s × C o l u m n s × N f r a m e s  of the original input image
  • P × P : size of a patch
  • N = R C P 2 : resulting number of patches
  • P 2 N f : dimension of flattened patch
ViT next applies a linear projection to the flattened patches to create a low-dimensional linear embedding. The weights of this linear layer are learned during the training phase. We obtain a linear embedding of the sequences of flattened patches, which, are furthermore positionally encoded:
X R N × D
where D is the dimension of the embedded space.
The embedded patches are sent into the transformer encoder block, which implements the Multi-Head Self Attention (MHSA) mechanism composed of multiple SA blocks. A single transformer’s architecture can be constituted by a sequence of multiple transformer blocks connected in a cascade.
The SA mechanism captures the interaction between all the N embeddings to determine the importance of each embedded patch with respect to all the others.
The input sequence of patches X is projected into three different spaces:
  • Queries:  Q = X W Q
    -
    W Q R D × D q
  • Keys:  K = X W K
    -
    W K R D × D k
  • Values:  V = X W V
    -
    W V R D × D v
All the weights matrices  W Q , W K , W V  are matrices of learnable weights and  D q = D k .
Then, we use this projection to generate the corresponding attention matrix and compute the output of a SA layer.
Attention Matrix,  A R N × N :  
A = s o f t m a x Q K T D q
Output of the SA layer,  Z R N × D v  
Z = A V
An MHSA block is composed of a channel-wise concatenation of multiple SA blocks  [ Z 0 , Z 1 , , Z h 1 ] . Each SA block (head) has its learnable weight matrices  { W Q i , W K i , W V i }  with  i = 0 , , ( h 1 )  and where h denotes the number of heads in the MHSA block.
The MHSA block returns as the final output a linear projection of the multiple heads:
M H S A ( Q , K , V ) = [ Z 0 , Z 1 , , Z h 1 ] W O
where  W O R h · D v × N  computes linear transformation of heads and a single head  Z i  can be written as:
Z i = s o f t m a x X W Q i ( X W K i ) T D q / h X W V i
Each transformer encoder block implements first a layer normalization followed by an MHSA operation.
Then, the output of the MHSA is added to the input of the normalization layer via a skip connection. The output of the skip-connection is further normalized and sent to a Multi-Layer Perceptron (MLP) block (multiple layers of fully connected units). Finally, the output of the MLP is added to the input of the last normalization layer via a skip connection.
The output of the transformer encoder goes to an MLP which returns as output the four dimensions representing the bounding box coordinates of an object, in our case the  ( x , y )  coordinates of the top left and bottom right angles of the bounding box.

2.5. Model Training and Validation

We selected the first, fifth, and ninth echo times from the original sequence to train the network. Each scan given as input to the network is composed of  256 × 256  slices at the 3 different selected echo times; hence, the input image sequence dimensions are  R = 256 C = 256 N f = 3 . No further pre-processing operations were applied in input sequences.
We selected the ViT architecture from the Keras implementation available on GitHub trained on the Caltech 101 dataset [31]. We selected this architecture as it results in a reduced number of parameters to be optimized if compared to the original ViT-Base version [22]. This may help in preventing overfitting considering the limited dimension of our dataset. The model is characterized by the following hyperparameters:
  • Patch size  p = 32  (resulting #patches  N = 64 );
  • Dimension of embedded space  D = 64 ;
  • #heads in each MHSA layer  h = 4  with dropout  = 0.1 ;
  • An MLP in each transformer encoder with 2 layers of, respectively, 128 and 64 units and dropout  = 0.1 ;
  • A final MLP with 5 layers of, respectively, 2048, 1024, 512, 64, and 32 units and dropout  = 0.3 .
To avoid data leakage problems we split the dataset into training, validation, and test sets so that basal, middle, and apical slices of the same subjects go into the same set [32].
The entire dataset employed is composed of 1590 images (3 slices of 530 different subjects). We used 10% and 20% as respective validation and test split, resulting in 1143 images in the training set (3 slices of 381 different subjects), 129 in the validation set (3 slices of 43 different subjects), 318 in the test set (3 slices of 106 different subjects). We applied on-the-fly data augmentation on the training set using random shift along both the x and y axes (range  [ 20 ÷ 20 ]  pixels). We followed a stratified k-fold cross-validation (k = 5) strategy for the training + validation/test split and we reported the average performances on the test set over the 5 folds.
We trained from scratch our neural network by using an Adam optimizer to minimize the  1 D I C E  index between the GT and the predicted bounding box as the loss function. We used a mini-batch strategy with batch size  = 32 . We set an initial learning rate of  10 4  and we decreased it during the training with a decay rate  = 0.96  and a decay steps =100,000. We set a maximum number of epochs  = 300 . All the other optimizer’s parameters were left at the default value.
We evaluated the model’s performances over the 5 folds on the test set, computing the following metrics:
  • Intersection over Union ( I o U ) or Jaccard Index:
    I o U = | B B G T B B p r e d | | B B G T B B p r e d |
  • D I C E  Index:
    D I C E = 2 | B B G T B B p r e d | | B B G T | + | B B p r e d |
    where  B B G T  and  B B p r e d  represent, respectively, the ground truth and predicted bounding boxes.
  • Centre Point Absolute Error,  ϵ C P , A : Euclidean distance between the center point of the ground truth and predicted bounding boxes;
  • Centre Point Epicardial Fractional Error,  ϵ C P , e p i ϵ C P , A  normalized by the radius of a circle having the same area as the epicardial mask;
  • Centre Point Endocardial Fractional Error,  ϵ C P , e n d o ϵ C P , A  normalized by the radius of a circle having the same area as the endocardial mask;
  • Correct Identification Rate,  C I R : Rate that the Centre Point of the predicted bounding box falls within the endocardial mask on the entire test set.
We selected the best model obtained over the five folds and we evaluated the performances on the three different independent test sets.

2.6. Statistical Analysis

The normality of the distribution of the parameters was assessed by using the Kolmogorov–Smirnov test or the Shapiro–Wilk test for a sample size ≤ 50.
One-way repeated measures ANOVA or the Friedman test were used to evaluate if there was a significant difference among parameters in different slices and the Bonferroni adjustment was used in all pairwise comparisons.
For continuous values with normal distribution, comparisons between groups were made by independent-sample t-test (for 2 groups) or one-way ANOVA (for more than 2 groups). Wilcoxon’s signed rank test or Kruskal–Wallis test were applied for continuous values with non-normal distribution. The Bonferroni post hoc test was used for multiple comparisons between pairs of groups.
We also evaluated if performances in the independent T2*, T2, and T1 series significantly differ compared to the ones reported on the test set considering the best fold.
In all tests, a 2-tailed probability value of  p = 0.05  was considered statistically significant.

2.7. Hardware and Software Specification

The proposed deep learning model was implemented using Python utilities (version 3.9), with Keras framework on the Tensorflow backend (version 2.6.0). The training was performed on an Intel Core i7 5.1 MHz PC, 32 Gb RAM, equipped with an NVIDIA RTX3090 GPU with 24 Gb of embedded RAM. Developed code is available at https://github.com/desantilisa/ViT_LeftVentricle_Object_Detection (accessed on 26 January 2023).
Statistical analysis was performed using the SPSS version 27.0 statistical package.

3. Results

In Figure 3, we reported examples of the model’s output of three different subjects belonging to the test set in basal, middle, and apical slices. In Figure 4, we reported examples of the model’s output of a selected subject for all three independent test sets in basal, middle, and apical slices.
In Table 1, we reported the averaged performances evaluated in the test set over the five folds differentiated by short-axis views of the AHA model (basal, middle, and apical) and the global ones. Averaged training times was 1 h ± 5 min; averaged single prediction time once the network was trained was 20 ± 4 ms.
In Table 2, we reported the averaged performances over the test set using the model that achieves the best performances (fold #1). Averaged performances on the independent T2*, T2, and T1 test sets were also reported. Statistical difference in performances between slices was observed in the test set between apical vs. basal and apical vs. middle  I o U D I C E ϵ C P , e p i  and  ϵ C P , e n d o  with a reported  p 0.0001  in all pairwise comparisons. No relevant difference in performances was observed according to slice type in the T2* independent test sets ( I o U p = 0.405 D I C E p = 0.405 ϵ C P , A p = 0.495 ϵ C P , e p i p = 0.066 ϵ C P , e n d o p = 0.066 ), T2 independent test sets ( I o U p = 0.827 D I C E p = 0.827 ϵ C P , A p = 0.867 ϵ C P , e p i p = 0.097 ϵ C P , e n d o p = 0.066 ), and T1 independent test sets ( I o U p = 0.717 D I C E p = 0.717 ϵ C P , A p = 0.717 ϵ C P , e p i p = 0.565 ϵ C P , e n d o p = 0.172 ).
The statistical analysis highlighted no significant difference between performances in the test set and the T2* independent test set. We reported a significant difference in  I o U  basal and middle ( p < 0.0001 ),  D I C E  basal and middle ( p < 0.0001 ), and  ϵ C P , A  in all three slice types in the test set and T2 independent test set (basal:  p = 0.024 , middle:  p < 0.0001 , apical:  p = 0.036 ). Differences in  I o U  basal ( p = 0.024 ) and  D I C E  basal ( p = 0.024 ) were also observed in the test set and T1 independent test set.

4. Discussion

In our study, we applied a ViT architecture to automatically localize the LV blood pool from CMR multi-echo T2* relaxometry sequences collected from the Italian MIOT. We also tested the detector’s performances in an independent test set of different relaxometry sequences to check its generalization capability to different types of acquisition.
Transformer networks are novel architectures that obtained state-of-the-art performances in NLP tasks. Due to their appealing properties, these models are gaining ground even in the medical imaging community; however, at present, there are only a limited number of applications that perform OD tasks in this field. To the best of our knowledge, this is the first attempt to train a ViT network to localize LV from CMR relaxometry sequences [23,25].
LV detection is an important step to facilitate subsequent operations to perform quantitative measures of myocardial structure, anatomy, and function. We found several works in the literature that deal with the detection of LV from fast-cine CMR using different strategies exploiting both handcrafted and deep learning-based approaches. A direct comparison of our results with the ones reported in the literature is difficult, as most of the studies are focused on LV segmentation, so the efficiency of the preliminary LV location task was not quantitatively assessed. Other recent studies focused on heart structure localization do not include comparison with ground truth data as in [33]. We selected the studies that, to our knowledge, assessed quantitatively the efficiency of the LV localization task in Table 3. The studies in which available results can be compared with ours are highlighted in the table. All of them perform the LV localization task in fast-cine CMR sequences, while we do not find any attempts to localize LV from relaxometry acquisitions. Fast-cine sequences are designed to assess cardiac function by capturing the heart movement during the cardiac cycle. In these sequences, the LV shape changes over time while the other anatomical structures remain almost fixed. Hence, an OD algorithm designed to detect LV from fact-cine sequences resembles a motion detection algorithm able to identify a moving object on a fixed background. Other important CMR protocols, such as first-pass perfusion and relaxometry, are designed to characterize the myocardium tissue, with or without the use of a contrast medium. In these sequences, the acquisition is typically performed in the diastolic phase, avoiding heart motion. Hence, an OD algorithm designed to detect LV from these sequences should detect the signal change in LV over time due to the passing of a contrast medium (perfusion sequences) or the change of an MR acquisition parameter (TE for T2* and T2 relaxometry and TI for T1 relaxometry). LV detection from a relaxometry sequence represents a challenging task because the signal of all tissues in the field of view changes over time. Several papers in the literature deal with DL-based segmentation of CMR relaxometric sequences [34,35,36], but to the best of our knowledge, none of them included a preliminary LV localization task. The inclusion of an accurate cropping step based on LV localization may bring benefits even in segmentation due to the reduction in the computational space for subsequent operation.
We collected a global  I o U  of 0.68 and a  C I R  of 0.99. This means that the center of the bounding box predicted falls within the LV blood pool 99% of times. This could be particularly useful to automatically define a seed for any subsequent segmentation step in near real-time (20 ± 4 ms). The average  I o U  indexes in the three independent test sets were, respectively, 0.68 (T2*), 0.62 (T2), and 0.67 (T1), and the global  C I R  0.94 (T2*), 0.95 (T2), and 0.98 (T1).
Several authors evaluated their performances according to the ability of the algorithm to predict the position of a centroid that falls within the LV blood pool, which we named  C I R . By comparing our results reported in Table 1 with  C I R  obtained in relevant works we observed that our algorithm achieves state-of-the-art performances in localizing the LV centroid, both in global performances and in that ones differentiated by slice position [4,5,6]. We obtained a comparable  D I C E  index, and a slightly worse  ϵ C P , A , and  ϵ C P , e n d o  with the ones reported by Tan et al. [6]. The authors localized the LV exploiting the heart’s motion. Cinetic information could not be exploited in our dataset as the T2* mapping protocol produces static sequences. The findings obtained highlight the great potential of deep learning models compared to the hand-crafted approaches.
As our dataset contains LV in all the input images, the developed model always returns as output a bounding box labeled as LV, and this would boost true and false positive and false negative rates and performances based on them such as Precision, Recall, and Average Precision (AP). The absence of slices without LV depends on the protocol employed to acquire the T2* sequences, which requires the acquisition of the three parallel short-axis views (basal, middle, and apical slices) of LV. We decided to not compute these metrics reporting only measurements of how close the bounding box predicted is to the GT bounding box, so our results are not comparable with the ones of Niu et al. [19], Wang et al. [21] and Shaaf et al. [8]. The only result comparable with the work of Shaaf et al. [8] is an averaged  I o U  over a cardiac cycle of 20 slices of 0.83, which authors reported as an Ovelap Ratio. We obtained a global  I o U  lower than the aforementioned work (0.68), but the authors reported only the results relative to a single acquisition, while our global  I o U  was obtained averaging 318 different images.
We used statistical analysis to verify if the performances evaluated differ according to slice type (basal, middle, or apical), as a smaller object (typically apical slices) could be potentially harder to be localized. Statistical tests confirm that ViT reported worse performances in performing localization in apical slices.
We selected the trained network which achieved the best results over the five different trials, and we applied it to an independent dataset of three different relaxometry sequences (T2*, T2, and T1). Statistical analysis did not highlight any significant difference in performances reported in the test set and in the T2* independent test set. This is reassuring since the two sets use the same mapping sequences. When applied to the T2 and T1, the independent test set statistical analysis highlighted lower performances as expected since the network had been training with different types of input. However, it should be noted that overall, the performances could be considered acceptable and  C I R  values remain consistent with relevant literature, with performances particularly promising for the T1 mapping.
Some limitations can be recognized in the present study. Firstly, training was performed on a dataset of limited size composed of a single scanner, a single bright-blood sequence dataset. Black-blood T2* images [37] were not included. Moreover, to match ViT’s architectures, we selected only 3 TEs from the entire multi-echo input sequence. All these aspects may affect the generalizability power of the developed model and the statistical significance of the results obtained. In the future, an extension of the study to increase the dataset’s dimension and also including multi-scanner acquisition could be advisable to overcome these limitations. A comparison of detection results obtained was performed with works that use a different type of CMR sequences (fast cine CMR), as we did not find any attempts to localize LV from relaxometry acquisitions in the literature. As our dataset contains LV in all the input images due to the standardized acquisition protocol, we could not assess the presence of LV differently from other works that performed both identification and location. Even if from a statistical point of view performances cannot be considered as perfectly generalizable to T2 and T1 mapping, considering its promising results, the present architecture could constitute a good starting point to apply transfer learning to T2 and T1 relaxometry sequences. So, future development of the work will include the fine-tuning of the present architecture on different types of relaxometry sequences.

5. Conclusions

In our research, we applied a ViT to automatically predict an ROI in CMR T2* mapping multi-echo sequences collected by the MIOT network. We evaluated detection performances performing k-fold cross-validation, and we differentiated them by slice view according to the AHA model. Once trained, we tested the model on an independent test set which included different types of mapping (T2*, T2, T1).
To the best of our knowledge, this is the first attempt to automatically detect LV from CMR relaxometry images using a ViT, a recently introduced deep learning architecture that may overcome different limitations of CNN models. All the other works in the literature performed the detection task on different types of imaging modalities, typically fast cine MR images. This study defines a benchmark in LV detection from CMR relaxometry sequences; with a  C I R  higher than those obtained in works that performed the same localization task but on different types of CMR acquisitions. The statistical analysis highlighted differences in the network’s performances according to the short axes view. In particular, we reported worse performances in apical slices; this may be due to the typically smaller size of the apical regions, which may result in being harder to be detected.
As results were collected on a dataset of relatively limited dimensions, in the future, an extension of the study will be performed to enhance the generalizability and statistical significance of the developed model. Performances were statistically different in the T2 and T1 independent sets but still promising considering that our network had been trained to detect LV from different types of imaging modalities.
The final trained model has been made publicly available and suitable for transfer learning. Future development will include the fine-tuning of the model on different relaxometric sequences.

Author Contributions

Conceptualization, L.A.D.S. and V.P.; methodology, L.A.D.S., A.M. and M.F.S.; software, L.A.D.S. and V.P.; validation, A.M.; data curation, L.P., A.S., T.C., M.C.P. and L.C.; writing—original draft preparation, L.A.D.S.; writing—review and editing, V.P., A.M. and M.F.S.; supervision, F.C.; project administration, F.C. All authors have read and agreed to the published version of the manuscript.

Funding

The E-MIOT project receives “no-profit support” from industrial sponsorships (Chiesi Farmaceutici S.p.A. and Bayer). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Ethics Committee of Area Vasta Nord Ovest (protocol code 56664, date of approval 8 October 2015).

Informed Consent Statement

Informed consent was obtained from all patients involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Acknowledgments

We would like to thank all the colleagues involved in the E-MIOT project (https://emiot.ftgm.it/ accessed on 26 January 2023) and all patients for their cooperation.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SAVShort Axis Views
LVLeft Ventricle
CMRCardiac Magnetic Resonance
AHAAmerican Heart Association
CVComputer Vision
ODObject Detection
ROIRegion Of Interest
DLDeep Learning
AIArtificial Intelligence
ACDC-2017Automated Cardiac Diagnosis Challenge
FCNFully Convolutional Neural network
CAPCardiac Atlas Project
SSAEdeep-stacked sparse auto-encoder
R-CNNRegion-based Convolutional Neural Networks
YOLOYou Only Look Once
SSDSingle Shot Detector
SCDSunnybrook Cardiac Dataset
SSFPSteady-State Free Precession
ViTsVision Transformers
NLPNatural Language Processing
SASelf Attention
DETRDEtection TRansformer
COTRCOnvolutional and TRansformer layers
MIOTMyocardial Iron Overload in Thalassemia
GTGround Truth
BTBlood Pool
MHSAMulti Head Self Attention
MLPMulti-Layer Perceptron
  I o U Intersection over Union
  C I R Correct Identification Rate

References

  1. Pennell, D.J.; Sechtem, U.P.; Higgins, C.B.; Manning, W.J.; Pohost, G.M.; Rademakers, F.E.; van Rossum, A.C.; Shaw, L.J.; Yucel, E.K. Clinical indications for cardiovascular magnetic resonance (CMR): Consensus Panel report. Eur. Heart J. 2004, 25, 1940–1965. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Schulz-Menger, J.; Bluemke, D.A.; Bremerich, J.; Flamm, S.D.; Fogel, M.A.; Friedrich, M.G.; Kim, R.J.; Knobelsdorff-Brenkenhoff, F.V.; Kramer, C.M.; Pennell, D.J.; et al. Standardized image interpretation and post-processing in cardiovascular magnetic resonance—2020 update: Society for Cardiovascular Magnetic Resonance (SCMR): Board of Trustees Task Force on Standardized Post-Processing. J. Cardiovasc. Magn. Reson. 2020, 22, 1–22. [Google Scholar] [CrossRef] [PubMed]
  3. Cerqueira, M.D.; Weissman, N.J.; Dilsizian, V.; Jacobs, A.K.; Kaul, S.; Laskey, W.K.; Pennell, D.J.; Rumberger, J.A.; Ryan, T.; Verani, M.S. Standardized Myocardial Segmentation and Nomenclature for Tomographic Imaging of the Heart. Circulation 2002, 105, 539–542. [Google Scholar] [CrossRef] [PubMed]
  4. Pednekar, A.S.; Muthupillai, R.; Lenge, V.V.; Kakadiaris, I.A.; Flamm, S.D. Automatic identification of the left ventricle in cardiac cine-MR images: Dual-contrast cluster analysis and scout-geometry approaches. J. Magn. Reson. Imaging 2006, 23, 641–651. [Google Scholar] [CrossRef] [PubMed]
  5. Kurkure, U.; Pednekar, A.; Muthupillai, R.; Flamm, S.D.; Kakadiaris, I.A. Localization and segmentation of left ventricle in cardiac cine-MR images. IEEE Trans. Biomed. Eng. 2009, 56, 1360–1370. [Google Scholar] [CrossRef] [PubMed]
  6. Tan, L.K.; Liew, Y.M.; Lim, E.; Aziz, Y.F.A.; Chee, K.H.; McLaughlin, R.A. Automatic localization of the left ventricular blood pool centroid in short axis cardiac cine MR images. Med. Biol. Eng. Comput. 2018, 56, 1053–1062. [Google Scholar] [CrossRef] [PubMed]
  7. Attar, R.; Pereañez, M.; Gooya, A.; Albà, X.; Zhang, L.; de Vila, M.H.; Lee, A.M.; Aung, N.; Lukaschuk, E.; Sanghvi, M.M.; et al. Quantitative CMR population imaging on 20,000 subjects of the UK Biobank imaging study: LV/RV quantification pipeline and its evaluation. Med. Image Anal. 2019, 56, 26–42. [Google Scholar] [CrossRef]
  8. Shaaf, Z.F.; Jamil, M.M.A.; Ambar, R.; Alattab, A.A.; Yahya, A.A.; Asiri, Y. Detection of Left Ventricular Cavity from Cardiac MRI Images Using Faster R-CNN. Comput. Mater. Contin. 2023, 74, 1819–1835. [Google Scholar] [CrossRef]
  9. Papetti, D.M.; Van Abeelen, K.; Davies, R.; Menè, R.; Heilbron, F.; Perelli, F.P.; Artico, J.; Seraphim, A.; Moon, J.C.; Parati, G.; et al. An accurate and time-efficient deep learning-based system for automated segmentation and reporting of cardiac magnetic resonance-detected ischemic scar. Comput. Methods Programs Biomed. 2023, 229, 107321. [Google Scholar] [CrossRef]
  10. Xue, H.; Brown, L.A.; Nielles-Vallespin, S.; Plein, S.; Kellman, P. Automatic in-line quantitative myocardial perfusion mapping: Processing algorithm and implementation. Magn. Reson. Med. 2020, 83, 712–730. [Google Scholar] [CrossRef] [Green Version]
  11. Bhatt, N.; Ramanan, V.; Gunraj, H.; Guo, F.; Biswas, L.; Qi, X.; Roifman, I.; Wright, G.A.; Ghugre, N.R. Technical Note: Fully automatic segmental relaxometry (FASTR) for cardiac magnetic resonance T1 mapping. Med. Phys. 2021, 48, 1815–1822. [Google Scholar] [CrossRef] [PubMed]
  12. Positano, V.; Meloni, A.; Santarelli, M.F.; Gerardi, C.; Bitti, P.P.; Cirotto, C.; De Marchi, D.; Salvatori, C.; Landini, L.; Pepe, A. Fast generation of T2* maps in the entire range of clinical interest: Application to thalassemia major patients. Comput. Biol. Med. 2015, 56, 200–210. [Google Scholar] [CrossRef] [PubMed]
  13. Jiang, X.; Hadid, A.; Pang, Y.; Granger, E.; Feng, X. Deep Learning in Object Detection and Recognition; Springer: Singapore, 2019; pp. 1–224. [Google Scholar] [CrossRef]
  14. Awad, A.I.; Hassaballah, M. Studies in Computational Intelligence 630 Image Feature Detectors and Descriptors Foundations and Applications; Springer: Cham, Switzerland, 2016. [Google Scholar] [CrossRef]
  15. Lin, X.; Cowan, B.R.; Young, A.A. Automated Detection of Left Ventricle in 4D MR Images: Experience from a Large Study. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2006, Copenhagen, Denmark, 1–6 October 2006; Larsen, R., Nielsen, M., Sporring, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; pp. 728–735. [Google Scholar]
  16. Jolly, M.P. Automatic Recovery of the Left Ventricular Blood Pool in Cardiac Cine MR Images. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2008, New York, NY, USA, 6–10 September 2008; Metaxas, D., Axel, L., Fichtinger, G., Székely, G., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; pp. 110–118. [Google Scholar]
  17. Zhong, L.; Zhang, J.M.; Zhao, X.; Tan, R.S.; Wan, M. Automatic localization of the left ventricle from cardiac cine magnetic resonance imaging: A new spectrum-based computer-aided tool. PLoS ONE 2014, 9, e0092382. [Google Scholar] [CrossRef] [PubMed]
  18. Abdeltawab, H.; Khalifa, F.; Taher, F.; Alghamdi, N.S.; Ghazal, M.; Beache, G.; Mohamed, T.; Keynton, R.; El-Baz, A. A deep learning-based approach for automatic segmentation and quantification of the left ventricle from cardiac cine MR images. Comput. Med. Imaging Graph. 2020, 81, 101717. [Google Scholar] [CrossRef]
  19. Niu, Y.; Qin, L.; Wang, X. Myocardium detection by deep SSAE feature and within-class neighborhood preserved support vector classifier and regressor. Sensors 2019, 19, 1766. [Google Scholar] [CrossRef] [Green Version]
  20. Yang, R.; Yu, Y. Artificial Convolutional Neural Network in Object Detection and Semantic Segmentation for Medical Imaging Analysis. Front. Oncol. 2021, 11, 638182. [Google Scholar] [CrossRef]
  21. Wang, X.; Wang, F.; Niu, Y. A convolutional neural network combining discriminative dictionary learning and sequence tracking for left ventricular detection. Sensors 2021, 21, 3693. [Google Scholar] [CrossRef]
  22. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  23. Shamshad, F.; Khan, S.; Zamir, S.W.; Khan, M.H.; Hayat, M.; Khan, F.S.; Fu, H. Transformers in Medical Imaging: A Survey. arXiv 2022, arXiv:2201.09873. [Google Scholar]
  24. Pachetti, E.; Colantonio, S.; Pascali, M.A. On the Effectiveness of 3D Vision Transformers for the Prediction of Prostate Cancer Aggressiveness. In Proceedings of the Image Analysis and Processing, ICIAP 2022 Workshops, Lecce, Italy, 23–27 May 2022; Mazzeo, P.L., Frontoni, E., Sclaroff, S., Distante, C., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 317–328. [Google Scholar]
  25. Willemink, M.J.; Roth, H.; Sandfort, V. Toward Foundational Deep Learning Models for Medical Imaging in the New Era of Transformer Networks. Radiol. Artif. Intell. 2022, 4, e210284. [Google Scholar] [CrossRef]
  26. Pepe, A.; Pistoia, L.; Gamberini, M.R.; Cuccia, L.; Lisi, R.; Cecinati, V.; Maggio, A.; Sorrentino, F.; Filosa, A.; Rosso, R.; et al. National networking in rare diseases and reduction of cardiac burden in thalassemia major. Eur. Heart J. 2021, 43, 2482–2492. [Google Scholar] [CrossRef] [PubMed]
  27. Meloni, A.; Ramazzotti, A.; Positano, V.; Salvatori, C.; Mangione, M.; Marcheschi, P.; Favilli, B.; De Marchi, D.; Prato, S.; Pepe, A.; et al. Evaluation of a web-based network for reproducible T2* MRI assessment of iron overload in thalassemia. Int. J. Med. Inform. 2009, 78, 503–512. [Google Scholar] [CrossRef] [PubMed]
  28. Positano, V.; Pepe, A.; Santarelli, M.F.; Scattini, B.; De Marchi, D.; Ramazzotti, A.; Forni, G.; Borgna-Pignatti, C.; Lai, M.E.; Midiri, M.; et al. Standardized T2* map of normal human heart in vivo to correct T2* segmental artefacts. NMR Biomed. 2007, 20, 578–590. [Google Scholar] [CrossRef]
  29. Meloni, A.; Nicola, M.; Positano, V.; D’Angelo, G.; Barison, A.; Todiere, G.; Grigoratos, C.; Keilberg, P.; Pistoia, L.; Gargani, L.; et al. Myocardial T2 values at 1.5 T by a segmental approach with healthy aging and gender. Eur. Radiol. 2022, 32, 2962–2975. [Google Scholar] [CrossRef] [PubMed]
  30. Meloni, A.; Martini, N.; Positano, V.; D’Angelo, G.; Barison, A.; Todiere, G.; Grigoratos, C.; Barra, V.; Pistoia, L.; Gargani, L.; et al. Myocardial T1 Values at 1.5 T: Normal Values for General Electric Scanners and Sex-Related Differences. J. Magn. Reson. Imaging 2021, 54, 1486–1500. [Google Scholar] [CrossRef]
  31. Object Detection with Vision Transformers. Available online: https://github.com/keras-team/keras-io/blob/master/examples/vision/object_detection_using_vision_transformer.py (accessed on 1 October 2022).
  32. Varoquaux, G.; Cheplygina, V. Machine learning for medical imaging: Methodological failures and recommendations for the future. NPJ Digit. Med. 2022, 5, 48. [Google Scholar] [CrossRef]
  33. Amel, L.; Ammar, M.; El Habib Daho, M.; Mahmoudi, S. Toward an automatic detection of cardiac structures in short and long axis views. Biomed. Signal Process. Control. 2023, 79, 104187. [Google Scholar] [CrossRef]
  34. Fadil, H.; Totman, J.J.; Hausenloy, D.J.; Ho, H.H.; Joseph, P.; Low, A.F.H.; Richards, A.M.; Chan, M.Y.; Marchesseau, S. A deep learning pipeline for automatic analysis of multi-scan cardiovascular magnetic resonance. J. Cardiovasc. Magn. Reson. 2021, 23, 1–13. [Google Scholar] [CrossRef]
  35. Martini, N.; Meloni, A.; Positano, V.; Latta, D.D.; Keilberg, P.; Pistoia, L.; Spasiano, A.; Casini, T.; Barone, A.; Massa, A.; et al. Fully Automated Regional Analysis of Myocardial T2* Values for Iron Quantification Using Deep Learning. Electronics 2022, 11, 2749. [Google Scholar] [CrossRef]
  36. Howard, J.P.; Chow, K.; Chacko, L.; Fontana, M.; Cole, G.D.; Kellman, P.; Xue, H. Automated Inline Myocardial Segmentation of Joint T1 and T2 Mapping Using Deep Learning. Radiol. Artif. Intell. 2023, 5, e220050. [Google Scholar] [CrossRef]
  37. He, T.; Gatehouse, P.D.; Kirk, P.; Tanner, M.A.; Smith, G.C.; Keegan, J.; Mohiaddin, R.H.; Pennell, D.J.; Firmin, D.N. Black-blood T2* technique for myocardial iron measurement in thalassemia. J. Magn. Reson. Imaging 2007, 25, 1205–1209. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Examples of T2* (upper panel), T2 (middle panel) and T1 (lower panel) CMR image sequences.
Figure 1. Examples of T2* (upper panel), T2 (middle panel) and T1 (lower panel) CMR image sequences.
Sensors 23 03321 g001
Figure 2. Visual transformer general architecture. Our model receives as input multi-echo CMR sequences and predicts a bounding box to detect the left ventricle.
Figure 2. Visual transformer general architecture. Our model receives as input multi-echo CMR sequences and predicts a bounding box to detect the left ventricle.
Sensors 23 03321 g002
Figure 3. Examples of the model’s output in three different subjects belonging to the test set in basal, middle, and apical slices.
Figure 3. Examples of the model’s output in three different subjects belonging to the test set in basal, middle, and apical slices.
Sensors 23 03321 g003
Figure 4. Examples of the model’s output of a selected subject for all three independent test sets in basal, middle, and apical slices.
Figure 4. Examples of the model’s output of a selected subject for all three independent test sets in basal, middle, and apical slices.
Sensors 23 03321 g004
Table 1. Averaged object detection metrics (Mean ± St Dev) over the 5 folds evaluated in the test set.
Table 1. Averaged object detection metrics (Mean ± St Dev) over the 5 folds evaluated in the test set.
Slice   IoU   DICE   ϵ CP , A   ϵ CP , epi   ϵ CP , endo   CIR
Basal0.69 ± 0.030.81 ± 0.027.12 ± 0.970.27 ± 0.030.31 ± 0.041.00 ± 0.00
Middle0.71 ± 0.020.82 ± 0.026.52 ± 0.810.26 ± 0.030.31 ± 0.040.99 ± 0.00
Apical0.63 ± 0.010.77 ± 0.016.68 ± 0.770.33 ± 0.040.42 ± 0.050.97 ± 0.01
Global0.68 ± 0.020.80 ± 0.026.77 ± 0.810.29 ± 0.030.35 ± 0.040.99 ± 0.00
Table 2. Averaged object detection metrics (Mean ± St Dev) over T2* test set, T2* independent test set, T2 independent test set and T1 independent test set.
Table 2. Averaged object detection metrics (Mean ± St Dev) over T2* test set, T2* independent test set, T2 independent test set and T1 independent test set.
Slice   IoU   DICE   ϵ CP , A   ϵ CP , epi   ϵ CP , endo   CIR
T2* test set
Basal0.74 ± 0.100.84 ± 0.075.44 ± 2.890.21 ± 0.110.24 ± 0.131.00
Middle0.74 ± 0.100.85 ± 0.075.24 ±2.890.21 ± 0.120.25 ± 0.150.99
Apical0.65 ± 0.130.78 ± 0.115.65 ± 3.450.29 ± 0.190.36 ± 0.240.97
Global0.71 ± 0.120.82 ± 0.095.44 ± 3.100.23 ± 0.150.29 ± 0.190.99
T2* independent test set
Basal0.70 ± 0.150.81 ± 0.127.59 ± 6.330.25 ± 0.200.30 ± 0.231.00
Middle0.69 ± 0.140.81 ± 0.127.42 ± 6.470.26 ± 0.210.30 ± 0.240.95
Apical0.63 ± 0.210.76 ± 0.178.44 ± 7.080.36 ± 0.310.45 ± 0.380.86
Global0.68 ± 0.170.79 ± 0.147.82 ± 6.540.29 ± 0.240.35 ± 0.300.94
T2 independent test set
Basal0.62 ± 0.120.75 ± 0.108.56 ± 5.670.29 ± 0.180.34 ± 0.211.00
Middle0.62 ± 0.130.76 ± 0.109.32 ± 6.160.33 ± 0.210.39 ± 0.240.95
Apical0.63 ± 0.140.76 ± 0.118.90 ± 6.020.38 ± 0.300.47 ± 0.380.90
Global0.62 ± 0.130.76 ± 0.108.93 ± 5.860.33 ± 0.240.40 ± 0.290.95
T1 independent test set
Basal0.66 ± 0.090.79 ± 0.076.51 ± 2.900.22 ± 0.110.26 ± 0.121.00
Middle0.69 ± 0.100.81 ± 0.076.38 ± 3.370.24 ± 0.160.28 ± 0.191.00
Apical0.67 ± 0.130.80 ± 0.097.02 ± 4.570.31 ± 0.270.38 ± 0.340.95
Global0.67 ± 0.110.80 ± 0.086.64 ± 3.630.26 ± 0.190.31 ± 0.240.98
Table 3. Performances of LV localization tasks in relevant works in literature. Performances comparable with ours are highlighted.
Table 3. Performances of LV localization tasks in relevant works in literature. Performances comparable with ours are highlighted.
AuthorDatasetLV OD ApproachPerformances
Pednekar et al. [4]Vector Cardiographic Gating (VCG)-gated cine Steady-State Free-Precession SSFP cardiac MR images (locally acquired dataset)Two different algorithms:
  • Dual-contrast (DC) algorithm
  • Scout-geometry (SG) algorithm
Best results obtained with SG algorithm. C I R S G = 93 %
Kurkure et al. [5]Vector Cardiographic Gating VCG-gated cine SSFP cardiac MR images (locally acquired dataset)Scout-view geometry, blood-to-myocardial tissue contrast, and geometrical continuity constraint   C I R = 98.3 %
Zhong et al. [17]Cardiac cine MR images (locally acquired dataset), algorithm tested on 10 volunteersFourier Transform followed by anisotropic weighted circular Hough transformBy visual inspection, the cropped ROI of all cases contains the LV
Average Cropping Ratio = 0.165
Tan et al. [6]Cardiac cine MRI data:
  • 161 clinical cine datasets (locally acquired)
  • 1140 datasets from the Kaggle Datascience Bowl Cardiac Challenge
  • 45 datasets from the STACOM 2009 Cardiac MR LV Segmentation Challenge
Fourier transform over time to highlight regions of significant motion D I C E :  from 0.67 to 0.88
ϵ C P , A :  from 2.8 to 4.7 mm
ϵ C P , e n d o :  from 12 to  22 %  
C I R = 97.3 %
Niu et al. [19]Cardiac cine MR images from Cardiac Atlas Project (CAP) dataset 27 subjects test setRegion proposal + Deep-Stacked Sparse Auto-Encoder (SSAE) + Soft margin support vector (C-SVC) classifier and multiple-output support vector ( ϵ -SVR) regressor F 1 = 0.924
True positive rate,  T p r = 0.936
Positive predicted value  P p v = 0.916
A U C = 0.89
Abdeltawab et al. [18]Cardiac cine MR images:
  • Locally acquired datasets
  • ACDC-2017 MICCAI dataset
Fully Convolutional Neural network (FCN) ϵ C P , A  = 1.41 ± 1.65
Wang et al. [21]Cardiac cine MR images collected from CAP datasetTwo stage R-CNN based on VGG-16 backboneThree different R-CNN version:
  • Discriminant dictionary + Faster:
    • AP50 = 91.86;
    • AP75 = 84.17.
  • Correlation filter + Faster:
    • AP50 = 92.95;
    • AP75 = 84.19.
  • Correlation filter + Proposal:
    • AP50 = 92.32;
    • AP75 = 85.21.
Shaaf et al., 2023 [8]Cardiac cine MR images collected from SCD MICCAI workshop 2009Faster R-CNN I o U  = 0.83
Accuracy = 0.91
Recall = 0.95
Precision = 0.94
F1 = 0.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

De Santi, L.A.; Meloni, A.; Santarelli, M.F.; Pistoia, L.; Spasiano, A.; Casini, T.; Putti, M.C.; Cuccia, L.; Cademartiri, F.; Positano, V. Left Ventricle Detection from Cardiac Magnetic Resonance Relaxometry Images Using Visual Transformer. Sensors 2023, 23, 3321. https://doi.org/10.3390/s23063321

AMA Style

De Santi LA, Meloni A, Santarelli MF, Pistoia L, Spasiano A, Casini T, Putti MC, Cuccia L, Cademartiri F, Positano V. Left Ventricle Detection from Cardiac Magnetic Resonance Relaxometry Images Using Visual Transformer. Sensors. 2023; 23(6):3321. https://doi.org/10.3390/s23063321

Chicago/Turabian Style

De Santi, Lisa Anita, Antonella Meloni, Maria Filomena Santarelli, Laura Pistoia, Anna Spasiano, Tommaso Casini, Maria Caterina Putti, Liana Cuccia, Filippo Cademartiri, and Vincenzo Positano. 2023. "Left Ventricle Detection from Cardiac Magnetic Resonance Relaxometry Images Using Visual Transformer" Sensors 23, no. 6: 3321. https://doi.org/10.3390/s23063321

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop