Next Article in Journal
Virtual Training System for Unmanned Aerial Vehicle Control Teaching–Learning Processes
Previous Article in Journal
FOSSBot: An Open Source and Open Design Educational Robot
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multimodal CT Image Synthesis Using Unsupervised Deep Generative Adversarial Networks for Stroke Lesion Segmentation

College of Information and Computer, Taiyuan University of Technology, Taiyuan 030024, China
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(16), 2612; https://doi.org/10.3390/electronics11162612
Submission received: 15 July 2022 / Revised: 12 August 2022 / Accepted: 17 August 2022 / Published: 20 August 2022
(This article belongs to the Section Computer Science & Engineering)

Abstract

:
Deep learning-based techniques can obtain high precision for multimodal stroke segmentation tasks. However, the performance often requires a large number of training examples. Additionally, existing data extension approaches for the segmentation are less efficient in creating much more realistic images. To overcome these limitations, an unsupervised adversarial data augmentation mechanism (UTC-GAN) is developed to synthesize multimodal computed tomography (CT) brain scans. In our approach, the CT samples generation and cross-modality translation differentiation are accomplished simultaneously by integrating a Siamesed auto-encoder architecture into the generative adversarial network. In addition, a Gaussian mixture translation module is further proposed, which incorporates a translation loss to learn an intrinsic mapping between the latent space and the multimodal translation function. Finally, qualitative and quantitative experiments show that UTC-GAN significantly improves the generation ability. The stroke dataset enriched by the proposed model also provides a superior improvement in segmentation accuracy, compared with the performance of current competing unsupervised models.

1. Introduction

Stroke is the problem with blood supply blocking in cerebral vessels, and it is the most prevalent cause of mortality and acquired handicap [1,2]. Among various types of strokes, ischemic stroke is reported in a large proportion, and it mainly induces brain cell death and fatal paralysis. Hence, early diagnosis and quantification of the lesions could help stroke patients achieve effective recovery, which also benefits clinicians in optimizing the therapeutic schedules. Quantitative stroke lesion segmentation from medical imaging is a necessary procedure for the doctor to make decisions. Additionally, computed tomography (CT) is a typical effective non-incursion technique to evaluate the lesion regions of stroke patients [3,4]. It also has the merits of speediness, wide availability and inexpensiveness in detecting brain structure by ionizing radiation. Moreover, CT perfusion modalities including cerebral blood volume (CBV), cerebral blood flow (CBF), mean transit time (MTT) and time to peak of the residue function (Tmax) are also successfully used to assess stroke infarct core size [5]. However, precise segmentation based on these diagnostic means requires rich experience and a significant amount of time from physicians.
Recently, deep neural network-based methods have shown a remarkable impact on the segmentation accuracy for various medical images [6,7]. However, the scarcity of labelling multimodal pictures due to the enormous time cost and complex acquisition procedures often leads to low segmentation accuracy. Although previous image augmentation strategies such as rotation, flipping and elastic deformation have been applied widely to expand the volume of the dataset [8,9,10], it is impossible to produce a wide diversity of new features from the aspects of texture, shape and location. The disadvantages also limit the learning capacity of the deep medical image segmentation model. Generative Adversarial Network (GAN) is a solution for verisimilar image generation and domain translation [11,12]. Nonetheless, this technique always relies on fully annotated paired images and supervised training, which is impractical to collect all modalities for each patient. Additionally, they always have an arduous process for the cross-modality translation. Several recent GAN variants try to tackle the problem by encouraging the extra encoders to capture the domain features [13,14,15,16]. However, the quality of the synthetic image may be negatively influenced when the difference between domains increases.
In this work, we proposed a GAN-based data enhancement architecture for CT ischemic stroke lesion segmentation. An unsupervised translation cycle generative adversarial network (UTC-GAN) is presented for the sake of the segmentation accuracy improvement. The main contribution includes the following:
  • We develop an image augmentation architecture that is capable of synthesizing CT images and automatic learning translation from CT to its perfusion domains. It allows us to tackle the data scarcity problem for stroke segmentation.
  • A Gaussian mixture–translation representation module (GM-TRM) is proposed to learn the transformations between CT modalities automatically, and its corresponding translation loss function is defined. The module guarantees the model could flexibly learn the various translations.
  • Experiments demonstrate the proposed model offers a better quality of synthesis imaging and lesion segmentation accuracy on popular datasets than its counterparts.

2. Related work

2.1. Medical Image Segmentation

Numerous deep network designs have been exploited for medicine image segmentation in recent years [17,18,19]. Albert Clèrigues et al. [20] introduced symmetrical residual auto-encoding U-Net to perform lesion segmentation on CT images. Meanwhile, modality augmentation is utilized to provide more symmetric samples. Liu et al. [21] embeds an attention component in the deep CNN architecture to improve the predictive quality for white matter hypertension lesions. Furthermore, Zhang et al. [22] employed a 3D DenseNet model with dense block and multi-scale unit to localize the stroke lesions with harsh noise and low picture quality. Among those convolutional neural networks, U-Net is an influential architecture for biological image segmentation. For example, in [23], a U-Net model was applied to complete interwoven neurons and neurites segmentation tasks. Instead of only including contracting and expanding paths, Cui et al. [24] exploited a Bi-Directional ConvLSTM U-Net for blood vessel segmentation to fuse higher resolution features and semantic information.
Traditional works mainly focus on supervised learning. However, it requires sufficient training images with pixel-wise annotations. Therefore, some segmentation studies based on data augmentation are reported [25,26]. In these works, the supervised generative adversarial networks (GAN) are well used to expand the data amount by modelling proper data distribution in a two-player game framework. Jelmer et al. [27] employed DCGAN to complete brain CT image synthesis from MRI. A similar method has been employed in [28] to convert the T1 MRI scans into the T2 modality. Moreover, a multi-stage GAN method is designed to form image–mask pairs for segmentation task, by using the U-Net-like WGAN-GP as the central architecture [29]. To gain a higher augmentation quality and segmentation performance, a multi-scaled GAN framework, which also preserves the boundary of the tumor core, is composed to collaborate with the U-Net [12]. In addition, GAN is utilized for the unbalanced semantic segmentation task to balance data distribution [30]. Most GANs are hard to translate more than two domains or need to obtain the domain label for all training images.

2.2. Unsupervised Generative Adversarial Network

Several works use GANs to shift domain by unsupervised transferring of information between multiple modalities. Andrade et al. [31] trained Cycle-GAN to transform skin images from the macroscopic domain into the dermoscopic domains, which helps to acquire better segmentation capacity. Chen et al. [32] adapted a formation and appearance-detached augmentation GAN for unannotated cardiac CT image segmentation, synthesizing data from the annotated MRI slices. A unified generative adversarial network is established to execute 3D multimodal segmentation [33]. Such ideas above use other information to guide the generator in GAN to master the domain translation mapping. However, it is hard to capture semantic information across multimodal domains. To improve shape transformation and focus on the difference among the domains, U-GAT-IT [13] introduces an attentional component and a new layer instance-normalization technique in the least-squares GAN to complete the unsupervised domain transfer task. More recently, a tuple of concurrent GANs are designed to perform multi-class unsupervised image domain translation through conditional image generator and multi-task adversarial discriminator [14,15,16], where the generator is used for encoding content and class. In other words, the figure from a certain domain is composed of content and type information simultaneously.

3. Network Implementation

In this section, the overall UTC-GAN framework is firstly given. After that, we then offer the details of the translation representation module (GM-TRM) and objective loss functions accordingly.

3.1. Model Architecture

The image synthesizing is realized by a UTC-GAN module, which allows translating the CT slices from one modality domain to another. This model inherits the benefits of Cycle-GAN [31], which consists of a forward and backward cycle. The forward cycle of UTC-GAN is illustrated in Figure 1 and vice versa. It is noted that generator G A is designed to transfer CT images to its perfusion modalities on the dimension of latent space, and G B provides the reversed mapping images. The discriminators D B try to discriminate whether the synthesized images are fact or fiction. Inspired by the observation in the auto-encoding transformations approach [34], a paralleled auto-encoding structure is embedded in discriminator D B to extract the representation of modality transformation automatically. Therefore, the discriminator is partitioned into diverse components: encoder E , decoder D e and classifier C . Additionally, the architecture of each part is illustrated in Figure 2. Firstly, two encoders E A , E B are trained to extract the desired parameters of the training image from different domains, which uses a Siamese structure to share weight and obtain a co-training information. Each encoder introduces the AlexNet [17] as the backbone, and the Inception v1 block is embedded to enhance the network convergence. The decoder D e , a network with one convolutional layer and one fully connected layer, is coupled with the encoders to estimate the modality translation from the fused features. Finally, one classifier C A d v is added upon the encoder E A from generated domain to decide whether the synthesized image is real or not, and another C T r a n is built upon the decoder D e to distinguish which transformation is inputted. Each classifier contains two convolutional layers followed by two other fully connected layers.

3.2. Gaussian Mixture–Translation Representation Module

As in the UTC-GAN, the auto-encoding structure, also called the Gaussian mixture–translation representation module (GM-TRM), is proposed for learning the modality transformation representations automatically. The GM-TRM module first encodes the CT images from different modalities into a latent space. Then, the translation representation is extracted (self-supervised) in the decoding part via estimating the mutual information across the latent features. The GM-TRM module would lead the generator to build highly entangled translation representations, and would enforce the discriminator to obtain more additional supervision on image generation. Here, we elaborate on the principle of the GM-TRM as shown in Figure 3.
Let d(x) presents the data distribution of an image x from the original modality A. When it is translated to another modality B, a translation function t is assumed to turn d ( x ) into t ( d ( x ) ) . The encoder E A is considered to produce a representation from the image t ( x ) with network parameters θ , which also maps the low-dimensional input data t ( x ) to a high-level latent variable r . The classical reflection of the latent variable r from encoder is a statistical formulation, which is specified by the mean m θ and variance σ θ 2 of a normal distribution N ( ε | 0 , I ) , such that:
r = m θ ( t ( x ) ) + v θ ( t ( x ) )
To better approximate the transformation in latent space, the probabilistic density of encoding images can be assumed to follow the Gaussian mixture model instead of a single Gaussian. The resultant probabilistic representation of true posterior can be formulated as:
d θ ( r | t , x ) = i = 1 k ϕ i N ( r i | m i θ ( t ( x ) ) , σ i θ 2 t ( x ) ) ) i = 1 k ϕ i = 1
where N denotes a standard normal distribution, k is the total count of transformations, ϕ i is a weight vector, and m i θ and σ i θ 2 are defined as the mean and variance of the i th Gaussian component, respectively.
Meanwhile, we train a decoder D e to estimate the applied transformation parameters t by comparing the representations from encoded image features of original and target modality. The probabilistic density of decoder is defined as p φ ( t | r , r ˜ ) with the parameter φ , and r ˜ represents the probabilistic density from sample of original domain.
From an information-theoretic viewpoint [35], the transformation t can be equivalent to the joint mutual information I ( t , ( r , r ˜ ) ) between itself and the encoding latent space ( r , r ˜ ) , and the maximum I ( t , ( r , r ˜ ) ) can be considered the optimized representation. However, the posterior p φ ( t | r ) cannot be calculated directly. So, a moment matching approximation approach is introduced to compute p φ ( t | r ) conveniently, which can be derived to the following formulation:
I ( t , r ) = H ( t ) H ( t | r ) = H ( t ) + i = 1 k ϕ i E l o g d i θ ( t i | r i ) = H ( t ) + j = 1 k δ j E l o g p j φ ( t j | r j ) + i = 1 k j = 1 k E ( K L ( d i θ ( t j | r i ) | | p j φ ( t j | r i ) ) ) H ( t ) + j = 1 k δ j E d θ ( t , r | x ) l o g p j φ ( t j | r i )
where K L ( d i θ ( t j | r i ) | | p j φ ( t j | r i ) represents a non-negative Kullback–Leibler divergence between density d i θ and p j φ , δ j is the mixture weight. From this representation, the variational posterior distribution d θ ( r | t , x ) can be tractable and replaced approximated by an upper-bounded parameterized model p φ ( t | r ) . In addition, the entropy H ( t ) is independent of parameters θ and φ with respect to GAN model. So, we can maximize the lower variational bound I ( t , r ) through only calculating l o g p φ ( t | r , x ) .
In the meantime, the decoded transformation vector can be associated with the input images pairs, which enforces the generator to utilize classified transformation information as well. Hence, the corresponding generated image will learn different attributes from the other domains.

3.3. Loss Function

The UTC-GAN not only relies on the model architecture mentioned above but also the appropriate loss function to perfect the model performance. Our loss includes three components, of which the translation representation loss is newly proposed to learn the style transferring from the CT slices pair automatically.

3.3.1. Adversarial and Cycle-Consistency Loss

The adversarial and cycle-consistency loss from Cycle-GAN [30] are employed both in the generation cycle and discriminator. For the images translation A B , the adversarial loss can be expressed as:
L A d v ( G A , D B ) = E v ~ p B ( v ) [ l o g D B ( v B ) ] + E u ~ p A ( u ) [ l o g ( 1 D B ( G A ( u A ) ) ) ]
where u and v are the training slices come from source and target modality. Similarly, the adversarial loss of the translation B A is denoted as:
L A d v ( G B , D A ) = E u ~ p A ( u ) [ l o g D A ( u A ) ] + E v ~ p B ( v ) [ l o g ( 1 D A ( G B ( v B ) ) ) ]
In order to ensure the generated slices can be reconstructed to their previous modality simultaneously, a cycle-consistency loss L C y c ( G A , G B ) is utilized into architecture to associate the reconstructed image G A ( G B ( u A ) ) with the input image u . Thus, the loss function with forward–backward consistency is defined as:
L ( G A , G B , D A , D B ) = L A d v ( G A , D B ) + L A d v ( G B , D A ) + λ 1 ( L C y c ( G A , G B ) + L C y c ( G B , G A ) )
where λ 1 is the relative importance of GAN loss concering cycle loss.

3.3.2. Translation Representation Loss

Since the adversarial loss cannot detect the transformation directly from encoders, we have used the joint mutual information I ( t , v | u ) as the decoder to predict domain translation. Additionally, the lower variational bound of I ( t , v | u ) can be maximized by learning the expectation over posterior distribution p φ ( t | v ) according to (2). Thus, the translation forecasting loss L T r a n can be described as:
L T r a n = m a x I ( t , v | u ) = m a x θ , φ E l o g p φ ( t | r )                                                                                                 = m a x θ , φ δ j j = 1 k l o g N ( t | u j ( r ) , v j 2 ( r ) )
where the mean u j and variance v j are derived from the encoder, respectively, and δ j is a weight vector. This loss function, L T r a n , is added to the discriminator D B to learn the transformation functions between different domains. Then, the whole augmented objective function is given by:
L T o t a l = L A d v + λ 1 L C y c + λ 2 L T r a n
where λ 2 is a hyper-parameter applied for affecting the significance proportion in total loss.

4. Experiments

4.1. Experimental Settings

4.1.1. Dataset

The Ischemic Stroke Lesion Segmentation Challenge (ISLES) 2018 dataset is used to execute the training and assessment of our augmenting-based segmentation. The dataset contains multiple modalities, including CT and four derived perfusion maps, i.e., mean transit time (MTT), time to peak of the residue function (Tmax), cerebral blood flow (CBF) and cerebral blood volume (CBV) [36,37]. In our experiments, the 94 labelled cases with the CT and its perfusion modalities serve as input for the UTC-GAN network.
Additionally, all images in the dataset are preformed augmentation via skull stripping and traditional operations such as flipping, scaling and rotating. Finally, eighty percent of all the image scans are treated as the training and validation set, and the remainder is the testing set. That is, we uses 10,980 slices for training, 3660 slices for validating and 3660 slices for testing, respectively.

4.1.2. Baseline Model and Evaluation Measures

For comparison in augmentation, we use the ISLES2018 dataset to compare UTC-GAN to five existing unsupervised baseline models: Cycle-GAN [30], DRIT++ [14], EGSC-IT [15], FUNIT [16] and U-GAT-IT [13]. The Cycle-GAN and U-GAT-IT model have a similar mechanism and effectiveness to the proposed method. DRIT++ introduces the disentangled representation model to learn the mapping from the CT to other perfusion modalities. EGSC-IT realizes unsupervised image-to-image translation by adopting weight-sharing architecture and feature masks. FUNIT is chosen as another comparable baseline model to perform image translation in that it has a multi-task synthesis structure with reconstruction and feature-matching loss function. At the segmenting phase, a multiple-scale minus network (MSNet) [38] is trained to evaluate the improvement by the generative model. To input the multi-modalities CT slices, the encoder path of the MSNet is duplicated five times and concatenated to the decoding part. Due to computational constraints, every slice image is resized to 256 × 256 .
We verify the synthesis methods by adopting four widely used metrics: Peak Signal to Noise Ratio (PSNR), Normalized Mean Squared Error (NMSE) and Structural Similarity Index Measurement (SSIM) [39].
Six performance metrics are chosen to analyze the improvement in the segmentation accuracy for the testing set, including Dice Coefficient (DC), Intersection-over-Union (IoU) score, Precision, Accuracy, Recall and Hausdorff Distance (HD) [40].

4.1.3. Implementation Details

In the data augmentation stage, we adopt the ResNet architectures from Miyato et al. [20] as the backbone of the UTC-GAN generator. For UTC-GAN discriminator, encoder E A and E B in each branch consists of InceptionV1 block, then the output features from decoder D e are concatenated to a convolutional classifier C T r a n . Classifier C A d v is framed as same as C T r a n . During the segmentation stage, a standard U-Net framework is trained for segmenting stroke lesion areas. The back-propagation of both networks were completed by adopting ADAM optimization algorithm [41] with β 1 = 0.5 and β 2 = 0.999 . The initial learning rate was positioned at 2 × 10 4 for all networks. The loss-balancing weight parameters λ 1 and λ 2 were determined as 15 and 10, respectively. The exponential decay rate is used at a rate of 0.001 every 30 epochs, training the model for a total of 200 epochs with batch size of 24. Each paired input contains the original CT slice and their perfusion modality counterpart.
The synthesis and segmentation networks were implemented by Pytorch on the NVIDIA Titan XP GPU device. Furthermore, the overall experiments are conducted five times with varying random seeds, and the average value is reported.

4.2. The Impact of GM-TRM Module

First, we compare the generation quality to assess the effectiveness of components in our UTC-GAN model. The main component including the GM-TRM module and the classifier C T r a n concatenated to decoder D e are replaced in the architecture and compared with other existing schemes in sequence. As the competitors for comparison, the AET module also employs the auto-encoding architecture to learn the unsupervised transformation representation. Additionally, the AVT module creates the translation information by applying a constrained variational approach to the similar auto-encoders network. Table 1 firstly presents the quantitative synthesis performance with GM-TRM and other modules. We observe that the GM-TRM enhances the synthesized performance by leading to 1% improvement on average over PSNR, and 0.3% over NMSE. The evaluation result demonstrates that the mixture density can derive better transformation representation. Moreover, we also quantitatively compare the results by varying different classifiers upon the GM-TRM. The traditional non-linear classifier adopts fully connected layers to discrete transformations, whereas the convolutional type takes advantage of the convolution kernel to narrow the range of the transformation prediction. From the results, we can see that the convolutional classifier is consistently better than the non-linear classifier, and our synthesis model can almost achieve the best PSNR no matter which classifier is used.

4.3. Comparison with Other Unsupervised Data Augmentation Methods

Moreover, we next evaluate the UTC-GAN compared with various corresponding synthesis models.
Table 2 reports the quantitative synthesis results for all baselines. From the results underlined in Table, the UTC-GAN surpasses the other five unsupervised models by raising the PSNR from 23.13 to 26.40, and the NMSE approximates 0.096 when the SSIM achieves 0.918. This indicates our model attains the transformation characteristic across different modalities, and that it contributes to achieving superior synthesis effectiveness.
Figure 4 shows a visualization comparison under different types of modalities between the proposed UTC-GAN model and other unsupervised baselines. As we can see, the UTC-GAN generates much more realistic synthesis images, while samples from Cycle-GAN and U-GAT-IT generate some unclear regions or fail to create a detailed feature. In contrast, samples from FUNIT, EGSC-IT and DRIT++ show similar results, which supply more complex attributes of the brain to a certain extent but also have some unwanted artifacts that can be found in the image. Overall, the proposed method yields higher visual realism results for all CT modalities than the others, as indicated by qualitative and quantitative measures.

4.4. Segmentation Using Data Augmentation

To investigate the influence of UTC-GAN on the segmentation improvement, we first use the real and synthetic data to provide for the MSNet segmentation model, the performance evaluation is described in Figure 5. The synthetic images have a positive effect on the segmentation accurateness where the synthetic data ratio is less than 60%. The increasing proportion of the synthetic images leads to 7% and 8% improvement on the dice score and precision than the results obtained only by the real images. Moreover, we found that too many synthetic images could degrade the performance.
The second experiment is conducted on the task of segmentation by applying both all the data augmentation methods mentioned earlier and the MSNet model on ISLES2018 data. The quantitative results of segmentation can be viewed in Table 3. According to the demonstration in the table, all of the synthesis models provide more significant improvement on all evaluation metrics than only using the MSNet segmentation method with the randomness added by data pre-processing. Additionally, our process appears to produce the most significant improvement when generating transformation detectable images. For example, our proposed UTC-GAN allows the Dice score to improve from 0.675 to 0.768, increasing the precision by and 10% and reducing the Hausdorff distance by 9.3 mm, respectively. The IoU score and accuracy also achieved the highest value. This improvement may benefit the most due to the high contrast and variation from the generated images. Furthermore, the boxplot of the Dice coefficient is also used to analyze the segmentation robustness in Figure 6. Overall, the mean value and the median line of the UTC-GAN expresses relatively higher improvement than the other six competitors.
Additionally, three representative cases of visual segmentation scan for variant combinations are displayed in Figure 7. Compared with other data augmentation means, the UTC-GAN produces considerable improvements in the segmentation mask, allowing segmentation regions to be closer to their manual annotation counterpart. This intuitively demonstrates that the stroke CT images augmented by the UTC-GAN model help obtain a more accurate lesion region.

5. Conclusions

In this paper, a GAN-based data augmentation paradigm is presented to promote the exactness of ischemic stroke segmentation. By integrating Siamesed auto-encoders and information-theoretic loss into a Cycle-generative adversarial framework, the architecture can learn sufficient representations about transformation from the original to the target domain in an unsupervised fashion. The sampling problem on latent data space is further solved by introducing a Gaussian mixture probability distribution to better approximate the characteristics of the transformation. Based on the experimental evaluation and comparison, it demonstrates that the proposed method outperforms alternative structures and provides better high-quality generated stroke images. Meanwhile, the augmentation method could yield higher segmentation quality improvement by cooperating with traditional segmentation methods.

Author Contributions

Conceptualization and methodology, S.W. and X.Z.; formal analysis, S.W., X.Z., H.H., F.L. and Z.W.; software, validation and writing—original draft, S.W., X.Z. and H.H.; writing—review and editing and data curation, S.W., X.Z., H.H. and F.L.; supervision and funding acquisition, S.W., X.Z. and F.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the General Program under grant funded by the National Natural Science Foundation of China (NSFC) (No. 62171307), and the Basic Research Program of Shanxi Province under grant funded by the Department of Science and Technology of Shanxi Province (China) (No. 202103021224113).

Data Availability Statement

The ISLES 2018 dataset is publicly available at http://www.isles-challenge.org/ (accessed on 15 July 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Ethical Approval

Dataset was obtained from ISLES challenge, there is no live interaction with subjects and the dataset is anonymous. Hence this article does not contain any studies with live human participants or animals performed by any of the authors.

References

  1. Soltanpour, M.; Greiner, R.; Boulanger, P.; Buck, B. Improvement of automatic ischemic stroke lesion segmentation in CT perfusion maps using a learned deep neural network. Comput. Biol. Med. 2021, 137, 104849. [Google Scholar] [CrossRef] [PubMed]
  2. Gautam, A.; Raman, B. Segmentation of ischemic stroke lesion from 3D MR images using random forest. Multimed. Tools Appl. 2019, 78, 737–738. [Google Scholar] [CrossRef]
  3. Ostman, C.; Garcia, C.; Lillicrap, T.; Tomari, S. Multimodal Computed Tomography Increases the Detection of Posterior Fossa Strokes Compared to Brain Non-contrast Computed Tomography. Front. Neurol. 2020, 11, 588064. [Google Scholar] [CrossRef] [PubMed]
  4. Maegawa, T.; Sasahara, A.; Ohbuchi, H.; Chernov, M. Cerebral vasospasm and hypoperfusion after traumatic brain injury: Combined CT angiography and CT perfusion imaging study. Surg. Neurol. Int. 2021, 12, 361. [Google Scholar] [CrossRef] [PubMed]
  5. Honda, M.; Ichibayashi, R.; Yokomuro, H.; Yoshihara, K.; Masuda, H.; Haga, D.; Seiki, Y.; Kudoh, C.; Kishi, T. Early cerebral circulation disturbance in patients suffering from severe traumatic brain injury (TBI): A xenon CT and perfusion CT study. Neurol. Med. Chir. 2016, 56, 501–509. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [Green Version]
  7. Chilamkurthy, S.; Ghosh, R.; Tanamala, S.; Biviji, M.; Campeau, N.; Venugopal, V.; Mahajan, V.; Rao, P.; Warier, P. Deep learning algorithms for detection of critical findings in head CT scans: A retrospective study. Lancet 2018, 392, 2388–2396. [Google Scholar] [CrossRef]
  8. Oliveira, A.; Pereira, S.; Silva, C.A. Augmenting data when training a CNN for retinal vessel segmentation: How to warp. In Proceedings of the 2017 IEEE 5th Portuguese Meeting on Bioengineering (ENBENG), Coimbra, Portugal, 16–18 February 2017; pp. 1–4. [Google Scholar]
  9. Noguchi, S.; Nishio, M.; Yakami, M.; Nakagomi, K.; Togashi, K. Bone segmentation on whole-body CT using convolutional neural network with novel data augmentation techniques. Comput. Biol. Med. 2020, 121, 103767. [Google Scholar] [CrossRef]
  10. Koh, Y.J.; Kim, C.S. Primary object segmentation in videos based on region augmentation and reduction. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–27 July 2017; pp. 7417–7425. [Google Scholar]
  11. Majurski, M.; Manescu, P.; Padi, S.; Schaub, N.; Hotaling, N.; Simon, C., Jr.; Bajcsy, P. Cell image segmentation using generative adversarial networks, transfer learning, and augmentations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019; pp. 1–9. [Google Scholar]
  12. Mok, T.C.; Chung, A. Learning data augmentation for brain tumor segmentation with coarse-to-fine generative adversarial networks. In Proceedings of the International MICCAI Brainlesion Workshop, Granada, Spain, 16–17 September 2018; pp. 70–80. [Google Scholar]
  13. Kim, J.; Kim, M.; Kang, H.; Lee, K. U-gat-it: Unsupervised generative attentional networks with adaptive layer-instance normalization for image-to-image translation. arXiv 2019, arXiv:1907.10830. [Google Scholar]
  14. Lee, H.Y.; Tseng, H.Y.; Mao, Q.; Huang, J.B.; Lu, Y.D.; Singh, M.; Yang, M.H. Drit++: Diverse image-to-image translation via disentangled representations. Int. J. Comput. Vis. 2020, 128, 2402–2417. [Google Scholar] [CrossRef] [Green Version]
  15. Ma, L.; Jia, X.; Georgoulis, S.; Tuytelaars, T.; Van Gool, L. Exemplar guided unsupervised image-to-image translation with semantic consistency. arXiv 2018, arXiv:1805.11145. [Google Scholar]
  16. Liu, M.Y.; Huang, X.; Mallya, A.; Karras, T.; Aila, T.; Lehtinen, J.; Kautz, J. Few-shot unsupervised image-to-image translation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 10551–10560. [Google Scholar]
  17. Dvořák, P.; Menze, B. Local structure prediction with convolutional neural networks for multimodal brain tumor segmentation. In Proceedings of the International MICCAI Workshop on Medical Computer Vision, Munich, Germany, 9 October 2015; pp. 59–71. [Google Scholar]
  18. Rodríguez Colmeiro, R.G.; Verrastro, C.A.; Grosges, T. Multimodal brain tumor segmentation using 3D convolutional networks. In Proceedings of the International MICCAI Brainlesion Workshop, Quebec City, QC, Canada, 14 October 2017; pp. 226–240. [Google Scholar]
  19. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Van Leemput, K. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef] [PubMed]
  20. Clèrigues, A.; Valverde, S.; Bernal, J.; Freixenet, J.; Oliver, A.; Lladó, X. Acute and sub-acute stroke lesion segmentation from multimodal MRI. Comput. Methods Programs Biomed. 2020, 194, 105521. [Google Scholar] [CrossRef] [PubMed]
  21. Liu, L.; Kurgan, L.; Wu, F.X.; Wang, J. Attention convolutional neural network for accurate segmentation and quantification of lesions in ischemic stroke disease. Med. Image Anal. 2020, 65, 101791. [Google Scholar] [CrossRef] [PubMed]
  22. Zhang, R.; Zhao, L.; Lou, W.; Abrigo, J.M.; Mok, V.C.; Chu, W.C.; Shi, L. Automatic segmentation of acute ischemic stroke from DWI using 3-D fully convolutional DenseNets. IEEE Trans. Med. Imaging 2018, 37, 2149–2160. [Google Scholar] [CrossRef]
  23. Wang, W.; Yu, K.; Hugonot, J.; Fua, P.; Salzmann, M. Recurrent U-Net for resource-constrained segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 2142–2151. [Google Scholar]
  24. Azad, R.; Asadi-Aghbolaghi, M.; Fathy, M.; Escalera, S. Bi-directional ConvLSTM U-Net with densely connected convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Korea, 27–28 October 2019; pp. 1–10. [Google Scholar]
  25. Long, M.; Zhu, H.; Wang, J.; Jordan, M.I. Unsupervised domain adaptation with residual transfer networks. Adv. Neural Inf. Process. Syst. 2016, 29, 136–144. [Google Scholar]
  26. Bousmalis, K.; Silberman, N.; Dohan, D.; Erhan, D.; Krishnan, D. Unsupervised pixel-level domain adaptation with generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3722–3731. [Google Scholar]
  27. Wolterink, J.M.; Dinkla, A.M.; Savenije, M.H.; Seevinck, P.R.; van den Berg, C.A.; Išgum, I. Deep MR to CT synthesis using unpaired data. In Proceedings of the International Workshop on Simulation and Synthesis in Medical Imaging, Strasbourg, France, 27 September 2017; pp. 14–23. [Google Scholar]
  28. Welander, P.; Karlsson, S.; Eklund, A. Generative adversarial networks for image-to-image translation on multi-contrast mr images-a comparison of cyclegan and unit. arXiv 2018, arXiv:1806.07777. [Google Scholar]
  29. Pandey, S.; Singh, P.R.; Tian, J. An image augmentation approach using two-stage generative adversarial network for nuclei image segmentation. Biomed. Signal Process. Control 2020, 57, 101782. [Google Scholar] [CrossRef]
  30. Liu, S.; Zhang, J.; Chen, Y.; Liu, Y.; Qin, Z.; Wan, T. Pixel level data augmentation for semantic image segmentation using generative adversarial networks. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 1902–1906. [Google Scholar]
  31. Andrade, C.; Teixeira, L.F.; Vasconcelos, M.J.M.; Rosado, L. Data Augmentation Using Adversarial Image-to-Image Translation for the Segmentation of Mobile-Acquired Dermatological Images. J. Imaging 2020, 7, 2. [Google Scholar] [CrossRef]
  32. Chen, X.; Lian, C.; Wang, L.; Deng, H.; Kuang, T.; Fung, S.H.; Yap, P.T. Diverse data augmentation for learning image segmentation with cross-modality annotations. Med. Image Anal. 2021, 71, 102060. [Google Scholar] [CrossRef]
  33. Yuan, W.; Wei, J.; Wang, J.; Ma, Q.; Tasdizen, T. Unified generative adversarial networks for multimodal segmentation from unpaired 3D medical images. Med. Image Anal. 2020, 64, 101731. [Google Scholar] [CrossRef] [PubMed]
  34. Qi, G.J.; Zhang, L.; Chen, C.W.; Tian, Q. Avt: Unsupervised learning of transformation equivariant representations by autoencoding variational transformations. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 8130–8139. [Google Scholar]
  35. Agakov, D.B.F. The IM algorithm: A variational approach to information maximization. Adv. Neural Inf. Process. Syst. 2004, 16, 201–210. [Google Scholar]
  36. Cereda, C.W.; Christensen, S.; Campbell, B.C.; Mishra, N.K.; Mlynash, M.; Levi, C.; Lansberg, M.G. A benchmarking tool to evaluate computer tomography perfusion infarct core predictions against a DWI standard. J. Cereb. Blood Flow Metab. 2016, 36, 1780–1789. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Hakim, A.; Christensen, S.; Winzeck, S.; Lansberg, M.G.; Parsons, M.W.; Lucas, C.; Zaharchuk, G. Predicting infarct core from computed tomography perfusion in acute ischemia with machine learning: Lessons from the ISLES challenge. Stroke 2021, 52, 2328–2337. [Google Scholar] [CrossRef] [PubMed]
  38. Zhao, X.; Zhang, L.; Lu, H. Automatic polyp segmentation via multi-scale subtraction network. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Singapore, 18–22 September 2021; pp. 120–130. [Google Scholar]
  39. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  40. Nielsen, A.; Hansen, M.B.; Tietze, A.; Mouridsen, K. Prediction of tissue outcome and assessment of treatment effect in acute ischemic stroke using deep learning. Stroke 2018, 49, 1394–1401. [Google Scholar] [CrossRef]
  41. Sage, A.; Agustsson, E.; Timofte, R.; Van Gool, L. Logo synthesis and manipulation with clustered generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5879–5888. [Google Scholar]
  42. Zhang, L.; Qi, G.J.; Wang, L.; Luo, J. Aet vs. aed: Unsupervised representation learning by auto-encoding transformations rather than data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2547–2555. [Google Scholar]
Figure 1. Illustration of the UTC-GAN architecture.
Figure 1. Illustration of the UTC-GAN architecture.
Electronics 11 02612 g001
Figure 2. Illustration of the component architecture in discriminator for (a) Encoder; (b) Decoder; (c) Classifier.
Figure 2. Illustration of the component architecture in discriminator for (a) Encoder; (b) Decoder; (c) Classifier.
Electronics 11 02612 g002
Figure 3. The workflow of Gaussian mixture translation representation module (GM-TRM).
Figure 3. The workflow of Gaussian mixture translation representation module (GM-TRM).
Electronics 11 02612 g003
Figure 4. The visualizations of real and generated CT slices under different modalities. From left to right: CT, Tmax, CBF, MTT, CBV modality. The top-down results: Input ground truth, Cycle-GAN, U-GAT-IT, FUNIT, EGSC_IT, DRIT++ and our augmentation method.
Figure 4. The visualizations of real and generated CT slices under different modalities. From left to right: CT, Tmax, CBF, MTT, CBV modality. The top-down results: Input ground truth, Cycle-GAN, U-GAT-IT, FUNIT, EGSC_IT, DRIT++ and our augmentation method.
Electronics 11 02612 g004
Figure 5. The performance of stroke segmentation by changing the percentage of synthetic images.
Figure 5. The performance of stroke segmentation by changing the percentage of synthetic images.
Electronics 11 02612 g005
Figure 6. The Dice coefficient in boxplot by different methods, where the asterisk * and number in black represent the respective mean value.
Figure 6. The Dice coefficient in boxplot by different methods, where the asterisk * and number in black represent the respective mean value.
Electronics 11 02612 g006
Figure 7. Visual results of lesion segmentation processed via the MSNet and data augmentation models. The boundaries of expertise-based and unsupervised learning-based segmentation are painted with red and green curves, respectively.
Figure 7. Visual results of lesion segmentation processed via the MSNet and data augmentation models. The boundaries of expertise-based and unsupervised learning-based segmentation are painted with red and green curves, respectively.
Electronics 11 02612 g007
Table 1. Synthesizing performance of the UTC-GAN composed of different components.
Table 1. Synthesizing performance of the UTC-GAN composed of different components.
MethodPSNRNMSESSIM
AET + non-linear [42]24.120.1250.8654
AET + conv [40]24.860.1250.8675
AVT + non-linear [34]24.470.1220.8590
AVT + conv [34]25.190.1190.8678
GM-TRM + non-linear25.970.1090.8816
GM-TRM + conv26.400.0960.918
Table 2. Synthesizing performance comparison with different baselines.
Table 2. Synthesizing performance comparison with different baselines.
MethodPSNRNMSESSIM
Cycle-GAN23.130.1530.8594
U-GAT-IT23.250.1490.8673
FUNIT23.090.2330.8516
EGSC-IT23.860.1460.8624
DRIT++24.720.1150.8946
UTC-GAN26.400.0960.918
Table 3. Segmentation performance comparison with different strategies.
Table 3. Segmentation performance comparison with different strategies.
MethodDice
Coefficient
IoUPrecisionAccuracyRecallHausdorff
Distance (mm)
MSNet0.6750.5780.7210.8940.69927.03
Cycle-GAN + MSNet0.7090.6330.7730.9030.72525.00
U-GAT-IT + MSNet0.6840.6020.7450.9070.68724.21
FUNIT + MSNet0.7180.6450.7860.9120.73220.34
EGSC-IT + MSNet0.7510.6830.8090.9250.77334.11
DRIT++ + MSNet0.7350.6690.7810.9150.75819.99
UTC-GAN + MSNet0.7680.6850.8200.9260.78117.73
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, S.; Zhang, X.; Hui, H.; Li, F.; Wu, Z. Multimodal CT Image Synthesis Using Unsupervised Deep Generative Adversarial Networks for Stroke Lesion Segmentation. Electronics 2022, 11, 2612. https://doi.org/10.3390/electronics11162612

AMA Style

Wang S, Zhang X, Hui H, Li F, Wu Z. Multimodal CT Image Synthesis Using Unsupervised Deep Generative Adversarial Networks for Stroke Lesion Segmentation. Electronics. 2022; 11(16):2612. https://doi.org/10.3390/electronics11162612

Chicago/Turabian Style

Wang, Suzhe, Xueying Zhang, Haisheng Hui, Fenglian Li, and Zelin Wu. 2022. "Multimodal CT Image Synthesis Using Unsupervised Deep Generative Adversarial Networks for Stroke Lesion Segmentation" Electronics 11, no. 16: 2612. https://doi.org/10.3390/electronics11162612

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop