Abstract
The significance of medical image synthesis has exponentially grown due to constrained medical resources, making it a critical component in numerous clinical applications. This process facilitates the generation of high-quality, multi-modal medical images, ultimately enhancing medical image diagnostics. Currently, prevailing medical image synthesis methodologies primarily rely on voxel-based or GAN-based strategies to address substantial challenges arising from disparities in various imaging principles and significant noise. However, these methodologies rarely consider the intensity distribution difference among multi-modal or multi-parameters medical images, which generates unstable and unexplainable results. In response to these limitations, we propose a novel approach-a general variation-driven neural network for medical image synthesis that considers explicit data distribution. Within this method, we introduce the concept of a variation-based distance metric, providing a quantitative framework for capturing distribution disparities between medical images originating from both the source and target domains. Subsequently, guided by this variation-based distance metric, we introduce a robust end-to-end neural network architecture carefully designed to synthesize target medical images. Our proposed method has undergone extensive experimentation across various medical image synthesis tasks, including cross-modality transformations between CT and MRI, high-dose CT synthesis from low-dose CT, and the conversion of multi-parameters in MRI, including T1, T2, T1ce, and Flair sequences. In comparative assessments against existing methods, our approach consistently outperforms them across three publicly available datasets: Gold Atlas, LDCT, and BraTS2018. Additionally, we have successfully applied our model to generate high-quality Micro-CT images in dental clinics from CBCT data, significantly enhancing diagnostic capabilities in this clinical setting.
Similar content being viewed by others
Availability of data and materials
The data and materials used during the current study is available from the corresponding author on resonable request.
References
Wang Y, Zhou L, Yu B, Wang L, Zu C, Lalush DS, Lin W, Wu X, Zhou J, Shen D (2019) 3D Auto-Context-Based Locality Adaptive Multi-Modality GANs for PET Synthesis. IEEE Trans Med Imaging 38(6):1328–1339
Chen Y, Yue X, Fujita H, Fu S (2017) Three-way decision support for diagnosis on focal liver lesions. Knowl-Based Syst 127:85–99
Yang X, Chen Y, Yue X, Ma C, Yang P (2021) Local linear embedding based interpolation neural network in pancreatic tumor segmentation. Appl Intell
Zhang Y, Li H, Du J, Qin J, Wang T, Chen Y, Liu B, Gao W, Ma G, Lei B (2021) 3D Multi-Attention Guided Multi-Task Learning Network for Automatic Gastric Tumor Segmentation and Lymph Node Classification. IEEE Trans Med Imaging 40(6):1618–1631
Gao S, Zhuang X (2022) Bayesian Image Super-Resolution with Deep Modeling of Image Statistics. IEEE Trans Pattern Anal Mach Intell pp 1–1
Huang Y, Shao L, Frangi AF (2018) Cross-Modality Image Synthesis via Weakly Coupled and Geometry Co-Regularized Joint Dictionary Learning. IEEE Trans Med Imaging 37(3):815–827
Escobar M, Castillo A, Romero A, Arbeláez P (2020) UltraGAN: Ultrasound Enhancement Through Adversarial Generation. In: International Workshop on Simulation and Synthesis in Medical Imaging, pp. 120–130
Isola P, Zhu J-Y, Zhou T, Efros AA (2017) Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1125–1134
Nie D, Trullo R, Lian J, Petitjean C, Ruan S, Wang Q, Shen D (2017) Medical image synthesis with context-aware generative adversarial networks. In: International conference on medical image computing and computer-assisted intervention, pp 417–425
Zhu J-Y, Park T, Isola P, Efros AA (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2223–2232
Chen X, Lian C, Wang L, Deng H, Fung SH, Nie D, Thung K-H, Yap P-T, Gateno J, Xia JJ, Shen D (2020) One-Shot Generative Adversarial Learning for MRI Segmentation of Craniomaxillofacial Bony Structures. IEEE Trans Med Imaging 39(3):787–796
Wang G, Gong E, Banerjee S, Martin D, Tong E, Choi J, Chen H, Wintermark M, Pauly JM, Zaharchuk G (2020) Synthesize high-quality multi-contrast magnetic resonance imaging from multi-echo acquisition using multi-task deep generative model. IEEE Trans Med Imaging 39(10):3089–3099
Yang H, Sun J, Carass A, Zhao C, Lee J, Prince JL, Xu Z (2020) Unsupervised MR-to-CT Synthesis Using Structure-Constrained CycleGAN. IEEE Trans Med Imaging 39(12):4249–4261
Yu B, Zhou L, Wang L, Shi Y, Fripp J, Bourgeat P (2020) Sample-Adaptive GANs: Linking Global and Local Mappings for Cross-Modality MR Image Synthesis. IEEE Trans Med Imaging 39(7):2339–2350
Hoshen Y, Li K, Malik J (2019) Non-adversarial image synthesis with generative latent nearest neighbors. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 5811–5819
Zhou T, Fu H, Chen G, Shen J, Shao L (2020) Hi-Net: Hybrid-Fusion Network for Multi-Modal MR Image Synthesis. IEEE Trans Med Imaging 39(9):2772–2781
Zhu J-Y, Zhang R, Pathak D, Darrell T, Efros AA, Wang O, Shechtman E (2017) Toward multimodal image-to-image translation. In: Proceedings of the international conference on neural information processing systems, pp 465–476
Mao Q, Lee H-Y, Tseng H-Y, Ma S, Yang M- (2019) Mode seeking generative adversarial networks for diverse image synthesis. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 1429–1437
Dar SU, Yurt M, Karacan L, Erdem A, Erdem E, Çukur T (2019) Image Synthesis in Multi-Contrast MRI With Conditional Generative Adversarial Networks. IEEE Trans Med Imaging 38(10):2375–2388
Yue Z, Yong H, Zhao Q, Meng D, Zhang L (2019) Variational Denoising Network: Toward Blind Noise Modeling and Removal. In: Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada., pp. 1688–1699
Qu L, Zhang Y, Wang S, Yap P-T, Shen D (2020) Synthesized 7T MRI from 3T MRI via deep learning in spatial and wavelet domains. Med Image Anal 62:101663
Meng M, Li S, Yao L, Li D, Zhu M, Gao Q, Xie Q, Zhao Q, Bian Z, Huang J (2020) Semi-supervised learned sinogram restoration network for low-dose CT image reconstruction. In: Medical Imaging 2020: Physics of Medical Imaging, vol. 11312, p 113120
Xie Q, Zhou M, Zhao Q, Meng D, Zuo W, Xu Z (2019) Multispectral and hyperspectral image fusion by MS/HS fusion net. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 1585–1594
Menze BH, Jakab A, Bauer S, Kalpathy-Cramer J, Farahani K, Kirby J, Burren Y, Porz N, Slotboom J, Wiest R (2014) The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans Med Imaging 34(10):1993–2024
Moen TR, Chen B, Holmes DR III, Duan X, Yu Z, Yu L, Leng S, Fletcher JG, McCollough CH (2021) Low-dose CT image and projection dataset. Med Phys 48(2):902–911
Nyholm T, Svensson S, Andersson S, Jonsson J, Sohlin M, Gustafsson C, Kjellén E, Söderström K, Albertsson P, Blomqvist L (2018) MR and CT data with multiobserver delineations of organs in the pelvic area-Part of the Gold Atlas project. Medical Phys 45(3):1295–1300
Balwant M (2022) A review on convolutional neural networks for brain tumor segmentation: Methods, datasets, libraries, and future directions. IRBM 43(6):521–537
Dequidt P, Bourdon P, Tremblais B, Guillevin C, Gianelli B, Boutet C, Cottier J-P, Vallée J-N, Fernandez-Maloigne C, Guillevin R (2021) Exploring radiologic criteria for glioma grade classification on the brats dataset. IRBM 42(6):407–414
Khairandish MO, Sharma M, Jain V, Chatterjee JM, Jhanjhi N (2022) A hybrid cnn-svm threshold segmentation approach for tumor detection and classification of mri brain images. Irbm 43(4):290–299
Jiang L, Mao Y, Chen X, Wang X, Li C (2023) Cola-diff: Conditional latent diffusion model for multi-modal mri synthesis. arXiv preprint arXiv:2303.14081
Rombach R, Blattmann A, Lorenz D, Esser P, Ommer B (2022) High-resolution image synthesis with latent diffusion models. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 10684–10695
Chartsias A, Joyce T, Giuffrida MV, Tsaftaris SA (2017) Multimodal MR synthesis via modality-invariant latent representation. IEEE Trans Med Imaging 37(3):803–814
Yurt M, Özbey M, Dar SU, Tinaz B, Oguz KK, Çukur T (2022) Progressively volumetrized deep generative models for data-efficient contextual learning of mr image recovery. Med Image Anal 78:102429
Peng B, Liu B, Bin Y, Shen L, Lei J (2021) Multi-modality mr image synthesis via confidence-guided aggregation and cross-modality refinement. IEEE J Biomed Health Inform 26(1):27–35
Sharma A, Hamarneh G (2019) Missing mri pulse sequence synthesis using multi-modal generative adversarial network. IEEE Trans Med Imaging 39(4):1170–1183
Huang Z, Zhang J, Zhang Y, Shan H (2021) DU-GAN: Generative Adversarial Networks with Dual-Domain U-Net Based Discriminators for Low-Dose CT Denoising. IEEE Trans Instrum Meas pp 1–1
Chen H, Zhang Y, Kalra MK, Lin F, Chen Y, Liao P, Zhou J, Wang G (2017) Low-Dose CT With a Residual Encoder-Decoder Convolutional Neural Network. IEEE Trans Med Imaging 36(12):2524–2535
Fan F, Shan H, Kalra MK, Singh R, Qian G, Getzin M, Teng Y, Hahn J, Wang G (2020) Quadratic Autoencoder (Q-AE) for Low-Dose CT Denoising. IEEE Trans Med Imaging 39(6):2035–2050
Yang Q, Yan P, Zhang Y, Yu H, Shi Y, Mou X, Kalra MK, Zhang Y, Sun L, Wang G (2018) Low-Dose CT Image Denoising Using a Generative Adversarial Network With Wasserstein Distance and Perceptual Loss. IEEE Trans Med Imaging 37(6):1348–1357
Mäkinen Y, Azzari L, Foi A (2020) Collaborative filtering of correlated noise: Exact transform-domain variance for improved shrinkage and patch matching. IEEE Trans Image Process 29:8339–8354
Shan H, Zhang Y, Yang Q, Kruger U, Kalra MK, Sun L, Cong W, Wang G (2018) 3-D Convolutional Encoder-Decoder Network for Low-Dose CT via Transfer Learning From a 2-D Trained Network. IEEE Trans Med Imaging 37(6):1522–1534
Bera S, Biswas PK (2021) Noise conscious training of non local neural network powered by self attentive spectral normalized markovian patch gan for low dose ct denoising. IEEE Trans Med Imaging 40(12):3663–3673
Geng M, Meng X, Yu J, Zhu L, Jin L, Jiang Z, Qiu B, Li H, Kong H, Yuan J, Yang K, Shan H, Han H, Yang Z, Ren Q, Lu Y (2021) Content-Noise Complementary Learning for Medical Image Denoising. IEEE Trans Med Imaging, pp 1–1
Marcos L, Alirezaie J, Babyn P (2022) Low dose ct denoising by resnet with fused attention modules and integrated loss functions. Front Signal Process 1:812193
Gholizadeh-Ansari M, Alirezaie J, Babyn P (2020) Deep learning for low-dose ct denoising using perceptual loss and edge detection layer. J Digit Imaging 33:504–515
Acknowledgements
This work was supported by National Natural Science Foundation of China (No. 62173252, 61976134).
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Ethics approval
Not applicable.
Consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The original online version of this article was revised: The word “image” was accidentally removed.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Chen, Y., Yang, X., Yue, X. et al. A general variation-driven network for medical image synthesis. Appl Intell 54, 3295–3307 (2024). https://doi.org/10.1007/s10489-023-05017-1
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-05017-1