Next Article in Journal
Rat 50 kHz Trill Calls Are Tied to the Expectation of Social Interaction
Previous Article in Journal
The Association between the Binding Processes of Working Memory and Vascular Risk Profile in Adults
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Segmentation of the Olfactory Bulb

1
Smell & Taste Clinic, Department of Otorhinolaryngology, Technische Universität, 01307 Dresden, Germany
2
Department of Informatics, School of Engineering, University of Minho, 4704-553 Braga, Portugal
3
Center for Music in the Brain, Department of Clinical Medicine, Aarhus University, Noerrebrogade 44, 1A, 8000 Aarhus, Denmark
4
Center of Functionally Integrative Neuroscience, Aarhus University, Noerrebrogade 44, 1A, 8000 Aarhus, Denmark
5
Flavour Institute, Department of Clinical Medicine, Aarhus University, Noerrebrogade 44, 1A, 8000 Aarhus, Denmark
*
Author to whom correspondence should be addressed.
Shared last authorship.
Brain Sci. 2021, 11(9), 1141; https://doi.org/10.3390/brainsci11091141
Submission received: 6 July 2021 / Revised: 24 August 2021 / Accepted: 25 August 2021 / Published: 28 August 2021
(This article belongs to the Section Computational Neuroscience and Neuroinformatics)

Abstract

:
The olfactory bulb (OB) has an essential role in the human olfactory pathway. A change in olfactory function is associated with a change of OB volume. It has been shown to predict the prognosis of olfactory loss and its volume is a biomarker for various neurodegenerative diseases, such as Alzheimer’s disease. Thus far, obtaining an OB volume for research purposes has been performed by manual segmentation alone; a very time-consuming and highly rater-biased process. As such, this process dramatically reduces the ability to produce fair and reliable comparisons between studies, as well as the processing of large datasets. Our study aims to solve this by proposing a novel methodological framework for the unbiased measurement of OB volume. In this paper, we present a fully automated tool that successfully performs such a task, accurately and quickly. In order to develop a stable and versatile algorithm and to train the neural network, we used four datasets consisting of whole-brain T1 and high-resolution T2 MRI scans, as well as the corresponding clinical information of the subject’s smelling ability. One dataset contained data of patients suffering from anosmia or hyposmia (N = 79), and the other three datasets contained data of healthy controls (N = 91). First, the manual segmentation labels of the OBs were created by two experienced raters, independently and blinded. The algorithm consisted of the following four different steps: (1) multimodal data co-registration of whole-brain T1 images and T2 images, (2) template-based localization of OBs, (3) bounding box construction, and lastly, (4) segmentation of the OB using a 3D-U-Net. The results from the automated segmentation algorithm were tested on previously unseen data, achieving a mean dice coefficient (DC) of 0.77 ± 0.05, which is remarkably convergent with the inter-rater DC of 0.79 ± 0.08 estimated for the same cohort. Additionally, the symmetric surface distance (ASSD) was 0.43 ± 0.10. Furthermore, the segmentations produced using our algorithm were manually rated by an independent blinded rater and have reached an equivalent rating score of 5.95 ± 0.87 compared to a rating score of 6.23 ± 0.87 for the first rater’s segmentation and 5.92 ± 0.81 for the second rater’s manual segmentation. Taken together, these results support the success of our tool in producing automatic fast (3–5 min per subject) and reliable segmentations of the OB, with virtually matching accuracy with the current gold standard technique for OB segmentation. In conclusion, we present a newly developed ready-to-use tool that can perform the segmentation of OBs based on multimodal data consisting of T1 whole-brain images and T2 coronal high-resolution images. The accuracy of the segmentations predicted by the algorithm matches the manual segmentations made by two well-experienced raters. This method holds potential for immediate implementation in clinical practice. Furthermore, its ability to perform quick and accurate processing of large datasets may provide a valuable contribution to advancing our knowledge of the olfactory system, in health and disease. Specifically, our framework may integrate the use of olfactory bulb volume (OBV) measurements for the diagnosis and treatment of olfactory loss and improve the prognosis and treatment options of olfactory dysfunctions.

1. Introduction

As one of the five basic human senses, the sense of smell plays an important role in our daily life. It is essential for the detection of dangers, such as gas, fire, smoke, or hazardous chemicals, and the quality of our everyday social life. Moreover, smelling odors makes up a major part of our pleasant experiences, whether eating a favorite meal, walking outside smelling blooming flowers and trees, or during intimacy with one’s partner. Last but not least, the sense of smell plays a fundamental role in some professions such as chef (approximately 0.5% of the German workforce), baker, or perfumer [1].
Clinical reviews have shown that 3–20% of the general population are affected by anosmia (complete loss of sense of smell) or hyposmia (reduced sense of smell) [2,3]. Olfactory deficits are associated with numerous neurodegenerative disorders such as Parkinson’s disease or Alzheimer’s disease and appear as prodromal symptoms. Moreover, patients with complete or partial olfactory loss have a higher risk of exhibiting symptoms of depression. Inversely, depressive symptoms are correlated with lower activation in structures involved in the olfaction perception pathway or with a decreased olfactory bulb (OB) volume [4]. Therefore, there is a need to understand the functioning of the human olfactory system.
Olfactory perception begins as the volatile odor molecules inhaled from the air bind olfactory receptor proteins in the cilia of olfactory sensory neurons housed in the neuroepithelium of the nasal cavity. This neuroepithelium contains 6–10 million neurons [5]. The axons of these neuroepithelium cells ascend through the cribriform plate as fila olfactoria to the OB located in the olfactory fossa of the ethmoid bone. The OB is the first stage of the olfactory signal processing system and, therefore, an essential part of the olfactory pathway. Signals from activated neuroepithelium cells are transmitted to the OB and then to primary olfactory regions of the brain such as the piriform cortex, entorhinal cortex, and amygdala. The outputs from the primary olfactory areas are then sent to other brain structures such as the orbitofrontal cortex (OFC), insula, and hippocampus [6].
It has been shown that OB volume correlates with olfactory sensitivity and it is decreased in patients with olfactory disorders [7]. Moreover, previous studies have shown that the OB volume correlates with the volume and grey matter density of the primary olfactory region [8]. From the range of known causes of primary olfactory loss, patients with post-traumatic and post-infectious anosmia or hyposmia consistently present a reduced OB volume when compared to healthy individuals [9]. About two thirds of the patients with congenital anosmia do not have a detectable OB on magnetic resonance imaging (MRI) scans, and one third present hypoplastic OB [10,11]. It has been shown that the decrease in OB volume correlates with decreased olfactory sensitivity. Reverse to this, an increase in OB volume, e.g., after endoscopic nasal surgery, leads to increased olfactory sensitivity [12]. All these findings suggest that the volume of the OB is an important marker for olfactory function and a predicting factor for the treatment of olfactory disorders [13].
The OB is a very small structure within the human brain with a volume ranging between 35–100 mm3 in normosmic individuals [13,14]. Therefore, to assess the volume of such a small structure, specialized high-resolution MR sequences are needed. However, the location of the OB within the olfactory fossa of the ethmoid bone makes it vulnerable to susceptibility artifacts. Susceptibility artifacts are distortions due to local magnetic field inhomogeneities and often arise at interfaces of tissues with different proton densities. Especially in fast sequences with lower spatial resolution, it is extremely challenging to achieve a satisfactory level of image quality necessary for performing image segmentation of small anatomical structures and estimating its volume. In clinical practice and research, the T2 coronal sequence is most commonly used because of its contrast between the OB’s tissue and the cerebrospinal fluid (CSF) in the surrounding area and because it is less vulnerable to susceptibility artifacts. Currently, the gold standard strategy for the measurement of OB volumes involves the manual segmentation of the visible OB in the coronal plane view slice by slice. Different segmentation techniques are known, such as manually tracing the outlines of the bulbs or highlighting the entire visible area slice by slice [15]. Still, manual segmentation is an extremely time-consuming process exhibiting variations in the degree of inter-observer and intra-observer reliability. It takes a well-experienced rater about 10–15 min to complete the OB segmentation of a single subject.
Over the last few years, machine learning (ML) algorithms have provided efficient solutions for automatic image segmentation. These algorithms have the potential to efficiently process more data and increase the reliability and repeatability of the results. Here, we applied ML to neuroimaging data to produce one of the first models that can automatically and accurately segment the OB, a very small anatomical region in the human brain [16,17].
In short, our algorithm starts by performing multimodal data preprocessing to localize the OB and compute the bounding box. Subsequently, a 3D U-Net model segments the OB within the bounding box. The performance of the algorithm was evaluated using established metrics [18]. Additionally, an experienced independent and blinded rater rated the quality of the agreement between the segmentations produced by both the algorithm and the human raters. To prove that the algorithm performs well independently of the field of view (FOV), image orientation, angulation, or other acquisition parameters, we tested our model on randomly selected subjects from neuroimaging datasets from previous studies.

2. Methods

2.1. Study Population

In this study, we used data from four different datasets [19,20,21]. All data were obtained at the University Hospital Carl-Gustav Carus in Dresden, Germany. Within the context of the respective studies, all participants signed informed consent on the use of their data for research purposes. All of these studies had been approved by the Ethics Committee at the University Clinic of the TU Dresden.
Data were separated into the following two sub-datasets: anosmia and healthy controls. The first dataset was collected on 79 patients between August 2015 and July 2017. Patients were diagnosed with anosmia or hyposmia. Healthy controls’ (n = 91) sub-dataset contained data from three different MRI studies.
In both datasets, participants’ smell ability was measured psychophysically using the Sniffin’ Sticks test [22]. The scores of the test (range: 1–48) were used to categorize participants’ smelling ability in the following categories: functional anosmia (TDI ≤ 16), hyposmia (16 > TDI < 30.75), or normosmia (TDI ≥ 30.75) [23] (Table 1).

2.2. MRI Image Acquisition

All MRI scans were acquired on a 3T Siemens Verio scanner (Siemens, Erlangen, Germany). All acquisitions were made using a 32-channel head coil. The following two modalities were used in this study: T1-weighted axial whole-brain scans and T2-weighted high-resolution coronal scans for OB imaging.
T1-weighted MPRAGE sequence was acquired using the following parameters: repetition time: 2300 ms; echo time 2.98: ms; flip angle: 9°; field of view: 240 × 256; acquisition matrix: 250 × 256 voxel size: 1 × 1 × 1 mm3; slice thickness: 1 mm; slices: 176. T2-weighted high-resolution sequence was acquired using following sequence parameters: repetition time: 5500 ms; echo time: 110 ms; flip angle: 150°, field of view: 120 × 120; acquisition matrix: 256 × 256; voxel size: 0.47 × 0.47 × 1.2 mm3; slice thickness: 1.2 mm (no gap).

2.3. Manual Segmentations of OB Volume

For both datasets, manual segmentation of the OBs was performed on T2-weighted images in coronal plane view using ITK-SNAP Software v. 3.6 [24]. First, both T1-weighted and T2-weighted scans were converted from DICOM format to g-zipped NIFTI format (nii.gz) using the dcm2niix conversion tool [25]. All measurements were performed by two raters independently. Both raters used the protocol for performing manual segmentations. Voxels belonging to the left and right OB were labeled with values 1 and 2, respectively. The manual measurements were saved as binary segmentation masks in g-zipped NIFTI format (nii.gz). To access the information for each label, respectively, the binary segmentation masks were converted to a NumPy array with nibabel python library [26,27]. The volumes of the labels for left and right OBs were calculated by multiplying the number of voxels of the label by voxel dimensions using the NumPy python library [28]. For all manual segmentation masks, the DC was calculated using the MedPy python library [29] to estimate the level of overlap between the two raters.

2.4. Automated Localization of the OBs

OBs have very small volumes, especially compared to the entire scanned brain volume. This leads to highly imbalanced data due to class imbalance between voxels labeled as foreground (voxels with values one/two for left/right OB) and voxels labeled as background (zero value voxels for any other tissue). To solve this issue, we opted for using a template-based approach. To the best of our knowledge, no earlier studies have normalized individual OBs to a standard stereotactic Montreal Neurological Institute (MNI) [30,31] space. Here, we developed a pipeline that allows the automatic transformation of all manual segmentations from T2 native space to MNI space. First, the T1 whole-brain image was co-registered to ICBM 2009c Nonlinear Asymmetric T1 MNI template image using ANTs SyN nonlinear registration tool [32]. Subsequently, the T2 image was co-registered to the T1 image by ANTs affine registration function. For both steps, the transformation matrices were saved and applied to the manual segmentations in native T2 space. Applying the inverted registration matrices allowed a two-step transformation of manual segmentation masks from T2 native space to MNI space. This procedure was extended to the manual segmentations produced by the two raters. The resulting binary masks in MNI space were transformed to NumPy arrays using nibabel. The NumPy arrays were added to one NumPy array and divided by the highest value in the array.
This process resulted in a probability map for the OBs in MNI space. To calculate the coordinates of the center of gravity (COG) of the OB, we applied a threshold of 0.5 to the resulting OB probability map and calculated the coordinates in MNI space (xyz: −4/44/−36) using the SciPy python library [33], and saved the result as a binary image (Figure 1).

2.5. Preprocessing Pipeline

The preprocessing pipeline of the segmentation tool performs all necessary steps to create normalized data for 3D U-Net input. As described above, the first step consists in co-registering the T1 image to MNI template image and the T2 image to T1 image to obtain the resulting transformation matrices. These will subsequently be inverted, and the resulting inverted transformation matrices applied to the COG binary map resulting in the COG in T2 native space. To reduce the number of features (voxels) and the data imbalance of the input images for processing in 3D U-Net, we constructed a bounding box extraction algorithm based on the COG in T2 native space. To ensure that all images have the same orientation in three-tridimensional space, all T2 images and the COG binary mask in T2 native space were reoriented using nibabel python library to canonical orientation. The bounding box edges were defined as points in 3D space shifted from the COG by +/−10 mm in the x-direction, +/−15 mm in the y-direction, and +/−5 mm in the z-direction. Afterwards, the array inside the calculated bounding box was extracted from the T2 image and the corresponding manual segmentation. The resulting images were resampled to a common voxel dimension of (0.5, 1, 0.5) and image shape of (4, 32, 32) using cubic interpolation function from fslpy python library (Figure 2) [34].

2.6. Training of the 3D U-Net Model

The entire model training process was carried out using the Monai Python Library [35]. Model training was performed on the dataset resulting from preprocessing pipeline described above. The dataset included the preprocessed data from 159 T2 images and 318 manual segmentations (159 manually created binary masks by two raters). Seven subjects from the anosmia dataset were excluded, due to not having a visible OB. Therefore, the manual segmentations made by both raters were empty, containing only the background. First, the dataset was randomly split into a training dataset (n = 191 subjects; 60% of the dataset), a validation dataset (n = 64 subjects; 20% of the dataset,), and a test dataset (n = 64 subjects; 20% of the dataset). First, all data underwent intensity normalization, a standard step of the monai preprocessing pipeline. To perform data augmentation, a random affine transformation was applied to the normalized data from the training dataset only. These transformations were automatically applied by RandAffined function from the monai python package during each epoch of the model training process. Specifically, the transformation features included a translation range of −10 to +10 voxels in each direction and rotation of −30° to +30° degrees using bilinear interpolation function for T2 images and nearest-neighbor interpolation for binary label masks.
The U-Net model was imported from monai python library and specified using following parameters: dimensions: 3; input channels: 1; output channels: 2; channels: 16, 32, 64, 128, 256; strides: 2, 2, 2, 2; num of residual units = 2.
The model was trained over 300 epochs. Each epoch contained 86 iterations of mini batches containing two corresponding image and binary mask pairs of size 64 × 32 × 32 voxels. We opted for the Tversky function to measure loss aversion. Network weights were optimized using the Adam optimization function (Ir = 0.001). One input channel and two output channels were defined for the input–output data stream of the model, containing three dimensions (Figure 3).

2.7. Postprocessing Pipeline

We designed and implemented a novel postprocessing pipeline to perform data cleanup of the primary output of the trained 3D U-Net model and transform the output to the original T2 image’s size, resolution, and orientation. As a first step, all clusters are relabeled to unique values. Secondly, clusters are thresholded based on their size to eliminate spurious clusters. Finally, the resulting mask is resampled to the resolution, voxel size, and orientation of the corresponding T2 image. This post-processed binary mask is then saved into an empty NumPy array of the same size as the T2 image—the final binary segmentation mask.

2.8. Evaluation of Training and Testing

The performance of the automatic segmentation was evaluated on the testing dataset. This dataset was kept separate before the model training step and, therefore, contained only data “unseen” by the model, consisting of 64 subjects and the corresponding manual segmentations of two raters. Performance was evaluated by computing an extensive set of statistical metrics—the dice coefficient (DC) and average symmetric surface distance (ASSD)—to evaluate the similarity between the model’s predicted segmentations and the manual segmentations, for both left and right OB.
DC is an overlay similarity index that reflects size and localization agreement and ranges from 0 (no overlap) to 1 (complete overlap). ASSD represents the mean distance of the binary objects in two images. Therefore, an ASSD of 0 mm represents a perfect match between both segmentations.
For subjects with diagnosed congenital anosmia and absence of the OBs, it was not possible to compute the DC because empty segmentation masks contain only the label corresponding to the image background.
Furthermore, to ensure the quality of predicted segmentations, all data were randomized and anonymized to ensure unbiased rating of the manual and predicted segmentations. A well-experienced rater, who had not been involved in any stage of the data segmentation process, was asked to rate the overlap of both manual segmentations and predicted masks using a performance scale. The rating criteria were defined as a scale: 1: “No congruency”, 2–3: “Poor congruency”, 4–5: “good congruency”, 6–7: “very good congruency”, 8–10: “excellent congruency”.

3. Results

Firstly, we compared the volumes of the manual segmentations made by rater one and two and calculated the DC for the left and right OBs individually (Table 2). For the dataset containing the anosmia patients’ data, the average DC was 0.77 ± 0.07 and 0.74 ± 0.10 mm3 for the left and right OB, respectively. The mean volume of the left OB was 37.82 ± 11.48 and 47.82 ± 14.78 mm3 measured by raters one and two, respectively. For the right OB, the mean volume of the segmentations produced by raters one and two was 34.32 ± 10.93 and 46.47 ± 15.43 mm3, respectively.
For the dataset containing the healthy control subjects’ data, the average DC for the left OB was 0.81 ± 0.06 mm3 and 0.80 ± 0.05 mm3 for the right OB. The mean volume for the left OB was 44.14 ± 12.38 mm3 measured by rater one and 56.01 ± 16.92 mm3 measured by rater two. The mean volume for the right OB was 42.47 ± 13.54 mm3 measured by rater one and 54.15 ± 17.66 mm3 measured by rater two (Figure 4).
For the evaluation of the overlap between the predicted segmentations of the algorithm and the manual segmentations of the OB for the subjects in the test dataset (N = 64), various metrics were calculated for the left and right OBs, separately, as well as for the entire segmentation mask. As it can be found in Table 3, the mean DC was 0.77 ± 0.05 (left OB: 0.78 ± 0.06, right OB: 0.75 ± 0.08) and the mean symmetric surface distance (ASSD) was 0.43 ± 0.10 (left OB: 0.41 ± 0.10, right OB: 0.44 ± 0.14) (Table 3).
Moreover, we compared the volumes of the manual segmentations and the predicted segmentations generated by our algorithm, for the left and right OBs individually, for the test dataset. For the left hemisphere, the mean OB volume of the predicted mask was 44.80 ± 8.59 mm3 and 46.14 ± 12.70 mm3 for the manually segmented mask. For the right OB, the predicted OB volume was 46.73 ± 8.86 mm3 for the predicted mask and 42.63 ± 14.19 mm3 for the manual masks (Table 4).
Subsequently, an independent, unbiased, and well-experienced rater rated the manual segmentations produced by both raters, as well as the predicted segmentation of the algorithm (Figure 5) (Table 5). The mean score was 6.23 ± 0.87 and 5.92 ± 0.81 for rater one and two, respectively, and 5.95 ± 0.87 for the segmentations generated by our model.
Additionally, we tested the algorithm on subjects from the olfactory dysfunction dataset with diagnosed congenital anosmia (N = 7). As aforementioned, given that these subjects do not have a visible OB, the resulting manual segmentation masks were empty, containing only the background label. Our model was able to correctly identify all these extraordinary datasets as non-existent OB cases, with the generated masks only containing the background label, as expected.
In addition to the extended quality evaluation, we developed two interfaces for our algorithm. The tool is fully available via the pip python package management system, and the command-line interface is accessible in the terminal on UNIX-based operating systems (OS). For use in a Microsoft native OS, a Linux sub-system needs to be further installed. A visual interface is also available, allowing the user to select the paths to folders containing data comfortably at the distance of only a few mouse clicks.

4. Discussion

In this study, we have developed a ready-to-use solution for the automatic segmentation of human OBs using 3D U-Net. The algorithm localizes the COG of OBs, performs the segmentation, and calculates the volumes with nearly the same accuracy (DC = 0.77 ± 0.05) as two experienced raters independently (inter-rater DC = 0.79 ± 0.08).
Previous studies have shown a high inter-rater reliability using manual segmentation to measure OBV [36]. Although the convergence of volumetric measurements based on absolute volume can be high, the segmentation labels do not necessarily overlap. This fact makes the reproducibility of manual volumetric segmentations dependent on the observer and his expertise. Moreover, manual segmentation is a very time-consuming process. It takes an experienced rater about 15 min per subject. Our automated solution of OB segmentation takes about 3–5 min per subject, dependent on the system properties. Our study offers solid evidence that the OB can indeed be segmented automatically and quickly, matching the level of expertise of trained raters, while minimizing potential sources of bias. We have tested our algorithm on a dataset (N = 7) containing only patients with congenital anosmia and our model was able to accurately detect the absence of the OBs.
In this study, manual segmentations made by two independent raters were considered as ground truth. However, as it can be found in Figure 5, given the relatively low inter-rater DC of 0.79 ± 0.08, we can reasonably assume that there is an approximately 20% variation in the “ground truth” (gold standard method). Therefore, it is difficult to assess how precise each of these measurements are in representing the real volume of the OB. A potential reason for this may lie in the limited resolution of the MRI acquisitions, which are additionally biased by signal noise and artifacts. These may, therefore, constitute major methodological limitations with direct impact on the precision of the manual and automated measurements of the OB volume. This issue can only be addressed by developing better MRI hardware and improving the acquisition techniques. One of the main known limiting factors of the acquisition quality is the acquisition time, as longer acquisition times are needed to improve the spatial resolution and signal-to-noise ratio of the images.
Overall, our results showed that the segmentations produced by our algorithm for OB volume measurements could be used for performing clinical measurements both on healthy subjects and on different clinical populations suffering from various olfactory dysfunctions. To further validate our method, we aim to expand our dataset using additional MRI scans from upcoming studies at our department and other research institutions. This improvement would prospectively increase the accuracy of the model’s predictions. Hopefully, it would lead to establishing the described tool as a clinical standard in the diagnosis of olfactory diseases.
At the current stage of development and validation, our algorithm for measuring the OB volume may have immediate application to the clinical setup, and reliably offer support to the diagnosis and treatment of patients with olfactory dysfunction. The unbiased OB detection offers a faster, easier, and more reliable (less prone to rater-dependent/image-quality bias) strategy for producing morphological characterizations of the OB. Furthermore, it offers the opportunity to perform reliable follow-up studies, centered in the investigation of volumetric changes to the OB that may occur over time.

5. Conclusions

In this study, we developed a tool that can perform multimodal automatic segmentation of the OBs based on whole-brain T1 and T2 coronal high-resolution images. We demonstrated that our method provides reliable results, matching the accuracy of the gold-standard technique for OB segmentation (i.e., manual segmentations made by experienced raters). As such, our methodological framework could be directly implemented in clinical practice, as well as provide a crucial contribution to olfactory research by offering the opportunity to investigate large neuroimaging datasets of healthy and unhealthy populations.
This improvement may open up new important possibilities for characterizing and predicting the absolute OBV changes induced by treatment to olfactory disfunction, as well as provide new insights into important dynamic alterations occurring in the olfactory system, which is essential for advancing our fundamental knowledge of the olfactory system and enhancing the diagnosis and treatment of olfactory dysfunction. Furthermore, it may facilitate the integration of the use of OB volume measurements in the daily clinical routine practice for the diagnosis and treatment of olfactory loss, which may be essential to improve the prognosis and, in essence, treatment options.

Author Contributions

Conceptualization, D.D., H.F., and T.H.; methodology, D.D., H.F., and T.H.; software, D.D., H.F., and V.A.; validation, D.D., F.A., and X.Y.; formal analysis, D.D.; investigation, D.D.; resources, T.H.; data curation, D.D.; writing—original draft preparation, D.D., H.F., T.H.; writing—review and editing, D.D., T.H., H.F., and V.A.; visualization, D.D.; supervision T.H., H.F., and V.A.; project administration, T.H. and H.F.; funding acquisition, none. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Within the context of the respective studies, all participants signed informed consent on the use of their data for research purposes. All of these studies had been approved by the Ethics Committee at the University Clinic of the TU Dresden.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The described Software is publically available at: https://doi.org/10.5281/zenodo.5283132 (accessed on 5 July 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Branche. Available online: https://www.vkd.com/verband/branche/ (accessed on 9 June 2021).
  2. Boesveldt, S.; Postma, E.M.; Boak, D.; Welge-Luessen, A.; Schöpf, V.; Mainland, J.D.; Martens, J.; Ngai, J.; Duffy, V.B. Anosmia—A Clinical Review. Chem. Senses 2017, 42, 513–523. [Google Scholar] [CrossRef]
  3. Hummel, T.; Whitcroft, K.L.; Andrews, P.; Altundag, A.; Cinghi, C.; Costanzo, R.M.; Damm, M.; Frasnelli, J.; Gudziol, H.; Gupta, N.; et al. Position Paper on Olfactory Dysfunction. Rhinology 2016, 56, 1–30. [Google Scholar] [CrossRef] [PubMed]
  4. Croy, I.; Hummel, T. Olfaction as a Marker for Depression. J. Neurol. 2017, 264, 631–638. [Google Scholar] [CrossRef]
  5. Lie, G.; Wilson, A.; Campion, T.; Adams, A. What’s That Smell? A Pictorial Review of the Olfactory Pathways and Imaging Assessment of the Myriad Pathologies That Can Affect Them. Insights Imaging 2021, 12, 7. [Google Scholar] [CrossRef] [PubMed]
  6. Lundström, J.N.; Boesveldt, S.; Albrecht, J. Central Processing of the Chemical Senses: An Overview. ACS Chem. Neurosci. 2011, 2, 5–16. [Google Scholar] [CrossRef] [PubMed]
  7. Huart, C.; Rombaux, P.; Hummel, T. Plasticity of the Human Olfactory System: The Olfactory Bulb. Molecules 2013, 18, 11586–11600. [Google Scholar] [CrossRef] [Green Version]
  8. Han, P.; Winkler, N.; Hummel, C.; Hähner, A.; Gerber, J.; Hummel, T. Alterations of Brain Gray Matter Density and Olfactory Bulb Volume in Patients with Olfactory Loss after Traumatic Brain Injury. J. Neurotrauma 2018, 35, 2632–2640. [Google Scholar] [CrossRef]
  9. Rombaux, P.; Huart, C.; Deggouj, N.; Duprez, T.; Hummel, T. Prognostic Value of Olfactory Bulb Volume Measurement for Recovery in Postinfectious and Posttraumatic Olfactory Loss. Otolaryngol. Head Neck Surg. 2012, 147, 1136–1141. [Google Scholar] [CrossRef]
  10. Abolmaali, N.D.; Hietschold, V.; Vogl, T.J.; Huttenbrink, K.-B.; Hummel, T. MR Evaluation in Patients with Isolated Anosmia Since Birth or Early Childhood. Am. J. Neuroradiol. 2002, 7, 157–164. [Google Scholar]
  11. Weiss, T.; Soroka, T.; Gorodisky, L.; Shushan, S.; Snitz, K.; Weissgross, R.; Furman-Haran, E.; Dhollander, T.; Sobel, N. Human Olfaction without Apparent Olfactory Bulbs. Neuron 2020, 105, 35–45.e5. [Google Scholar] [CrossRef] [Green Version]
  12. Gudziol, V.; Buschhüter, D.; Abolmaali, N.; Gerber, J.; Rombaux, P.; Hummel, T. Increasing Olfactory Bulb Volume Due to Treatment of Chronic Rhinosinusitis—A Longitudinal Study. Brain 2009, 132, 3096–3101. [Google Scholar] [CrossRef]
  13. Buschhüter, D.; Smitka, M.; Puschmann, S.; Gerber, J.C.; Witt, M.; Abolmaali, N.D.; Hummel, T. Correlation between Olfactory Bulb Volume and Olfactory Function. NeuroImage 2008, 42, 498–502. [Google Scholar] [CrossRef] [PubMed]
  14. Herzallah, I.R.; Askar, S.M.; Amer, H.S.; Ahmed, A.F.; El-Anwar, M.W.; Eesa, M.H. Olfactory Bulb Volume Changes in Patients with Sinonasal Polyposis: A Magnetic Resonance Imaging Study. Otolaryngol. Head Neck Surg. 2013, 148, 689–693. [Google Scholar] [CrossRef] [PubMed]
  15. Joshi, A.; Thaploo, D.; Yan, X.; Herrmann, T.; Khabour, H.A.; Hummel, T. A Novel Technique for Olfactory Bulb Measurements. PLoS ONE 2020, 15, e0243941. [Google Scholar] [CrossRef] [PubMed]
  16. Estrada, S.; Lu, R.; Diers, K.; Zeng, W.; Ehses, P.; Stöcker, T.; Breteler, M.M.B.; Reuter, M. Automated Olfactory Bulb Segmentation on High Resolutional T2-Weighted MRI. NeuroImage 2021, 242, 118464. [Google Scholar] [CrossRef]
  17. Noothout, J.M.H.; Postma, E.M.; Boesveldt, S.; Vos, B.D.D.; Smeets, P.A.M.; Išgum, I. Automatic Segmentation of the Olfactory Bulbs in MRI. In Proceedings of the Medical Imaging 2021: Image Processing, Online, 15 February 2021; p. 115961J. [Google Scholar]
  18. Taha, A.A.; Hanbury, A. Metrics for Evaluating 3D Medical Image Segmentation: Analysis, Selection, and Tool. BMC Med. Imaging 2015, 15, 29. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Han, P.; Mann, S.; Raue, C.; Warr, J.; Hummel, T. Pepper with and without a sting: Brain processing of intranasal trigeminal and olfactory stimuli from the same source. Brain Res. 2018, 1700, 41–46. [Google Scholar] [CrossRef]
  20. Croy, I.; Drechsler, E.; Hamilton, P.; Hummel, T.; Olausson, H. Olfactory modulation of affective touch processing—A neurophysiological investigation. Neuroimage 2016, 135, 135–141. [Google Scholar]
  21. Janine, G.; Pengfei, H.; Cedric, M.; Amelie, B.; Anne, J.; Claudia, R.; Valentin, A.S.; Thomas, H. Brain volume changes in hyposmic patients before and after olfactory training. Laryngoscope 2018, 128, 1531–1536. [Google Scholar]
  22. Hummel, T.; Sekinger, B.; Wolf, S.R.; Pauli, E.; Kobal, G. “Sniffin” Sticks’: Olfactory Performance Assessed by the Combined Testing of Odor Identification, Odor Discrimination and Olfactory Threshold. Chem. Senses 1997, 22, 39–52. [Google Scholar] [CrossRef] [PubMed]
  23. Oleszkiewicz, A.; Schriever, V.A.; Croy, I.; Hähner, A.; Hummel, T. Updated Sniffin’ Sticks Normative Data Based on an Extended Sample of 9139 Subjects. Eur. Arch. Otorhinolaryngol. 2019, 276, 719–728. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Yushkevich, P.A.; Piven, J.; Hazlett, H.C.; Smith, R.G.; Ho, S.; Gee, J.C.; Gerig, G. User-Guided 3D Active Contour Segmentation of Anatomical Structures: Significantly Improved Efficiency and Reliability. NeuroImage 2006, 31, 1116–1128. [Google Scholar] [CrossRef] [Green Version]
  25. Li, X.; Morgan, P.S.; Ashburner, J.; Smith, J.; Rorden, C. The First Step for Neuroimaging Data Analysis: DICOM to NIfTI Conversion. J. Neurosci. Methods 2016, 264, 47–56. [Google Scholar] [CrossRef]
  26. Brett, M.; Markiewicz, C.J.; Hanke, M.; Côté, M.-A.; Cipollini, B.; McCarthy, P.; Cheng, C.P.; Halchenko, Y.O.; Cottaar, M.; Ghosh, S.; et al. Nipy/Nibabel: 3.0.0; Zenodo: Geneve, Switzerland, 2019. [Google Scholar]
  27. van Rossum, G.; Drake, F.L. The Python Language Reference. In Python Documentation Manual/Guido van Rossum; Drake, F.L., Ed.; Release 3.0.1 [Repr.]; Python Software Foundation: Hampton, NH, USA, 2010; ISBN 978-1-4414-1269-0. [Google Scholar]
  28. Harris, C.R.; Millman, K.J.; van der Walt, S.J.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N.J.; et al. Array Programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef]
  29. Oskar, M.; Alex, R.; Pradeep, R.R.; Rémi, B.; Fabian, I.; Michael, A.M.; Vincent, X.W.D.; Jay, J. Loli/Medpy: MedPy 0.4.0; Zenodo: Geneve, Switzerland, 2019. [Google Scholar]
  30. Fonov, V.S.; Evans, A.C.; Botteron, K.; Almli, C.R.; McKinstry, R.C.; Collins, D.L.; BDCG. Unbiased average age-appropriate atlases for pediatric studies. Neuroimage 2011, 54, 313–327. [Google Scholar] [CrossRef] [Green Version]
  31. Fonov, V.S.; Evans, A.C.; McKinstry, R.C.; Almli, C.R.; Collins, D.L. Unbiased nonlinear average age-appropriate brain templates from birth to adulthood. In Organization for Human Brain Mapping, San Francisco, USA, 15 July 2009; Elsevier: Amsterdam, The Netherlands, 2009. [Google Scholar] [CrossRef]
  32. Avants, B.B.; Tustison, N.J.; Song, G.; Cook, P.A.; Klein, A.; Gee, J.C. A Reproducible Evaluation of ANTs Similarity Metric Performance in Brain Image Registration. Neuroimage 2011, 54, 2033–2044. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. SciPy 1.0 Contributors; Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef] [Green Version]
  34. McCarthy, P.; Cottaar, M.; Webster, M.; Fitzgibbon, S.; Craig, M. Fslpy; Zenodo: Geneve, Switzerland, 2019. [Google Scholar]
  35. The MONAI Consortium. Project MONAI; Zenodo: Geneve, Switzerland, 2020. [Google Scholar]
  36. Hummel, T.; Urbig, A.; Huart, C.; Duprez, T.; Rombaux, P. Volume of Olfactory Bulb and Depth of Olfactory Sulcus in 378 Consecutive Patients with Olfactory Loss. J. Neurol. 2015, 262, 1046–1051. [Google Scholar] [CrossRef] [PubMed]
Figure 1. OB probability map: background image MNI ICBM 2009c Nonlinear Asymmetric template, threshold 0.5. The coordinates of the COG are xyz: −4/44/−36. (x: sagittal, y: frontal, z: axial orientations). The colormap indicates the probability from 0 to 1.
Figure 1. OB probability map: background image MNI ICBM 2009c Nonlinear Asymmetric template, threshold 0.5. The coordinates of the COG are xyz: −4/44/−36. (x: sagittal, y: frontal, z: axial orientations). The colormap indicates the probability from 0 to 1.
Brainsci 11 01141 g001
Figure 2. Automatic segmentation of the OBs. 1. T1 whole-brain image to MNI2009casym MNI template co-registration. 2. T2 high-resolution image to T1 whole-brain image co-registration. 3. Appling of inversed transformation matrix from step 1 to OB COG mask in MNI space. 4. Appling of inversed transformation matrix from step 2 to OB COG mask in T1 space. 5. Creating a 3D bounding box based on COG coordinates in T2 space from step 4. 6. Cut out the bounding box from the T2 image. 7. Performing image segmentation using trained 3D U-net.
Figure 2. Automatic segmentation of the OBs. 1. T1 whole-brain image to MNI2009casym MNI template co-registration. 2. T2 high-resolution image to T1 whole-brain image co-registration. 3. Appling of inversed transformation matrix from step 1 to OB COG mask in MNI space. 4. Appling of inversed transformation matrix from step 2 to OB COG mask in T1 space. 5. Creating a 3D bounding box based on COG coordinates in T2 space from step 4. 6. Cut out the bounding box from the T2 image. 7. Performing image segmentation using trained 3D U-net.
Brainsci 11 01141 g002
Figure 3. Model training process. Left: Training loss function. Right: DC plot for training (green curve) and validation (red curve) datasets. The DC and average loss values are both arbitrary units. The curves reach a plateau at epoch 100. The model at the epoch 297 was selected as the highest DC (0.84) of the training process. The unit (au) means arbitrary unit.
Figure 3. Model training process. Left: Training loss function. Right: DC plot for training (green curve) and validation (red curve) datasets. The DC and average loss values are both arbitrary units. The curves reach a plateau at epoch 100. The model at the epoch 297 was selected as the highest DC (0.84) of the training process. The unit (au) means arbitrary unit.
Brainsci 11 01141 g003
Figure 4. Top: Boxplot showing the OBVs of manual segmentations measured by two independent raters and the 3D U-Net model on test dataset. Bottom left and right: XY-Plot of the left and right OBV measured by two independent raters and 3D U-Net on test dataset. The trend of the graphs shows a high level of congruency of the measurements.
Figure 4. Top: Boxplot showing the OBVs of manual segmentations measured by two independent raters and the 3D U-Net model on test dataset. Bottom left and right: XY-Plot of the left and right OBV measured by two independent raters and 3D U-Net on test dataset. The trend of the graphs shows a high level of congruency of the measurements.
Brainsci 11 01141 g004
Figure 5. Automatic segmentation of the left (red) and right (green) OB in coronal T2 MRI scan performed manually by two independent raters and the algorithm.
Figure 5. Automatic segmentation of the left (red) and right (green) OB in coronal T2 MRI scan performed manually by two independent raters and the algorithm.
Brainsci 11 01141 g005
Table 1. Participant’s TDI Scores (sum score (TDI) for threshold (T), discrimination (D), and identification (I) of the odors; means (M), standard deviations (SD).
Table 1. Participant’s TDI Scores (sum score (TDI) for threshold (T), discrimination (D), and identification (I) of the odors; means (M), standard deviations (SD).
Smell Dysfunction Patients (N = 79)
TDITDI
M18.002.677.817.52
SD6.722.503.073.01
Normosmic Controls (N = 91)
TDITDI
M36.399.9712.9213.50
SD2.102.241.651.27
Table 2. The table shows convergence of manual segmentations performed by two raters, based on DC and left and right olfactory bulb volume (LOBV and ROBV) for each rater, respectively. The unit (au) means arbitrary unit.
Table 2. The table shows convergence of manual segmentations performed by two raters, based on DC and left and right olfactory bulb volume (LOBV and ROBV) for each rater, respectively. The unit (au) means arbitrary unit.
Smell Dysfunction Patients
Left OB DC (au)Right OB DC (au)LOBV Rater 1 (mm3)LOBV Rater 2 (mm3)ROBV Rater 1 (mm3)ROBV Rater 2 (mm3)
M0.770.7437.8247.6434.3246.47
SD0.070.1011.4814.7810.9315.43
Healthy Controls
Left OB DC (au)Right OB DC (au)LOBV Rater 1 (mm3)LOBV Rater 2 (mm3)ROBV Rater 1 (mm3)ROBV Rater 2 (mm3)
M0.810.8044.1456.0142.4754.15
SD0.060.0512.3816.9213.5417.66
Table 3. Model performance metrics for the test dataset. DC (DC), Average Symmetric Surface Distance (ASSD). The unit (au) means arbitrary unit.
Table 3. Model performance metrics for the test dataset. DC (DC), Average Symmetric Surface Distance (ASSD). The unit (au) means arbitrary unit.
DC (au)ASSD (au)
Left OBRight OBMeanLeft OBRight OBMean
M0.780.750.770.410.440.43
SD0.060.080.050.100.140.10
Table 4. OB volumes of manual segmentations and predicted segmentations of the algorithm for the test dataset.
Table 4. OB volumes of manual segmentations and predicted segmentations of the algorithm for the test dataset.
Manual Segmentation3D U-Net Segmentations
Left OB (mm3)Right OB (mm3)Left OB (mm3)Right OB (mm3)
M46.1442.6344.8046.73
SD12.7014.198.598.86
Table 5. Human validation scale. This table shows the manual rating criteria used by blinded rater for human validation of the segmentations. The rating scale was defined as: 1: “No congruency”, 2–3: “Poor congruency”, 4–5: “good congruency”, 6–7: “very good congruency”, 8–10: “excellent congruency.” The unit (au) means arbitrary unit.
Table 5. Human validation scale. This table shows the manual rating criteria used by blinded rater for human validation of the segmentations. The rating scale was defined as: 1: “No congruency”, 2–3: “Poor congruency”, 4–5: “good congruency”, 6–7: “very good congruency”, 8–10: “excellent congruency.” The unit (au) means arbitrary unit.
Human Rating Results
Rater 1 (au)Rater 2 (au)U-Net (au)
M6.235.925.95
SD0.870.810.87
Minimum4.004.004.00
Maximum8.007.007.00
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Desser, D.; Assunção, F.; Yan, X.; Alves, V.; Fernandes, H.M.; Hummel, T. Automatic Segmentation of the Olfactory Bulb. Brain Sci. 2021, 11, 1141. https://doi.org/10.3390/brainsci11091141

AMA Style

Desser D, Assunção F, Yan X, Alves V, Fernandes HM, Hummel T. Automatic Segmentation of the Olfactory Bulb. Brain Sciences. 2021; 11(9):1141. https://doi.org/10.3390/brainsci11091141

Chicago/Turabian Style

Desser, Dmitriy, Francisca Assunção, Xiaoguang Yan, Victor Alves, Henrique M. Fernandes, and Thomas Hummel. 2021. "Automatic Segmentation of the Olfactory Bulb" Brain Sciences 11, no. 9: 1141. https://doi.org/10.3390/brainsci11091141

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop