1 Introduction

Coronavirus (COVID-19) is a highly contagious, severe respiratory disease that spreads more rapidly than the traditional flu. On March 11, 2020, the World Health Organization (WHO) declared COVID-19 as a global pandemic [21]. Since its outbreak, COVID-19 has affected millions of people worldwide. By September 2022, it has sadly caused the death of over 6.5 million persons around the globe [19].

COVID-19 symptoms include fever, dry cough, shortness of breath, chest pain, fatigue, loss of taste, loss of smell, headaches, sore throat, runny nose, diarrhea, abdominal pain, and nausea [13]. Most COVID-19 patients tend to have mild or moderate flu-like symptoms, whereas few suffer from severe respiratory problems. Some patients, however, are healthy or lucky enough not to notice any symptoms. In mild and moderate cases, lung lesions are typically absorbed after treatment. Yet in advanced cases, lung damage may be severe and irreversible. In addition, moderate disease symptoms left untreated were sometimes found to rapidly progress causing severe complications that lead to hospitalization, lengthy recovery periods, and problems with lung and heart functions [20]. Early detection and proper treatment of COVID-19 is thus essential to speed up recovery and lessen the odds of developing severe complications, by that reducing the demand on the healthcare system and saving lives.

Shortly after the start of the outbreak, polymerase chain reaction (PCR) tests became the gold standard for COVID-19 diagnosis. PCR works by detecting the genetic material of the virus on a swab taken from a person’s nose and/or throat. PCR tests require highly trained professionals and sophisticated lab equipment which can be a challenge in suburban or rural regions. In developing countries, PCR tests are also considerably expensive and not abundantly available. In addition, PCR tests suffer from low sensitivity, meaning that there is high probability that a COVID-19 positive case would be missed by the test [64].

Chest computer tomography (CT) has thus rapidly emerged as a valuable tool for the diagnosis and follow-up of COVID-19 patients, especially those exhibiting acute respiratory symptoms. CT scan equipment are already readily available and commonly utilized in many medical institutions worldwide. In addition, CT scans were shown to have a higher sensitivity than PCR tests [40]. However, accurate diagnosis using chest CT scans requires their thorough examination by an experienced personal, a task that would cause a huge overload on the limited available medical experts. Computer-aided-diagnosis (CAD) has the potential to reduce the load on the healthcare system by providing an automatic, fast, and reliable method for COVID-19 screening from the chest CT images.

CT image abnormalities in COVID-19 patients vary according to the disease stage and severity. Ground-glass opacites (GGO) are considered the earliest and most common radiological abnormality found in up to 98% of the chest CT scans of COVID-19 patients [9]. GGO are usually described as peripheral and bilateral hazy areas with increased lung opacity through which vessels and bronchial structures can still be seen. At later stages, these opacities become patchy with an irregular paving stones appearance (crazy paving). In addition, consolidations may start to appear which are characterized by their white-dense appearance masking vessels and bronchial structures. Other chest CT manifestations of COVID-19 include reticular pattern, air bronchogram, bronchus deformation, fibrosis, vascular dilation, pulmonary nodules, and halo signs. Figure 1 shows examples of chest CT scans of normal and COVID-19 patients at different disease stages.

Fig. 1
figure 1

Examples of chest CT images showing a no COVID-19 symptoms [52], b ground glass opacity (GGO) [26], c crazy paving (pointed to by the yellow arrows) [51], and d consolidations [26]

Wavelet and contourlet transforms have the advantage of being localized in both space and time, which makes them well suited for medical images analysis. Multiresolution decomposition separates an image’s edge and luminance information in its different subbands, as well as brings out different image related information in the different decomposition levels. In addition, the downsampling step implemented within both transforms allows features to be rapidly computed from the downscaled subbands. The aim of the present study is to investigate the reliability of the wavelet and contourlet transforms for COVID-19 detection from chest CT images. Several transform-based features were thus computed from the approximation and detail subbands in order to fully capture the COVID-19 symptoms within the chest CT images. Various analyses were then performed to determine the transform and decomposition level most suitable for COVID-19 classification. Next, feature selection was implemented to find the best set of features for optimal classification performance. The introduced transform-based COVID-19 detection method is shown to outperform several state-of-the-art deep learning approaches from literature, with the added advantage of reduced computational complexity. The merits of the proposed method can be attributed to the efficiency of the implemented transform-based features in capturing COVID-19 manifestations in the chest images, in addition to them being rapidly computed from the downsampled transform subbands.

The rest of the paper is organized as follows: Section 2 describes several related works considering both traditional and deep learning approaches. Section 3 gives a brief summary of multiresolution analysis techniques, specifically the wavelet and contourlet transforms considered in this work. Section 4 introduces the two public chest CT images datasets utilized in the present study, as well as the transform-based features computed for COVID-19 detection. Section 5 summarizes the performance metrics used to evaluate the performance of the introduced algorithm. In Section 6, the results of all experiments are presented and discussed in addition to being compared to various state-of-the-art methods from literature. Section 7 presents a discussion of the merits and limitations of the introduced transform-based COVID-19 detection method. Finally, Section 8 draws the conclusions of this study, then gives directions for future work.

2 Literature review

Artificial intelligence (AI) approaches for the detection of COVID-19 from chest CT images can be divided into traditional and deep learning approaches. Traditional COVID-19 detection methods consider handcrafted features computed either from the spatial or transform domain of the chest CT images. Examples of features used include statistical (e.g., mean, variance, skewness, kurtosis, energy, entropy, etc.) and textural features (e.g., gray level co-occurrence matrices (GLCM), local binary patterns (LBP), etc.). Chen [14] considered GLCM features for COVID-19 detection from chest CT images using a support vector machine (SVM) classifier. For a small dataset of 296 images, an accuracy of 75.69% was reported. Tan et al. [59] initially calculated 1,688 textural, statistical, and shape features. Next, they implemented feature selection to remove irrelevant features. They tested their algorithm on a dataset consisting of 319 chest CT images achieving accuracies of approximately 98 ± 1% in their different experiments. Both of these approaches have the limitation of being tested on very small datasets (< 350 images), in addition to achieving limited performance. Although promising, handcrafted-based COVID-19 detection methods, so far, leave a huge room for improvement.

Unlike traditional methods that rely on an expert for the extraction of relevant handcrafted features, deep learning methods incrementally learn high-level distinctive features from the data without the need for any human intervention. Convolutional neural networks (CNNs) are typically used deep learning networks for image processing and classification. Deep networks used for image classification can be categorized into standard pre-trained networks and customized networks.

Pre-trained networks refer to standard CNNs that have been previously introduced and trained on a large generic dataset. ImageNet [22] is the most commonly used dataset for pretraining standard CNNs as it consists of 1.2 million images within 1,000 categories. These pre-trained networks are then fine-tuned using the target dataset for the final classification. This process is commonly referred to as transfer learning (TL). TL has the advantage leveraging knowledge from one domain where abundant data is available to another domain having limited data hence boosting the latter’s performance. Due to the complexity, cost, and ethical issues associated with data collection, medical datasets are typically much smaller than those used for computer vision tasks (e.g., hundreds/thousands vs. millions) [43]. TL is thus widely applied in the medical domain where limited data is available.

COVID-19 detection approaches relying on TL techniques having been widely implemented in literature considering various networks such as AlexNet, VGG, ResNet, MobileNet, Inception, and DenseNet [6, 45]. Ahuja et al. [7] compared three pre-trained variants of ResNet to SqueezeNet for COVID-19 detection. ResNet18 achieved the best results with an area under the receiver operating characteristics curve (AUC) of 0.996 and an accuracy of 99.4%. Jaiswal et al. [38] compared TL approaches using VGG16, Inception-ResNetV2, ResNet152V2, and DenseNet. They concluded that, for the considered networks, DenseNet gave the highest results (AUC = 0.97 and accuracy = 96.25%). Gaur et al. [31] also considered TL, but the novelty of their work was that they utilized the empirical wavelet transform for decomposing the chest CT images. Wavelet subbands from levels 1 to 5 were applied as inputs to several DenseNet networks. Best results achieved were for the DenseNe121 (AUC = 0.96 and accuracy = 85.5%). Several other works have also compared different deep architectures [30, 44, 57] where, interestingly, the best performing pre-trained network was found to vary across the different research. Another limitation of TL approaches is that they are somewhat generic. TL methods consider pre-trained deep networks that were originally designed for computer vision tasks, not for medical image analysis. By that, these networks may not efficiently take into account the anatomical structures within the analyzed medical images, which could in turn affect its overall performance.

Although there has been an abundant body of research relying on TL for COVID-19 detection from chest CT images, limited works implemented customized networks for the task. Among these works, Pathak et al. [49] used a deep bidirectional long short-term memory network with mixture density network (DBM). In their work, they tune the DBM’s hyperparameters using a Memetic Adaptive Differential Evolution (MADE) algorithm resulting in an accuracy of 98.37% for a large public dataset having 2,482 chest CT images. Another work by Islam et al. [36] introduced a novel deep CNN for extracting prominent features from the chest CT images. They used an ensemble learner based on five different classifiers and soft voting to detect COVID-19 resulting in accuracy of 99.73%. Both these works showed that their deep learning methods outperformed various TL-based approaches from literature. These findings are in agreement with previous research that showed that in medical applications, a well-developed deep network can significantly outperform generic TL based methods [3, 43].

An interesting study by Dina et al. [53] combines both deep and handcrafted features for COVID-19 disease diagnosis. Pre-trained AlexNet, GoogleNet, ShuffleNet, and ResNet18 were all employed to extract the deep features. Statistical and textural features along with different wavelet subbands were considered as the handcrafted features. The authors then compared using (1) deep features only (6,176 feature), (2) handcrafted features only (772 feature), and (3) FUSI-CAD were both the deep and handcrafted features were combined. FUSI-CAD was shown to give the highest results considering a public dataset of 2,482 chest CT images. Interestingly, the difference in performance between the three compared scenarios was minimal. For example, FUSI-CAD’s AUC and accuracy were higher than for the case of using only the handcrafted features by merely 1% and 0.5%, respectively. Their work thus showed that handcrafted features have great potential for the development of reliable COVID-19 detection algorithms with the added advantage of reduced computation complexity and less data requirements [60].

COVID-19 automatic detection methods in literature mainly considered the spatial chest images for feature computation. Multiresolution decomposition is another approach for image representation that has been successfully implemented for various disease classifications from CT images [41, 50]. Multiresolution analysis has the ability to separate an image’s luminance (low frequency) and edge (high frequency) information in its different subbands, such that each can be analyzed and processed separately. Although Dina et al. [53] have previously considered wavelet features in their algorithm, the wavelet subband coefficients were directly used as features as opposed to extracting relevant handcrafted measures from them. In addition, wavelet coefficients were used alongside other spatial and deep features by that not fully exploring the potential of transform-based features for the implementation of reliable automatic COVID-19 diagnosis algorithms.

In the present study, a transform-based COVID-19 detection algorithm from chest CT images is proposed that relies solely on handcrafted features. This work was inspired by two main research papers: The first is the work of Dina et al. [53] which showed that for COVID-19 detection from chest CT images, handcrafted features gave results that were almost as good as those attained by deep features. The second was a previous work by the author in which transform-based features were found to significantly outperform several other spatial features for disease diagnosis [2]. The aim of the present study is thus to investigate the usefulness of handcrafted transform-based features for the development of a computationally efficient, fast, and reliable COVID-19 detection algorithm from chest CT images. The main contributions of this paper can be summarized as follows:

  1. (1)

    Different handcrafted transform-based features were calculated from the approximation and detail subbands in order to fully capture COVID-19 symptoms in the CT images.

  2. (2)

    A comprehensive analysis was made to determine the transform (wavelet vs. contourlet) and decomposition level better suited for COVID-19 classification.

  3. (3)

    Following the previous analysis, feature selection was implemented to find the best set of transform-based features for optimal performance.

  4. (4)

    A computationally efficient, fast, and reliable transform-based COVID-19 diagnosis algorithm was finally presented and shown to outperform several state-of-the art deep learning methods with the advantage of being computationally inexpensive.

3 Multiresolution transforms

Multiresolution decomposition has the merit of being similar to the human visual system (HVS) which indicates that the eye decomposes images into low and high frequency subbands then analyzes each differently [29]. The decomposition of images into several scales and directions assists in their deep analysis, leading in turn to relevant feature extraction from the different subbands. In addition, multiresolution techniques typically involve image dimensionality reduction allowing for reduced computation complexity and faster feature computation [4]. Among the most widely used transforms in image processing applications are the wavelet and contourlet transforms, both which have the advantage of being localized in both space and time.

3.1 Wavelet transform

Wavelet transform (WT) [5] applies a series of low (L) and high (H) pass filters along the rows and columns of an image to separate its illumination and edge information in the approximation (LL) and details subbands, respectively. The high frequency information is distributed among three different detail subbands representing an image’s horizontal (LH), vertical (HL), and diagonal (HH) edge information. Multiresolution analysis can be implemented by reapplying the WT to an image’s approximation subband by that bringing out its finer details in higher decomposition levels. More details concerning the mathematical formulations of the WT can be found in Ref. [27]. Figure 2a shows the block diagram of a single-level wavelet decomposition, whereas Fig. 2b demonstrates different wavelet subbands for single and multilevel wavelet decompositions of a sample image.

Fig. 2
figure 2

a Wavelet transform multiresolution decomposition scheme [33], b From left to right: Original image—single wavelet decomposition indicating the approximation and detail (horizontal – vertical – diagonal) subbands—three level multiresolution wavelet decomposition [1]

3.2 Contourlet transform

A drawback of the wavelet transform is its limited capability to efficiently capture directional information within the images, specifically for curved structures. Contourlet transform [24] uses a double filter bank structure to get smooth image contours. Initially, the image is decomposed into low and high frequency subbands. Next, directional filter banks (DFBs) are applied to the high frequency subbands in order to further bring out the image’s finer details. Furthermore, multiresolution analysis can be implemented by reapplying the contourlet transform to the approximation subband by that bringing out its finer details in the higher decomposition levels. More details on the mathematical representation of the contourlet transform can be found in Ref. [24]. Figure 3a illustrates the block diagram of a Contourlet filter bank, whereas Fig. 3b shows the different contourlet subbands for the 3-level decomposition of a sample image.

Fig. 3
figure 3

a Contourlet transform multiresolution decomposition scheme [24], b Original image along with its three level contourlet decomposition, where the detail subbands in all levels are decomposed into eight different directions

Contourlet transform has several advantages over WT such as directionality and anisotropy. The DFBs oriented at several directions implemented within the contourlet transform allows for an image’s decomposition in more than just the three directions offered by the WT. In addition, the elongated shapes and different aspect ratios of the basic elements within the contourlet transform facilitate the capturing of the smooth contours within images. Such merits generally make the contourlet transform more suitable for curve representation than the WT [24].

4 Materials & methods

4.1 Chest CT datasets

Two chest CT datasets were used for the development and testing of the proposed COVID-19 detection algorithm (Table 1):

  1. 1)

    CoDS1 (SARS-CoV-2) [57]: includes 1,230 positive and 1,252 negative COVID-19 chest CT scans. Images were collected from several hospitals across Sao Paulo, Brazil.

  2. 2)

    CoDS2 (subset of iCTCF [48]): consists of 3,777 positive and 3,910 negative COVID-19 chest CT scans. Images were collected from two hospitals: Union Hospital (HUST-UH) and Liyuan hospital (HUST-LH) in China.

Table 1 Summary of the two considered COVID-19 chest CT datasets

Figure 4 shows sample images from the CoDS1 and the CoDS2 chest CT datasets. The first dataset (CoDS1) was used for the preliminary analysis in which the different transforms and decomposition levels were compared. The second larger dataset (CoDS2) was considered for the final classification experiments.

Fig. 4
figure 4

Chest CT sample images of COVID-19 negative  (column 1) and COVID-19 positive (columns 2&3) patients from the CoDS1 (row1) and the cropped CoDS2 (row 2) datasets

4.2 Preprocessing

Chest CT images are typically surrounded by a large black area that includes no relevant diagnostic information. The CoDS1 chest CT images were already cropped by its providers. As for the CoDS2 dataset images, the black region surrounding the chest was cropped using Algorithm 1. Figure 5 shows a sample image from CoDS2 before and after cropping. For both datasets, the cropped chest CT images were resized to 256 × 256 pixels to make them suitable for the contourlet decomposition as well as to assure consistent information extraction from the different images [5]. Finally, contrast stretching was applied to all images in order to enhance their contrast prior to feature extraction. Contrast stretching works by stretching an image’s intensity values to the full range of pixel values [54]. Specifically, the chest CT image intensity values between 0.1 and 0.9 were stretched to the range from 0 to 1 by that accenting the image details for enhanced classification performance.

figure a

Algorithm 1: Chest CT Image Cropping

Fig. 5
figure 5

Sample chest CT image from CoDS2 before (left) and after (right) cropping

4.3 Feature extraction

COVID-19 significantly alters the overall appearance and texture of a patient’s chest CT images. Multiresolution decomposition has the advantage of separating an image’s luminance and edge information in its approximation and detail subbands, respectively. In this work, a transform-based COVID-19 detection algorithm is presented in which different handcrafted features are computed from the approximation and detail subbands in order to capture these changes within the chest CT images. Figure 6 shows the flowchart of the introduced transform-based COVID-19 detection algorithm.

Fig. 6
figure 6

Flowchart of the proposed transform-based COVID-19 detection algorithm from chest CT images

4.3.1 Singular Value Decomposition (SVD) features

SVD is a robust and reliable orthogonal matrix factorization method. It is considered a linear algebra technique that represents any matrix (A) of size m × n as the product of three matrices as follows:

$$A=U\;S\;V^T$$
(1)

where Um×m and Vn×n are unitary matrices, and Sm×n is a diagonal matrix containing the square roots of the non-zero Eigen values of matrices U and V. The diagonal values in the S matrix are referred to as the singular values. They are typically arranged in descending order, where the largest values are the most relevant.

SVD has been widely adopted for dimensionality reduction in image analysis techniques [8], in addition to being used to extract salient features from either the images’ spatial [28] or wavelet subbands [25]. In this work, the first k singular values were considered in order to extract meaningful information from the chest CT images’ approximation subbands, where k was empirically chosen to be 20.

4.3.2 GLCM features

GLCM features, also known as Haralick features, are widely implemented for image texture classification. Initially, the GLCM are created to represent how often specific pixel pairs occur within an image. Next, different order statistical features are computed from these matrices in order to thoroughly describe the image texture. Extracting the GLCM features from the approximation subbands was previously shown to result in reliable performance with the added advantage of reduced time complexity [2]. In this work, twenty-two GLCM features [15, 17, 34] were calculated from the approximation subbands of the chest CT images. Each feature was initially computed from the 0°, 45°, 90°, and 135° directional matrices then averaged in these four directions in order to thoroughly capture textural variations within the image. This approach is similar to that performed in Ref. [34] in which the Haralick features were originally introduced. Table 2 lists the GLCM features considered in this work.

Table 2 Summary of the considered GLCM features computed from an image’s approximation subbands

4.3.3 Statistical features

Chest CT images including COVID-19 symptoms typically have a cloudier appearance (i.e., less edge related information) than healthy images. Multiresolution analysis separates an image’s edge (high frequency) information within its detail subbands. Three statistical features were thus computed from the detail subbands of the decomposed chest CT images to evaluate their overall cloudiness which are Shannon entropy [16], mean, and standard deviation (stdev) as given by the following equations:

$${Entropy}_{sb}= - {\sum }_{i=1}^{N}{\left|{C}_{i}\right|}^{2}{\mathit{log}}_{e} {\left|{C}_{i}\right|}^{2}$$
(2)
$${Mean}_{sb} = \frac{1}{N} {\sum }_{i=1}^{N}{\left|{C}_{i}\right|}$$
(3)
$${Stdev}_{sb} = \sqrt{\frac{1}{N-1} {\sum }_{i=1}^{N}({Mean}_{sb}- {\left|{C}_{i}\right|)}^{2}}$$
(4)

where N is the number of coefficients in the considered detail subband (sb) and Ci is the detail subband coefficient having index i.

In this work, wavelet and contourlet transform-based features are investigated in order to determine which is more suitable for reliable COVID-19 detection from the chest CT images. The introduced features were computed from three decomposition levels of both transforms, then classification performance was compared for the different levels. Specifically, SVD and GLCM features were extracted from the approximation subbands, whereas the statistical features were calculated from the detail subbands. For each image, 20 SVD and 22 GLCM features were extracted from its approximation subband. In addition, three statistical features were calculated from the three and eight detail subbands of the wavelet and contourlet transforms, respectively. The feature vector for each image thus includes a total of 51 features when considering the images’ wavelet decomposition and 66 features when considering their contourlet decomposition.

5 Performance metrics

The proposed transform-based COVID-19 detection algorithm was evaluated using six performance metrics that are commonly used in medical AI applications which are [35]:

  1. a)

    Accuracy: the percentage of correctly classified cases, both positive and negative, with respect to the entire dataset.

    $$Accuracy=\frac{TP+TN}{TP+FP+TN+FN}\times100$$
    (5)
  2. b)

    Sensitivity (Recall): the percentage of correctly classified positive cases with respect to all the positive cases (i.e., the accuracy of correctly detecting COVID-19 cases).

    $$Sensitivity=\frac{TP}{TP+FN}\times100$$
    (6)
  3. c)

    Specificity: the percentage of correctly classified negative cases with respect to all the negative cases (i.e., the accuracy of correctly detecting non-COVID-19 cases).

    $$Specificity=\frac{TN}{TN+FP}\times100$$
    (7)
  4. d)

    AUC: In the receiver operating characteristic (ROC) curve, sensitivity and specificity are plotted against each other for different threshold values. A classifier whose performance is no better than random guessing would have an AUC of around 0.5. Better classifiers are those having AUC values that are closer to one.

  5. e)

    Precision: the percentage of correctly classified positive cases divided by the total number of positively classified cases.

    $$Precision=\frac{TP}{TP+FP}\times100$$
    (8)
  6. f)

    F-score: is the harmonic mean of the model’s precision and sensitivity (recall).

    $$F-score=\frac{2(Precision\times Recall)}{Precision+Recall}$$
    (9)

For Eqs. (59), the TP (true positive) and TN (true negative) values indicate the number of COVID-19 and non-COVID-19 images that are properly classified, respectively. The FP (false positive) and FN (false negative) are the number of wrongly classified COVID-19 and non-COVID-19, respectively.

6 Experiments & results

Two experimental phases were performed in the present study which are: (1) transform analyses and (2) final experiments. In the first phase, several analyses were made to determine the transform and decomposition level most suitable for capturing relevant information related to COVID-19 within the chest CT images. In the second phase, the complete feature vector was first used to compare the performance of several classifiers for the required task. Next, feature selection was implemented to find the best feature set for optimal COVID-19 classification performance. CoDS1 and CoDS2 datasets were utilized for the first and second experimental phases, respectively. The reason why two different datasets were used for each phase was to ensure the generalized and unbiased results of the final experiments. The summary of the performed experiments is given in Table 3.

Table 3 Summary of the two experimental phases performed for the development of the proposed transform-based COVID-19 detection algorithm

MATLAB (MathWorks, Inc., Natick, MA, USA) was used for image preprocessing, multiresolution decomposition, and for the computation of the SVD, GLCM [61], and statistical features. Specifically, wavelet and contourlet decompositions were performed using the Wavelet Toolbox (dwt command) [23] and the Contourlet Toolbox (pdfbdec command) [18], respectively. Classification and feature selection were performed within the WEKA platform [32]. All computations and processing were made using an Intel Core i7-5500U CPU @2.4 GHz with 16 GB of RAM.

6.1 Transform analyses

6.1.1 Wavelet filter analysis

Several wavelet filters (basis functions) exist in literature, each being more suited for a different task [10]. In this work, several wavelet basis functions were compared which are: Haar, Daubechies, and Symlets. Haar is the first and simplest wavelet basis that is represented by a step function. Haar wavelet does not have overlapping windows, by that only reflecting changes between adjacent pixels. Daubechies wavelets are a family of non-symmetric orthogonal functions that are among the most commonly used bases for image texture analysis [56]. They employ overlapping windows, hence are better at capturing edge information within images than the simple Haar wavelet. Symlets are the least-asymmetric wavelets because they are nearly but not exactly symmetrical. Symlets have smoother functions and higher capabilities than the Daubechies wavelets making them more attractive in several image processing applications. For the Daubechies (dbN) and Symlets (symN) wavelet families, N denotes the number of vanishing moments (where 2N is the number of filter taps) [46]. The Haar wavelet is also referred to as the Daubechies1 (db1) wavelet. More information on the different wavelet families can be found in Refs. [56, 58].

Table 4 compares the classification performance of fifteen different wavelet basis functions considering the CoDS1 dataset. For all experiments, the wavelet features were computed from the first decomposition level’s subbands. The k-nearest neighbor (kNN) with tenfold cross validation was used to classify the images. The kNN classifier was chosen as it is easy to tune and gives consistent results. Overall, slight variation in performance was observed among the different wavelet functions. Best performance (i.e., AUC and accuracy) was attained by the db4 and the sym10 wavelet basis functions. Since sym10 achieved a higher sensitivity, i.e., more positive COVID-19 cases were correctly identified, the sym10 wavelet basis function will be utilized for wavelet decomposition in the next experiments.

Table 4 Performance comparison between different wavelet basis functions using the kNN classifier (CoDS1)

6.1.2 Decomposition level analysis

In this analysis, wavelet and contourlet transforms were compared to determine which would be more suitable for COVID-19 detection from chest CT images. For each transform, three-level decomposition was performed, then the complete feature vector was computed from each decomposition level. KNN with tenfold cross validation was used to classify the images as positive or negative COVID-19.

Table 5 summarizes the performance metrics for the six feature vectors computed from the different wavelet and contourlet decomposition levels. Experimental results indicated that considering a specific decomposition level, better performance was attained by the features computed from the contourlet transform. And then for the contourlet transform, better performance was achieved for features calculated from level 1 (L1) followed by those from level 2 (L2). Level 3 (L3) contourlet features, however, gave the lowest results with accuracies being approximately 13% and 10% less than those achieved by the L1 and L2 features, respectively. Level analysis thus showed that the L1-Contourlet feature vector resulted in the best classification performance (AUC = 0.992 and accuracy = 96.3%).

Table 5 COVID-19 classification results considering different wavelet and contourlet levels (CoDS1)

In order to preliminary assess the robustness of the proposed L1-Contourlet features, the classification performance was compared to several state-of-the-art deep learning methods from literature (Table 6). Soares et al. [57] and Gaur et al. [31] COVID-19 detection approaches both relied on pre-trained deep networks (TL). Accuracies attained by these methods for the CoDS1 dataset were found to be in the range between 85 and 95%. The L1-Contourlet features implemented in this study gave an accuracy of 96% by that outperforming TL methods by up to 11%. TL approaches rely on standard networks pre-trained on general image classes. These networks are not specifically designed to capture anatomical structures related to the considered disease within the processed medical images. On the other hand, the handcrafted features implemented in this work were carefully chosen to capture the appearance changes within the chest CT images. Specifically, image decomposition into low and high frequency subbands allows for the efficient representation of these changes as a result of the disease progression.

Table 6 Performance of several deep learning COVID-19 detection from literature (CoDS1)

Pathak et al. [49] did not rely on pre-trained networks in their work, instead they used a deep bidirectional long short-term memory network with mixture density network (DBM) whose hyperparameters were tuned using an adaptive evolution algorithm. Although they attain higher accuracies than the L1-Contourlet features implemented in this work, the L1-Contourlet approach achieves a higher AUC, which indicates the robustness of the devised method. Transform-based handcrafted features thus show great potential for the development of reliable automatic COVID-19 detection algorithms.

The aim of the analysis performed in this section was to compare between wavelet and contourlet transforms, as well as to determine the most suitable decomposition level for feature extraction. From this analysis, L1-Contourlet features were found to give the best performance, hence they will be considered for all the upcoming experiments. In the next section, different classifiers will be compared to determine the one best suited for COVID-19 detection. Next, feature selection will be employed to find the best set of features for optimal performance, as well as to investigate the relevance of the different feature types implemented in this work. For these two experiments, L1-Contourelet features will be extracted from a different dataset (i.e., CoDS2) in order to asses the generalizability of the introduced transform-based features.

6.2 Final experiments

6.2.1 Classifier selection

Table 7 compares different classifiers for the detection of COVID-19 considering L1-Contourlet features computed from the CoDS2 dataset. The results are presented for five different classifiers which are: kNN, random forest (RF), neural network (NN), support vector machine (SVM) with a linear kernel, and SVM with a radial basis function (rbf) kernel. The parameters of each classifier were tuned for best performance using a grid search algorithm. In all experiments, tenfold cross validation was used. Best performance is shown to be achieved by the kNN and SVM (rbf) classifiers resulting in an AUC of 1.0 and an accuracy of 99.97%. The SVM classifier with an rbf kernel was chosen as it is less computationally demanding the kNN.

Table 7 COVID-19 classification results for different classifiers considering the L1-Contourlet features (CoDS2)

Timing analysis was performed considering the compete feature vector in order to check whether the presented L1-Contourlet method would be suitable for integration in real-time systems. Table 8 summarizes these results considering the CoDS2 dataset which consists of 7,687 images resized to 256 × 256 pixels. The timing analysis showed that in order to compute the complete feature vector, an average of 25 ms were required per image. The tenfold cross validation experiments using the SVM (rbf) classifier required less than 8 s indicating that a single training session was completed in less than a second. The proposed algorithm thus has the advantages of providing reliable performance while remaining computationally inexpensive.

Table 8 Timing analysis of the introduced COVID-19 detection algorithm (CoDS2)

6.2.2 Feature selection

Particle swarm optimization (PSO) is an evolutionary search method that was first developed by Kennedy and Eberhart in 1995 [39]. Later, it was adapted to be used for feature selection (FS) and has since been widely used in several machine learning applications. PSO starts with a swarm of random solutions (particles) moving in the search space to find the optimal solution. Each particle’s position is updated based on its own experience in addition to the experience of its neighboring particles. PSO has the advantages of being powerful, having few parameters, as well as being less computationally expensive than other evolutionary search algorithms [62].

L1-Contourlet features include a total of sixty-six textural and statistical features computed from the approximation (42 features) and detail (24 features) subbands. The PSO algorithm was applied to find the most relevant feature set for optimal performance. Twenty-four different features were selected by the PSO algorithm which are summarized in Table 9. The reduced feature set is shown to consist of all feature types including the SVD and GLCM features computed from the approximation subbands, as well as the statistical features calculated from the detail subbands. This indicates relevance of the different transform-based feature types considered in the present study.

Table 9 Reduced feature set for COVID-19 detection based on the particle swarm optimization (PSO) technique

Table 10 summarizes the classification results for different combinations of the reduced feature vector. In all experiments, classifications were performed using an SVM (rbf) classifier with tenfold cross validation. Results show that, used separately, statistical features gave the highest results (accuracy = 98.47%) closely followed by the GLCM features (accuracy = 92.14%). The SVD features used alone, however, resulted in the poorest performance (accuracy = 78.31%). Combining two of the feature groups together (i.e., GLCM + SVD, GLCM + statistical, SVD + statistical) led to better performance than when one feature group was solely considered. These results indicate the importance of employing different feature types for reliable COVID-19 diagnosis. Finally, considering all features within the reduced feature vector is shown to give the best performance where an AUC of 1.0 along with accuracy, sensitivity, specificity, precision, and F-score all having a value 100% were achieved.

Table 10 Classification results for different combinations of the reduced feature set using the SVM (rbf) classifier (CoDS2)

Table 11 summarizes the results of different deep learning approaches from literature that considered the iCTCF dataset. Liu et al. [44] used anomaly weighted images to train several DenseNet networks. A baseline VGG19 network resulted in an accuracy of 95.9%, whereas the different DenseNet networks gave accuracies that were in the range between 96.38% and 97.88%. Liang et al. [42] integrated ResNet and Vision Transformer (ViT) to extract local and global features, respectively. In their work, a weakly supervised module assisted the ViT to better converge. They reported an accuracy of 98.19%. The introduced L1-Contourlet based features resulted in an accuracy of 100% indicating the superiority of the proposed method.

Table 11 Performance of several deep learning COVID-19 detection algorithms from literature

7 Discussion

Transform-based features were scarcely used in literature for the development of COVID-19 detection algorithms from chest CT images. Among the limited traditional works, Yao et al. [63] implemented a wavelet-based COVID-19 detection algorithm in which they only used wavelet-entropy as a feature. They reported an accuracy of 73.95% and an F-score of 73.66%. As for TL based approaches, Matsuyama et al. [47] used the wavelet subbands of the chest CT images as inputs for fine-tuning the ResNet50 network resulting in an accuracy of 92.2% and an F-score of 91.5%. Similarly, Guar et al. [31] input the wavelet subbands to several pre-trained DenseNet networks achieving best accuracy and F-score of 85.5% and 85.25%, respectively. In these works, performance was somewhat limited (i.e., accuracies barely exceeded 92%) indicating that the potential of the transform-based features for COVID-19 detection was not yet fully reached. An interesting work by Dina et al. [53] combined wavelet coefficients alongside textural and deep features derived from several pre-trained networks for chest CT classification. They were able to achieve accuracies of 98.6%, 98%, and 99% for the deep, handcrafted, and combined features, respectively. By that, they showed that deep features (attained from pre-trained networks) and handcrafted features (including the wavelet coefficients) achieved very close performance. Their work shed the light on the relevance of spatial and transform-based handcrafted features, and how they can result in comparable performance to deep learning algorithms.

In this work, a comprehensive study was performed in order to investigate the usefulness of transform-based features for COVID-19 detection from chest CT images. Several analyses were made to determine the transform and decomposition level better suited for COVID-19 detection. L1-Contourlet statistical and textural features were shown to result in accuracies of 96.3% and 100% for two different chest CT datasets. To the best of the author’s knowledge, this work is the first to implement contourlet-based handcrafted features for the development of an automatic COVID-19 diagnosis from chest CT images. The proposed transform-based algorithm was found to outperform several deep learning approaches from literature, specifically those employing TL techniques. Similar findings were reported in Refs. [53] and [12] comparing handcrafted and deep features for COVID-19 detection using CT and X-ray images, respectively. In addition, Ismael et al. [37] showed that simple Shearlet-based features (energy and entropy) computed from chest X-ray images gave better COVID-19 classification performance than ResNet50 regardless of whether or not TL was employed. All these works thus point to the importance of well-designed features for reliable performance.

Multiresolution analysis allows for the separation of an image’s high and low frequency information in its approximation and detail subbands, respectively. Features computed from the different subbands can thus thoroughly capture the changes in the chest CT images due to COVID-19 progression. In addition, contourlet transform provides a high degree of directionality which allows for robust representation of the image characteristics. Accordingly, the superiority of the proposed handcrafted L1-Contourlet features can be attributed to the relevance of the textural and statistical features computed from the different contourlet subbands. Deep pretrained networks commonly employed in medical applications are designed without consideration of the characteristics of the medical images or disease symptoms [3, 43]. Features extracted by these generic deep networks might thus inefficiently represent the different classes resulting in less-than-optimal classification performance.

Another advantage of the introduced transform-based COVID-19 detection algorithm is the reduced computational complexity, specifically in comparison to the deep learning methods. L1-Contourelt features presented in this work were shown to require only a few milliseconds for feature extraction, and less than a second for training. On the other hand, deep networks typically require hours or even days for their training [55]. For example, Liu et al. [44] reported training their deep networks for COVID-19 detection from chest CT images in around 80 h using a Nvidia GeForce GTX 1080 GPU machine. The proposed algorithm thus has the advantage of providing reliable results while remaining computationally inexpensive. Consequently, it can be easily and rapidly re-trained whenever more labeled data is made available. In the COVID-19 pandemic era, this is very useful as every hour counts in saving the lives of hundreds or even thousands of patients [11]. Nevertheless, a limitation of the presented method is that it only classifies the chest CT images into two classes: positive or negative COVID-19. In medical diagnosis, it is typically more desirable to determine the stage of the undergoing disease in order to decide on a suitable treatment plan. In addition, it would be desirable to further investigate its performance over a wide variety of chest CT datasets including real-time environmental data.

8 Conclusions

A contourlet-based COVID-19 detection algorithm from chest CT images was introduced and shown to outperform several state-of-the-art methods from literature. A comprehensive study was preliminary performed to determine the transform and decomposition level better suited for the required task. Experiments showed that the handcrafted features computed from the first contourlet decomposition level resulted in the most reliable performance. Feature computations required ~ 25 ms, whereas the rbf SVM classifier was trained in less than a second. Transform-based handcrafted features are thus highly efficient in comparison deep networks which typically require hours for their training. Feature selection was then performed to find the most relevant features, which were shown to include the three different feature types considered in this work. In the final experiments, accuracy, sensitivity, specificity, precision, and F-score of 100% were attained. The proposed contourlet-based approach was shown to outperform various state-of-the-art deep learning methods from literature by up to 11% indicating the superiority of the proposed method. Unlike generic pre-trained deep networks that do not consider specific disease symptoms in the analyzed medical images, handcrafted features have the advantage of efficiently capturing disease manifestations in the images which results in reliable performance. Future work involves designing a deep network that is specifically tailored to capture disease-related features within chest CT images, then comparing its performance to the transform-based method introduced in this study. Furthermore, it would be desirable for clinical purposes to determine the stage of COVID-19, i.e., mild, moderate, or progressive.