Improved graph neural network-based green anaconda optimization for segmenting and classifying the lung cancer

: Normal lung cells incur genetic damage over time, which causes unchecked cell growth and ultimately leads to lung cancer. Nearly 85% of lung cancer cases are caused by smoking, but there exists factual evidence that beta-carotene supplements and arsenic in water may raise the risk of developing the illness. Asbestos, polycyclic aromatic hydrocarbons, arsenic, radon gas, nickel, chromium and hereditary factors represent various lung cancer-causing agents. Therefore, deep learning approaches are employed to quicken the crucial procedure of diagnosing lung cancer. The effectiveness of these methods has increased when used to examine cancer histopathology slides. Initially, the data is gathered from the standard benchmark dataset. Further, the pre-processing of the collected images is accomplished using the Gabor filter method. The segmentation of these pre-processed images is done through the modified expectation maximization (MEM) algorithm method. Next, using the histogram of oriented gradient (HOG) scheme, the features are extracted from these segmented images. Finally, the classification of lung cancer is performed by the improved graph neural network (IGNN), where the parameter optimization of graph neural network (GNN) is done by the


Introduction
The unregulated proliferation of aberrant pulmonary cells represents a symptom of lung cancer.According to histology, prognosis, biological behavior and therapy, there are two main forms of lung cancer [1].SCLC and NSCLC represent the two types [2].NSCLC makes up 85% of cases and SCLC accounts for the other 15%.The most prevalent kind of NSCLC, adenocarcinoma, develops in epithelial cells that exude mucus or fluids [3].Squamous cells that line several organs, including the bladder, lung, intestines, kidney, and stomach, are where squamous cell carcinoma develops [4].
Lung cancer may be diagnosed using a number of techniques, including CT scans, X-rays, bronchoscopies, PET-CT scans and biopsies [5].H and E staining, in which the staining is performed on the tissue extracted from a biopsy, is commonly utilized to determine the subtype of lung cancer on the basis of the tissue type.E includes a pink color and stains proteins, whereas H involves a deep purple color and stains nucleic acids in cells [6].According to research, lung cancer therapy as well as survival are related to the kind of histology [7].Recent developments in genetic research have cleared the way for personalized treatment for people with lung cancer [8].
It is also crucial to analyze the histopathological images of the disease [9].Manually analyzing histopathology results, nevertheless, takes time and is not objective.Pathologists are struggling to handle the burden associated with histopathologic cancer detection [10].In order to lessen the workload on pathologists and accelerate the crucial procedure of lung cancer detection, deep learning approaches are employed [11].The effectiveness of these methods has increased when used to examine cancer histopathology slides.
The paper contribution is as follows: • To employ a deep learning approach to quicken the crucial procedure of diagnosing lung cancer.
• To accomplish the pre-processing of the collected images using Gabor filter method.
• To do the segmentation of the pre-processed images through the MEM algorithm method.
• To extract the features from the segmented images using the HOG scheme.
• To perform the classification of the lung cancer by the IGNN, where the parameter optimization of GNN is done by the GAO in order to derive the accuracy maximization as the major objective function.• To classify the lung cancer into normal, adenocarcinoma and squamous cell carcinoma as the final output by the IGNN.The paper organization is as follows: Section 1 is the introduction of the lung cancer classification model.Section 2 is a literature survey about it.Section 3 is the proposed methodology in consideration of the IGNN and GAO algorithms.Section 4 shows the results in detail for the lung cancer classification model.Section 5 ends with the conclusion.

Related work
In 2020, Masood et al. [12] suggested an improved mRFCN-based lung nodule identification as well as classification.We used a median intensity extension to make use of the 3-D data and we included a deconvolutional layer to incorporate the suggested mLRPN in the framework to choose the possible RoI.The simulation findings demonstrated good detection efficiency in compared to the conventional nodule detection/categorization approaches.The framework was trained and tested utilizing the LIDC dataset.
In 2020, Bicakci et al. [13] thoroughly examined deep learning-oriented classification techniques to distinguish between the two NSCLC subtypes known as squamous cell carcinoma and adenocarcinoma.38 slices were of patients having ADC, while the remaining slices were of SqCC patients.To investigate the impact of peritumoral regions in PET scans on the subtype categorization of tumors, three trials were conducted.In these simulations, we evaluated three CNN methods, including VGG16, SqueezeNet, and VGG19, employing three different types of images: entire slices without segmentation or cropping, square subimages cropped to involve the tumor and random walk technique segmented image parts respective to tumors.Every method was optimized for the diagnostic categorization using a variety of optimizers and regularization techniques.F-score and AUC measures were employed to assess the effectiveness of the project.The findings allow us to conclude that peritumoral areas and tissues are both required for the development of methods and eliminate the need for segmentation.
In 2021, Chen et al. [14] suggested LDNNET, which incorporated Dense-Block, dropout and BN.LDNNET, an adaptable framework on the basis of convnets incorporating softmax classifier, was used to address the issues with deep convnet training.The primary work is as below: First, we classified lung nodules using the database LDNNET on LUNA16, and lung cancers using the database Kaggle DSB 2017.Second, the comparative tests were made to assess how well lung CT images with thick connections, pooling layers and larger input pixels operate.Third, LDNNET used dense connection, dropout layer and data augmentation to minimize overfitting.Fourth, to examine the pre-processing on the categorization of images, several pre-processing techniques were contrasted to the no-processing technique.Fifth, the indices of specificity, accuracy and sensitivity on LUNA16 were 0.994585, 0.988396 and 0.982072, respectively, while they were 0.999652, 0.999480 and 0.998974 on Kaggle DSB 2017.Additionally, the AUC for the two datasets was greater than 0.98.Since lung images were used as input images without any preprocessing, this research runs simulations to demonstrate that LDNNET was still the more enhanced method compared to various existing methods.Additionally, a number of comparison tests were carried out to further demonstrate, via verification and discussion, that the suggested method has superior accuracy and resilience.
In 2021, Li et al. [15] involved very uncommon lung adenosquamous carcinoma specimens for the initial time.The relief method was utilized after the multidimensional characteristics of 121 LC histopathology images were initially retrieved.The ROC curve and AUC were utilized to more easily assess the classifier's capacity for generalization while the SVMs classifier was employed to categorize LC subtypes.Lastly, investigations demonstrate that the Relief-SVM method achieves the optimal classification impact via a horizontal comparison with a range of widely used classification methods.The accuracy of the classifications using the LUSC-SCLC, LUSC-ASC and ASC-SCLC systems was 83.91%, 73.91% and 73.67%, respectively.The effectiveness of the auxiliary diagnosis method created by ML in the detection of LC is confirmed by the simulation findings.
In 2023, Ahmad et al. [16] suggested a fresh approach to categorizing lung CT scans.We combine Gabor filters with an E-DBN and several classification techniques in the suggested strategy for classifying lung cancer.We employ the GB and BB RBMs as two cascaded RBMs in this E-DBN.Of all the applicable approaches, a SVM provides the optimal performance metrics.
In 2022, Ajni and Anitha [17] created a technique known as SSSO-based MORAN to accurately categorize lung cancer illness.The SSD algorithm and the SSOA have been combined to create the SSSO algorithm.The pre-processing stage receives the input CT image, which is then used to preprocess the image using Gaussian filtering.As a result, the input image's RoI was obtained.After that, the suggested DREFC is used to segment the lung lobes.The nodule region was recognized from the lung image using the segmented lung lobes, and cancer was classified on the basis of characteristics.GLCM features, LGXP, Tetrolet transform, GBP and statistical characteristics were among the features taken into account.
In 2023, Bushara et al. [18] put forth a brand-new deep learning methodology referred as LCD-CapsNet to reduce the enormous quantity of information and attain spatial invariance for lung cancer detection.The main goal of the suggested approach was to construct methods that would categorize and analyze images.This deep learning methodology was evaluated using the LIDC datasets.
In 2019, Kavitha et al. [19] suggested a region-oriented FCM clustering method, which significantly improves mortality in clinical practice.CT lung images are used for processing in the developed ECM-CSD, because they offer greater imaging resolution, sensitivity, and excellent isotopic collection in lung nodule detection.These images have undergone pre-processing using a Gabor filter for enhancing and a Gaussian filter for smoothing.Next, utilizing FCM-oriented clustering, an efficient segmentation of lung nodules was carried out on the basis of the retrieved image characteristics.The SVM classification method is often used to determine the phases of cancer.Additionally, the method was examined using the MATLAB platform and the clinical dataset of LIDC-IDRI lung CT scans.The comparison studies demonstrate the suggested method's effectiveness with respect to performance assessment parameters such as improved accuracy and decreased error rate.
In 2023, Alsadoon et al. [20] looked into, found and carefully examined DL methods that performed well in the categorization of lung cancer.37 articles out of the 338 we assessed satisfied the requirements for consideration in the suggested paradigm.In addition, we provide and assess a structure to control the procedure for choosing and deploying DL methods in actual systems.DFCV are the framework's four primary parts.On the basis of 37 cutting-edge research publications, we examine the effectiveness and significance of the developed DFCV method.The DFCV framework could serve as a manual for choosing and implementing DL-oriented solutions in lung cancer treatment facilities.
In 2023, Braveen et al. [21] put up a unique ALbAE method for accurate lung cancer and pneumonia categorization.Initially, noise artefacts were removed from CT images employing median filters to enhance the image quality.Therefore, the ALbAE approach is used to extract the pertinent characteristics, like pixel rates, image edges and blood clots.Lastly, utilizing the RF approach, the lung CT scans were divided into three groups, including lung affected by malignancy, normal lung and lung impacted by pneumonia.Distinct metrics, including recall, accuracy, precision and F1-measure, are used to determine how successful the executed methodology was.The created methodology achieves 98% recall as well as F-measure rate, whereas the developed methodology achieves 96% precision score.The developed technique achieves 97% accuracy.According to simulation results, the suggested methodology was more accurate at identifying pneumonia and lung cancer than the ELM, SVM and MLP models that are already in use.
In 2023, Chen et al. [22] proposed a methodology that extended VLP to 3D medical images without depending on associated textual descriptions called GTGM.LLM are specifically used by GTGM to produce medical-style text from 3D medical images.The learning of 3D visual representations was next supervised using this synthetic text.In order to foster consistent visual descriptions amongst enhanced 3D medical image patches, a negative-free contrastive learning goal technique was also implemented, which successfully reduces the biases linked with strict positivenegative sample pairings.The GTGM was tested on 13 datasets from MRI, CT and EM, three imaging modalities.As a result of facilitating VLP expansion into 3D medical imagery without the requirement for associated text, GTGM's exceptional efficiency across a variety of medical image segmentation tasks highlights its efficacy and adaptability.
In 2023, Qin et al. [23] extensively investigated the knowledge transferability associated with pre-trained VLMs to the medical area, in which it was demonstrated that the key to eliciting information from pre-trained VLMs is to use well-modelled medical cues.It was shown that the VLM could transport the knowledge across domains and enhance its generalization by suggesting with expressive features that were shared across domains.The ability to recognize novel things with fewer or no image specimens was provided by this process.In addition, three methods were studied for automatic development of medical cues that may use expert level medical knowledge and imagespecific data for fine-grained grounding in order to bypass the time-consuming human modelling procedure.We undertake comprehensive studies on thirteen distinct medical datasets spanning a variety of modalities, demonstrating that the carefully crafted cues significantly outperform the standardized cues in terms of zero-shot effectiveness.
In 2023, Wan et al. [24] introduced a revolutionary system called Med-UniC, which aims to combine multimodal medical information from the two major common languages, Spanish and English.To particularly unify cross-lingual semantic descriptions of medical reports coming from several language groups, CTR was developed.By making the optimization aim independent of negative specimens, latent language disentanglement was able to maximize CTR while greatly reducing the bias that results from identifying positive-negative specimen pairings among comparable medical reports.Additionally, it makes sure that no particular language community was overrepresented in the cross-lingual description.Med-UniC provides a flexible methodology for combining multi-modal medical information across many language groups.It achieves exceptional effectiveness across five medical imaging tasks and ten datasets covering more than thirty disorders.The results linked with the simulation demonstrate the existence of community bias in cross-lingual VLP.The effectiveness of both uni-modal visual tasks and vision-language was improved by minimizing this bias.

Proposed methodology
The proposed lung cancer classification model includes various phases such as data gathering, pre-processing, segmentation, feature extraction and classification.The common benchmark dataset is where the data is initially acquired.Additionally, the acquired images are pre-processed utilizing the Gabor filter technique.These pre-processed images are segmented via the MEM algorithm technique.The characteristics are then retrieved from these segmented images employing the HOG technique.The IGNN is used to classify lung cancer, and the GAO method is used to optimize the parameters of the GNN in order to maximize accuracy as the main objective function.The final output of this IGNN categorizes lung cancer as normal, adenocarcinoma and squamous cell carcinoma.The overall model of the developed lung cancer classification is shown in Figure 1.

Data collection
The LC25000 Lung and Colon Histopathological Image collection, which includes 5000 images in each among the three groups of benign (normal cells), squamous and adenocarcinoma carcinoma cells (both malignant cells), is where the data is taken from.The dataset has been verified and is HIPAA appropriate.By rotating left as well as right and flipping horizontally and vertically, augmentation is accomplished.

Pre-processing
The pre-processing for the lung cancer classification model is accomplished by the Gabor filter method.Owing to its localization in the spatial as well as spatial frequency domains, it may be characterized as a Gaussian modulated sinusoid.A 2-D sinusoidal plane wave having orientation  as well as radial center frequency  modulated by a Gaussian envelope containing standard deviations  and  correspondingly along the  and  axes has the real impulse response provided by: Here,  =     +     and  = −    +     .The Gabor filter should be described as two symmetrically spaced Gaussians in the relevant spatial-frequency domain as shown.
,  =  −2  −   +   +  −2  +   +   . ( After establishing the frequency cut-off as -6 db and the frequency and orientation bandwidths  ,  to constant values corresponding to psycho visual information, the Gaussian envelope unknowns  and  may be calculated as in Eq (3).The optimum distance among the radial frequencies is specifically one octave.The octave-specified frequency bandwidth (i.e., the distance among  and its double  ) grows in a logarithmic manner as indicated by .This was motivated by research that revealed basic cells in the visual cortex had a frequency bandwidth of around one octave.A circular Gaussian with equal spatial coverage in all directions and a 45° orientation bandwidth was selected for this experiment by setting  =  .
The influence of aliasing should be lessened, and data should be properly captured with precise filter characteristic tuning.This is accomplished by choosing the filter location  ,  and bandwidth  ,  appropriately.Additionally, it is important to ensure that the channel filters' core frequencies are near to the frequencies that are typical in order to avoid the filter response falling off too quickly.

Overall, 𝐵 ×
, in which  shows the image width and  shows the count of orientation separations, gives the quantity of dyadic Gabor filter banks needed.
The process of employing a number of banks is similar to how the scaling function and mother wavelet are scaled and converted for the development.The basis created by the Gabor function is nonorthogonal and results in redundant decompositions, yet it is nonetheless accepted as an acceptable wavelet.Additionally, the count of needed frequencies for locating the centers must be set beforehand on the basis of the size of the converted image, in a manner identical to determining the count of decomposition stages for the wavelet packets.
The energy  ,  = 1, 2, in which  and  represents the size of the subband intensity  ,  , shows the traditional approach for extracting the Gabor filter signature.

Segmentation
The segmentation is performed by the Modified Expectation Maximization (MEM) [25] method for the developed lung cancer classification model.The weights used in the MEM method determine how closely the estimated value is connected to each segment of the separated component.A replicated matrix is used to compare single pixel to another in order to calculate the concealed information.In comparison to various procedures, the segmentation outcome is therefore more appropriate.
Stages in the MEM algorithm are as follows: Initialize the settings for the expectation and maximization steps first.The maximization will next be increased by one stage.Concatenating the supplied image represents the following stage.Swap out the matrices for expectation and maximization.Find the differences between the input image and the duplicated image.Determine the novel expectation and stage location for maximization.Reiterate the expectation and stage value maximization.

Feature extraction
The features are gathered from the segmented images of the developed lung cancer classification model utilizing the HOG scheme.In order to extract typical features from the segmented images, the HOG scheme is investigated.The image is initially partitioned into cells for the purpose of implementing HOG, after which the HOG orientations for the pixels inside the cells are generated.To describe the image description, the resultant histograms are next concatenated.Nevertheless, local histograms are contrast normalized by measuring an intensity metric across a wider portion of the image known as a block, which improves the effectiveness of the descriptor.The block's cells are next completely normalized based on their intensity values, creating a description that is more resistant to variations in lighting and shadows.To measure HOG, there exists four main phases.
The input image  is filtered through the kernels in both the horizontal and vertical directions employing the 1D derivative mask.
Here, a transposition vector • is used.The convolution process is next used to derive the  and  derivatives of the grayscale or silhouette image  as The below formulas are used to calculate the gradient of 's magnitude as well as direction, accordingly: A process known as orientation binning, which requires the production of cell histograms, represents the second stage in the calculation of HOG.The histogram channels for HOG are either signed or unsigned, and the HOG cells are rectangular (or circular in certain genuine applications).Unsigned channels in a histogram are scattered over a range of 0 to 360 degrees, whereas signed channels are distributed over a range of 0 to 180 degrees.
Descriptor block formation represents the third phase in the HOG computing process.This process is used to take variations in lighting and contrast into consideration.Half the size of a block's worth of blocks are positioned on top of one another.
Block normalization represents the last stage of the HOG calculation.One among the below norms serves as the normalization factor for a nonnormalized vector  that includes the histogram in a particular block: Here,  shows a constant whose value has no bearing on the outcome.Here, the size, shape and position of the lung cancer is mostly considered as extracted features for the classification of lung cancer.

Classification by IGNN
The classification of the developed lung cancer model is accomplished by the IGNN model.Here, the parameters of GNN are tuned by green anaconda optimization (GAO) in order to enhance the classification performance as well as to derive the fitness function.Adjacency matrix , the Laplacian , or their normalized variants are examples of GNN  ∈  × that illustrate the sparsity as well as connectedness of a graph .The GNNs are symmetric because we limit the graph to being unweighted and undirected.Signals may be moved and averaged throughout their neighborhoods employing GNN.The graph signal's single time shift is denoted by  , and its value at node . Utilizing polynomials of GNN and the filter weights  = { in the graph convolutional filter,   having  + 1 taps is created.
Here,   =   ,   , ⋯ ,   signifies the weight of local data following  -hop information exchanges and describes a shift-invariant graph filter having  + 1 taps.
Layers of graph convolutional filters are accompanied by nonlinearities in a generic GNN.The input graph signal having  features is denoted by matrix , where  =  , ⋯ ,  , and the feature propagation in a single GNN layer is represented as Here,  shows the count of output features,  ∈  × shows the collection of the  entries of the  ×  learnable graph filters, and  signifies the pointwise nonlinear activation function.

GAO Algorithm
The GAO is selected to perform the parameter tuning in GNN with the intention of enhancing the performance of classification thereby attaining the appropriate objective function.The GAO mimics green anaconda behavior in its native state.The process used by male species to locate female species during the breeding season as well as the green anaconda's hunting method served as the foundation for GAO.On the basis of the simulation of these two green anaconda strategy stages of exploration as well as exploitation, GAO's analytical modelling is provided.
A population-oriented metaheuristic algorithm, the developed GAO includes green anacondas as members of its population.Equation ( 13) is used to randomly create every green anaconda's starting location within the search space before the method is run.12) Here,  shows the GAO population matrix,  represents the  green anaconda (possible solution),  , represents the search space's thickness in  dimension (decision variable), number of green anacondas are referred to as , decision variables as , random numbers as  , and lower and upper bounds of the  decision variable as  and  , correspondingly.The objective function of the issue may be assessed in accordance with the recommended values associated with every green anaconda for the choice variables.According to Eq (14), a vector may be used to mathematically describe this collection of determined values for the goal function.
Here,  shows the computed objective function on the basis of the  green anaconda and shows the vector related to the computed objective function.Equation ( 15) is used to identify the group of possible female species for every green anaconda.
Here,  shows the green anaconda row count in the GAO population matrix and the position count of the respective element in the objective function vector that involves a better objective function value than the  green anaconda, and  shows the group of candidate females' positions for the  green anaconda.
The motion of green anacondas is significantly influenced by the pheromone content.The values of the goal function have been utilized to replicate the pheromone intensity.As a result, the likelihood that a member will be chosen by the green anaconda increases with the value of that member's goal function.
Here,  shows the vector related to the group of objective function values of candidate females for the  green anaconda,  shows its  value,  shows its maximum value, and  shows the count of candidate females for the  green anaconda. shows the probability of the pheromone intensity related to the  female for the  green anaconda.Equation ( 17) is to replicate this selection procedure.Equation ( 18) next determines the chosen female species for the green anaconda.
Here,  represents the chosen female for the  green anaconda,  shows the cumulative probability function associated with the  candidate female for the  green anaconda and  shows a random integer having a normal distribution between [0, 1].Equation ( 19) is used to determine a random location after finding the chosen female species.Equation (20), states that the location of the matching green anaconda is modified to the novel location if the goal function's value is better in the novel place; else, it will stay in the old location.
shows the count of green anacondas,  shows the count of decision variables,  shows the  dimension of the chosen female for the  green anaconda,  , represents random numbers from the group {1,2 ,  shows the novel recommended location of the  green anaconda on the basis of the initial stage of GAO,  , shows its  dimension,  shows its objective function value and  , represents the random numbers having a normal distribution in [0, 1] range.Utilizing Eq (21), a random location is initially created next to every green anaconda in relation to the approaching prey.Equation (22), on the other hand, states that it is okay to alter the position of the green anaconda if the goal function's value is enhanced at this novel place. , =  , + 1 − 2 ,  = 1,2, ⋯ , ,  = 1, 2, ⋯ , ,   = 1, 2, ⋯ , .
Here,  shows the algorithm's iteration counter,  shows the maximum count of algorithm iterations,  shows the novel proposed location of the  green anaconda on the basis of the second stage of GAO,  , shows its  dimension,  shows its objective function value, and so forth.In Algorithm 1, many GAO stages are shown as pseudocode.

Experimental setup
The performance rate associated with the introduced strategy may be computed in the results part.Various parameters have been considered for the evaluation of the developed methodology against state-of-the-art works such as MLP [26], SVM [27], ALBAE [21], and ELM [28].Also, the segmentation algorithm MEM has been compared with conventional techniques such as SV-net [29], self-supervised DL method [30], DREFC [17], IGJO [2], MDE [4], mRSA [6], COVIDOA+HHOA [8], and LCWOA [10].Better performance has been returned with the proposed model than the other considered works across several iterations.The population size is considered as 10, chromosome length is considered as 1 and the number of iterations is considered as 100.These are the specific parameter settings and optimization strategies utilized for the GAO and IGNN algorithm.The performance of the developed methodology is superior to the other considered state-of-the-art works with respect to different measures.

Accuracy analysis
Figure 2 and Table 1 show that the suggested IGNN-GAO technique is more accurate than conventional networks like MLP, SVM, ALBAE and ELM.The suggested methodology has a greater accuracy rate when compared to the conventional techniques.As a result, the recommended strategy functions better than conventional techniques.Accuracy in lung cancer categorization refers to the percentage of correctly predicted cases within the input data set.The accuracy formula can be specified as below: In the above equation, the terms true positive, true negative, false positive, and false negative are shown by , ,  and , respectively.Nevertheless, a system's capacity is determined by its level of accuracy.The count of true and false positives and true and false negatives is used to calculate the accuracy rate.The suggested method's accuracy rate is 99.2%.It suggests that the technology would identify a specific illness more accurately than the conventional techniques.

Specificity analysis
Figure 3 and Table 2 provide comparison and specificity analysis examples.The suggested IGNN-GAO approach achieves great specificity when compared to remaining conventional approaches.The formula for specificity is given as below: The suggested method's specificity rate may be evaluated in comparison to several existing methods.The introduced methodology has achieved 93.8% specificity.It shows that the algorithm may categorize lung cancer on the basis of the images that are provided.

Scalability analysis
A system's capacity to scale up or down may be determined by measuring how its effectiveness varies as input size and the number of processors increase.The increase in input size may be caused by one of three objectives: retaining a certain execution time, employing all available memory or handling a predetermined effectiveness.The Figure 4 and Table 3 below exhibit the scalability study of the introduced IGNN-GAO and the cutting-edge techniques.

Run time analysis
Theoretically, run-time analysis predicts and forecasts how much an algorithm's running time (or execution time) will rise as the size of its input increases.Figure 5 and Table 4 serve as comparison and runtime analysis examples.The suggested IGNN-GAO approach achieves minimal runtime compared to remaining conventional methods.

Segmentation analysis
The segmentation analysis of different algorithms in terms of their respective accuracy is shown in the Table 5 below.It can be seen clearly that the accuracy of segmentation of the MEM algorithm is better than the other state-of-the-art segmentation methods, being considered respectively.Hence, it can be demonstrated that the proposed MEM method of segmentation is better in segmenting the lung cancer than the other methods, respectively.Table 5. Segmentation analysis.

Conclusions
To expedite the critical process of identifying lung cancer and reduce the stress on pathologists, deep learning algorithms were used.Examining cancer histology slides has improved the efficiency of these techniques.The common benchmark dataset was where the data were initially acquired.Additionally, the acquired images were pre-processed using the Gabor filter technique.These preprocessed images were segmented via the MEM algorithm technique.The characteristics were then retrieved from these segmented images using the HOG technique.The IGNN was used to classify lung cancer, and the GAO method was used to optimize the parameters of the GNN in order to maximize accuracy as the main objective function.The final output of this IGNN categorized lung cancer as normal, adenocarcinoma and squamous cell carcinoma.The simulation results showed that the new approach was superior to existing methods when compared to several performance measures.In the future, the proposed model can be extended to various concerns such as prediction of lung cancer at an early stage and it can be determined using some enhanced form of hybrid deep learning algorithms being tuned by some form of novel nature-inspired metaheuristic optimization algorithms.

Use of AI tools declaration
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

Figure 1 .
Figure 1.Overall model of the developed lung cancer classification.

Table 4 .
Run time analysis.