Abstract
For effective treatment options and better patient outcomes, early and accurate diagnosis of brain tumors is essential. This research introduces an innovative strategy to improving brain tumor diagnosis accuracy by combining deep learning with traditional machine learning classifiers. This research investigation employs the Xception Convolutional Neural Network (CNN) through a transfer learning approach as a feature extractor via two distinct strategies: (1) pruning the CNN’s classification layers while freezing the remaining layers, and (2) utilizing feature extraction with all CNN layers frozen. The extracted features are subsequently classified utilizing five traditional classifiers: Support Vector Machine (SVM), Decision Tree (DT), K-Nearest Neighbors (KNN), Random Forest (RF), and Logistic Regression (LR). The suggested approaches are assessed using the BR35H: Brain Tumor Detection 2020 dataset, which is publicly accessible on Kaggle and includes a thorough collection of labeled MRI scans of the brain for both training and testing purposes. Results show that the hybrid models achieve exceptional performance, with both transfer learning-based strategies providing highly accurate tumor classification. Specifically, the Xception model with frozen CNN layers and feature extraction yielded testing accuracies of 0.9900 for Logistic Regression (LR) and 0.9850 for K-Nearest Neighbors (KNN). In comparison, pruning the CNN layers and freezing the remaining layers also resulted in comparable high performance, with testing accuracies of 0.9883 for KNN and 0.9900 for Logistic Regression (LR). According to these results, brain tumor diagnosis may be made much more efficient and accurate by combining deep learning feature extraction with standard machine learning classifiers.
Introduction
With high rates of death and morbidity across the globe, brain tumours rank among the most serious health issues facing contemporary medicine1,2. Both benign and malignant forms of these aberrant cell proliferations in the brain are possible, but the latter are particularly dangerous because of their quick development and ability to infiltrate healthy tissues3. Recent research indicates that primary brain and central nervous system (CNS) tumours are becoming more common worldwide, which emphasises the significance of creating effective and precise diagnostic methods4,5. Magnetic Resonance Imaging (MRI) is often regarded as the most effective technique for identifying brain tumours due to its remarkable contrast resolution and ability to detect complex anatomical characteristics. However, accurately interpreting MRI images is a difficult process that greatly depends on the knowledge of radiologists. Manual diagnosis can be laborious and prone to mistakes due to inter-observer variability, visual fatigue, and human perception limits, particularly when differentiating distinct tumours that share similar visual features. These difficulties emphasise the necessity of automated, impartial, and repeatable diagnostic instruments.
Artificial intelligence (AI), especially machine learning and deep learning, has transformed the discipline of medical imaging. Of the several deep neural network models, Convolutional Neural Networks (CNNs) stand out for their handling of visual input. Convolutional neural networks are superior to conventional, hand-crafted feature extraction techniques in many image classification applications because they are able to learn hierarchical features from unprocessed image data6,7. On the other hand, especially when applied to well specified feature sets, classic machine learning classifiers—such as Support Vector Machines (SVM), Random Forests, and Logistic Regression—are acknowledged for their clarity and dependability. Improving the accuracy, efficiency, and generalisability of automated brain tumour detection systems by means of the combination of CNN feature learning skills with standard ML classifier decision-making strengths offers a convincing strategy.
The development of artificial intelligence (AI) has transformed the medical imaging industry especially in machin Although CNNs have shown progress in the field of medical image classification, their direct use for brain tumour diagnosis presents some difficulties. First of all, privacy concerns and the difficulties finding captioned medical photographs sometimes limit the expansion of medical databases8,9. To avoid overfitting and guarantee adequate generalisation across different tumour types and imaging settings, convolutional neural networks often require large quantities of labelled data10,11. Moreover, in clinical environments, where reasonable choices are crucial for developing confidence and enabling acceptance, CNN models’ clarity remains a major concern.
Though typically more interpretable and less keen on data, traditional machine learning classifiers cannot independently extract spatial and hierarchical imagine characteristics. They depend on either externally derived characteristics or manually created ones, which might not be sufficient to capture the intricate patterns shown in brain tumour photos. Thus, in performance and robustness, a solely CNN-based or ML-based method might be inadequate. A hybrid model combining the best of both worlds—deep CNN-based feature extraction and conventional ML-based classification—is needed to solve these constraints.
This work proposes a hybrid framework combining the Xception Convolutional Neural Network’s (CNN) strong feature extraction capabilities as the basic model with a method based on transfer learning and various traditional ML classifiers in order to enable reliable and exact identification of brain tumours and feature extraction. The Xception architecture is selected because of its use of depthwise separable convolutions, which enable effective computation and capture of rich and varied characteristics from medical images—a necessary ability when handling the complexity of MRI scans. Starting with transfer learning and using the pretrained Xception model first trained on the vast ImageNet collection, the proposed approach All convolutional layers of the model are maintained in a fixed state in order to maintain the integrity of the obtained visual characteristics and remove the need for considerable retraining. Eliminated are the higher classification layers, designed for the first training job. Instead, concentrate on preserving and freezing an intermediary layer, identified for its capacity to encode high-level, abstract information very useful for medical imaging analysis. Each MRI scan is encapsulated in the deep feature vector derived from this layer.
Combining knowledge and deep learning. Among the several kinds of deep neural network models, CNNs stand out for their visual input handling ability. Convolutional neural networks have a leg up in many image categorisation activities as they can learn hierarchical features from raw picture data, unlike conventional, manually-crafted feature extraction techniques6,7. On the other hand, especially when used on well specified feature sets, classic machine learning classifiers—such as Support Vector Machines (SVM), Random Forests, and Logistic Regression—are acknowledged for their clarity and dependability. Improving the accuracy, efficiency, and generalisability of automated brain tumour detection systems by means of the combination of CNN feature learning skills with standard ML classifier decision-making strengths offers a convincing path.
These CNN-derived features are then used as inputs for a suite of classical ML classifiers. The selected classifiers—Support Vector Machine (SVM), K-Nearest Neighbors (KNN), Decision Tree, Random Forest, and Logistic Regression—are each trained to distinguish between different types of brain tumors using the extracted feature vectors. The proposed framework combines the deep feature learning capabilities of CNNs with the structured decision-making power of ML classifiers, resulting in several advantages. This approach captures essential spatial abstractions vital for medical diagnosis, boosts generalization and robustness across various datasets, and enhances model interpretability, thereby making the system more adaptable and reliable for clinical applications.
The following contributions are included in this study:
-
Utilizing Xception Convolutional Neural Networks (CNNs), with all convolutional layers frozen enables the model to leverage deep, hierarchical features learned from large-scale datasets like ImageNet. This preserves general visual representations that are effective for various image classification tasks.
-
Trimming the fully connected classification layers of the CNN and feeding the output into conventional machine learning classifiers like Support Vector Machines and Random Forests enhances model generalization. These classifiers typically demonstrate superior performance with high-dimensional feature vectors and exhibit a reduced tendency to overfit, particularly in scenarios where training data is scarce.
-
Freezing all CNN layers and removing the dense classification head reduces computational load, resulting in a lighter and more resource-efficient model.
The structure of this paper is outlined as follows. Section “Related work” examines current studies on brain tumor detection, emphasizing the difficulties that motivate the suggested hybrid approach. Section “Materials and visualization” outlines the datasets and visualizations employed, offering valuable insights into the characteristics of brain tumors. Section “Dataset description” details the image pre-processing techniques utilized to ready the data for the model. The fifth section delineates the hybrid methodology that integrates CNN feature extraction with various machine learning classifiers. Section “Augmentation and data generator” outlines the criteria for evaluating the performance of the model. Section Normalization step details the experimental findings, juxtaposing the proposed method against current methodologies. Section “Methodology adapted” presents the findings, examining the strengths, weaknesses, and practical implications. In conclusion, Sect. “Xception model as a base model” provides an outline of the study’s contributions and offers recommendations for future inquiries.
Related work
Khan et al.12 introduced a multimodal classification approach that effectively combines deep learning models with robust feature selection strategies to enhance brain tumor detection. Utilizing a combination of MRI modalities such as T1, T2, and FLAIR, the study harnessed the strengths of convolutional neural networks (CNNs) in learning spatial hierarchies of features from medical images. They supplemented these with a filter-based feature ranking approach, which selected the most discriminative features across modalities, ultimately boosting classification performance. Their model outperformed traditional ML techniques, demonstrating that integrating multimodal data not only enriches the input space but also significantly improves diagnostic accuracy. This framework supports radiologists by providing highly discriminative features for tumor identification.
Using explainable artificial intelligence techniques, Rasool and colleagues developed CNN-TumorNet, a deep convolutional neural network architecture13, to get accurate brain tumour classification. By including Grad-CAM into the model, the scientists enabled doctors to view how the model categorised photos; this helped them ascertain which areas of the brain images were most significant to the model’s results. In a healthcare context, where interpretability is extremely helpful, openness and faith in artificial intelligence systems are very vital. Reaching state-of- the-art accuracy in tumour type classification, the network passed all of the strict assessments using enormous MRI datasets. The paper emphasises the need of medical artificial intelligence solutions being both efficient and understandable.
Asiri et al. examined in their work14 CNN hyperparameter tuning’s impact on classification accuracy in brain tumour detection challenges. To maximise important model parameters like kernel size, number of layers, dropout rate, and activation functions, they investigated many tuning techniques including grid search and random search. Their tests revealed that optimal hyperparameters produce notable generalisation across datasets and performance gains. The work shows how a well-calibrated CNN may preserve accuracy on unseen data, therefore addressing one of the fundamental difficulties in DL—overfitting. This paper offers practitioners hoping to create strong CNN architectures for clinical use a comprehensive road map.
Sarkar et al.15 presented a hybrid paradigm for brain tumour categorisation. Deep features obtained from a pretrained convolutional neural network (CNN) named AlexNet form the basis of this approach. These characteristics subsequently feed SVM, k-NN, and random forest among other machine learning classifiers. This combination’s concept is to use CNN’s enhanced feature representation capabilities as well as the generalisability of conventional ML techniques. Their results lead support vector machines (SVMs) to be the most successful classifiers for different types of tumours. This work shows the feasibility of merging deep feature embeddings with lightweight ML classifiers in order to build efficient and cheap diagnostic tools.
Mathivanan et al. explored the utility of transfer learning in brain tumor detection using limited labeled MRI data16. By employing pre-trained CNN models like VGG16 and InceptionV3 and fine-tuning them on brain tumor datasets, they effectively addressed the common issue of data scarcity in medical imaging. The transfer learning models not only reduced training time but also enhanced classification performance due to their rich pre-learned features from large-scale image datasets. The study presents transfer learning as a highly practical and resource-efficient approach for medical applications where high-quality labeled data are difficult to obtain.
Nahiduzzaman and colleagues17 proposed a hybrid and explainable model for MRI-based brain tumor classification, combining convolutional neural networks with ensemble machine learning techniques. They incorporated SHAP (SHapley Additive exPlanations) and LIME (Local Interpretable Model-agnostic Explanations) to provide insight into the model’s predictions. Their system was validated on multiple datasets and outperformed baseline methods in accuracy and interpretability. The integration of explainable AI ensures that clinicians can trace back the AI’s decisions, adding a layer of reliability necessary for real-world deployment.
This work18 by Bhimavarapu et al. created a two-phase brain tumour analysis pipeline comprising an initial segmentation stage using enhanced unsupervised clustering, then classification using a supervised ML model. By lowering non-informative background data, the unsupervised step separated tumour areas more efficiently, thereby improving the accuracy of the next classification job. Their approach showed good results in multiclass tumour classification and proved that exact presegmentation may greatly improve diagnosis results.
To improve MRI-based brain tumour classification, Mohanty et al. presented an original CNN model enhanced with soft attention mechanisms19. Dynamic emphasis of the attention layers on tumor-relevant areas of the input pictures enables the network to provide top priority for significant spatial information during training. Higher accuracy and more strong model predictions resulted from this approach addressing problems including background clutter and intra-class volatility. The soft attention component further enhanced interpretability by pointing out which picture portions were most important for categorisation.
Agarwal et al.20 developed a unique CNN architecture with convolutional and pooling operations especially for MRI scan-based brain tumour identification. They brought optimisation in layer connection and filter use, thereby enhancing feature abstraction capacity and lowering overfitting. On benchmark datasets, its design was assessed for good sensitivity and specificity in tumour area identification. Emphasising the need of model modification for clinical relevance, this study helps to create customised CNN architectures for medical imaging uses.
Basthikodi et al. concentrated on improving multiclass brain tumour diagnosis by means of a hybrid approach combining sophisticated feature extraction techniques21 with Support Vector Machine (SVM) classifiers. To raise MRI image quality, their work built a strong preprocessing pipeline comprising skull stripping, noise filtering, and contrast enhancement. Gray-Level Co-occurrence Matrix (GLCM) and Local Binary Pattern (LBP) were used for feature extraction to grab picture texture and spatial patterns. These characteristics loaded into an SVM for classification produced better accuracy over several tumour types. The methodological rigidity of the study and its focus on handmade elements underline its relevance to effective and interpretable ML-driven diagnostics.
Using MRI data, Saeedi and colleagues22 carried a comparison research combining convolutional neural networks (CNNs) with conventional machine learning models to identify brain tumours. Beginning with data augmentation and normalisation, they used a multi-stage pipeline to feature learn using a CNN then classify using models like SVM, Decision Trees, and k-NN. Their findings showed that while hybrid models achieved the optimal balance between accuracy and computational economy, CNNs shone in capturing spatial hierarchies. This method emphasises, especially in clinical settings with limited resources, the pragmatic advantage of combining DL and ML techniques.
Emphasising the importance of model tweaking for clinical use, Sadr et al. developed a deep learning model for precise brain cancer classification from MRI data. Designed23 as a convolutional neural network (CNN) trained using enhanced datasets in order to solve class imbalance. Using layer-wise tuning in concert with batch normalisation and dropout, their approach prevented overfitting. Under tests on pituitary tumours, meningioma, and glioma, the model outperformed the human one. By assessing the model over several datasets, they also demonstrated its durability and generalisability. This approach considerably increases the validity of DL models in medical diagnosis.
In their work24, Wageh et al. offered a complete system including traditional ML classifiers together with deep feature extraction and genetic algorithm-driven feature selection. Deep features were obtained using a pre-trained CNN and then optimised using a genetic algorithm to minimise feature dimensionality while maintaining discriminating ability. Using random forests and support vector machines among other machine learning models, these characteristics were categorised. While simultaneously reducing computing requirements, the approach shown notable improvements in classification accuracy. Using evolutionary algorithms for feature selection presents a creative way to increase the accuracy of diagnosis.
Kang and colleagues created an ensemble framework25 combining various CNN architectures and deep features collected from MRI images. Aggregated and supplied into ensemble classifiers including voting and gradient boosting classifiers, these elements Their approach underlined the complimentary character of deep features from many CNNs, which when aggregated greatly enhanced classification resilience. Showcasing its promise for real-world deployment where model consistency is crucial, the ensemble technique reduced model variation and improved diagnostic accuracy.
Using MRI images, Asiri et al. investigated a dual-network architecture26 integrating ResNet50 with U-Net to concurrently identify and categorise brain tumours. After localising the tumour area, the U-Net segmenter fed the fine-tunized ResNet50 for classification. The technique was evaluated for segmentation as well as classification using TCGA-LGG and TCIA datasets. Focussing categorisation on tumor-specific areas helped this integrated design not only raise classification accuracy but also enhance interpretability. The work distinguishes itself with its full-pipeline approach, which simplifies tumour detection and localisation.
Dixon and colleagues put up a hybrid learning architecture27 combining CNN-based learning with manual feature extraction. Their solution uses CNN-learned representations combined with radiomics-inspired elements, which are then classified using a fully connected neural network. The hybridisation sought to grab deep-learned patterns as well as domain knowledge. Results revealed improved tumour classification over several datasets, therefore stressing the synergy between traditional image analysis and deep learning methods in challenging diagnosis problems.
Examining interpretability in deep learning models for brain tumour diagnosis, Nhlapho et al. addressed the black-box question in artificial intelligence. On CNN architectures used on MRI data, they employed28 a set of interpretability methods including Grad-CAM and SHAP. The study assessed how various layers support final predictions and how doctors may profit from such understanding. The work significantly advances explainable artificial intelligence and opens the path for more honest and open medical artificial intelligence uses.
Semwal et al. developed a hybrid approach29 optimising a convolutional neural network with a support vector machine (CNN-SVM) by use of particle swarm optimisation (PSO). While the SVM supplied classification, the CNN managed feature extraction; PSO tuned hyperparameters for maximum accuracy. Their method improved tumour subtype classification and addressed overfitting and feature redundancy. A potential approach for precision diagnostics is provided by this new integration of evolutionary optimisation with hybrid modelling. To automatically detect brain tumours from radiological pictures, Natha and colleagues30 built a multi-model ensemble deep learning system. To increase resilience and lower model bias, they aggregated the results of many CNNs using a majority voting system. Tested over several datasets, the ensemble proved generalisability and low false positives. The capacity of the model to combine strengths of several architectures qualifies it as a trustworthy choice for implementation in varied clinical environments.
Based on the EfficientNet architecture tuned for speed and precision, Islam et al. presented BrainNet31, a brain tumour classification system. To improve convergence the model used sophisticated optimisation techniques like weight decay regularisation and learning rate warm-up. BrainNet exceeded traditional CNNs in accuracy and inference speed according evaluations using MRI datasets. Important for real-time clinical decision support systems, this study helps to create efficient artificial intelligence models. Remzan et al.32 presented a new ensemble learning-based feature extraction and categorisation system. They derived deep features using CNNs and used an ensemble system combining several ML classifiers. Their use of cross-validation driven feature selection—which guarantees just the most relevant features—was a major breakthrough. Emphasising the efficiency of ensemble approaches in reducing model variance and improving generalisation, their approach obtained good classification metrics over several datasets.
Raza and colleagues33 classified brain tumours from MRI images using transfer learning with DenseNet121. Changing the pre-trained model allowed the network to use the generic characteristics it had learnt before while acquiring domain-specific ones. The results confirmed that, particularly in cases with limited medical datasets, transfer learning has the ability to surpass models created from the ground up in general. The results stress the need of using pre-trained networks in contexts when annotated data availability is restricted.
Güler and Namlı34 investigated methods of classifier optimisation for DL-based brain tumour identification. They tweaked hyperparameters like learning rates, batch sizes, and dropout rates and evaluated many CNN designs. Training efficiency and classification accuracy were much raised by the optimisation approach. Their findings emphasises how crucial architectural choice and model tweaking are to get best performance in medical picture categorisation problems.
Liu et al. developed a CNN-based feature extraction approach35 coupled with traditional ML classifiers to classify MRI brain tumours. They found deep features extracted from a CNN following SVM, k-NN, and decision tree models. Their findings showed that hybrid DL-ML models might mix CNN’s high-level abstraction capacity with the interpretability and efficiency of traditional classifiers. This fusion model shows success in cases demanding both accuracy and explainability.
Malakouti et al.36 examined machine learning and transfer learning methods for brain tumour classification using MRI images. Combining pre-trained networks as ResNet and DenseNet with standard classifiers like SVM and XGBoost, they The models were evaluated both balanced and unbalanced using AUC and F1-score criteria. While traditional models offer interpretability and fast training, transfer learning performed very well on limited data. This dual point of view offers flexibility for different therapeutic needs.
Vimala et al.37 introduced a hybrid deep learning architecture integrating convolutional neural networks (CNNs) with recurrent neural networks (RNNs), thereby effectively collecting the spatial and temporal properties of MRI images particularly in long-term memory (LSTM) units. While the convolutional layers distinctly identified spatial properties across the image slices, the recurrent layers studied the evolution of spatial features over consecutive slices. Especially for infiltrative gliomas, the dual-level processing allows the model to identify between tumour grades with higher accuracy. The work reveals how effectively hybrid architectures enable to grasp intricate patterns in volumetric medical imaging data.
Based on an in-depth review of the existing literature, the following research gaps have been identified:
-
There is a lack of systematic evaluation of feature fusion and selection strategies that can best leverage CNN-extracted representations for traditional ML classifiers.
-
There is a need to enhance domain generalization and robustness in hybrid CNN-ML pipelines through techniques such as domain adaptation or dataset-agnostic feature learning.
-
Few studies utilize or analyse the explainability potential of combining CNN feature extraction with interpretable classifiers to assist clinical decision-making.
-
There is a need for standardized evaluation protocols and benchmark datasets to assess the robustness, efficiency, and scalability of hybrid CNN-ML frameworks.
-
There is a lack of focus on designing lightweight and fast CNN-ML models suitable for real-time use in resource-constrained healthcare settings.
In response to the identified research gaps, this study explored a strategy of a robust and interpretable hybrid framework that leverages convolutional neural network (CNN) features in combination with traditional machine learning (ML) classifiers for brain tumor detection. The study will systematically evaluate various feature fusion and selection strategies to determine the most effective methods for optimizing CNN-extracted features for ML classification. Table 1 is showing the literature details.
Materials and visualization
Dataset description
Specifically concentrating on the diagnosis of brain tumours using MRI (Magnetic Resonance imaging) images, the Br35H: Brain Tumour diagnosis Dataset is a freely accessible dataset carefully curated to promote research and development in medical image processing. In binary classification problems, where the objective is to differentiate MRI images with brain tumours from those free of them, this dataset is absolutely vital. There are 3,000 brain MRI scans overall, split into two categories: “yes”, or images showing a tumour; “no”, or images lacking a tumour [Figure 1; Table 2]. The collection offers routinely used image formats together with high-resolution T1-weighted MRI pictures. Usually greyscale, these photos may have RGB representations depending on the preprocessing techniques applied by contributors. Though their dimensions differ, the photos in the dataset are generally scaled to consistent forms like 128 × 128 or 224 × 224 pixels for model training needs. Images from the “yes” class clearly show one or more tumours with variable forms, sizes, locations, and intensities, therefore guiding models to identify various patterns of aberrant tissue development. Conversely, the “no” class consists of normal MRI images devoid of any obvious tumour, thereby offering a necessary baseline for differentiating healthy from sick brain tissue. Separate folders for these classes help to make the dataset easily available and useable for machine learning applications. Although the Br35H dataset is strong for classification, its applicability for exact tumour border identification is limited until reinforced with hand annotations as it lacks segmentation masks. Along with basic preprocessing tasks include picture resizing, pixel value normalizing—usually ranging from 0 to 255 to 0–1—and cautious train-test splits to prevent data leaking.
Figuring out the dataset into discrete subsets for the goals of training and testing is a regular and necessary habit in deep learning systems [Table 3]. This approach involves separating the input features—such as pre-processed MRI scans used as the model’s learning input—from their matching labels, which denote the expected classification outcomes (e.g., cancer or not). 80% of the data is usually set aside for the training set, which helps the model to be fit and helps to adjust its parameters using learning methods. The last 20% is kept aside as the test set, intended to evaluate the model’s capacity for objective generalisation to fresh, untested data.
This partitioning is performed using a randomized sampling mechanism to ensure that the data distribution remains statistically representative across both subsets, thereby minimizing sampling bias. To ensure reproducibility and consistency of the results across different runs of the experiment, a fixed random seed (random state) is specified. This ensures that the random splitting process yields the same data partitioning each time the code is executed. Accurate train-test splitting is essential for mitigating overfitting and underfitting, and for generating reliable performance metrics. Table 4 is showing characteristics of the Dataset Used in This Study.
Pre-processing step
Augmentation and data generator
The purpose of this section is to provide a computationally efficient and technically sound envision pre-processing and augmentation pipeline for use in medical image investigation, particularly MRI for the purpose of detecting brain tumors38. In order to do real-time picture augmentation and pre-processing while training, the implementation makes use of the ImageDataGenerator class from Keras. By subjecting the input photos to a sequence of controlled, random modifications, data augmentation creates the illusion of a larger and more diverse training dataset38. Particularly helpful in medical imaging, where expert annotation makes the acquisition of huge labeled datasets time-consuming and expensive. By mimicking a range of real-world circumstances, augmentation helps reduce the risk of overfitting, strengthens the model, and enhances its capability to generalize. During training, the pipeline makes sure the model sees a wide variety of enhanced picture samples, which improves its capacity to generalize to new data. Normalizing pixel values is the first step in the preparation phase. Since MRI images are typically stored with 8-bit intensity values in the range [0, 255], the normalization operation rescales pixel intensities to the range [0.0, 1.0], using the transformation:
This normalization step is critical for ensuring numerical stability and accelerating the convergence of deep neural networks by homogenizing the input distribution. Following normalization, various augmentation strategies are employed to synthetically increase the size and diversity of the training data. Random image rotations are permitted up to ± 90◦, modelled using the 2D rotation matrix:
where θ is a randomly sampled angle. This transformation simulates different anatomical orientations, improving the model’s rotational invariance.
To simulate perspective variations, shear transformations are applied using a shear factor λ. Horizontal shearing is given by the affine matrix:
Zoom augmentation is also implemented to mimic spatial scaling effects. This is represented mathematically by:
where \(\:{s}_{x}\) and \(\:{s}_{y}\) are randomly chosen scaling factors. This permits the model to learn from features at manifold resolutions.
Flipping operations further enhance spatial invariance. Horizontal and vertical flips are represented by the matrices:
Horizontal Flip:
Vertical Flip:
These augmentation techniques collectively help the model become invariant to positional and structural variations in the input images, which is particularly valuable for detecting tumors of varying shape, size, and location.
To manage the input pipeline efficiently, the model utilizes data generators. In Keras, a data generator is an iterator that loads and pre-processes image data in real-time, yielding it in mini-batches during training. This approach is memory-efficient, as it eliminates the need to load the entire dataset into memory at once. It also supports dynamic augmentation, meaning that each epoch may see a different version of the same image, further enriching the diversity of the training process.
For validation and testing, a separate data generator is instantiated with only the normalization operation, ensuring that evaluation metrics are computed on original, unaltered images. The parameter validation_split = 0.2 internally reserves 20% of the dataset for validation, facilitating stratified sampling without the need for manual directory restructuring.
Image loading is handled via the flow_from_directory() method, which streams images directly from disk in real time. The images are resized to 224 × 224 pixels, consistent with the input dimensions required by widely used convolutional neural network architectures. Although MRI data are typically grayscale, color_mode=’rgb’ ensures compatibility with models pretrained on RGB datasets like ImageNet by replicating single-channel data across three channels.
To avoid overfitting, images are randomly shuffled (shuffle = True) before each training epoch. The batch size is set to 32 for training and 16 for validation. The labels are processed in categorical mode, meaning they are one-hot encoded as follows:
This encoding format is compatible with the categorical cross-entropy loss function:
where C denotes the numeral of classes, \(\:{y}_{i}\:\)is the ground truth label, and \(\:\widehat{{y}_{i}}\:\)is the foreseen probability for class i.
Normalization step
Normalising numerical input values within a predefined range helps to increase the accuracy and efficiency of the learning process39, hence it is among the most crucial preprocessing actions in deep learning. Working with picture data, where pixel values are sometimes expressed as integers within a range of 0 to 255, this procedure is very crucial. With 8-bit representation per colour channel, standard picture formats like PNG or JPEG contain pixel intensities; each pixel takes values from 0 (totally dark) to 255 (entirely brilliant).
Normalisation allows one to execute this pixel value transition to a floating-point range of 0.0 to 1.0 for numerous purposes. It first increases ML model efficiency by ensuring homogeneity and standardise of all input properties. By rescaling pixel values, make models more effective and thus shorten training time and speed up convergence. Gradient-based optimisation methods such as gradient descent considerably helps from this similarity in order to reduce the loss function during model training. Appropriate scaling of the input data helps to stabilise the optimisation process, thereby preventing problems such exploding or disappearing gradients, which could arise from either too big or too small input values for the activation functions of deep neural networks39. Normalising also helps to guarantee improved numerical stability during training as it helps to avoid unduly affecting the model by differences between input characteristics. This method improves model performance, accelerates convergence, and raises prediction accuracy by translating pixel values from integers within the 0–255 range into a floating-point range of 0.0 to 1.0. Therefore, normalisation is very important in the pre-processing pipeline especially for high-dimensional and complicated data, including medical imaging, where exact and reliable model training is absolutely vital. Figure 2 is shows the working flow of pre-processing.
Why Normalize the Data?
-
Faster Convergence: Neural networks tend to converge faster when input data is normalized. When the pixel values are divided by 255, you transform the range of input data to a uniform scale between 0 and 1, which allows the model’s optimization algorithm (like gradient descent) to progress more smoothly. Large values can cause the optimization process to struggle, and smaller, more uniform values help avoid that.
-
Improved Numerical Stability: In deep learning models, particularly those involving activation functions like sigmoid or tanh, large input values can lead to gradients that are too small or too large, causing problems such as exploding or vanishing gradients. By normalizing the data, we mitigate this risk and make the optimization process more stable.
-
Consistent Input Across Features: When working with data that has different scales, such as images where pixel intensities might range from 0 to 255, it’s important to have consistency in the input range. By dividing by 255, we ensure that all pixel values fall in the same range, making the model more efficient and less likely to be biased toward certain features due to their range.
The normalization operation:
can be broken down as follows:
-
\(\:{X}_{train}\:\)represents the pixel values of the training images.
-
Each pixel value initially lies within the range [0, 255], where 0 represents the darkest possible pixel (black) and 255 represents the brightest (white).
-
Dividing the pixel values by 255.0 scales them to the range [0.0, 1.0], where 0.0 is black and 1.0 is white. Intermediate values represent varying intensities of gray in a grayscale image, or different intensities for each of the RGB channels in a color image.
In other words:
This operation applies to each pixel in every image in adapted dataset, effectively normalizing the entire image dataset.
Methodology adapted
Xception model as a base model
Xception is a deep convolutional neural network architecture [Figure 3] introduced by Google researchers, which leverages depthwise separable convolutions to improve efficiency and performance over traditional convolutional approaches40. It can be viewed as an extension and refinement of the Inception architecture. Inception modules are designed to capture multi-scale features by applying multiple convolutional operations (e.g., 1 × 1, 3 × 3, 5 × 5 convolutions) in parallel—these parallel branches, often referred to as towers, are then concatenated to form a rich feature representation. This architecture effectively factorizes convolutions spatially and across channels, enabling deeper and more efficient networks40. The Xception architecture interprets the Inception module as an intermediary between ordinary convolutions and depthwise separable convolutions, therefore stretching this notion to an extreme. Originally, a depthwise convolution that performs spatial filtering on each input channel independently, followed by a pointwise convolution (1 × 1 convolution) that combines the results of the depthwise step across channels in a linear manner41, divides a standard convolution into two distinct steps. While keeping expressive ability, this decomposition significantly reduces processing costs and the number of parameters. Depthwise separable convolutions in this sense can be seen as Inception modules with several towers, where each input channel is processed separately before being combined back together. Inspired by this knowledge, Xception presents a novel architecture whereby depthwise separable convolutions completely replaces Inception modules, producing a model that is not only more efficient but also clearly more effective over several large-scale image classification benchmarks. By means of architectural simplicity, Xception can get exceptional performance while preserving or lowering computing complexity relative to past models such as Inception-v341.
Xception (Extreme Inception) is a Convolutional Neural Network (CNN) architecture that rethinks convolutional blocks by fully decoupling spatial and channel-wise correlations—something traditional convolution layers entangle. This design is grounded in two principles:
-
Depthwise Separable Convolution – Fully Decoupling Representation Learning.
-
Residual Connections – Facilitating Deep Feature Learning.
Depthwise separable convolution - fully decoupling representation learning
In a standard 2D convolutional layer:
-
You apply a set of 3D kernels (height × width × input_channels) to the input tensor.
-
Each filter interacts with all channels simultaneously and learns both spatial features (patterns in height × width) and inter-channel dependencies in one operation.
-
For an input of shape (H, W, \(\:{C}_{in}\)) and \(\:{C}_{out}\) output channels, the computational cost is:
where K is the kernel size.
Xception’s Insight: François Chollet (creator of Xception) proposed that spatial and cross-channel correlations need not be learned jointly. Instead, they can be learned sequentially and independently—which is what depthwise separable convolutions do. Figure 4 shows the working of Pointwise and Depthwise Convolution.
-
1.
Step 1: Depthwise Convolution.
-
Applies one filter per input channel, not across all channels.
-
Each filter is responsible only for spatial feature extraction.
-
For \(\:{C}_{in}\) channels, there are \(\:{C}_{in}\) separate spatial filters.
The computational cost is:
-
2.
Step 2: Pointwise Convolution (1 × 1 Convolution).
-
Applies a 1 × 1 convolution across the depth (channels).
-
Learns inter-channel correlations by linearly combining the channel-wise outputs.
-
Number of filters = output channels \(\:{C}_{out}\).
The computational cost is:
-
3.
Total Cost of Depthwise Separable Convolution
This is significantly cheaper than standard convolution, especially when K = 3 and 541.
\(\:{C}_{in}\) is large.
Residual connections – facilitating deep feature learning
Residual connections were introduced in ResNet to tackle the degradation problem, where deeper models tend to perform worse due to optimization difficulties.
In the Xception architecture:
-
Each convolutional block is wrapped in a shortcut connection, defined as:
where F(x) is the output of a series of depthwise separable convolutions.
-
If the input and output dimensions do not match (e.g., due to strided convolution or channel expansion), a 1 × 1 projection is applied to the shortcut to align their shapes before the addition.
Integrated Xception model with transfer learning layers
In the proposed transfer learning architecture, a pretrained Xception model is integrated as a robust feature extractor. Xception, short for “Extreme Inception”, is a convolutional neural network trained on the ImageNet dataset, capable of extracting rich and semantically meaningful features from input images. The output of the Xception model yields a high-dimensional feature tensor of shape (7, 7, 2048), where each of the 2048 channels corresponds to an abstract feature map capturing spatial and contextual patterns.
The model architecture leverages depthwise separable convolutions, which drastically reduce computation while maintaining representational power. The total number of parameters in the Xception backbone is approximately 20, 861, 480. For transfer learning purposes, these parameters are typically frozen to preserve the pretrained feature representations and minimize overfitting on small target datasets.
After constructing the convolutional base, a 2D Global Average Pooling (GAP) layer is used to flatten the feature vector of length 2048 by reducing each 7 × 7 feature map to a single scalar value by average pooling. This method streamlines the design by omitting complex layers with many parameters, while also preserving the core of each feature map by activation summarization.
The GAP layer’s output is subsequently sent via a dense, fully linked layer with 64 hidden units to enable a lower-dimensional embedding. This layer’s number of trainable parameters is given by:
where the first term corresponds to the weight matrix and the second term accounts for the bias vector. This transformation enables more effective learning of task-specific features.
A Dropout layer is added after the dense layer to avoid overfitting. To improve the model’s generalizability without adding more parameters, this regularization strategy randomly deactivates certain neurons during training.
A second dense layer with two output units, representing binary classification, makes up the last step of the design. This layer produces class probabilities based on the assumption of a softmax activation function. In this stratum, there are a total of:
The end-to-end forward propagation of the model can be mathematically expressed as:
where F(x) denotes the output feature map from the Xception base, W1 and W2 are the weight matrices of the two dense layers, b1 and b2 are the corresponding bias vectors, and σ represents a non-linear activation function, typically ReLU.
In summary, this architecture exemplifies an effective use of transfer learning by combining a powerful, pretrained convolutional base with a lightweight, task-specific classifier [Figure 5]. It enables rapid convergence, reduced overfitting, and improved accuracy on limited datasets by leveraging the generality of features learned from large-scale data.
Feature extraction with a frozen all CNN layers
In this approach, a pre-trained Convolutional Neural Network (CNN) is adapted for feature extraction by freezing all of its layers. Let the CNN be parameterized by
where each \(\:{\theta\:}_{i}\) represents the learnable weights of layer i, and L is the total number of layers in the network.
During training, weight updates are typically governed by gradient descent:
where η is the learning rate and L is the loss function.
Freezing the layers implies:
This ensures that the pre-trained knowledge (typically learned from a large dataset such as ImageNet) is preserved and not overwritten during further training or inference. In implementation, this is achieved by marking all layers as non-trainable.
Next, instead of using the full CNN including the classification head—which generally consists of fully connected (dense) layers that map the learned features to a set of class probabilities—we discard this portion.
Mathematically, if the original CNN can be decomposed as:
where f (X) denotes the feature extractor (the convolutional base) and g(·) denotes the classifier, we retain only f (X) for our purpose.
In this specific case, the feature extractor is taken to be the portion of the CNN up to the layer at index − 4, giving:
where \(\:X\in\:{R}^{\text{n}\times\:\text{H}\times\:\text{W}\times\:\text{C}}\) represents a batch of n input images, each of spatial dimension H ×W with C color channels.
Passing the training and testing sets through f (X) yields high-dimensional feature tensors:
Here, h, w, d represent the spatial dimensions and depth (i.e., number of channels) of the extracted feature maps from the selected CNN layer.
Since most classical machine learning models (e.g., Support Vector Machines, Logistic Regression, Random Forests) expect 2D input matrices of shape \(\:{R}^{n\times\:m}\), the 3D feature maps for each image must be flattened. This transformation is applied as:
Thus, each image is now represented by a single feature vector of length:
capturing hierarchical spatial features learned by the CNN.s.
Pruning CNN classification layers and freezing remaining layer
In this method, a pre-trained CNN is adjusted by freezing the convolutional layers that remain after manually eliminating the classification layers. Then, the network is used as a fixed feature extractor. Reducing the size of input photos so they may be utilized in traditional ML models is the main objective. Layers for classification, which usually make up the final few of a convolutional neural network (CNN), include:
-
Global average pooling.
-
Fully connected (dense) layers.
-
Dropout layers.
-
Output layer (e.g., softmax or sigmoid).
The code sequentially removes the last four layers from the CNN. Mathematically, we can represent the original CNN as a function:
where \(\:f\left(X\right)\) is the convolutional base (feature extractor) and \(\:g\left(f\right(X\left)\right)\) is the classifier (fully connected layers). By removing the classification layers \(\:g\left(f\right(X\left)\right)\), we isolate \(\:f\left(X\right)\), which maps the input image X to high-dimensional feature representations.
Freezing a layer means preventing its weights from being updated during training. Mathematically, for each trainable parameter \(\:{\theta\:}_{i}\) in the model, we set:
where L(θ) is the loss function. This ensures:
-
The pre-trained weights \(\:{\theta\:}_{i}\) remain unchanged.
-
The model acts purely as a fixed transformation: \(\:X\to\:f\left(X\right)\)
Next, a new Keras Model is created that maps:
This model uses the same input layer as the original model but stops at the fourth-to-last layer (after the classification layers have been removed). The result is a truncated CNN that outputs a tensor of shape:
where:
-
n is the number of samples,
-
h, w are the spatial dimensions,
-
d is the number of feature channels.
Each input image \(\:X\left(i\right)\in\:{R}^{H\times\:W\times\:C}\) is passed through the truncated CNN, resulting in a feature map:
The output tensor contains local features (from convolutional filters) aggregated spatially. This transformation captures the semantic information learned during pre-training.
Each 3D feature map \(\:{T}^{\left(i\right)}\in\:{R}^{h\times\:w\times\:d}\) is flattened into a 1D vector:
This converts the full set into a 2D feature matrix:
Given input images \(\:X\in\:{R}^{n\times\:H\times\:W\times\:C}\), the pipeline transforms them as follows:
The final feature matrix can then be used directly for classification, clustering, or other downstream tasks.
Machine learning classifier
Support vector machine classifier
When it comes to classification, few supervised learning algorithms are as effective as Support Vector Machine (SVM). Finding the optimal boundary (a hyperplane) that divides data points into multiple groups is the fundamental premise of support vector machines (SVM). A machine learning model called support vector machine (SVM) will not pick a border at random but will instead look for the one that maximizes the distance, or the margin, across it and the nearest points of data in every category42. Because they specify the location and direction of the ideal border, these nearest points—called support vectors—are crucial. Support vector machines (SVMs) may keep running with non-linearly separable data by utilizing kernel functions42. This occurs when the data is elevated to a higher-dimensional space, making the possibility of a dividing line more realistic. When dealing with high-dimensional data, support vector machines (SVMs) find widespread use in bioinformatics, picture classification, and text classification.
The first approach uses a pre-trained CNN as a full feature extractor with all CNN layers frozen. A Support Vector Machine (SVM) classifier is trained using the output feature vectors. Since the CNN layers are frozen, no part of the feature extraction network is updated during training. The SVM then learns a hyperplane (or multiple, in multi-class settings) that best separates the classes in this fixed high-dimensional space. The decision boundary is defined mathematically as:
where x is the feature vector, w is the weight vector, and b is the bias term.
The second approach involves removing all but the convolutional base from the CNN, which includes all of its classification layers (including fully connected layers, softmax, and global average pooling). In addition to serving as a general-purpose feature extractor, these frozen convolutional layers are not changed during training. The last convolutional layer produces a tensor with many dimensions that includes both local and geographic data. The SVM classifier receives this tensor after it has been flattened into a one-dimensional vector. Despite their increased dimensionality, these feature vectors can keep richer spatial characteristics than the first technique and may be more effective for some applications. Once again, the best hyperplane for class separation is determined by training the SVM on these vectors that have been flattened.
Decision tree classifier
For classification problems, a supervised learning technique known as a Decision Tree Classifier can be employed. To work, it uses input features to recursively divide the feature space into subsets, forming a tree-like structure with class labels at the leaf nodes and decisions at the inside nodes43. A decision tree’s main purpose is to partition the data such that distinct classes are as far apart as possible, with the objective of creating progressively more homogenous branches with regard to the dependent variable.
The technique utilizes impurity metrics like Gini impurity or entropy (information gain based) to find the optimal feature and threshold for data splitting at each node. The Gini impurity at a given node t may be determined by:
and for entropy, it is
C is the number of classes, and \(\:{p}_{i}\) is the fraction of samples belonging to class i at that node. By iteratively building branches, the method finds the feature that reduces impurity the most and continues doing so until either a maximum depth is reached, a minimum number of samples per node is reached, or the node becomes pure.
The first approach involves using a Decision Tree Classifier after flattening all of the retrieved features. This feature space is iteratively divided into smaller, more homogenous parts by the decision tree using these fixed feature vectors. By utilizing metrics such as Gini impurity or entropy, the model chooses a feature and a threshold at each internal node of the tree to partition the data in a manner that optimizes the decrease of impurity.
In this second strategy, feature vectors are often higher-dimensional and richer in spatial semantics than those obtained from the fully frozen CNN. Because of this, the input to the decision tree may contain more detailed information about local structures in the image (like object parts, edges, or texture variations).
The Decision Tree Classifier then takes these rich, fixed features and learns a sequence of if-else rules that split the feature space to classify samples. The training goal is the same: to maximize class separation using a series of feature-based splits.
At each node, the decision tree selects a feature xj and a threshold θ such that the data is split as:
The best split is chosen to maximize information gain or minimize impurity. The decision path for each input leads to a leaf node corresponding to a predicted class.
K-Nearest neighbors classifier
When it comes to classification and regression, the K-Nearest Neighbors (K-NN) algorithm is your go-to non-parametric option. For classification purposes, it finds the k feature points that are geographically nearest to a test sample and uses the most frequent class among those k to predict the label for the test sample44. The K-NN model is unique in that it uses the stored training data and distances between points as its only inputs, rather than explicitly establishing a decision boundary.
Mathematically, for a test feature vector x, the predicted label \(\:\widehat{y}\) is:
where:
-
\(\:{N}_{k}\left(x\right)\) is the set of the k-nearest neighbours to x,
-
\(\:{y}_{i}\:\)are the labels of those neighbours.
In the first approach, a pre-trained CNN is used to extract deep feature vectors from images. The final output is typically a dense feature vector stored as part of the training dataset. During prediction, a test image is passed through the same CNN, producing a new feature vector. The Euclidean distance (or other metrics like cosine similarity) between the test vector and all training vectors is computed to find the k closest matches:
The class label is then determined by majority vote among the k nearest training vectors.
In this second approach, the classification layers of the CNN (global average pooling, fully connected layers, dropout, softmax) are removed, and only the convolutional base is retained. The key difference here is the richer feature representation. The output from the final convolutional layer is typically a 3D tensor,
which is flattened into a 1D vector:
These higher-dimensional vectors retain more spatial information and can be more discriminative. Once flattened, the K-NN classifier proceeds identically as in the first approach: compute distances, find the k-nearest, and assign the majority class.
Random forest classifier
Classification and regression hitches are well-suited to the ensemble learning approach known as a Random Forest. During training, it builds a network of decision trees, or a “forest”, and then uses the mean of those trees’ predictions to determine the output class45. It improves accuracy and avoids overfitting by averaging the output of numerous decision trees learned on different data and subsets of features.
This initial approach involves feeding an input picture into a convolutional neural network (CNN). The CNN produces a compact feature vector, which is then fed into a Random Forest classifier. The Random Forest classifier learns to differentiate between classes by building several decision trees. Through the use of bagging and feature sub-sampling, each tree is trained using a randomly selected portion of the training data and features. Using majority voting, we compile all of the trees’ predictions and use them to make a final class prediction for a test picture.
Mathematically, for a feature vector \(\:x\in\:{R}^{d}\), and trees \(\:{T}_{1},{T}_{2},{T}_{3},{T}_{4},{T}_{5},\dots\:\dots\:..,{T}_{M}\), the prediction is:
where.
-
M is the number of trees in the forest.
-
\(\:{T}_{i}\left(x\right)\:\)is the predicted class label from the i-th decision tree.
In the second approach, the output from the final convolutional layer is typically a 3D tensor representing spatial and channel-wise features. This tensor is flattened into a 1D feature vector (often higher-dimensional than in the first approach due to preserved spatial features):
These rich, high-dimensional vectors are then used as inputs to train the Random Forest classifier. Because these features preserve more local information (such as edges, textures, and spatial positions), the Random Forest can potentially perform better on tasks where spatial patterns are crucial.
Logistic regression classifier
Logistic Regression, a model that uses linear regression, is utilized for binary and multi-class classification tasks. It employs the logistic (sigmoid) or softmax (multi-class) functions to articulate the probability that an input vector is associated with a certain class46. The logistic regression function is employed for the classification of binary data.
Mathematically, for binary classification:
For multi-class classification with C classes:
where:
-
\(\:x\in\:{R}^{d}\) is the feature vector,
-
\(\:{w}_{c},{b}_{c}\) are the weight vector and bias term for class c,
-
\(\:\sigma\:\) is the sigmoid function.
The output of the model is the class with the highest predicted probability.
In the first approach, feature vectors are passed to the Logistic Regression classifier, which learns to linearly separate the classes in this feature space. Despite being a linear model, Logistic Regression often performs well on such CNN features because the convolutional network already projects images into a highly informative representation space.
Prediction is made using the softmax (or sigmoid for binary) function, and training is done by minimizing the cross-entropy loss.
The cross-entropy loss for multi-class classification using logistic regression is given by:
where:
-
\(\:{y}_{i,c}\in\:\left\{\text{0,1}\right\}\) is the ground truth indicator for sample i belonging to class c ,
-
N is the total number of samples,
-
\(\:P\left(y=c|{x}_{i}\right)\) is the predicted probability of class c given input \(\:{x}_{i}\).
In the second approach, feature vectors are then passed to the Logistic Regression classifier. The rest of the classification process is the same: the classifier tries to learn linear decision boundaries in this high-dimensional space to predict class probabilities via softmax.
Overall framework walkthrough
The proposed approach [Figure 6] combines deep neural network extraction of features with traditional machine learning classifiers to efficiently classify brain tumors using MRI data. The pipeline starts with a preliminary processing phase, during which brain MRI pictures are methodically retrieved from disk. Every picture is downsized to a standardized resolution to maintain uniformity in input dimensions for the deep neural network. These resized images are then transformed into NumPy arrays, facilitating efficient computational handling. Subsequently, each preprocessed image is appended to a dataset list, while the corresponding binary class label—indicating tumor presence (1) or absence (0)—is appended to a label list. This structured dataset serves as the foundation for feature learning and classification. Following preprocessing, images are passed into the Xception model, a convolutional neural network known for its use of depthwise separable convolutions, enabling efficient and scalable deep feature extraction. Two distinct strategies for feature extraction are employed in this framework:
-
In the first approach, the entire Xception model is retained in its pre-trained form, with all weights frozen. The model acts solely as a static feature extractor, and no backpropagation updates are performed on its layers. Deep feature vectors are obtained from the final output of the network, typically after the GlobalAveragePooling2D layer, resulting in compact yet semantically rich representations of the input images. These feature vectors, generally of fixed lower dimensionality are then used as input to a suite of conventional classifiers, including Support Vector Machine (SVM), Decision Tree (DT), K-Nearest Neighbors (K-NN), Random Forest (RF), and Logistic Regression (LR). Each classifier is trained to learn the optimal decision boundary that separates the tumor and non-tumor classes in the feature space.
-
The second strategy introduces a key variation by removing the classification layers of the Xception model, including the global pooling, dense, dropout, and softmax layers. Only the convolutional base is retained, and these layers are frozen to maintain the integrity of learned spatial features. The output from the last convolutional block is a 3D tensor of dimensions.
capturing fine-grained spatial and hierarchical feature information. This tensor is subsequently flattened into a 1D vector, where the final feature dimension is computed as.
These high-dimensional vectors, while more computationally intensive, retain greater spatial and structural information compared to global-pooled features. They are then supplied to the same ensemble of machine learning classifiers used in the first strategy.
This hybrid architecture effectively combines the representation power of pre-trained CNNs with the interpretability and flexibility of classical classifiers. The dual-strategy approach offers two complementary paradigms for feature extraction: one favoring computational efficiency (global pooled vectors), and the other emphasizing spatial richness (flattened convolutional outputs). Such a framework proves particularly beneficial in medical imaging domains, where diagnostic accuracy and model reliability are paramount. Table 5 showing the details of hyperparameter.
Measurement parameter
Accuracy: the overall verdict
Accuracy as the final score of the model — it’s the fraction of correct decisions out of all decisions made. In other words, it answers the question:
“Out of everything the model tried to classify, how often was it right?”
Accuracy is defined as:
where:
-
TP (True Positives): Correctly predicted tumor cases.
-
TN (True Negatives): Correctly predicted non-tumor cases.
-
FP (False Positives): Non-tumor cases incorrectly predicted as tumor.
-
FN (False Negatives): Tumor cases missed by the model.
Loss: the cost of being wrong
Loss is the model’s internal disappointment. It quantifies how far the predictions are from the actual labels, forming the core signal that guides model learning during training.
One popular function of loss for classification with multiple classes applications is categorical cross-entropy. It is a measure of how different the actual class distribution is from the projected probability distribution. As the estimated likelihood moves more away from the real label, the loss grows.
Let:
-
N be the number of samples,
-
C be the number of classes,
-
\(\:{y}_{i,c}\in\:\{0,\:1\}\)be the ground truth label for class c of sample i,
-
\(\:{\widehat{y}}_{i,c}\in\:[0,\:1]\:\)be the predicted probability that sample i belongs to class c.
The Categorical Cross-Entropy loss is given by:
This formulation ensures that the loss is minimized when the predicted probability for the correct class approaches 1, making it a powerful objective for training classifiers in deep learning.
Precision: how reliable are positive predictions?
Precision tells us: “When the model predicts positive, how often is it correct?” This metric is particularly important in scenarios where false positives carry a significant cost, such as in medical diagnostics or fraud detection.
Precision is defined as:
Recall: the sensitivity radar
Recall is like a safety net that ensures the model doesn’t miss anything important especially critical in high-stakes fields like medicine. It answers the question:
Out of all the actual positive cases, how many did the model successfully identify?
Recall is defined as:
In other words, recall measures how sensitive the model is in detecting positive cases. For instance, in tumor detection, it is preferable to raise a false alarm (which can be verified by further testing) than to miss a real tumor. A model with high recall helps minimize such dangerous false negatives, making it a crucial metric in medical diagnostics and other critical applications.
F1-score — balance between precision and recall
The F1-Score represents the harmonic mean of accuracy and recall. It offers a singular metric that reconciles both issues, particularly advantageous for unbalanced datasets.
ROC curve— threshold-independent performance
The Receiver Operating Characteristic (ROC) curve illustrates the True Positive Rate (TPR) in relation to the False Positive Rate (FPR) across different categorization thresholds.
Mathematically, these are defined as:
This curve illustrates the trade-off between sensitivity (recall or true positive rate) and the false positive rate (1 minus specificity) across several threshold settings. A curve that approaches the top-left corner signifies superior performance.
Results
Using loss and accuracy measures on training and validation datasets, Fig. 7; Table 6 show the training dynamics of a machine learning model across 70 epochs. With each passing epoch, the training loss is less, which means the error function is becoming minimized. At the same time, the validation loss decreases in a comparable pattern and settles at a low value, indicating that the model retains its capacity to generalize without substantial overfitting.
Training accuracy improves over time, eventually reaching a near-optimal level approaching convergence, when considering performance. Maintaining a high and consistent number, validation accuracy closely resembles training accuracy, suggesting strong performance on unseen data. High generalizability and low variance are results of an optimized model with well-regularized parameters, as shown by training and validation metrics that are in agreement with one another.
Using the Xception architecture linked with a transfer learning layer, this confusion matrix Fig. 8 offers an evaluation of a binary classification model meant to separate healthy from tumour samples. The model reportedly attained exceptional categorisation accuracy. Medical diagnosis depends on the elimination of false negatives as it reduces the possibility of missing a crucial diagnosis. The very remarkable accuracy of the model is further improved by the low frequency of false positives. Particularly in high-stakes domains like medical image analysis, the results reveal that integrating Xception with transfer learning lets the model collect strong and unique characteristics, hence producing outstanding classification performance and great generalisation.
Using the Xception model coupled with a transfer learning layer, this ROC (Receiver Operating Characteristic) curve [Figure 9] shows how well a binary classifier performs. Under various threshold settings, the figure evaluates whether the model can distinguish between two categories—probably “healthy” and “tumour”. Both Class 0 and Class 1 attain an Area Under the Curve (AUC) of 1.00, so the curve shows remarkable classification performance. With a value of 1.0 representing ideal performance, the AUC is a comprehensive statistic that captures the ability of the model to differentiate across classes.
With practically no false positive rate and a value of 1.0 almost instantaneous, the sensitivity shows that the model can effectively detect almost all positive cases while avoiding labelling negative instances mistakes. By using transfer learning with the Xception architecture, the model may use pre-trained feature representations, hence enhancing its ability to detect complex patterns in medical imaging data and producing remarkable discriminative performance.
Using Xception architecture in conjunction with a transfer learning layer, a classification model shows the cancer detection performance in Fig. 10—a collection of qualitative data. Grid displays from MRI brain images, each annotated to illustrate the ground truth diagnosis with the model’s forecast. Featuring a transfer learning layer and the Xception architecture, Model 4’s classification report shows the encouraging efficiency of this method for binary medical picture classification problems. Key measures of a classification model’s performance include recall, precision, and F1-score [Table 7]. While avoiding false positives, accuracy evaluates the model’s capacity to accurately detect pertinent events. Recall finds all relevant events, therefore helping to minimise false negatives in the model. Combining accuracy and recall, the F1-score provides a harmonic mean of these two measures.
The prediction results [Figure 11] reflect the model’s capability to assign high confidence to the correct class, showcasing the effectiveness of deep feature extraction achieved through Xception’s convolutional layers and the refinement brought by transfer learning. In the first case, the model strongly predicts the presence of a tumor, indicating high confidence for the tumor brain class and very low confidence for the healthy brain class. In contrast, the second case illustrates a scenario where the model confidently identifies the image as representing a healthy brain, assigning full confidence to that class while completely excluding the possibility of a tumor.
From the Fig. 12; Table 8, it is evident that all models achieved near-perfect training accuracy, indicating they were able to learn the patterns in the training set effectively. However, the testing accuracies vary, revealing the generalization ability of each model. While some models like CNN-SVM and CNN-KNN maintain high testing accuracy with minimal drop from training, others like CNN-DT show a more significant gap, suggesting possible overfitting.
This Fig. 13 demonstrates the effectiveness of using a CNN as a fixed feature extractor, where all layers of the CNN are frozen and not updated during training. In this setup, the CNN serves solely to extract deep features from the input data, which are then passed to traditional machine learning classifiers for final classification. On the left side, the confusion matrix displays the performance of the hybrid CNN-SVM model. Despite the CNN being entirely frozen, the SVM classifier performs strongly, indicating that the features captured by the pretrained CNN are highly informative and suitable for distinguishing between healthy and tumor categories. The right side of the figure includes confusion matrices for additional hybrid models that use different classifiers: Decision Tree, K-Nearest Neighbors, Random Forest, and Logistic Regression. Among these, Logistic Regression stands out with the most accurate classification performance, followed closely by K-Nearest Neighbors and Random Forest. The Decision Tree model shows comparatively lower accuracy, suggesting it may be less effective in utilizing the high-dimensional features extracted by the frozen CNN.
Each classifier in this set uses a combination of Convolutional Neural Networks (CNN) and one or more machine learning models, such as Support Vector Machine (SVM), Decision Tree (DT), K-Nearest Neighbors (KNN), Random Forest (RF), or Logistic Regression (LR). Table 9; Fig. 14 show the results of these comparisons. An improved fit between the training data and the model’s output is indicated by a greater training accuracy. However, the capacity of the model to generalize to previously unknown data is assessed by accuracy testing. Figure 15 shows the confusion matrix for each classifier.
The results show that classifiers, such as CNN-SVM, CNN-RF, and CNN-LR, achieve perfect training accuracy, meaning they fit the training data perfectly. However, their testing accuracy slightly decreases, suggesting that while these models perform excellently on the training set, they face some challenges in generalizing to the test set. The CNN-DT classifier also achieves perfect training accuracy, but its testing accuracy drops significantly, indicating overfitting, where the model performs well on the training data but struggles with unseen data. In contrast, the CNN-KNN classifier demonstrates a smaller gap between its training accuracy and testing accuracy, indicating better generalization and less overfitting.
Discussion
The efficiency comparison in Table 10 demonstrates that the pruned Xception configuration achieves substantial computational savings, supporting its potential integration into clinical decision-support systems where inference speed is critical.
Table 11 provides a comprehensive comparison of recent methodologies applied to brain tumor classification using MRI datasets.
Combining the Xception model with transfer learning layers produced remarkable performance with a peak accuracy of.9900 using linear regression (CNN-LR). This was similar across two setups: one with feature extraction using all CNN layers frozen and the other with pruned classification layers while maintaining the remaining layers frozen, hence stressing the power of the Xception architecture in transfer learning techniques. In a same line, Nizamani et al.48 presented the Feature-Enhanced Hybrid U-Net (FE3-HU-NET), which showed the success of including sophisticated feature fusion approaches into hybrid deep learning models with an outstanding accuracy of.9890. By contrast, Agarwal et al.17 used the Inception V3 model, which achieved an accuracy of.9889, higher than their Multi-Path CNN arrangement, which recorded 0.9600. Investigated by Shoaib et al.49, the DenseNet 201 model consistently attained an accuracy of 0.9800 when combined with principal component analysis (PCA) and machine learning classifiers like SVM and MLP. This highlights how effectively methods for lowering feature dimensionality when paired with deep features work. Furthermore, the improved ResNet152 model created by Kaur et al.46 shown remarkable classification performance with an accuracy of.9853, therefore underlining the need of fine-tuning models in transfer learning environments. Combining light GBM with transfer learning produced an amazing accuracy of 0.957047. By means of a much lower accuracy of.9000, the ANFIS-F-DBN model50 indicated that conventional neuro-fuzzy systems may have restrictions when compared to deep neural networks. Furthermore noted to have a significant effect on outcomes were variations in classifier choices, particularly in models based on Xception. For example, CNN combined with decision trees (CNN-DT) showed a performance level (about.9017), however using KNN and SVM produced noticeably higher accuracies (reaching up to.9883). The results show that adding advanced deep learning models—especially when combined with transfer learning and careful classifier selection—particularly helps to greatly increase the precision and dependability of brain tumour classification systems.While the proposed hybrid approach of leveraging Convolutional Neural Networks (CNNs) for feature extraction and traditional machine learning (ML) classifiers for brain tumor detection offers promising results, several limitations must be considered:
-
Data Dependency and Imbalance: The availability of big, high-quality labelled datasets greatly determines the performance of CNNs and ML classifiers. Particularly in the field of brain tumour identification, these sometimes-restricted datasets in the medical imaging domain might adversely affect model performance. Furthermore, aggravating the problem of model bias and producing less-than-optimal predictions for less-represented classes is the disparity in tumour classifications (e.g., more benign than malignant instances).
-
Computational Overhead: Particularly in large-scale MRI datasets, freezing CNN layers during transfer learning lowers the number of parameters to be taught, although the computational demand stays high. The demand for large computing resources, particularly GPU acceleration, could impede the general use of such models in settings with limited resources.
-
While CNNs are adept in extracting high-level features, conventional ML classifiers such as Support Vector Machines (SVM) or Random Forests can be prone to overfitting, particularly when considering high-dimensional feature spaces extracted from deep networks. This is especially true in cases when the available training data is either very unbalanced or inadequate, therefore compromising the generalising capacity of the model.
-
Lack of transparency and explainability is one of the main obstacles in using deep learning-based models in healthcare environments. Though conventional machine learning classifiers have some interpretability, the hybrid system—which consists of CNN feature extraction followed by ML classification—remains difficult overall. This might make clinical acceptance difficult as patient safety and trust depend on an awareness of the decision-making process.
-
Enhanced Model Complexity: Combining CNNs with ML classifiers gives the model design and training procedure even another degree of complexity. Longer training periods and more difficulty fine-tuning hyperparameters—especially in big, complicated datasets—can follow from this. Furthermore, causing questions about model scalability and maintenance in practical healthcare environments is their complexity.
Although the proposed XcepFusion framework achieved high accuracy, there remains a potential risk of overfitting due to certain experimental limitations. Specifically, the use of a single binary-labeled dataset may restrict the model’s ability to generalize to more complex or diverse clinical data. Table 12 summarizes the key sources of potential overfitting and the corresponding mitigation strategies applied in this study.
Conclusion
To enhance the accuracy and reliability of MRI-based brain tumor detection, we implemented a hybrid approach that integrates Xception CNNs with traditional machine learning classifiers. By freezing the convolutional layers of the pre-trained CNN, the system efficiently leverages robust feature extraction to obtain high-level representations of medical images without extensive retraining. To address the limitations of both deep learning and conventional approaches, the dense classification head was removed, and the extracted features were fed into classical classifiers such as SVM, KNN, Decision Tree, Random Forest, and Logistic Regression. Our experimental results demonstrate that this combined method achieves strong performance in tumor classification tasks, even with small or unbalanced datasets, while minimizing overfitting and enhancing model stability. Furthermore, the modular architecture allows for clinical personalization and potential integration into decision-support systems. Despite these advantages, the approach has limitations related to data dependency, model complexity, and adaptability to diverse imaging contexts. To further improve model reliability and clinical utility, future work will focus on multi-modal integration (e.g., combining MRI with CT or PET scans) to capture complementary diagnostic information, the incorporation of attention mechanisms and explainable AI (XAI) techniques to enhance interpretability, real-time deployment and clinical testing to evaluate practical performance, and expansion to larger and more diverse datasets to improve generalization across patient populations and imaging conditions. These steps aim to make the framework not only highly accurate but also interpretable, resource-efficient, and clinically applicable for brain tumor detection and diagnosis.
Data availability
The dataset used in this study is publicly available on Kaggle under the title BR35H :: Brain Tumor Detection 2020. It can be accessed at the following URL: https://www.kaggle.com/datasets/ahmedhamada0/brain-tumor-detection.
References
Remzan, N., Tahiry, K. & Farchi, A. Advancing brain tumor classification accuracy through deep learning: Harnessing radimagenet pre-trained convolutional neural networks, ensemble learning, and machine learning classifiers on MRI brain images. Multimedia Tools Appl. 83 (35), 82719–82747. https://doi.org/10.1007/s11042-024-18780-1 (2024).
Srinivasan, S. et al. A hybrid deep CNN model for brain tumor image multi-classification. BMC Med. Imaging. 24 (1), 21 (2024).
Meshram, P., Barai, T., Tahir, M. & Bodhe, K. The brain tumors identification, detection, and classification with AI/ML algorithm with certainty of operations. Lecture Notes Networks Syst. 611, 611–628 (2023).
Verma, A. & Singh, V. P. Design, analysis and implementation of efficient deep learning frameworks for brain tumor classification. Multimedia Tools Appl. 81 (26), 37541–37567. https://doi.org/10.1007/s11042-022-13545-0 (2022).
Batool, A. & Byun, Y. A lightweight multi-path convolutional neural network architecture using optimal features selection for multiclass classification of brain tumor using magnetic resonance images. Results Eng. 25, 104327 (2025).
Alshuhail, A. et al. Refining neural network algorithms for accurate brain tumor classification in MRI imagery. BMC Med. Imaging. 24 (1), 118 (2024).
Imam, R. & Alam, M. T. Optimizing Brain Tumor Classification: A comprehensive study on transfer learning and imbalance handling in deep learning models. In Lecture Notes in Computer Science (pp. 74–88). (2024). https://doi.org/10.1007/978-3-031-57963-9_6
Yadav, A. C., Shah, K., Purohit, A. & Kolekar, M. H. Computer-aided diagnosis for multi-class classification of brain tumors using CNN features via transfer-learning. Multimedia Tools Appl. https://doi.org/10.1007/s11042-025-20751-z (2025).
Jansi, R., Kowsalya, S., Seetha, S. & Yogadharshini, A. A. Deep Learning based Brain Tumour Detection using Multimodal MRI Images. In: 2023 2nd International Conference on Automation, Computing and Renewable Systems (ICACRS). https://doi.org/10.1109/icacrs58579.2023.10404952
Rajendirane, R., Kumar, T. A., Sandhya, S. G. & Hu, Y. A novel brain tumor segmentation and classification model using deep neural network over MRI-flair images. Multimedia Tools Appl. https://doi.org/10.1007/s11042-024-19487-z (2024).
Tuncer, I. et al. FlexiCombFE: A flexible, combination-based feature engineering framework for brain tumor detection. Biomed. Signal Process. Control. 104, 107538. https://doi.org/10.1016/j.bspc.2025.107538 (2025).
Khan, M. A. et al. Multimodal brain tumor classification using deep learning and robust feature selection: A machine learning application for radiologists. Diagnostics 10 (8), 565. https://doi.org/10.3390/diagnostics10080565 (2020).
Rasool, N. et al. CNN-TumorNet: leveraging explainability in deep learning for precise brain tumor diagnosis on MRI images. Front. Oncol. 15, 1554559 (2025).
Asiri, A. A. et al. Enhancing brain tumor diagnosis: an optimized CNN hyperparameter model for improved accuracy and reliability. PeerJ Comput. Sci. 10, e1878. https://doi.org/10.7717/peerj-cs.1878 (2024).
Sarkar, A., Maniruzzaman, M., Alahe, M. A. & Ahmad, M. An effective and novel approach for brain tumor classification using AlexNet CNN feature extractor and multiple eminent machine learning classifiers in MRIs. J. Sens. 2023 (1), 1224619. (2023).
Mathivanan, S. K. et al. Employing deep learning and transfer learning for accurate brain tumor detection. Sci. Rep. 14 (1), 7232 (2024).
Nahiduzzaman, M. et al. A hybrid explainable model based on advanced machine learning and deep learning models for classifying brain tumors using MRI images. Sci. Rep. 15 (1), 1649 (2025).
Bhimavarapu, U., Chintalapudi, N. & Battineni, G. Brain tumor detection and categorization with segmentation of improved unsupervised clustering approach and machine learning classifier. Bioengineering 11 (3), 266. https://doi.org/10.3390/bioengineering11030266 (2024).
Mohanty, B. C., Subudhi, P. K., Dash, R. & Mohanty, B. Feature-enhanced deep learning technique with soft attention for MRI-based brain tumor classification. Int. J. Inform. Technol. 16 (3), 1617–1626. https://doi.org/10.1007/s41870-023-01701-0 (2024).
Agarwal, M. et al. Deep learning for enhanced brain tumor detection and classification. Results Eng. 22, 102117. https://doi.org/10.1016/j.rineng.2024.102117 (2024).
Basthikodi, M., Chaithrashree, M., Shafeeq, B. M. A. & Gurpur, A. P. Enhancing multiclass brain tumor diagnosis using SVM and innovative feature extraction techniques. Sci. Rep. 14 (1), 26023 (2024).
Saeedi, S., Rezayi, S., Keshavarz, H. & Kalhori, S. R. N. MRI-based brain tumor detection using convolutional deep learning methods and chosen machine learning techniques. BMC Med. Inf. Decis. Mak. 23 (1), 16 (2023).
Sadr, H. et al. Enhancing brain tumor classification in MRI images: A deep learning-based approach for accurate classification and diagnosis. Image Vis. Comput. 159, 105555 (2025).
Wageh, M., Amin, K., Algarni, A. D., Hamad, A. M. & Ibrahim, M. Brain tumor detection based on deep features concatenation and machine learning classifiers with genetic selection. IEEE Access. 12, 114923–114939. https://doi.org/10.1109/access.2024.3446190 (2024).
Kang, J., Ullah, Z. & Gwak, J. MRI-Based brain tumor classification using ensemble of deep features and machine learning classifiers. Sensors 21 (6), 2222. 1062222 (2021).
Asiri, A. A. et al. Brain tumor detection and classification using Fine-Tuned CNN with ResNet50 and U-Net model: A study on TCGA-LGG and TCIA dataset for MRI applications. Life 13 (7), 1449. https://doi.org/10.3390/life13071449 (2023).
Dixon, J. et al. A hybrid Learning-Architecture for improved brain tumor recognition. Algorithms 17 (6), 221 (2024).
Nhlapho, W., Atemkeng, M., Brima, Y. & Ndogmo, J. Bridging the gap: exploring interpretability in deep learning models for brain tumor detection and diagnosis from MRI images. Inform 15 (4), 182 (2024).
Semwal, T., Jain, S., Mohanta, A. & Jain, A. A hybrid CNN-SVM model optimized with PSO for accurate and non-invasive brain tumor classification. Neural Comput. Appl. https://doi.org/10.1007/s00521-025-11078-9 (2025).
Natha, S. et al. Automated brain tumor identification in biomedical radiology images: A Multi-Model ensemble deep learning approach. Appl. Sci. 14 (5), 2210. https://doi.org/10.3390/app14052210 (2024).
Islam, M. M., Talukder, M. A., Uddin, M. A., Akhter, A. & Khalid, M. BrainNet: precision brain tumor classification with optimized EfficientNet architecture. Int. J. Intell. Syst. 2024 (1) 3583612. (2024).
Remzan, N., Hachimi, Y. E., Tahiry, K. & Farchi, A. Ensemble learning based-features extraction for brain MR images classification with machine learning classifiers. Multimedia Tools Appl. 83 (19), 57661–57684. https://doi.org/10.1007/s11042-023-17213-9 (2023).
Raza, A. et al. Enhancing brain tumor classification with transfer learning: Leveraging DenseNet121 for accurate and efficient detection. Int. J. Imaging Syst. Technol. 34 (1), e22957 (2023).
Güler, M. & Namlı, E. Brain tumor detection with deep learning methods’ classifier optimization using medical images. Appl. Sci. 14 (2), 642. https://doi.org/10.3390/app14020642 (2024).
Liu, Y. et al. MRI brain tumor classification based on CNN features and machine learning classifiers. J. Ambient Intell. Humaniz. Comput. 16 (1), 233–242. https://doi.org/10.1007/s12652-024-04928-2 (2024).
Malakouti, S. M., Menhaj, M. B. & Suratgar, A. A. Machine learning and transfer learning techniques for accurate brain tumor classification. Clin. eHealth. 7, 106–119. https://doi.org/10.1016/j.ceh.2024.08.001 (2024).
Vimala, M., Palanisamy, S., Guizani, S. & Hamam, H. Efficient GDD feature approximation based brain tumour classification and survival analysis model using deep learning. Egypt. Inf. J. 28, 100577. https://doi.org/10.1016/j.eij.2024.100577 (2024).
Xu, M., Yoon, S., Fuentes, A. & Park, D. S. A comprehensive survey of image augmentationtechniques for deep learning. Pattern Recogn. 137, 109347. https://doi.org/10.1016/j.patcog.2023.109347 (2023).
Huang, L. Normalization techniques in deep learning. In Synthesis Lectures on Computer Vision 2022. https://doi.org/10.1007/978-3-031-14595-7
Sharma, S. & Kumar, S. The Xception model: A potential feature extractor in breast cancer histology images classification. ICT Express. 8 (1), 101–108. https://doi.org/10.1016/j.icte.2021.11.010 (2021).
Rao, K. N. et al. An efficient brain tumor detection and classification using Pre-Trained convolutional neural network models. Heliyon 10 (17), e36773. https://doi.org/10.1016/j.heliyon.2024.e36773 (2024).
Singh, R. Support Vector Machines (SVM) - Rishabh Singh - Medium. Medium 2024, November 14. https://medium.com/@RobuRishabh/support-vector-machines-svm-27cd45b74fbb
Baladram, S. Decision Tree Classifier | TDS Archive. Medium November 30. (2024). https://medium.com/data-science/decision-tree-classifier-explained-a-visual-guide-with-code-examples-for-beginners-7c863f06a71e
Baladram, S. K. & Nearest Neighbor Classifier | TDS Archive. Medium. November 30. (2024). https://medium.com/data-science/k-nearest-neighbor-classifier-explained-a-visual-guide-with-code-examples-for-beginners-a3d85cad00e1
Baladram, S. Random Forest | TDS Archive. Medium November 30. (2024). https://medium.com/data-science/random-forest-explained-a-visual-guide-with-code-examples-9f736a6e1b3c
Logistic classification model (logit or logistic regression). Statlect. https://www.statlect.com/fundamentals-of-statistics/logistic-classification-model
Pande, Y. & Chaki, J. Brain tumor detection across diverse MR images: an automated triple module approach integrating reduced fused deep features and machine learning. Results Eng. 25, 103832 (2024).
Yoon, S. Brain tumor classification using a hybrid ensemble of Xception and parallel deep CNN models. Inf. Med. Unlocked. 54, 101629 (2025).
Malebary, S. J. Intelligent machine learning based brain tumor segmentation through Multi-Layer hybrid U-Net with CNN feature integration. Computers Mater. Continua. 79 (1), 1301–1317. https://doi.org/10.32604/cmc.2024.047917 (2024).
Kaur, P. & Mahajan, P. Detection of brain tumors using a transfer learning-based optimized ResNet152 model in MR images. Comput. Biol. Med. 188, 109790. https://doi.org/10.1016/j.compbiomed.2025.109790 (2025).
Nizamani, A. H., Chen, Z., Nizamani, A. A. & Bhatti, U. A. Advance brain tumor segmentation using feature fusion methods with deep U-Net model with CNN for MRI data. J. King Saud Univ. - Comput. Inform. Sci. 35 (9), 101793. https://doi.org/10.1016/j.jksuci.2023.101793 (2023).
Shoaib, M. R. et al. Improving brain tumor classification: an approach integrating Pre-Trained CNN models and machine learning algorithms. Heliyon. 11, 33471. (2024).
Tiwari, R. G. et al. Adaptive neuro-FUZZY inference system-fusion-deep belief network for brain tumor detection using MRI images with feature extraction. Biomed. Signal Process. Control. 103, 107387. https://doi.org/10.1016/j.bspc.2024.107387 (2024).
Khushi, H. M. T., Masood, T., Jaffar, A., Akram, S. & Bhatti, S. M. Performance analysis ofstate-of‐the‐art cnnarchitectures for brain tumour detection. Int. J. Imaging Syst. Technol. 34, e22949 (2023).
Acknowledgements
SeongKi Kim is the corresponding author.
Funding
This research was supported by Global-Learning & Academic research institution for Master’s·PhD students, and Postdocs(LAMP) Program of the National Research Foundation of Korea(NRF) grant funded by the Ministry of Education(No. RS-2023-00285353).
Author information
Authors and Affiliations
Contributions
Deependra Rastogi: Conceptualization, methodology design, manuscript drafting and ExperimentPrashant Johri: Data curation, overall supervision of the research work.Seifedine Kadry: Formal analysis, validation of results, and critical revision of the manuscript for important intellectual content.SeongKi Kim: Technical support, Funding acquisition.Lalit Kumar: Literature review, and contribution to writing the introduction and related work sections.Vishwadeepak Singh Baghela: Experimental validation, Review and EditingArfat Ahmad Khan: Editing, visualization, and contribution to refining the final draft for submission.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Ethics statement
This research utilizes a publicly available dataset hosted on Kaggle. The dataset was accessed in accordance with Kaggle’s Terms of Service and is intended solely for academic and non-commercial research purposes. The dataset does not contain personally identifiable information (PII) or any sensitive data that would compromise the privacy of individuals. Any identifiable data (if originally present) has been anonymized by the original dataset contributors prior to its publication. The use of this dataset complies with ethical research standards and applicable data protection regulations. No attempts were made to re-identify individuals or entities within the dataset. Proper attribution is given to the original authors and contributors of the dataset, in line with the licensing and citation requirements specified on the Kaggle platform.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Rastogi, D., Johri, P., Kadry, S. et al. XcepFusion for brain tumor detection using a hybrid transfer learning framework with layer pruning and freezing. Sci Rep 16, 3849 (2026). https://doi.org/10.1038/s41598-025-33970-z
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-025-33970-z














