Introduction

Lung cancer is one of the cancers with the highest mortality rate in the world1. Adenocarcinoma is the most prevalent histopathological subtype of lung cancer, and accounts for approximately one-third to one-half of all lung cancer cases, which presents as ground glass nodules (GGNs) on computed tomography (CT) scans.2,3,4. According to the latest guideline of the 2021 World Health Organization (WHO), lung adenocarcinoma (LUAD) includes precursor glandular lesions (PGL), minimally invasive adenocarcinoma (MIA), and invasive adenocarcinoma (IAC). Within PGL, there are atypical adenomatous hyperplasia (AAH) and adenocarcinoma in situ (AIS). It is generally believed that the evolution of GGNs follows the sequence AAH \(\rightarrow\) AIS \(\rightarrow\) MIA \(\rightarrow\) IAC5. Previous studies reported that different subtypes of early-stage LUAD correspond to different responses to treatments and prognoses4,6. PGL and MIA are considered non-invasive and suitable for sublobar resection, with a good prognosis and a 5-year disease-free survival (DFS) rate of almost 100%. However, for IAC patients, lobectomy is more suitable but with a poor prognosis of only 40% to 85% 5-year DFS, and the risk of recurrence is higher as the degree of invasion increases4,6,7. Therefore, accurate assessment of the invasiveness of LUAD with GGNs is essential for clinical decision-making.

With the widespread application of CT in early screening of lung cancer, an increasing number of GGNs, especially sub-centimeter GGNs, are being detected and resected. Consequently, pathologists face multiple challenges in distinguishing AIS, MIA, and IAC, for example, determining the degree of invasion of MIA and IAC in the context of alveolar collapse, as well as distinguishing the adenoid structures formed in collapsed alveoli, requires pathologists to have a high level of expertise through long-term professional training8. Histopathological examination of tissue slides is considered to be the gold standard in the diagnosis of LUAD, and hematoxylin and eosin (H&E) staining technique can capture morphological features of tumor tissue via high-resolution images. However, the manual microscopic examination of H&E staining slides also relies on the expertise of pathologists, and is labor-intensive and prone to errors9. Therefore, it is urgent to develop a reliable, convenient, and fast auxiliary diagnostic tool to provide a reference for the invasiveness degree of LUAD, which is crucial to improving the diagnostic efficiency and accuracy of pathology practitioners, especially junior pathologists in the early stages of their careers.

The emergence of digital pathology has made it possible to use artificial intelligence to perform quantitative analysis of H&E staining-based whole slide image (WSI), and has brought about a series of computer-aided diagnosis methods for pathological images10. For example, pathomics utilizes WSIs to quantitatively extract histological and cytological features, providing the high-throughput characterization of the tumor microenvironment11, e.g., pathomic features were used to reveal immune and molecular evolution from lung preneoplasia to invasive adenocarcinoma12. Unlike pathomics, which mainly includes lesion segmentation, feature extraction and screening, and machine learning classification, deep learning uses specific neural networks to build end-to-end classification models10,13. Previous studies have also demonstrated that deep learning-based histopathological analysis of LUAD has important potential for tumor segmentation, classification, and outcome prediction10,14,15. However, most classification efforts have focused on the main subtypes of LUAD, which have not yet been applied to clinical work, especially those related to the diagnosis of micropapillary subtypes and their prognosis. Moreover, there may be a risk of misleading pathologists if applied in clinical work, such as identifying invasive or non-invasive GGNs, especially for those with relatively limited experience.

Inspired by the success of artificial intelligence in helping the diagnosis of pulmonary nodules in CT scans16,17, we propose a simple and clinically practical method that combines deep learning and pathomics to detect the invasiveness of LUAD. This method can help pathologists identify abnormal tumor regions and preliminarily classify the invasiveness of GGNs from pathology sections, thereby reducing their workload. Unlike previous studies, the objective of this study was to help improve the diagnostic capabilities of junior and intermediate pathologists (those who diagnose fewer than 50 LUAD surgical resection specimens per year).

Materials and methods

Patients

LUAD patients diagnosed with GGNs, from January 2019 to May 2022 at the Affilliated Zhongshan Hospital of Dalian University and from January 2019 to December 2019 at the Affilliated Xinhua Hospital of Dalian University, were enrolled retrospectively and consecutively. The inclusion criteria were: (i) patients aged over 18 years old; (ii) patients who had not been received radiotherapy, chemotherapy, or other treatments before surgery; (iii) CT images and pathological images are of the same GGN; (iv) patients with surgically resected and pathologically confirmed as LUAD. The exclusion criteria were: (i) solid nodules on CT; (ii) incomplete clinical or pathological data; (iii) combined with primary cancer in other locations. Finally, a total of 289 LUAD patients were included in the study cohort. All patients were randomly assigned to training and test cohorts in a ratio of 8:2. Each cohort was further categorized into invasive (ICA) and non-invasive (AAH, AIS, and MIA) groups. The patient enrollment flow chart of this study is shown in Fig. 1.

This study was carried out in accordance with the principles stated in the Declaration of Helsinki and approved by the Ethics Committee of Affiliated Zhongshan Hospital of Dalian University (No. 2021029, Dalian, China). Due to the retrospective nature of the study, the Ethics Committee waived the need of obtaining informed consent.

Fig. 1
figure 1

Flowchart revealing study population according to inclusion and exclusion criteria.

WSI collection

The postoperative pathological tissues were first immersed in 10% formalin. After embedding in paraffin, they were sectioned to 5\(\mu\)m-thick pathological slides, and then stained with hematoxylin and eosin (H&E) for pathological evaluation. The diagnosis was conducted by two pathologists with more than 10 years of expertise. In case of dispute, the decision was made by the higher-level chief physician. The histopathology slides were digitized using the fully automatic digital pathology whole-slide scanner (Easyscan model, Motic Medical Diagnostic Systems Co., Ltd., Xiamen, China) at 20\(\times\) magnification, with predefined pixel resolution ( 0.5\(\mu\)m/pixel) and an effective tissue area of 15 mm \(\times\) 15 mm. For each patient, two pathologists with more than ten years of clinical experience selected one specific WSI of the largest and most typical tumor tissue under a microscope. Finally, a total of 289 WSIs were collected in this study.

WSI processing

WSI annotation

All WSIs were manually delineated as tumor or non-tumor areas by two pathologists (with 10 and 7 years of diagnostic experience) using QuPath-0.3.2 software18. In case of disagreements during cross-examination, the final decision was made by the higher-level chief pathologist. During the annotation process, areas with folds, bleeding, and blurriness were excluded. Finally, the tumor areas were annotated and labelled as 0 (non-invasive: AAH, AIS, and MIA) or 1 (invasive: IAC) according to the pathological results, respectively. An example of WSI data annotation is illustrated in Fig. 2.

Fig. 2
figure 2

An example of annotation of regions of interest for WSI data.

WSI preprocessing and augmentation

All WSIs were preprocessed as follows: (i) the WSIs were partitioned into non-overlapping patches of 512\(\times\)512 pixels to reduce the computer memory requirements; (ii) the patches were processed by thresholding method, followed by removing white regions and eliminating patches with bright pixel ratio exceeding 0.8 or without pathological tissues; (iii) color normalization was performed using Vahadane method to address issues of color variation, inconsistency, and deviation among pathological images19. After the preprocessing, 2,584,435 WSI patches were obtained. Then we adopted image augmentation to increase the diversity of training samples including random translation, random rotation, vertical flipping, and \(90^\circ\) rotation (an example of a patch with its augmented tiles can be found in Fig.  S1 in the supplementary file), and image augmentation technique was selected randomly for each epoch, thereby addressing the issue of limited data volume and enhancing generalization ability of the model.

Model specification

Deep learning model

This study used three deep learning models, Resnet18, Resnet50, and Resnet101, to construct a model of tumor region recognition20,21. Only the last fully connected layer in the network used Softmax as the activation function, while the remaining layers used ReLU. The learning rate employed cosine decay learning rate algorithm, and other parameter settings were as follows: batch_size = 128, epoch = 32, initial learning_rate = 0.01, optimizer = Adam. Finally, the optimal model among the three was selected to extract deep learning features.

Feature selection

After feeding all patch-level WSIs into the deep learning model, labels and probabilities for each patch can be obtained. In this study, two multiple instance learning methods were employed to aggregate patch-level features to the WSI-level for image classification: patch likelihood histogram (PLH) and bag of words (BoW)22. PLH and BoW pipelines utilized prediction and probability histograms and term frequency-inverse document frequency features to characterize WSI features, respectively. Then, we used Spearman correlation to remove the features with high repeatability (one was retained if the correlation coefficient between the two pairs was larger than 0.9). Finally, for all features, the least absolute shrinkage and selection operator (LASSO) regression model was used for dimensionality reduction, and the optimal regulation weight (\(\lambda\)) was determined using a 10-fold cross-validation method on the training cohort.

Machine learning-based pathomics model

The features after dimensionality reduction were used to build pathomics models based on machine learning models, including random forest (RF), extremely randomized trees (ExtraTrees), eXtreme gradient boosting (XGBoost), and light gradient boosting machine(LightGBM). The classifier parameters were set as described in Table S1 in the supplementary file. By comparing the predictive performance of different models, the best machine learning model for identifying tumor invasiveness was selected as an assisted tool for pathologists.

Assisting model

The predictive results of pathomics for tumor invasiveness were compared with 10 pathologists (who lacked extensive experience in pathological diagnosis) before and after assistance, and the washout period was set to one month. Among them, 5 pathologists were junior physicians with 1-3 years of experience, while the other 5 pathologists were mid-level physicians with 6-9 years of experience. Inexperience was defined as pathologists reviewing less than 50 LUAD surgical specimens per year.

Performance metrics

Receiver operator characteristic (ROC) curves, the area under ROC curve (AUC), accuracy (ACC), sensitivity (SE), specificity (SP), positive predictive values (PPV), and negative predictive values (NPV) are used to evaluate the predictive performance of the model in the training and testing cohorts. Decision curve analysis (DCA) was employed to assess the clinical utility. A flowchart of pathological feature extraction and model establishment is shown as Fig. 3.

Fig. 3
figure 3

Flowchart of extraction of features, model establishment, and performance evaluation.

Results

Patient characteristics

This study included 289 patients with LUAD, ranging in age from 26 to 80 years (57.73±10.87). Among them, 196 cases (67.82%) were in the invasive group and 93 cases (32.18%) were in the non-invasive group, including 52 cases of MIA, 32 cases of AIS, and 9 cases of AAH. The training cohort consisted of 230 patients with the mean age of 57.59±10.87, 23.48% males and 76.52% females. 59 cases in the testing cohorts had the mean age of 58.25±10.97 years, with 37.29% males and 62.71% females. Statistical analysis showed the significant difference in age, volume, and maximum diameter between the invasive and non-invasive groups (\(p < 0.001\)), whereas no significant differences were observed between the two groups in terms of CT value, gender, and GGN location (\(p > 0.05\)). Table 1 summarizes the demographic information and sematic CT features of the patients. Out of a total of 2,584,435 WSI patches, 2,135,006 WSI patches were used for the training cohort, 449,429 for the test cohort.

Table 1 Demographic and semantic CT features of patients.

Model evaluation of deep learning

As shown in Table 2, ResNet50 achieved the highest AUC value of 0.973 on the training cohort, while ResNet18 reached the highest AUC value of 0.956 on the testing cohort. Higher AUC values indicate good predictive performance of the model in distinguishing between tumor and non-tumor regions. Regarding ACC, ResNet101 performed well in both the training and test cohorts, achieving 0.924 and 0.904, respectively. Notably, ResNet18 also reached an ACC of 0.904 on the test cohort. In terms of SE, ResNet50 performed the best in the training cohort with a value of 0.836, while in the test cohort, ResNet18 had the highest SE of 0.832. Higher SE values suggest that the model can more accurately identify tumor regions (including invasive regions). Therefore, we selected the ResNet18 model as the best model for its excellent ACC and AUC values, as well as relatively high SE and SP. Figure 4 shows the comparison of ROC curves of three deep learning models for recognizing tumor regions.

Table 2 Performance in recognizing tumor regions for ResNet18, ResNet50 and ResNet101 models.
Fig. 4
figure 4

Comparison of ROC curves in recognizing tumor regions for ResNet18, ResNet50, and ResNet101 models.

Feature selection

After training the patch-level model, we obtained predictions and corresponding probabilities for each patch. Using two pipelines, PLH and BoW, the results at the patch-level were aggregated into WSI-level features. The PLH method yielded a total of 101 probability features and 2 prediction result features, while the BoW similarly produced 101 probability features and 2 prediction result features. Therefore, 206 WSI-level features were then subjected to dimensionality reduction. First, 70 of 206 features were retained via screening the features with high repeatability using Spearman correlation. Second, applying the least absolute shrinkage and selection operator (LASSO) regression model, 7 features with non-zero coefficient (comprising 6 probability features and 1 prediction result feature) were retained for model creation, and the parameter (\(\lambda\)=0.0222) in the LASSO model was selected via 10-fold cross validation based on Mean Squared Error (MSE) value, as shown in Fig. 5.

Fig. 5
figure 5

Dimensionality reduction of WSI-level features. (A) coefficients via LASSO. (B) MSE value. (C) 7 non-zero coefficient features.

Moreover, Fig. 6 shows the distribution of values for the retained 7 features across different groups. There are significant differences in the values among the groups (p<0.01), which proves that the features obtained through multiple instance learning have excellent discriminative ability for identifying the invasiveness of LUAD.

Fig. 6
figure 6

The distribution comparison of features in differentiating the invasiveness of lung adenocarcinoma with GGNs.

Model evaluation of pathomics

Table 3 and Fig. 7 give the predictive results of machine learning-based pathomics model for the invasiveness of LUAD. The RF model showed higher ACC and AUC on both cohorts. Specifically, RF achieved an ACC of 0.817 and an AUC of 0.897 on the training cohort, with a high SE of 0.795, SP of 0.865, PPV of 0.925 and NPV of 0.667. On the testing cohort, RF achieved an ACC of 0.814 and an AUC of 0.807, showing a slight decrease in performance compared to the training cohort. Figure 8 illustrates that, within most threshold probability ranges, the RF model yields high clinical net benefit on both the training and testing cohorts, as shown by the distribution of sample prediction and DCA curves.

Table 3 Performance in predicting the invasiveness of lung adenocarcinoma using machine learning-based pathomics models.
Fig. 7
figure 7

Comparison of ROC curves in predicting invasiveness of lung adenocarcinoma using machine learning-based pathomics models.

Fig. 8
figure 8

Distribution of different samples and DCA curves predicted by Random Forest-based pathomics model. (A-B) Prediction scores of training and test cohorts. (C-D) DCA curves of training and test cohorts.

Performance comparison of pathomics in assisting junior and intermediate pathologists

Table 4 and Fig. 9 show the performance of the pathomics model, pathologists, and pathology-assisted pathologists in discriminating the invasiveness of LUAD. In the test cohort, the ACC value of RF-based pathomics reached 81.4%, which was significantly higher than the independent review levels of pathologists (junior pathologists at an average of 56.6%, intermediate pathologists at 69.5%). After a one-month washout period, with the assistance of pathomics, the average ACCs of two-level pathologists improved to 79.7% and 80.7%, with increases of 0.231 and 0.112. Additionally, the average AUCs increased by 0.212 and 0.113 for the junior and intermediate pathologists, respectively. In the training cohort, ACCs of RF-based pathomics, junior and intermediate pathologists reached 81.7%, 55.0%, and 71.6%, respectively, whereas with pathomics assisted, ACCs for junior and intermediate pathologists reached 77.3% and 79.9%, respectively. Meanwhile, AUCs for junior and intermediate pathologists increased to 77.2% and 76.0%, respectively.

Table 4 Average diagnostic performance of inexperienced pathologists with/without pathomics assisted.
Fig. 9
figure 9

Comparison of ROC curves in predicting the invasiveness of lung adenocarcinoma by pathomics, pathologists, and pathology-assisted pathologists.

Discussion

GGNs on CT images usually represent a series of pathological progression, including AAH, AIS, MIA, and IAC. Considering their biological behaviour, AAH, AIS, and MIA are classified as non-invasive with a favorable prognosis compared to IAC.23. Accurate postoperative pathological diagnosis of the invasiveness of GGNs is crucial for subsequent treatment and follow-up. This study focused on postoperative patients with LUAD manifesting as GGNs on CT images. We retrospectively compared the pathological features of patients with invasive and non-invasive LUAD and established a deep learning model based on ResNet18 for predicting tumor regions. We found that the ResNet18 model had high predictive efficiency in identifying tumor regions of LUAD. For pathologists with less experience in hospitals with limited surgical cases, the incorporation of pathomics-assisted diagnosis has the potential to enhance diagnostic efficiency. We initially opted not to conduct a detailed subtype analysis to ensure clarity for pathologists with varying levels of experience. We aim to contribute to both work efficiency and diagnostic assistance, addressing the challenges faced by pathologists with high workload and potential misdiagnoses. We aspire to provide pathologists with positive prompts during the diagnostic process. Our contributions in this study are as follows: (1) Three end-to-end ResNet deep learning models in this study are applied and to accurately identify tumor regions; (2) Using the features from deep learning, we built four kinds of pathomics based on machine learning models to assess the invasiveness of lung adenocarcinoma; (3) We built a simple and clinically practical tool using deep learning and pathomics for tumor invasiveness, and explore its convenience in future clinical applications and its potential value in improving the diagnostic capabilities of junior and intermediate pathologists.

Significance of ResNet model in identifying tumor regions

In this study, convolutional neural network (CNN) based ResNet models were employed to classify WSIs into tumor and non-tumor regions, aiming to enhance the efficiency of pathologists in identifying tumor regions of LUAD and reducing diagnostic omissions. The ResNet model was initially proposed in 2015 by He et al., introducing residual learning units to solve the issues of vanishing gradients when increasing the network depth20. Although there are deeper ResNet variants such as ResNet50, ResNet101, and ResNet152, ResNet18 had satisfactory performance in the current task. In this study, we found the performance of the ResNet model does not change significantly with the increase in the number of layers in identifying the tumor regions. The AUC values on the training and testing cohorts were 0.957 and 0.956, demonstrating superior generalization performance of the ResNet18 model.

Furthermore, the ResNet18 model achieved an ACC of 90.4% on the test cohort, showing excellent SE, SP, PPV, and NPV values in identifying invasive tumor regions. These results indicate that despite the limited amount of tumor tissues, especially when GGNS are smaller than 1cm on CT images, we can still accurately localize tumor regions. Therefore, utilizing the ResNet18 model in identifying tumor regions of LUAD in WSIs is an effective and feasible strategy. It not only enhances the efficiency of pathologists in recognizing and evaluating tumor region, but also helps reduce diagnostic omissions and provides a valuable auxiliary tool for future clinical pathological diagnoses. While it would also be valuable to compare different CNN structures, such as VGG, Inception, or DenseNet, exploring a single family of architectures in depth can provide detailed insights into their specific advantages and limitations in the given context, thereby offering a robust foundation for future comparisons with other CNN structures.

Significance of machine learning-based pathomics in predicting the invasiveness

The morphological evaluation criteria for tissue specimens in lung cancer diagnosis are often considerably subjective, Thunnissen et al. pointed out that alveolar collapse and extensive epithelial proliferation were the main causes of interobserver variability24. The combination of digital pathology and artificial intelligence is expected to provide objective support for decision-making13. This study employed machine learning-based pathomics approaches to predict the invasiveness of GGN-type LUAD. First, the LASSO model was applied to reduce 206 WSI-level features to 7 significant features. Second, machine learning models including RF, ExtraTrees, XGBoost, and LightGBM were utilized to construct the predictive models. Among them, RF and ExtraTrees both achieved ACC value of 0.814 on the test cohort, while RF had a higher AUC value of 0.807. To the best of our knowledge, there is limited research specifically addressing deep learning-based pathomics for lung adenocarcinoma invasiveness classification, and this gap shows the novelty and significance of our study. While direct comparisons are rare, there are several studies in related areas that provide valuable context. Yu et al. combined traditional thresholding and image processing techniques with machine learning methods, achieving an AUC of 0.85 for distinguishing normal and tumor tissues using WSI data, and an AUC of 0.75 for differentiating LUAD from squamous cell carcinoma25. Wang et al. subdivided WSI slides of LUAD into 300\(\times\)300-pixel patches and employed a CNN model to classify malignant or non-malignant, with an overall classification accuracy of 89.8%26. In another study, the authors generated a deep learning pathomics (DLP) with WSI features for the automated segmentation of tissue, and achieved an AUC value of 0.802 to preoperatively distinguish between luminal and non-luminal breast cancers at early stages21.

Studies have shown that distinguishing MIA from invasive adenocarcinoma is often challenging when the maximum diameter of tumor invasiveness on frozen sections approaches 5mm27. This difficulty also occurs in postoperative paraffin sections, compounded with the tendency of pathologists to overestimate the diseases, leading to inevitable human errors28. Therefore, it is significantly important to accurately assess the extent of invasiveness. Furthermore, there is a lack of consistency in the diagnosis of LUAD histological subtypes, especially in the diagnosis of invasive and non-invasive cancer subtypes, evaluated by different pathologists or even the same pathologist at different times29,30,31. These practical clinical problems provide potential opportunities for artificial intelligence to make a significant contribution to diagnostic standardization and homogenization process in the future. Although the model demonstrated high accuracy in classifying non-invasive and invasive LUAD, we acknowledge its limitation in distinguishing the invasive component of MIA, i.e., the model failed to classify the invasive component of MIA as invasive. This may be attributed to the fact that although pathomic divides digitized pathological slides into multiple patches, the model is trained to differentiate between invasive and non-invasive lesions based on feature patterns derived from the entire WSIs. While MIA may contain small invasive components, our model primarily focuses on larger regions of invasion that exceed this threshold, as these have a more significant impact on prognosis. Therefore, the model does not classify the minor invasive component of MIA as invasive. To address this limitation, the refinement of the training process and the incorporation of multi-scale analysis or auxiliary modalities may be required to better differentiate between small-scale invasiveness in cases like MIA.

Significance of pathomics in assisting junior and intermediate pathologists

Traditional pathological diagnosis suffers from high subjectivity, low efficiency, and prolonged time consumption, and relies heavily on the practical experience and diagnostic proficiency of pathologists. These problems are particularly pronounced in hospitals that perform fewer surgeries and lack subspecialty pathologists. Therefore, this study combined deep learning and pathomics as an assistant tool to enhance the efficiency and accuracy of pathological diagnosis. The results demonstrated that pathologists with assistance can effectively adjust their diagnostic levels, thereby improving diagnostic accuracy. In addition, the deep learning model can help pathologists rapidly interpret tumor regions. Even though the accuracy of deep learning models may not be entirely ideal, pathologists can make effective adjustments to tumor regions by evaluating the model results.

This study focuses on how pathomics can assist pathologists in interpreting the invasiveness of LUAD, thereby improving work efficiency and reducing misdiagnosis and missed diagnosis. Fortunately, with the help of pathomics, both pathologists obtained improvements in ACC, AUC, and other metrics. Wei et al. built a deep learning model to distinguish subtypes of LUAD, and the model’s performance was comparable to pathologists, suggesting that it can be used clinically to help pathologists classify LUAD14. Pan et al. utilized artificial intelligence-based grading to predict disease-free survival and consistently improved prognosis in stage I tumors to further assist pathologists8. Unlike these studies, we aim to help improve the diagnostic performance of junior and intermediate pathologists (those who diagnose fewer than 50 surgical resection specimens of LUAD per year). The pathomics model performed better than the pathologists, whose ACC and AUC values both improved with pathomics’ assistance.

While achieving some anticipated results in this study, there are still some limitations and future directions. First, the sample sizes of both the training and testing cohorts were relatively small, potentially impacting the generalizability and robustness of the learning model to a certain extent, and next we plan to expand the cohort size and incorporate external datasets or to implement nested cross-validation for more comprehensive validation. Second, the WSIs were manually annotated by pathologists, in future work, we will explore weakly supervised or unsupervised methods for more objective classification. Third, this study only utilized the ResNet model, next we will try to conduct comparative studies with other models, such as AlexNet, VGGNet, InceptionNet, EfficientNet, ViT, HIPT, and GigaPath etc., to further improve and validate the diagnostic efficiency. Furthermore, we acknowledge that a more detailed subtype analysis could provide valuable insights, and this will be one of our future work.

Conclusion

This study retrospectively included 289 patients of lung adenocarcinoma (LUAD) with GGNs on CT and established a predictive model for tumor invasiveness based on deep learning and pathomics. The ResNet18 network can accurately locate tumor regions, while pathomics can effectively distinguish the invasiveness of LUAD using WSIs. We also confirmed the efficacy of the model in assisting inexperienced pathologists in identifying tumor regions and determining tumor invasiveness. Our future work will focus on enhancing the model’s accuracy and generalizability by incorporating a larger, multi-center dataset and exploring more state-of-the-art deep learning architectures. We plan to integrate multi-modal imaging data and clinical information to further improve the predictive capabilities of our model, and we hope to advance the capabilities of AI-assisted diagnosis on LUAD.