Introduction

Ischemic heart disease (IHD) remains a leading cause of global morbidity and mortality, necessitating accurate diagnostic and prognostic tools for optimal patient management1,2. Cardiovascular magnetic resonance (CMR) imaging, particularly through late gadolinium enhancement (LGE), has established itself as the gold standard for assessing myocardial viability and fibrosis. LGE enables precise discrimination between ischemic and non-ischemic etiologies, providing crucial information for clinical decision-making regarding revascularization strategies and risk stratification3,4,5. However, this technique presents notable limitations, including contraindications in patients with kidney disease or contrast allergies, extended scan times due to the required 10–15-minute delay for contrast retention, and inability to provide comprehensive functional insights2,5.

The growing need for alternative diagnostic approaches has sparked interest in contrast-free techniques that could potentially reduce scan times, decrease costs, and improve CMR accessibility. Novel methods such as feature tracking (FT) strain analysis and T1/T2 mapping have emerged as promising gadolinium-free approaches for cardiac phenotyping2,6,7. While preliminary studies have demonstrated encouraging results, with research by Tantawy et al. showing the capability of FT-CMR to differentiate between viable and non-viable segments6 and Dastidar et al. highlighting the potential of native T1 mapping for viability assessment7, most investigations have been limited by isolated parameter analysis or small sample sizes.

The rapid advancement of artificial intelligence (AI) in cardiovascular imaging has opened new avenues for improving diagnostic accuracy and efficiency5. AI-based approaches have demonstrated significant utility across various CMR applications, including automated image acquisition, reconstruction, and analysis, while providing valuable diagnostic and prognostic information8. Recent developments in non-contrast CMR examinations supported by AI models have shown promising results in detecting myocardial infarction2,9,10,11. However, the application and validation of these approaches in comprehensive myocardial viability assessment across different segments and coronary artery territories remain largely unexplored.

This study aims to evaluate the combined efficacy of FT strain analysis and T1/T2 mapping integrated with machine learning (ML) pattern recognition techniques for predicting myocardial viability. By conducting a systematic comparison between these non-contrast methods and the established LGE technique, we seek to determine their relative diagnostic accuracy and potential advantages in clinical practice. This investigation represents a crucial step toward developing more accessible, efficient, and comprehensive approaches to myocardial viability assessment, potentially transforming the current paradigm of cardiac imaging diagnostics.

Method and materials

Study design

This study was conducted as a retrospective analysis of patients with myocardial infarction (MI) who underwent viability studies using LGE CMR two to four weeks post-MI. The primary aim was to assess the efficacy of FT strain analysis and T1/T2 mapping, in conjunction with ML, for predicting myocardial viability compared to LGE. A summary of the study method is demonstrated in Fig. 1.

Fig. 1
Fig. 1The alternative text for this image may have been generated using AI.
Full size image

Graphical Abstract. Overview of the study methodology.

Study population

This retrospective study included patients with ST-segment elevation MI who underwent LGE CMR for viability assessment within the subsequent two to four weeks, from January 2020 to December 2021. The research was conducted in accordance with the Declaration of Helsinki and received approval from the Ethics Committee at the Rajaie Cardiovascular Medical and Research Center. The requirement of informed consent was waived by the Institutional Ethics Committee because it was a retrospective study. All methods were performed in accordance with the relevant guidelines and regulations. The study population was selected from a database of MI patients, ensuring a comprehensive collection of clinical and imaging data for robust analysis.

Inclusion and exclusion criteria

The inclusion criteria comprised patients aged 18 years or older with a diagnosis of ST-segment elevation MI confirmed by clinical, biochemical, and electrocardiographic criteria who completed an LGE CMR viability study two to four weeks post-MI.

Exclusion criteria included patients with contraindications to CMR, such as implanted metallic devices, incomplete or poor-quality imaging data, previous history of coronary artery bypass grafting (CABG) or other significant cardiac surgeries before the index MI and severe comorbidities that could impact survival or imaging outcomes. Patients were also excluded if they had undergone other forms of viability assessment, such as positron emission tomography (PET), to avoid confounding results.

CMR imaging acquisition

CMR acquisition using the Sola MRI, a 1.5 Tesla XA20 with a 32-channel setup, involves a detailed protocol to ensure high-quality imaging, particularly for LGE. Initially, contrast is administered with a gadolinium-based agent at a dose ranging from 0.1 to 0.2 mmol/kg. Imaging begins in multiple planes, including four-chamber, two-chamber, and short-axis (SA) views. The field of view (FOV) is set to 300 mm, with a slice thickness of 6 mm across six slices. Following contrast administration, initial images are captured within 1 to 3 min to assess early contrast distribution with an inversion time (TI) greater than 400 ms. Subsequently, LGE images are acquired more than 10 min post-contrast to highlight areas of myocardial fibrosis or scarring, ensuring optimal tissue characterization and diagnostic accuracy.

Data preprocessing by CVi42

After acquiring the CMR on the Sola MRI (1.5 Tesla XA20, 32-channel), the digital imaging and communications in medicine (DICOM) images are exported for further analysis. The images are imported into CVi42 (Circle Cardiovascular Imaging Inc, Calgary, Canada) for the assessment of global longitudinal strain (GLS), global circumferential strain (GCS), and global radial strain (GRS). These parameters can detect subtle abnormalities in cardiac mechanics. Additionally, CVi42 supports T1 and T2 mapping, which are vital for tissue characterization. T1 mapping enables the assessment of myocardial fibrosis and extracellular volume, while T2 mapping is used to detect edema and inflammation. These advanced analyses help diagnose various cardiac conditions and monitor disease progression or response to therapy. Images were excluded from the study if they were suboptimal or non-diagnostic due to artifacts or signal loss that impaired accurate interpretation by the observers.

T1 mapping

T1 mapping is a crucial parameter for myocardial tissue characterization in CMR imaging. It is a promising tool for detecting myocardial abnormalities without gadolinium-based contrast agents. The myocardial T1 relaxation time reflects changes in the intracellular and extracellular components of the myocardium. Various pathologies affect the T1 values, such as collagen deposition (myocardial fibrosis), water (edema and inflammation), lipids (Fabry disease), and iron (hemosiderosis). T1 maps are generated from a series of T1-weighted images, with each pixel representing a specific T1 value, allowing precise quantification of tissue characteristics. T1 mapping can be conducted using several acquisition methods, with the modified look-locker inversion recovery (MOLLI) sequence commonly used, as in our study. At least three SA images are acquired from the base to the apex of the heart. Dedicated CMR software is used for quantifying T1 map images, where contours can be drawn manually or using automated features. Native T1 values are obtained by manually drawing a region of interest (ROI) on specific myocardial areas or by presenting the data in an AHA segmentation format. Normal myocardial native T1 reference ranges acquired using the MOLLI method have been reported to be 930 ± 21 ms at 1.5 Tesla (Fig. 2).

Fig. 2
Fig. 2The alternative text for this image may have been generated using AI.
Full size image

Viability assessment in a 57-year-old man with LAD territory MI in STEMI. (A) Transmural LGE in base to midpart of the anteroseptal wall of LV. (B) Contouring of the LV wall for T1 and T2 evaluation. (C) High T1 values (T1 = 1043 ms) in the base to mid-segment of the anteroseptal LV wall due to septal fibrosis. (D) Polar map of T1-weighted LV wall showing an increased T1 value in the septal wall. (E) Polar map of T2-weighted LV wall showing normal T2 values in the septal wall (T2 = 48 ms).

T2 mapping

Similar to T1 values, T2 values represent a global signal from both intracellular and extracellular compartments. Each tissue type has a normal range of T2 values, and increased T2 values typically indicate elevated water content. Thus, myocardial edema (inflammation) is the main pathology resulting in increased T2 values. T2 mapping sequences are instrumental in detecting myocardial edema in patients with acute MI and myocarditis. Normal myocardial T2 values acquired using steady-state free precession (SSFP)-based sequences have been reported to be 52.18 ± 3.4 ms at 1.5 Tesla (Fig. 2).

Feature tracking for strain analysis

Myocardial strain measures the deformation of the myocardium between relaxed and contracted states12. In ischemic heart disease, strain outperforms known risk markers like left ventricular (LV) ejection fraction or infarct size in predicting major adverse cardiac events13.

Several CMR strain techniques are available, including Tagging, displacement encoding with stimulated echoes (DENSE), and strain encoding (SENC), which require dedicated sequences. However, the most commonly used technique is FT, which processes routinely acquired cine SSFP MR images13,14. FT post-processing involves defining the endocardial and epicardial borders on short- and long-axis cine images and delineating the ventricular long axis at end-systole and end-diastole13,15 This allows the measurement of circumferential, longitudinal, and radial strain parameters (GLS, GCS, GRS). These steps can be done rapidly using artificial intelligence (AI) features of post-processing software.

Strain values can be displayed for the entire heart (global), at the section level (basal, mid, apical), by AHA segments, or within a specific ROI. Polar maps can display strain metrics in all 17 segments, and color-coded strain data can be overlaid on cine SSFP images for enhanced visualization (Fig. 3).

Fig. 3
Fig. 3The alternative text for this image may have been generated using AI.
Full size image

A 57-year-old man with LAD territory MI. (A) Imaging depicting color-coded strain at mid-SAX. (B) Graph showing reduced circumferential strain in a specific ROI (mid-anteroseptal LV wall). (C) Polar map of circumferential strain values.

Statistical analysis

Excluding or imputing missing values

Prior to conducting the machine learning analysis, we evaluated T1 and T2 mapping alongside strain feature tracking to maintain optimal data quality. Patients were excluded from the analysis if more than one segment within a given territory contained missing values. For cases with a single missing value in T1, T2, longitudinal, radial, or circumferential strain, the missing value was imputed by averaging the values of adjacent segments within the same territory. This approach replaced missing data with the mean of available values from the corresponding group, as demonstrated below.

$$\:{\widehat{\text{x}}}_{\text{i}\text{j}}=\:\frac{1}{{\text{n}}_{\text{i}}}\sum\:{\text{x}}_{\text{i}\text{j}}\:\ne\:\text{N}\text{A}$$

\(\:{\widehat{\text{x}}}_{\text{i}\text{j}}\): imputed value for the missing value, \(\:{\text{n}}_{\text{i}}\): the number of non-missing values in group i, \(\:{\text{x}}_{\text{i}\text{j}}\): the observed value in group i variable j, NA denotes missing values.

Normalization

Normalization using StandardScaler in Python involves transforming the features of a dataset so that they have a mean of 0 and a standard deviation of 1. This standardization ensures that each feature contributes equally to the analysis, which can improve the performance of many ML algorithms. The normalization formula applied is:

$$\:\text{z}=\frac{x-\mu\:}{\sigma\:}$$

\(\:x\) represents the original feature value, µ is the mean of the feature, and σ is the standard deviation of the feature. This process adjusts the data to a common scale without distorting differences in the ranges of values, ensuring consistent and reliable analysis.

Synthetic minority over-sampling technique (SMOTE)

The synthetic minority over-sampling technique (SMOTE) is a technique for addressing class imbalance in datasets, particularly where one class predominates, potentially skewing classifier performance. SMOTE works by selecting a minority class instance, identifying its k-nearest minority class neighbors, and creating synthetic instances along the line segments connecting the instance and its neighbors. The formula for generating a synthetic instance is:

$$\:{X}_{new}={X}_{i}+\lambda\:\left({X}_{j}-{X}_{i}\right)$$

Here, Xi represents a minority class instance, Xj denotes one of its k-nearest neighbors (k = 5), and λ is a random number between 0 and 1.

In our study, SMOTE was applied to ensure the training dataset was balanced, thereby improving the classifier’s ability to handle imbalanced data effectively.

Data set splitting and ten-fold cross-validation

A 10-fold cross-validation approach was employed to ensure robust evaluation and prevent overfitting. This method involved dividing the dataset into ten subsets, using nine for training and one for testing in each iteration, ensuring that each subset was used for testing only once.

$$\:{CV}_{error}=\frac{1}{k}\sum\:_{i=1}^{k}{error}_{i}$$

Machine learning algorithms

The following ML algorithms were implemented and compared: decision tree, gradient boosting, logistic regression, naive bayes, random forest (RF), support vector machine (SVM), K-Nearest Neighbors (KNN), and Neural Network (fully connected). Each algorithm was tuned using grid search to optimize hyperparameters and enhance performance.

Resources and requirements

For our machine learning tasks, we utilized an M2 MacBook Pro device powered by the Apple M2 chip, featuring an 8-core CPU and a 10-core GPU, with 16GB of unified memory, which efficiently handled data-intensive operations, ensuring smooth performance during model training and evaluation. It demonstrated stable and smooth performance for all machine learning tasks without any drops or interruptions.

Decision tree

Decision trees recursively partition the feature space into regions that minimize impurity, often measured by entropy for classification tasks.

Entropy Formula

$$\:\text{H}\left(\text{T}\right)=-\sum\:_{i=1}^{c}p\left(i|T\right){\text{log}}_{2}p\left(i|T\right)$$

H(T): Entropy of node T, p(i|T): Proportion of samples of class I in node T and c: number of classes (c = 2 in this study).

In this classifier, we used the criterion of Gini score and best splitter without class weight.

Gradient boosting

Gradient boosting combines weak learners sequentially to minimize a loss function, adjusting predictions with a learning rate of 0.001

$$\:{\widehat{y}}_{i}^{(t+1)}={\widehat{y}}_{i}^{\left(t\right)}+\eta\:.{h}^{\left(t\right)}\left({x}_{i}\right)$$

\(\:{\widehat{y}}_{i}^{\left(t\right)}\): Prediction of the ensemble at iteration t.

\(\:{h}^{\left(t\right)}\left({x}_{i}\right)\): Weak learner (e.g., decision tree) prediction for sample for xi.

Logistic regression

Logistic regression models the probability of a binary outcome using a logistic function.

Logistic function

$$\:\text{P}\left(y=1|x\right)=\frac{1}{1+{e}^{-\left({w}^{T}x+b\right)}}$$

w: Coefficients (weights), x: input features, b: Bias term.

The default configuration of LogisticRegression utilized the lbfgs solver, which is well-suited for small to medium-sized datasets. It employed L2 regularization (penalty=’l2’) with a regularization strength of C = 1. The class_weight parameter was set to None, indicating that all classes were treated equally unless specified otherwise. The optimization process was capped at a maximum of 100 iterations (max_iter = 100) to identify the optimal solution.

Loss function (binary cross-entropy)
$$\:\mathbf{L}\left(\varvec{w},\varvec{b}\right)=-\frac{1}{\varvec{N}}\sum\:_{\varvec{i}=1}^{\varvec{N}}\left[{\varvec{y}}_{\varvec{i}}\mathbf{log}\left(\varvec{P}\left(\varvec{y}=1|\varvec{x}\right)\right)+1-{\varvec{y}}_{\varvec{i}}\mathbf{log}\left(1-\varvec{P}\left(\varvec{y}=1|\varvec{x}\right)\right)\right]$$

\(\:\mathbf{L}\left(\varvec{w},\varvec{b}\right)\): Binary cross-entropy loss, yi: Actual label of sample i, P(y = 1|x): Predicted probability of calss 1 for sample I, N: number of samples.

Random forest

RF constructs an ensemble of decision trees where each tree is built on a random subset of the features and a random subset of the data. In the random forest, the Gini score is used as the criterion to make the optimum efficiency of the classifier, which is demonstrated as follows:

$$\:Gini=1-\sum\:_{i=1}^{c}{p}_{i}*\left({1-p}_{i}\right)$$

c: The number of classes in the target variable. p: The proportion (or probability) of class i in the node.

Support vector machine (SVM) with linear kernel and probability

SVM finds a hyperplane that separates classes in the feature space, with the linear kernel indicating a linear decision boundary.

Decision function for linear SVM\(\:f\left(x\right)={w}^{T}x+b\)

w: weight vector, b: Bias term.

Probability estimation

$$\:P\left(y=1|x\right)=\frac{1}{1+{e}^{-f\left(x\right)}}$$

In this experiment, we used a regularization parameter (c = 1.0), rbf kernel, and a decision function shape of ‘ovr’ (one-vs-rest) to optimize the classification process.

K-nearest neighbors (KNN)

KNN classifies samples based on the majority class among its k nearest neighbors in the feature space. In the present study, we considered the K = 5.

Decision Rule

$$\:\widehat{y}=arg\:max\:{y}_{j}\sum\:_{i\in\:{N}_{k}\left(x\right)\:}I\left({y}_{i}={y}_{j}\right)$$

\(\:\widehat{y}\): Predicted class for the new sample x, Nk(x) set of k nearest neighbors of x, I: indicator function.

Neural network

Neural networks (NN) consist of interconnected layers of neurons that learn representations of the data. In this study, the Multilayer Perceptron Classifier (MLPClassifier) was used with the following details:

Input to Hidden Layer Transformation

$$\:{h}_{j}=f\left(\sum\:_{i=1}^{c}{w}_{ij}{x}_{i}+{b}_{j}\right)\:\:$$

xi​: The i-th input feature. wij​: Weight connecting the i-th input to the j-th neuron. bj​: Bias term for the j-th neuron. f : ReLU Activation function.

Hidden Layer to Output Layer Transformation

$$\:{z}_{k}=g(\sum\:_{j=1}^{m}{w}_{ik}{h}_{j}+{b}_{k})$$

hj​: Output of the j-th neuron in the hidden layer. wj: Weight connecting the j-th hidden neuron to the k-th output neuron. bk​: Bias term for the k-th output neuron. g: Output activation function.

The predicted class is the one with the highest probability:

$$\:\widehat{y}=argma{x}_{k}\left({\widehat{y}}_{k}\right)$$

We set max_iter = 1000 to define the maximum number of iterations for the optimization algorithm to converge. Additionally, we configured the model with a hidden layer size of 100 (hidden_layer_sizes=(100)), used ‘relu’ as the activation function, selected the Adam solver for optimization, and applied a constant learning rate to regulate the training process.

Evaluation metrics

The performance of the ML was assessed using several evaluation metrics, including area under the curve (AUC), Precision, recall, accuracy, and F1 score. These metrics provided a comprehensive evaluation of the models’ predictive capabilities.

AUC measures the overall performance of the classification model by evaluating the area under the receiver operating characteristic (ROC) curve. Although AUC itself does not have a simple closed-form formula, it is computed as the area under the ROC curve, which plots the true positive rate (TPR) against the false positive rate (FPR) at various threshold settings.

Precision measures the proportion of true positives among the predicted positives, and is calculated using the formula:

$$\:\text{Precision}=\frac{\text{TP}}{\text{TP+FP}}$$

Recall, also known as sensitivity or true positive rate, indicates the proportion of true positives among all actual positives. The formula for recall is:

$$\:\text{Recall=}\frac{\text{TP}}{\text{TP+FN}}$$

Accuracy reflects the overall correctness of predictions and is calculated using the formula:

$$\:\text{Accuracy=}\frac{\text{TP+TN}}{\text{TP+TN+FP+FN}}$$

The F1 Score provides a balance between precision and recall by considering the harmonic mean of both metrics. The formula for the F1 score is:

$$\:{\text{F1 = 2}} \times \frac{{{\text{Precision}} \times {\text{Recall}}}}{{{\text{Precision + Recall}}}}$$

These metrics together give a comprehensive picture of the models’ performance, with AUC assessing overall performance, precision measuring the accuracy of positive predictions, recall indicating the model’s ability to capture actual positives, Accuracy reflecting the correctness of predictions, and F1 Score balancing the trade-off between precision and recall.

Feature selection and ablation study

Feature importance was evaluated to identify the most influential parameters in predicting myocardial viability. Various techniques, such as permutation importance and SHapley Additive exPlanations (SHAP) values, were used to rank features. An ablation study was conducted to determine the impact of individual features on model performance. This involved systematically removing one feature at a time and observing changes in the evaluation metrics, thus highlighting the contribution of each feature to the model’s predictive accuracy.

Results

In this study, 90 patients were enrolled at the beginning, but 11 patients were excluded due to a history of previous ischemia or CABG, lack of imaging quality, inappropriate T1 and T2 mapping, and strain analysis. The remaining 79 patients underwent statistical and ML analysis. Among the enrolled patients, according to the LGE viability assessment, the left anterior descending artery (LAD) territory was viable in 52.8% of participants (47 individuals). The right coronary artery (RCA) territory was viable in 59.5% of participants (53 individuals). The left circumflex artery (LCX) territory was viable in 60.6% of participants (54 individuals). (Fig. 4; Table 1).

Fig. 4
Fig. 4The alternative text for this image may have been generated using AI.
Full size image

Flowchart depicting the enrolled study population and the distribution of viable territories.

Table 1 Distribution of viable myocardial segments across coronary territories.

Demographic characteristics of patients

The study involved participants with a mean age of 47.97 years with a standard deviation of 12.48 years. Among the participants, 83.5% (66 individuals) were male. Regarding medical conditions, 43.0% (34 individuals) had hypertension, 32.9% (26 individuals) had diabetes, and 26.5% (21 individuals) had hyperlipidemia. Additionally, 35.4% (28 individuals) were smokers.

For functional indexes, the mean value of left atrium area was 23.56 ± 11.97 cm2. The mean value of right atrium was 19.99 ± 18.31 cm2. The mean value of left ventricular ejection fraction (LVEF) was 28.13 ± 10.26. The mean value of end-diastolic volume index (EDVI) was 131.83 ± 62.43 ml/m2, while the mean value of end-systolic volume index (ESVI) was 98.95 ± 54.55. ml/m2. The mean value of left ventricular stroke volume (LVSV) was 58.91 ml ± 17.76.

Regarding the right ventricle, the mean value of right ventricular ejection fraction (RVEF) was 44.64 ± 12.02. The mean value of right ventricular end-diastolic volume index (RVEDVI) was 78.67 ml/m2 ± 38.27. The mean value of right ventricular end-systolic volume index (RVESVI) was 45.52 ml/m2 ± 31.30. Lastly, the mean value of right ventricular stroke volume (RVSV) was 54.73 ml ± 16.97. Table 2 summarizes the complete demographic and functional metrics.

Table 2 Demographic characteristics and anatomical and functional metrics of the study population.

Data adjusting

Analysis of individuals with non-viable versus viable territories in LAD, RCA, and LCX showed no significant differences in clinical factors (hypertension, diabetes, hyperlipidemia, smoking) or imaging parameters (atrial areas, ventricular functions, volumes). Only age in RCA territories approached significance (P = 0.05), with viable territories showing slightly higher mean age (Table 3). In LAD territory, non-viable segments showed reduced radial and circumferential strain with elevated T1 and T2 values. RCA territory demonstrated only higher T2 values in non-viable areas. LCX territory exhibited lower strain values and higher T1 and T2 values in non-viable segments (Table 4; Fig. 5).

Table 3 Comparison of the demographic characteristics and conventional imaging findings between viable and non-viable territories.
Table 4 T1/T2 mapping and feature tracking analysis for non-viable versus viable territories in the LAD, RCA, and LCX regions.
Fig. 5
Fig. 5The alternative text for this image may have been generated using AI.
Full size image

Alteration of features in non-viable versus viable areas for each coronary territory. (A) LAD, (B) RCA, and (C) LCX.

Evolution metrics of various models in the prediction of viability in each territory

In the LAD territory, the RF model demonstrated the highest AUC (0.87) and strong overall performance with accuracy, balanced accuracy, and F1 score of 0.74. Logistic regression also performed well, achieving an AUC of 0.83, the highest accuracy (0.78), and balanced accuracy (0.77). While the KNN model had a relatively high AUC (0.82) and accuracy (0.75), its recall was lower (0.65), though it had high precision (0.84). The NN, SVM, gradient boosting, and decision tree models showed moderate performance with varying strengths, with NN and SVM models achieving balanced accuracy and F1 scores of around 0.70–0.74. Gradient boosting and decision tree had the lowest AUCs (0.73 and 0.72, respectively), with corresponding lower metrics in other categories.

For the RCA territory, the RF model performed the best, achieving the highest AUC of 0.90 and strong overall metrics, including an accuracy of 0.79 and an F1 score of 0.78. The NN model also performed well with an AUC of 0.84, the highest accuracy of 0.81, and a balanced accuracy of 0.81, although its recall was lower at 0.69. The KNN model showed a lower AUC of 0.80 and an accuracy of 0.70, with a particularly low recall of 0.54. Logistic regression and SVM had moderate performance with AUCs of 0.76 and 0.74, respectively, and balanced accuracy around 0.72–0.73. Gradient boosting and decision tree had the lowest AUCs of 0.74 and 0.72, respectively, with corresponding lower performance metrics in accuracy and F1 score.

For the LCX territory, both the RF and NN models achieved the highest AUC of 0.92. The NN model also had the highest accuracy (0.87), balanced accuracy (0.87), recall (0.83), precision (0.90), and F1 score (0.86). The RF model followed closely with an accuracy of 0.83, balanced accuracy of 0.83, and F1 score of 0.81. Logistic regression and SVM models also performed well, with logistic regression achieving an AUC of 0.91, accuracy of 0.84, and an F1 score of 0.83, while the SVM had an AUC of 0.89, accuracy of 0.83, and an F1 score of 0.82. The KNN model had a lower AUC of 0.87 and an accuracy of 0.74, with a recall of 0.62. Gradient boosting and decision tree showed the lowest performance with AUCs of 0.76 and 0.74 respectively, and corresponding lower metrics across accuracy, balanced accuracy, and F1 scores (Table 5; Fig. 6).

Table 5 Evaluation metrics of various models used for predicting viability in each coronary territory.
Fig. 6
Fig. 6The alternative text for this image may have been generated using AI.
Full size image

AUC curves of various machine learning models for predicting LGE.

Model selection

Among the studied models, RF, KNN, and logistic regression demonstrated superior performance in assessing myocardial viability in the LAD territory. For the RCA territory, RF, NN, and KNN emerged as the top performers, while RF, NN, and logistic regression excelled in the LCX territory.

Analysis of feature importance, as detailed in Table 6; Fig. 7, revealed territory-specific patterns in predictive parameters. In the LAD territory, the most influential features were dominated by T2 mapping parameters from segments 13 and 14, along with segmental radial strain measurements from segments 7 and 1. The RCA territory showed a strong dependence on mapping parameters, with T2 mapping from segments 9 and 3, and T1 mapping from segment 9 emerging as key predictors. The LCX territory exhibited distinct characteristics where strain measurements played a crucial role, particularly the segmental circumferential strain of segment 11, T2 mapping of segment 6, and segmental longitudinal strain of segment 5 (Table 7).

Table 6 Normalized feature importance and the selection of the top 10 most important features based on the three best-performing models.
Fig. 7
Fig. 7The alternative text for this image may have been generated using AI.
Full size image

Nomogram importance of T1/ T2 mapping and feature tracking analysis in each segment compared to other segments and features.

Table 7 Importance of T1/ T2 mapping and feature tracking analysis in each segment compared to other segments and features.

SHAP evaluation

SHAP evaluation was conducted to identify the most important features influencing the prediction of LGE. The obtained results align closely with the average feature importance reported in the previous section (Fig. 8).

Fig. 8
Fig. 8The alternative text for this image may have been generated using AI.
Full size image

SHAP analysis shows the importance of each feature in the top three best-performing models for predicting LGE across coronary territories.

Discussion

In ischemic heart disease, detecting areas beyond saving is just as important as finding salvageable tissue. Thus, viability studies are now an integral part of the IHD management armamentarium16,17. Current viability studies are based on LGE as a representative marker of fibrosis and subsequent extracellular volume expansion. Other imaging properties such as T1/T2 mapping and strain analysis theoretically show changes in concert with tissue replaced by fibrosis, making them potential markers for viability assessment18,19,20. In contrast to LGE, in order to measure these markers, contrast injection is not needed, which enables viability study to be performed faster and safer, especially in individuals with IHD who are at increased risk of concomitant renal disease and are more likely to be unable to undergo prolonged imaging modalities21,22. In the current study, by employing multiple ML models, we tried to extract the best features for predicting the myocardial tissue viability–determined by LGE–from strain and T1 and T2 mapping values. This could provide a basis for building sophisticated prediction models of myocardial tissue viability without LGE sequences.

Prior research has shown promising results in using non-contrast CMR images as an alternative to LGE-CMR for detecting myocardial infarction, eliminating the need for gadolinium injection. Several key studies have demonstrated this potential; Baessler et al. assessed 72 patients (52 with myocardial infarction, 20 healthy controls) using radiomics features and logistic regression models, achieving an AUC of 0.93 for large infarctions and 0.92 for small infarctions9. Zhang et al.‘s study of 843 myocardial infarction patients using a deep learning virtual native enhancement model showed strong correlation with LGE in scar quantification (R = 0.89) and 84% accuracy in detecting scars2. Additional research includes Chen et al.‘s study of 150 patients using a combined physiological-clinical ML model, achieving 88.67% classification accuracy23, and Xu et al.‘s work with deep spatiotemporal generative adversarial networks on cine-CMR images, reaching 96.98% pixel classification accuracy24. Abdulkareem et al. analyzed 272 patients (108 with infarction, 164 controls) using U-Net and ResNet50 models, with SVM achieving accuracy scores of 0.68 and decision trees reaching 0.6225.

However, non-contrast assessment of myocardial viability across different segments and coronary territories has been limited to isolated parameter analysis and small sample sizes. Two notable studies explored this area: Dastidar et al. demonstrated excellent diagnostic accuracy using native T1 mapping compared to LGE (AUC of 0.88 in chronic MI and 0.83 in acute MI) in 60 patients7, while Tantawy et al. showed moderate diagnostic accuracy using Feature Tracking strain analysis (AUC of 0.7 for circumferential strain and 0.67 for radial strain) in 30 chronic ischemic patients6.

Hence, for the first time our study aimed to evaluate the combined efficacy of FT strain analysis and T1/T2 mapping integrated with ML pattern recognition techniques for predicting myocardial viability at both segmental and coronary artery territory levels. Notably, the assessment of coronary artery territory viability has not been addressed in prior studies. This analysis provides crucial information for clinical decision-making regarding revascularization strategies and risk stratification.

We found that within the LAD territory, RF, logistic regression, and KNN showed the highest discriminatory power for viability. The ten most valuable features extracted comprise various segmental radial and circumferential peak strains and T2 mapping, along with global T2 mapping. Concerning LCX territory, RF, NN, and logistic regression were the most fruitful methods. Circumferential strains of all segments were within the ten best features, while only one segment’s radial and longitudinal strain was within these features. Basal T1 and basal and mid T2 were also found to be within the most effective features for viability assessment. Regarding segments supplied by RCA, RF, NN, and KNN were found to be superior. The ten most effective features only comprised T1 and T2 mapping values except for the circumferential strain of one mid-segment.

Similar to prior literature, we found the highest discriminatory value between viable and non-viable regions were circumferential and radial strains rather than longitudinal ones6. GCS and GRS showed the highest correlations with LVEF, while GLS was the least correlated global strain6,26. Moreover, the circumferential global strain is the most reproducible global strain6,27. Thus, taking circumferential strains into account for viability prediction might be more practical compared to other strain values, especially in LAD territory.

Another merit of segmental peak strains is their discriminatory value for myocardial viability assessment of their respective segments6. To the best of our knowledge, this study is the first to pursue strain and T1 and T2 mapping features to determine the myocardial viability according to coronary circulation territories. Evaluating the segmental strains and T1 and T2 mapping values based on their respective coronary territory theoretically could enhance myocardial tissue viability prediction. We found that the most valuable features differ for each coronary territory. The best extracted features for RCA territory were mostly T1 and T2 mapping, while those of LAD and LCX were mostly segmental peak strains. Possible explanation for this difference is the fact that strain values of RCA territory which is mostly comprised of right ventricle, cannot be as precisely and accurately measured as LCX and LCA territories28.

A number of studies pursued cut-points for peak strain values to discriminate viable and non-viable myocardium6,29. However, we tried to build a basis to develop a ML model composed of the best discriminatory features, taking arterial territories into account in order to attain higher accuracy. Given the complexity of myocardial tissue hibernation/stunning and viability pathophysiology, the ultimate model hypothetically surpasses the discriminatory power of a single feature.

The findings of this study demonstrate the potential of ML techniques guided by FT strain analysis and T1/T2 mapping, to predict myocardial viability with high accuracy. This approach offers a non-contrast alternative to the traditional LGE-based method, addressing some of the limitations associated with LGE, such as contrast agents dependency and inability to enlighten functional features of myocardial tissue.

Our results indicate that ML models, particularly gradient boosting and RF, achieved high performance metrics in predicting myocardial viability, with AUC, precision, recall, accuracy, and F1 scores comparable to those obtained with LGE30,31. This finding is consistent with previous studies that have highlighted the effectiveness of ML algorithms in cardiac imaging and viability assessment32,33. The integration of FT strain analysis and T1/T2 mapping provides a more comprehensive evaluation of myocardial tissue characteristics, potentially enhancing the diagnostic capability beyond what is achievable with LGE alone34,35.

The deep learning model, utilizing a 5-layer fully connected NN, also demonstrated strong predictive performance. This supports the growing body of evidence that deep learning techniques can effectively handle complex, high-dimensional data and uncover intricate patterns beyond the traditional methods’ grasp36. Previous studies have shown that deep learning models can outperform conventional ML algorithms in various medical imaging tasks, including cardiac imaging. Our study adds to this literature by demonstrating the applicability of deep learning in the context of myocardial viability assessment.

One significant advantage of using FT strain analysis and T1/T2 mapping is the ability to provide functional and structural information about the myocardium without the need for contrast administration. This is particularly beneficial for patients with contraindications to gadolinium-based contrast agents as those with chronic kidney disease10. Moreover, these non-contrast techniques can be easily integrated into routine CMR protocols, potentially increasing their adoption in clinical practice and decrease health expenditure37. However, the application of ML and deep learning techniques requires substantial computational resources and expertise, which may limit their widespread use in clinical settings without appropriate support and infrastructure.

This study demonstrates that ML and deep learning techniques can effectively predict myocardial viability based on FT strain analysis and T1/T2 mapping data, offering a promising alternative to LGE. Future research should focus on the prospective validation of these findings, the integration of these techniques into clinical workflow, and the exploration of their potential to improve patient outcomes. This approach has the potential to provide a comprehensive, non-invasive, and patient-friendly alternative to current methods. Machine learning is increasingly essential in medical imaging diagnostics, yet it faces critical limitations that need addressing. These include the small sample size problem and the scarcity of labeled data, inefficiencies in managing multiple optimization objectives in complex diagnostic scenarios, and challenges in real-time processing of streaming medical data. Although recent studies have made significant strides in overcoming these issues, further research is necessary38,39,40,41. Besides, Deep learning architectures, especially convolutional neural networks (CNNs), have demonstrated remarkable capabilities in automatically extracting relevant features from medical images, often matching or exceeding human-level performance in specific diagnostic tasks42,43,44, which could be further used in myocardial viability assessment.

Limitations

The primary limitations of our study stem from the inherent biases associated with its retrospective design and the quality of the data collected retrospectively. Although the use of SMOTE helped mitigate class imbalance, it could not fully resolve this issue. Therefore, future prospective studies with larger, more diverse populations, utilizing a variety of MRI scanners and post-processing software, are essential to validate our findings. Notably, in this study, myocardial viability was categorized into two groups: viable and non-viable, with remote myocardial regions included in the viable category. Furthermore, the involvement of radiology specialists remains critical for accurately measuring strain and delineating ROI when applying this model.

Conclusion

In conclusion, the study demonstrated that RF and NN models provide the highest accuracy in predicting myocardial viability across different coronary territories, with notable performance in RCA and LCX regions. The integration of non-contrast imaging techniques, including T1/T2 mapping and strain analysis, proves valuable in enhancing viability assessment. Key features from these imaging techniques significantly contribute to predictive modeling, highlighting their potential to improve diagnostic accuracy for MI. This research underscores the promise of advanced imaging combined with ML to refine cardiac viability evaluation. Further studies should consider including a more diverse patient population from multiple centers and validating the promising findings of this study in various clinical settings, using multiple MR scanners and post-processing software. Additionally, performing radiomics analyses could potentially improve the model’s performance.