Table 3 Performance of models built based on different ML algorithms.

From: Development and interpretation of a machine learning risk prediction model for post-stroke depression in a Chinese population

Data sets

Models

Accuracy (95%CI)

Sensitivity (95%CI)

Specificity (95%CI)

F1 score (95%CI)

Training data

LR

0.821(0.806–0.837)

0.873(0.843–0.904)

0.798(0.763–0.833)

0.753(0.743–0.764)

LightGBM

0.800(0.759–0.840)

0.759(0.649–0.870)

0.818(0.756–0.880)

0.700(0.634–0.767)

GNB

0.810(0.791–0.828)

0.851(0.820–0.882)

0.791(0.751–0.830)

0.737(0.724–0.749)

XGBoost

0.964(0.948–0.979)

0.970(0.962–0.977)

0.961(0.937–0.984)

0.944(0.921–0.966)

AdaBoost

0.899(0.883–0.915)

0.926(0.901–0.951)

0.887(0.859–0.915)

0.851(0.833–0.870)

SVM

0.865(0.860–0.871)

0.880(0.855–0.904)

0.859(0.840–0.878)

0.803(0.799–0.806)

Validation data

LR

0.801(0.770–0.832)

0.848(0.797–0.899)

0.779(0.716–0.842)

0.728(0.703–0.753)

LightGBM

0.765(0.747–0.784)

0.702(0.585–0.819)

0.794(0.744–0.844)

0.647(0.602–0.692)

GNB

0.795(0.772–0.818)

0.817(0.749–0.885)

0.785(0.757–0.813)

0.712(0.678–0.746)

XGBoost

0.876(0.854–0.897)

0.822(0.717–0.926)

0.899(0.848–0.950)

0.802(0.762–0.843)

AdaBoost

0.864(0.831–0.897)

0.867(0.816–0.917)

0.862(0.817–0.908)

0.8(0.756–0.844)

SVM

0.811(0.784–0.837)

0.816(0.699–0.932)

0.808(0.766–0.850)

0.725(0.668–0.781)

  1. AUC: Area under the curve; XGBoost: eXtreme gradient boosting; LR: Logistic regression; LightGBM: Light gradient boosting machine; SVM: Support vector machines; GNB: Gaussian naive bayes.