Table 4 Database for prediction of the ultimate bearing capacity.

From: A comparative machine and deep learning approach for predicting ultimate bearing capacity of shallow foundations in cohesionless soil

Algorithm

Type

Strengths

Weaknesses

Best use cases

Performance

Computational cost

GPR70,56

Probabilistic

Probabilistic predictions with uncertainty estimates. Handles non-linear relationships. Strong theoretical foundation

Computationally expensive. Sensitive to kernel choice. Poor scalability for large datasets

Small datasets with uncertainty quantification. Scientific modeling, Bayesian optimization

High accuracy for small datasets, but slow for large datasets

High

XGBoost80,60

Ensemble (Boosting)

High accuracy and scalability. Handles missing values. Regularization to prevent overfitting

Requires careful hyperparameter tuning. Computationally expensive for very large datasets

Large datasets, complex prediction tasks

State-of-the-art performance for structured data

High

LGBM74,75

Ensemble (Boosting)

Extremely fast and memory efficient. Handles large datasets. Supports categorical features

May overfit on small datasets. Requires tuning for optimal performance

Large datasets, high-dimensional data, ranking tasks

Faster than XGBoost, with comparable accuracy

Moderate

GBM42,47

Ensemble (Boosting)

High accuracy. Handles non-linear relationships. Robust to outliers

Computationally expensive. Prone to overfitting without tuning

Diverse datasets, complex prediction tasks

High accuracy but slower than XGBoost and LGBM

High

RF42,69

Ensemble (Bagging)

Robust to overfitting. Handles high-dimensional data. Provides feature importance

Less interpretable than single trees. Computationally expensive for large datasets

Diverse datasets, feature selection

High accuracy, but slower than boosting methods

Moderate to High

CatBoost77,78

Ensemble (Boosting)

Handles categorical data natively. Robust to overfitting. GPU support for fast training

Requires tuning. Less flexible for custom loss functions

Tabular data with categorical features, click-through rate prediction

High accuracy, especially for categorical data

High

AdaBoost79,72

Ensemble (Boosting)

Improves weak learners. Simple to implement. Robust to overfitting

Sensitive to noisy data. Computationally expensive for large datasets

Binary classification, face detection

Good accuracy for small datasets but struggles with noise

Moderate

KNN65

Instance-based

Simple and intuitive. No training phase. Handles non-linear data

Computationally expensive for large datasets. Sensitive to feature scaling. Struggles with high-dimensional data

Image recognition, recommendation systems

Good for small datasets, but slow for large datasets

Moderate to High

BR47,76

Ensemble (Bagging)

Reduces variance and overfitting. Combines multiple models for robustness. variance, improves generalization, robust to noise

Computationally expensive. Limited bias reduction

Regression tasks, noise reduction

Improves stability and accuracy of base models

Moderate to High

DT73,65

Non-parametric

Easy to interpret and visualize. Handles mixed data types. No need for feature scaling

Prone to overfitting. High variance. Struggles with extrapolation

Classification and regression, interpretable models

Good for small datasets, but prone to overfitting

Low to Moderate

SVR55,81

Kernel-based

Handles non-linear relationships using kernels. Robust to outliers. Provides probabilistic predictions

Computationally expensive. Requires careful hyperparameter tuning. Poor scalability for large datasets

Regression tasks with non-linear relationships, outlier-resistant modeling

High accuracy for small datasets, but slow for large datasets

High