Table 1 Search scope and description for hyperparameter optimization.

From: Decoding student cognitive abilities: a comparative study of explainable AI algorithms in educational data mining

Model

Hyperparameter

Description

Range

Lasso

‘alpha’

Regularization strength: controls model complexity and prevents overfitting.

0.001–10 (loguniform)

RF

‘n_estimators’

Number of trees in the forest; more trees can improve performance but increase computation.

50–300 (quniform, step 10)

‘max_depth’

Maximum depth of each tree; controls tree size and helps to avoid overfitting.

5–20 (quniform, step 1)

‘min_samples_split’

Minimum number of samples required to split an internal node.

0.1–1.0 (uniform)

XGBoost

‘n_estimators’

Number of boosting rounds; influences model accuracy and computation time.

50–300 (quniform, step 10)

max_depth

Maximum depth of a tree; deeper trees capture more complexity but risk overfitting.

3–10 (quniform, step 1)

‘learning_rate’

Step size shrinkage used to prevent overfitting; smaller values provide better generalization.

0.01–0.3 (uniform)

NN

‘hidden_layer_sizes’

Number of neurons in each hidden layer; affects model capacity and complexity.

50 or 100 (choice)

‘activation’

Activation function applied to neurons; affects non-linearity in model learning.

logistic, tanh, relu (choice)

‘alpha’

L2 regularization term; prevents overfitting by penalizing large weights.

0.0001–0.1 (loguniform)

‘max_iter’

Maximum number of iterations; controls model convergence.

300–1000 (quniform, step 100)

SVM

‘C’

Regularization parameter; controls the trade-off between smooth decision boundaries and correctly classifying training points.

0.001–100 (loguniform)

‘epsilon’

Epsilon-insensitive loss function; specifies the margin within which no penalty is given.

0–1