Table 3 Hyperparameter settings for the LSTM multi-attention model.

From: A hybrid super learner ensemble for phishing detection on mobile devices

Hyperparameter

Value

Maximum sequence length

75

Vocabulary size

\(|word\_index| + 1\)

Embedding dimension

100

Number of CNN filters

64

CNN filter sizes

[3, 4, 5]

LSTM units

128

Number of attention heads

4

Attention key dimension

32

Dropout rate

0.2

Loss function

Binary cross-entropy

Optimizer

Adam

Train-test split

80–20%