Table 3 Summary of existing research studies employing DL models for clickbait detection.

From: Clickbait detection in news headlines using RoBERTa-Large language model and deep embeddings

Ref

Models

Dataset

Features

Results

Limitation

20

LSTM

News Headlines

POST

97

Limited to linguistic and POS features.

21

CNN

News headlines

Glove

95

Assessed only on a specific dataset, reducing generalizability.

22

LSTM

English news

Word embeddings

94

Combined DL models but lacked interpretability and scalability evaluation.

23

BERT

Webis-Clickbait-17

Word2vec

85

Computationally expensive and resource heavy.

24

ALBERT

Telugu clickbait headlines

Word2vec, Glove, fast text

93

Limited to Telugu dataset.

25

XGB

Twitter and Instagram post

FastText

88

Stacking classifiers effective but complex and less interpretable.

26

C-LSTM

News Headlines of US publishers

Word Embeddings

98

Proposed extension limited to malicious.

27

BERT

Indonesian news headlines

Sentence embedding

91

Lacks cross-domain testing.

28

SBERT

Webs Clickbait Corpus 2017

Sentence embedding

84

dataset specific.

29

BiLSTM

Romanian Clickbait Corpus

Word2Vec

89

Ensemble methods effective but dataset (Romanian) small and limited.

30

BiLSTM

News headlines clickbait

Word2Vec

93

Bi-LSTM effective but dataset size limited and domain-specific.

31

BERT

WELFake, Fake Newsnet

Fast Text

97

Lacks explanation of decision process.

32

DNN

Corpus fake news

Glove

98

Limited evaluation scope.

33

BERT

Fake news detection

Word2Vec embedding

84

Computationally heavy and dataset specific.

34

CNN

Arabic dataset.

Word2Vec embedding

77

Results are tied to specific optimizers and dataset.

35

XLNet

Tweets on 2020 U.S. election

Transformer-based embeddings

87

Domain-specific limited to text-only, scalability and real-world deployment not fully addressed