Table 2 Comparative analysis of prior models and identification of research gaps.

From: An attention-based multi-residual and BiLSTM architecture for early diagnosis of autism spectrum disorder

Study

Core Model

Architecture

Handles

Sequential/Temporal Dependencies?

Includes

Feature Attention/Weighting?

Parameter Count

(Approx.)

Identified Gap

Akter et al.

MobileNetV1

No

No

4.2M

Lacks focus on feature importance

and sequential relationships.

Alam et al.

Xception & ResNet50

No

No

25.6M

Standard architecture;

does not prioritize specific facial regions.

Cao et al.

ViT

Yes (Implicitly)

Yes (Self-Attention)

-

While powerful, ViT can be data-hungry

and computationally intensive.

Rabbi et al.

VGG19 & InceptionV3

No

No

143.7M

Relies on conventional feature extraction

without specialized focus.

Mujeeb Rahman et al.

Xception & EfficientNet

No

No

-

Focuses on architectural comparison,

not on modeling feature dependencies.

Our Proposed Model

Multi-Residual + BiLSTM + Attention

Yes (BiLSTM)

Yes (Attention Layer)

28.5M

Addresses all identified gaps in

a unified framework

with moderate efficiency.