Table 4 Comparaison of Distil-BERT-MRC performance with State-of-the-Art Methods.

From: Resolving passage ambiguity in machine reading comprehension using lightweight transformer architectures

Year

Ref

Model

Dataset

Exact match

F1 score

Key observations

2022

49

DistilBERT

Covid data

80.6

87.3

Achieved high accuracy on domain-specific data with efficiency.

2023

50

BERT

SQuAD 2.0

72.34

60.45

Larger model; resource-intensive and slower inference.

2023

51

RoBERTa

MultiRC

68.78

72.12

Demonstrates strong contextual understanding in multi-sentence reasoning tasks.

2024

52

XLNet

NewsQA

65.14

70.89

Excels in paragraph-level QA but require significant computational resources.

2024

53,54

GPT-3

TriviaQA

85.23

88.67

Outstanding performance demands high computational power.

2024

15

Parallel-DistilBERT

SQuAD 1.1

86.20

88.36

Enhanced exact match and F1-score with 31% reduced parameters.

2025

35

Few-shot-Neural-MRC

Chinese comprehension corpus

86.42

74.65

Prepared Chinese MRC dataset and set the baseline for further MRC research studies.

-

Proposed

Distil-BERT-MRC

WikiQA

90.23

91.42

Achieved state-of-the-art performance with minimal computational cost.