Table 1 Comparison scores of our method and other state-of-the-art methods on the NWPU-Captions dataset.

From: Feature refinement and rethinking attention for remote sensing image captioning

Methods

BLEU1

BLEU2

BLEU3

BLEU4

METEOR

ROUGE_L

CIDEr

SAT

0.7340

0.6120

0.5280

0.4690

0.3370

0.6010

1.1090

SM-Att

0.7390

0.6160

0.5340

0.4690

0.3380

0.5950

1.1370

Struc-Att

0.7442

0.6091

0.5193

0.4557

0.3087

0.6064

1.2270

MLCA-Net

0.7450

0.6240

0.5410

0.4780

0.3370

0.6010

1.2640

MC-Net

0.7410

0.6260

0.5440

0.4780

0.3470

0.6110

1.1590

VRTMM

0.8116

0.7033

0.6213

0.5570

0.3660

0.6845

1.5885

GLCM

0.5536

0.4228

0.3353

0.2720

0.2789

0.5042

1.2774

P-to-H

0.7571

0.6291

0.5457

0.4828

0.3187

0.5858

1.2071

DiffNet

0.7972

0.6635

0.5604

0.4793

0.3059

0.6182

1.2324

Ours

0.8490

0.7620

0.6957

0.6441

0.4198

0.7468

1.8167

  1. Significant values are given in bold.