Table 6 Final training parameter settings

From: Evaluating clinical AI summaries with large language models as judges

Model

Strategy

Batch Size

Learning Rate

Beta

Mixtral 8 × 22B

SFT

16

1e-7

Mixtral 8 × 22B

DPO

1

1e-7

0.7

Llama 3.1 8B

SFT

32

1e-4

Llama 3.1 8B

DPO

1

1e-7

0.5