Table 5 QLoRA settings for training

From: Evaluating clinical AI summaries with large language models as judges

Model

Strategy

Quantization

LoRA Rank

LoRA Alpha

LoRA Drop-Out

Mixtral 8 × 22B

SFT

4-bit

8

8

0.1

Mixtral 8 × 22B

DPO

4-bit

8

8

0.1

Llama 3.1 8B

SFT

4-bit

64

256

0.0

Llama 3.1 8B

DPO

4-bit

8

8

0.1