The project uses the Qwen2.5-7B-Instruct model by default and applies LoRA for fine-tuning during the SFT (Supervised Fine-Tuning) phase. This requires approximately 16GB of VRAM.
You may also use other models and methods supported by LLaMA Factory.
Full (bf16 or fp16)
bf16
fp16
32
120GB
240GB
600GB
1200GB
18xGB
18x
Full (pure_bf16)
pure_bf16
16
60GB
300GB
8xGB
8x
Freeze/LoRA/GaLore/APOLLO/BAdam
16GB
32GB
64GB
160GB
2xGB
2x
QLoRA
8
10GB
20GB
40GB
80GB
xGB
x
4
6GB
12GB
24GB
48GB
x/2GB
x/2
2
4GB
8GB
x/4GB
x/4
Last updated 8 months ago