* Fintune ChatGLM with Deepspeed Zero3 LoRA * add deepspeed zero3 config * rename config * remove offload_param * add save_checkpoint parameter * Update lora_deepspeed_zero3_finetune_chatglm3_6b_arc_2_card.sh * refine
16 lines
387 B
JSON
16 lines
387 B
JSON
{
|
|
"zero_optimization": {
|
|
"stage": 3,
|
|
"contiguous_gradients": true,
|
|
"overlap_comm": true,
|
|
"offload_optimizer": {"device": "cpu"}
|
|
},
|
|
"bf16": {
|
|
"enabled": true
|
|
},
|
|
"world_size":2,
|
|
"train_batch_size": 2,
|
|
"train_micro_batch_size_per_gpu": 1,
|
|
"gradient_accumulation_steps": 1,
|
|
"stage3_gather_16bit_weights_on_model_save":true
|
|
}
|