LLM: add resume_from_checkpoint related section (#9705)

This commit is contained in:
binbin Deng 2023-12-18 12:27:02 +08:00 committed by GitHub
parent 320110d158
commit 12df70953e

View file

@ -97,7 +97,8 @@ bash qalora_finetune_llama2_7b_arc_2_card.sh
bash qalora_finetune_llama2_7b_pvc_1550_1_tile.sh bash qalora_finetune_llama2_7b_pvc_1550_1_tile.sh
``` ```
**Important: If you fail to complete the whole finetuning process, it is suggested to resume training from a previously saved checkpoint by specifying `resume_from_checkpoint` to the local checkpoint folder as following:** ### 4. (Optional) Resume Training
If you fail to complete the whole finetuning process, it is suggested to resume training from a previously saved checkpoint by specifying `resume_from_checkpoint` to the local checkpoint folder as following:**
```bash ```bash
python ./alpaca_qlora_finetuning.py \ python ./alpaca_qlora_finetuning.py \
--base_model "meta-llama/Llama-2-7b-hf" \ --base_model "meta-llama/Llama-2-7b-hf" \
@ -106,7 +107,7 @@ python ./alpaca_qlora_finetuning.py \
--resume_from_checkpoint "./bigdl-qlora-alpaca/checkpoint-1100" --resume_from_checkpoint "./bigdl-qlora-alpaca/checkpoint-1100"
``` ```
### 4. Sample Output ### 5. Sample Output
```log ```log
{'loss': 1.9231, 'learning_rate': 2.9999945367033285e-05, 'epoch': 0.0} {'loss': 1.9231, 'learning_rate': 2.9999945367033285e-05, 'epoch': 0.0}
{'loss': 1.8622, 'learning_rate': 2.9999781468531096e-05, 'epoch': 0.01} {'loss': 1.8622, 'learning_rate': 2.9999781468531096e-05, 'epoch': 0.01}