From 2dfcc368257b0d03c455c62b24e5f56118c0915c Mon Sep 17 00:00:00 2001 From: Qiyuan Gong Date: Fri, 8 Nov 2024 16:05:17 +0800 Subject: [PATCH] Fix trl version and padding in trl qlora example (#12368) * Change trl to 0.9.6 * Enable padding to avoid padding related errors. --- .../llm/example/GPU/LLM-Finetuning/QLoRA/trl-example/README.md | 2 +- .../GPU/LLM-Finetuning/QLoRA/trl-example/qlora_finetuning.py | 3 +++ 2 files changed, 4 insertions(+), 1 deletion(-) diff --git a/python/llm/example/GPU/LLM-Finetuning/QLoRA/trl-example/README.md b/python/llm/example/GPU/LLM-Finetuning/QLoRA/trl-example/README.md index 498eb8a9..949b211e 100644 --- a/python/llm/example/GPU/LLM-Finetuning/QLoRA/trl-example/README.md +++ b/python/llm/example/GPU/LLM-Finetuning/QLoRA/trl-example/README.md @@ -19,7 +19,7 @@ conda activate llm pip install --pre --upgrade ipex-llm[xpu] --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ pip install transformers==4.36.0 datasets pip install peft==0.10.0 -pip install bitsandbytes scipy "trl<0.12.0" +pip install bitsandbytes scipy trl==0.9.6 ``` ### 2. Configures OneAPI environment variables diff --git a/python/llm/example/GPU/LLM-Finetuning/QLoRA/trl-example/qlora_finetuning.py b/python/llm/example/GPU/LLM-Finetuning/QLoRA/trl-example/qlora_finetuning.py index dc78ae33..2e2551cd 100644 --- a/python/llm/example/GPU/LLM-Finetuning/QLoRA/trl-example/qlora_finetuning.py +++ b/python/llm/example/GPU/LLM-Finetuning/QLoRA/trl-example/qlora_finetuning.py @@ -105,6 +105,9 @@ if __name__ == "__main__": gradient_checkpointing=True, # can further reduce memory but slower ), dataset_text_field="instruction", + data_collator=transformers.DataCollatorForSeq2Seq( + tokenizer, pad_to_multiple_of=8, return_tensors="pt", padding=True + ), ) model.config.use_cache = False # silence the warnings. Please re-enable for inference! result = trainer.train()