From 86743fb57bc6bb3f6dfba9f4ec0b702e617bf204 Mon Sep 17 00:00:00 2001 From: binbin Deng <108676127+plusbang@users.noreply.github.com> Date: Wed, 22 Nov 2023 15:53:07 +0800 Subject: [PATCH] LLM: fix transformers version in CPU finetuning example (#9511) --- .../llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/README.md b/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/README.md index b86acc9d..a256fcc2 100644 --- a/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/README.md +++ b/python/llm/example/CPU/QLoRA-FineTuning/alpaca-qlora/README.md @@ -8,8 +8,8 @@ This example ports [Alpaca-LoRA](https://github.com/tloen/alpaca-lora/tree/main) conda create -n llm python=3.9 conda activate llm pip install --pre --upgrade bigdl-llm[all] -pip install transformers==4.34.0 -pip install fire datasets peft==0.5.0 +pip install datasets transformers==4.34.0 +pip install fire peft==0.5.0 pip install accelerate==0.23.0 ```