From 1de878bee10305c0024101c34bb348d1ed21242d Mon Sep 17 00:00:00 2001 From: "Wang, Jian4" <61138589+hzjane@users.noreply.github.com> Date: Tue, 7 May 2024 09:25:20 +0800 Subject: [PATCH] LLM: Fix speculative llama3 long input error (#10934) --- python/llm/example/CPU/Speculative-Decoding/llama3/README.md | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/python/llm/example/CPU/Speculative-Decoding/llama3/README.md b/python/llm/example/CPU/Speculative-Decoding/llama3/README.md index bf7875ed..35d7eec0 100644 --- a/python/llm/example/CPU/Speculative-Decoding/llama3/README.md +++ b/python/llm/example/CPU/Speculative-Decoding/llama3/README.md @@ -18,7 +18,8 @@ We suggest using conda to manage environment: conda create -n llm python=3.11 conda activate llm pip install --pre --upgrade ipex-llm[all] -pip install intel_extension_for_pytorch==2.1.0 +# transformers>=4.33.0 is required for Llama3 with IPEX-LLM optimizations +pip install transformers==4.37.0 ``` ### 2. Configures high-performing processor environment variables