From 877229f3be43e6f7a0ea5612b07ab79a3411e06c Mon Sep 17 00:00:00 2001 From: ZehuaCao <47251317+Romanticoseu@users.noreply.github.com> Date: Thu, 14 Dec 2023 09:55:47 +0800 Subject: [PATCH] [LLM]Add Yi-34B-AWQ to verified AWQ model. (#9676) * verfiy Yi-34B-AWQ * update --- .../Advanced-Quantizations/AWQ/README.md | 1 + .../Advanced-Quantizations/AWQ/README.md | 1 + 2 files changed, 2 insertions(+) diff --git a/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md b/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md index ebafd827..736126c2 100644 --- a/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md +++ b/python/llm/example/CPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md @@ -12,6 +12,7 @@ This example shows how to directly run 4-bit AWQ models using BigDL-LLM on Intel - [vicuna-13B-v1.5-AWQ](https://huggingface.co/TheBloke/vicuna-13B-v1.5-AWQ) - [llava-v1.5-13B-AWQ](https://huggingface.co/TheBloke/llava-v1.5-13B-AWQ) - [Yi-6B-AWQ](https://huggingface.co/TheBloke/Yi-6B-AWQ) +- [Yi-34B-AWQ](https://huggingface.co/TheBloke/Yi-34B-AWQ) ## Requirements diff --git a/python/llm/example/GPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md b/python/llm/example/GPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md index cd386860..32960122 100644 --- a/python/llm/example/GPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md +++ b/python/llm/example/GPU/HF-Transformers-AutoModels/Advanced-Quantizations/AWQ/README.md @@ -12,6 +12,7 @@ This example shows how to directly run 4-bit AWQ models using BigDL-LLM on Intel - [vicuna-13B-v1.5-AWQ](https://huggingface.co/TheBloke/vicuna-13B-v1.5-AWQ) - [llava-v1.5-13B-AWQ](https://huggingface.co/TheBloke/llava-v1.5-13B-AWQ) - [Yi-6B-AWQ](https://huggingface.co/TheBloke/Yi-6B-AWQ) +- [Yi-34B-AWQ](https://huggingface.co/TheBloke/Yi-34B-AWQ) ## Requirements