Update llm-readme (#9176)

This commit is contained in:
Jason Dai 2023-10-16 10:54:52 +08:00 committed by GitHub
parent a164c24746
commit b192a8032c
2 changed files with 25 additions and 26 deletions

View file

@ -15,7 +15,7 @@
- **[New]** `bigdl-llm` now supports QLoRA fintuning on Intel GPU; see the the example [here](python/llm/example/GPU/QLoRA-FineTuning).
- `bigdl-llm` now supports Intel GPU (including Arc, Flex and MAX); see the the latest GPU examples [here](python/llm/example/GPU).
- `bigdl-llm` tutorial is released [here](https://github.com/intel-analytics/bigdl-llm-tutorial).
- Over 20 models have been optimized/verified on `bigdl-llm`, including *LLaMA/LLaMA2, ChatGLM/ChatGLM2, MPT, Falcon, Dolly, StarCoder, Whisper, InternLM, QWen, Baichuan, Aquila, MOSS,* and more; see the complete list [here](python/llm/README.md#verified-models).
- Over 20 models have been optimized/verified on `bigdl-llm`, including *LLaMA/LLaMA2, ChatGLM/ChatGLM2, Mistral, Falcon, MPT, Dolly, StarCoder, Whisper, Baichuan, InternLM, QWen, Aquila, MOSS,* and more; see the complete list [here](python/llm/README.md#verified-models).
### `bigdl-llm` Demos
See the ***optimized performance*** of `chatglm2-6b` and `llama-2-13b-chat` models on 12th Gen Intel Core CPU and Intel Arc GPU below.

View file

@ -34,32 +34,31 @@ See the ***optimized performance*** of `chatglm2-6b` and `llama-2-13b-chat` mode
</table>
### Verified models
Over 20 models have been optimized/verified on `bigdl-llm`, including *LLaMA/LLaMA2, ChatGLM/ChatGLM2, MPT, Falcon, Dolly-v1/Dolly-v2, StarCoder, Whisper, InternLM, QWen, Baichuan, MOSS,* and more; see the complete list below.
Over 20 models have been optimized/verified on `bigdl-llm`, including *LLaMA/LLaMA2, ChatGLM/ChatGLM2, Mistral, Falcon, MPT, Dolly, StarCoder, Whisper, Baichuan, InternLM, QWen, Aquila, MOSS,* and more; see the complete list below.
| Model | CPU Example | GPU Example |
|------------|----------------------------------------------------------------|-----------------------------------------------------------------|
| LLaMA *(such as Vicuna, Guanaco, Koala, Baize, WizardLM, etc.)* | [link1](example/CPU/Native-Models), [link2](example/CPU/HF-Transformers-AutoModels/Model/vicuna) |[link](example/GPU/HF-Transformers-AutoModels/Model/vicuna)|
| LLaMA 2 | [link1](example/CPU/Native-Models), [link2](example/CPU/HF-Transformers-AutoModels/Model/llama2) | [link](example/GPU/HF-Transformers-AutoModels/Model/llama2) |
| ChatGLM | [link](example/CPU/HF-Transformers-AutoModels/Model/chatglm) | |
| ChatGLM2 | [link](example/CPU/HF-Transformers-AutoModels/Model/chatglm2) | [link](example/GPU/HF-Transformers-AutoModels/Model/chatglm2) |
| Mistral | [link](example/CPU/HF-Transformers-AutoModels/Model/mistral) | [link](example/GPU/HF-Transformers-AutoModels/Model/mistral) |
| Falcon | [link](example/CPU/HF-Transformers-AutoModels/Model/falcon) | [link](example/GPU/HF-Transformers-AutoModels/Model/falcon) |
| MPT | [link](example/CPU/HF-Transformers-AutoModels/Model/mpt) | [link](example/CPU/HF-Transformers-AutoModels/Model/mpt) |
| Dolly-v1 | [link](example/CPU/HF-Transformers-AutoModels/Model/dolly_v1) | [link](example/CPU/HF-Transformers-AutoModels/Model/dolly_v1) |
| Dolly-v2 | [link](example/CPU/HF-Transformers-AutoModels/Model/dolly_v2) | [link](example/CPU/HF-Transformers-AutoModels/Model/dolly_v2) |
| Replit Code| [link](example/CPU/HF-Transformers-AutoModels/Model/replit) | [link](example/CPU/HF-Transformers-AutoModels/Model/replit) |
| RedPajama | [link1](example/CPU/Native-Models), [link2](example/CPU/HF-Transformers-AutoModels/Model/redpajama) | |
| Phoenix | [link1](example/CPU/Native-Models), [link2](example/CPU/HF-Transformers-AutoModels/Model/phoenix) | |
| StarCoder | [link1](example/CPU/Native-Models), [link2](example/CPU/HF-Transformers-AutoModels/Model/starcoder) | [link](example/GPU/HF-Transformers-AutoModels/Model/starcoder) |
| Baichuan | [link](example/CPU/HF-Transformers-AutoModels/Model/baichuan) | [link](example/CPU/HF-Transformers-AutoModels/Model/baichuan) |
| Baichuan2 | [link](example/CPU/HF-Transformers-AutoModels/Model/baichuan2) | [link](example/GPU/HF-Transformers-AutoModels/Model/baichuan2) |
| InternLM | [link](example/CPU/HF-Transformers-AutoModels/Model/internlm) | [link](example/GPU/HF-Transformers-AutoModels/Model/internlm) |
| Qwen | [link](example/CPU/HF-Transformers-AutoModels/Model/qwen) | [link](example/GPU/HF-Transformers-AutoModels/Model/qwen) |
| Aquila | [link](example/CPU/HF-Transformers-AutoModels/Model/aquila) | [link](example/GPU/HF-Transformers-AutoModels/Model/aquila) |
| MOSS | [link](example/CPU/HF-Transformers-AutoModels/Model/moss) | |
| Whisper | [link](example/CPU/HF-Transformers-AutoModels/Model/whisper) | [link](example/GPU/HF-Transformers-AutoModels/Model/whisper) |
<details><summary>Table of verified models</summary>
| Model | Example |
|-----------|----------------------------------------------------------|
| LLaMA *(such as Vicuna, Guanaco, Koala, Baize, WizardLM, etc.)* | [link1](example/CPU/Native-Models), [link2](example/CPU/HF-Transformers-AutoModels/Model/vicuna) |
| LLaMA 2 | [link](example/CPU/HF-Transformers-AutoModels/Model/llama2) |
| MPT | [link](example/CPU/HF-Transformers-AutoModels/Model/mpt) |
| Falcon | [link](example/CPU/HF-Transformers-AutoModels/Model/falcon) |
| ChatGLM | [link](example/CPU/HF-Transformers-AutoModels/Model/chatglm) |
| ChatGLM2 | [link](example/CPU/HF-Transformers-AutoModels/Model/chatglm2) |
| Qwen | [link](example/CPU/HF-Transformers-AutoModels/Model/qwen) |
| MOSS | [link](example/CPU/HF-Transformers-AutoModels/Model/moss) |
| Baichuan | [link](example/CPU/HF-Transformers-AutoModels/Model/baichuan) |
| Baichuan2 | [link](example/CPU/HF-Transformers-AutoModels/Model/baichuan2) |
| Dolly-v1 | [link](example/CPU/HF-Transformers-AutoModels/Model/dolly_v1) |
| Dolly-v2 | [link](example/CPU/HF-Transformers-AutoModels/Model/dolly_v2) |
| RedPajama | [link1](example/CPU/Native-Models), [link2](example/CPU/HF-Transformers-AutoModels/Model/redpajama) |
| Phoenix | [link1](example/CPU/Native-Models), [link2](example/CPU/HF-Transformers-AutoModels/Model/phoenix) |
| StarCoder | [link1](example/CPU/Native-Models), [link2](example/CPU/HF-Transformers-AutoModels/Model/starcoder) |
| InternLM | [link](example/CPU/HF-Transformers-AutoModels/Model/internlm) |
| Whisper | [link](example/CPU/HF-Transformers-AutoModels/Model/whisper) |
| Aquila | [link](example/CPU/HF-Transformers-AutoModels/Model/aquila) |
</details>
### Working with `bigdl-llm`