Update readme (#10214)
This commit is contained in:
parent
f315c7f93a
commit
40584dec6d
2 changed files with 6 additions and 5 deletions
|
|
@ -28,7 +28,7 @@
|
||||||
- [2023/10] `bigdl-llm` now supports [FastChat serving](python/llm/src/bigdl/llm/serving) on on both Intel CPU and GPU.
|
- [2023/10] `bigdl-llm` now supports [FastChat serving](python/llm/src/bigdl/llm/serving) on on both Intel CPU and GPU.
|
||||||
- [2023/09] `bigdl-llm` now supports [Intel GPU](python/llm/example/GPU) (including iGPU, Arc, Flex and MAX).
|
- [2023/09] `bigdl-llm` now supports [Intel GPU](python/llm/example/GPU) (including iGPU, Arc, Flex and MAX).
|
||||||
- [2023/09] `bigdl-llm` [tutorial](https://github.com/intel-analytics/bigdl-llm-tutorial) is released.
|
- [2023/09] `bigdl-llm` [tutorial](https://github.com/intel-analytics/bigdl-llm-tutorial) is released.
|
||||||
- [2023/09] Over 30 models have been optimized/verified on `bigdl-llm`, including *LLaMA/LLaMA2, ChatGLM2/ChatGLM3, Mistral, Falcon, MPT, LLaVA, WizardCoder, Dolly, Whisper, Baichuan/Baichuan2, InternLM, Skywork, QWen/Qwen-VL, Aquila, MOSS,* and more; see the complete list [here](#verified-models).
|
- [2023/09] Over 40 models have been optimized/verified on `bigdl-llm`, including *LLaMA/LLaMA2, ChatGLM2/ChatGLM3, Mistral, Falcon, MPT, LLaVA, WizardCoder, Dolly, Whisper, Baichuan/Baichuan2, InternLM, Skywork, QWen/Qwen-VL, Aquila, MOSS,* and more; see the complete list [here](#verified-models).
|
||||||
|
|
||||||
### `bigdl-llm` Demos
|
### `bigdl-llm` Demos
|
||||||
See the ***optimized performance*** of `chatglm2-6b` and `llama-2-13b-chat` models on 12th Gen Intel Core CPU and Intel Arc GPU below.
|
See the ***optimized performance*** of `chatglm2-6b` and `llama-2-13b-chat` models on 12th Gen Intel Core CPU and Intel Arc GPU below.
|
||||||
|
|
@ -70,10 +70,11 @@ See the ***optimized performance*** of `chatglm2-6b` and `llama-2-13b-chat` mode
|
||||||
#### CPU INT4
|
#### CPU INT4
|
||||||
##### Install
|
##### Install
|
||||||
You may install **`bigdl-llm`** on Intel CPU as follows:
|
You may install **`bigdl-llm`** on Intel CPU as follows:
|
||||||
|
> Note: See the [CPU installation guide](https://bigdl.readthedocs.io/en/latest/doc/LLM/Overview/install_cpu.html) for more details.
|
||||||
```bash
|
```bash
|
||||||
pip install --pre --upgrade bigdl-llm[all]
|
pip install --pre --upgrade bigdl-llm[all]
|
||||||
```
|
```
|
||||||
> Note: `bigdl-llm` has been tested on Python 3.9
|
> Note: `bigdl-llm` has been tested on Python 3.9, 3.10 and 3.11
|
||||||
|
|
||||||
##### Run Model
|
##### Run Model
|
||||||
You may apply INT4 optimizations to any Hugging Face *Transformers* models as follows.
|
You may apply INT4 optimizations to any Hugging Face *Transformers* models as follows.
|
||||||
|
|
@ -100,7 +101,7 @@ You may install **`bigdl-llm`** on Intel GPU as follows:
|
||||||
# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
|
# below command will install intel_extension_for_pytorch==2.1.10+xpu as default
|
||||||
pip install --pre --upgrade bigdl-llm[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu
|
pip install --pre --upgrade bigdl-llm[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu
|
||||||
```
|
```
|
||||||
> Note: `bigdl-llm` has been tested on Python 3.9
|
> Note: `bigdl-llm` has been tested on Python 3.9, 3.10 and 3.11
|
||||||
|
|
||||||
##### Run Model
|
##### Run Model
|
||||||
You may apply INT4 optimizations to any Hugging Face *Transformers* models as follows.
|
You may apply INT4 optimizations to any Hugging Face *Transformers* models as follows.
|
||||||
|
|
|
||||||
|
|
@ -17,7 +17,7 @@ Please refer to [Environment Setup](#environment-setup) for more information.
|
||||||
|
|
||||||
.. important::
|
.. important::
|
||||||
|
|
||||||
``bigdl-llm`` is tested with Python 3.9, which is recommended for best practices.
|
``bigdl-llm`` is tested with Python 3.9, 3.10 and 3.11; Python 3.9 is recommended for best practices.
|
||||||
```
|
```
|
||||||
|
|
||||||
## Recommended Requirements
|
## Recommended Requirements
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue