diff --git a/README.md b/README.md index ae2935a2..73cb0aed 100644 --- a/README.md +++ b/README.md @@ -17,12 +17,34 @@ - Over 20 models have been optimized/verified on `bigdl-llm`, including *LLaMA/LLaMA2, ChatGLM/ChatGLM2, MPT, Falcon, Dolly-v1/Dolly-v2, StarCoder, Whisper, QWen, Baichuan, MOSS,* and more; see the complete list [here](python/llm/README.md#verified-models). ### `bigdl-llm` Demos -See the ***optimized performance*** of `chatglm2-6b`, `llama-2-13b-chat`, and `starcoder-15.5b` models on a 12th Gen Intel Core CPU below. +See the ***optimized performance*** of `chatglm2-6b` and `llama-2-13b-chat` models on 12th Gen Intel Core CPU and Intel Arc GPU below. -
+| 12th Gen Intel Core CPU | +Intel Arc GPU | +||
+
+ |
+
+
+ |
+
+
+ |
+
+
+ |
+
chatglm2-6b |
+ llama-2-13b-chat |
+ chatglm2-6b |
+ llama-2-13b-chat |
+
- BigDL-LLM is a library for running LLM (large language model) on your Intel laptop or GPU using INT4 with very low latency [1] (for any PyTorch model). -
+
+ bigdl-llm is a library for running LLM (large language model) on Intel XPU (from Laptop to GPU to Cloud) using INT4 with very low latency [1] (for any PyTorch model).
+
| 12th Gen Intel Core CPU | +Intel Arc GPU | +||
+
+ |
+
+
+ |
+
+
+ |
+
+
+ |
+
chatglm2-6b |
+ llama-2-13b-chat |
+ chatglm2-6b |
+ llama-2-13b-chat |
+
| 12th Gen Intel Core CPU | +Intel Arc GPU | +||
+
+ |
+
+
+ |
+
+
+ |
+
+
+ |
+
chatglm2-6b |
+ llama-2-13b-chat |
+ chatglm2-6b |
+ llama-2-13b-chat |
+