ipex-llm/python/llm/example/CPU/PyTorch-Models/Model
Zheng, Yi 2347f611cf Add cpu and gpu examples of Mamba (#9797)
* Add mamba cpu example

* Add mamba gpu example

* Use a smaller model as the example

* minor fixes

---------

Co-authored-by: Shengsheng Huang <shengsheng.huang@intel.com>
2024-02-28 11:33:29 +08:00
..
aquila2 Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
bark Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
bert Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
bluelm Add cpu and gpu examples for BlueLM (#9589) 2023-12-05 13:59:02 +08:00
chatglm Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
chatglm3 Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
codellama LLM: fix installation of codellama (#9813) 2024-01-02 14:32:50 +08:00
codeshell Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
deciLM-7b Add CPU and GPU examples for DeciLM-7B (#9867) 2024-02-27 13:15:49 +08:00
deepseek-moe Add DeepSeek-MoE-16B-Chat (#10155) 2024-02-28 10:12:09 +08:00
distil-whisper Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
flan-t5 Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
fuyu Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
internlm-xcomposer Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
internlm2 Add HF and PyTorch example InternLM2 (#10061) 2024-02-04 10:25:55 +08:00
llama2 Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
llava Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
mamba Add cpu and gpu examples of Mamba (#9797) 2024-02-28 11:33:29 +08:00
meta-llama rebase (#9104) 2024-02-28 11:18:21 +08:00
mistral Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
mixtral [LLM] Mixtral CPU examples (#9673) 2023-12-14 10:35:11 +08:00
openai-whisper Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
phi-1_5 Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
phi-2 Add CPU and GPU examples of phi-2 (#10014) 2024-02-23 14:05:53 +08:00
phixtral add phixtral and optimize phi-moe (#10052) 2024-02-05 11:12:47 +08:00
qwen-vl Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
qwen1.5 Add Qwen1.5-7B-Chat (#10113) 2024-02-21 13:29:29 +08:00
skywork Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
solar Fix README.md for solar (#9957) 2024-01-24 15:50:54 +08:00
wizardcoder-python Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
yi Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00
yuan2 Add CPU and GPU examples for Yuan2-2B-hf (#9946) 2024-02-23 14:09:30 +08:00
ziya Add ziya CPU example (#10114) 2024-02-20 13:59:52 +08:00
README.md Uing bigdl-llm-init instead of bigdl-nano-init (#9558) 2023-11-30 10:10:29 +08:00

BigDL-LLM INT4 Optimization for Large Language Model

You can use optimize_model API to accelerate general PyTorch models on Intel servers and PCs. This directory contains example scripts to help you quickly get started using BigDL-LLM to run some popular open-source models in the community. Each model has its own dedicated folder, where you can find detailed instructions on how to install and run it.

To run the examples, we recommend using Intel® Xeon® processors (server), or >= 12th Gen Intel® Core™ processor (client).

For OS, BigDL-LLM supports Ubuntu 20.04 or later, CentOS 7 or later, and Windows 10/11.

Best Known Configuration on Linux

For better performance, it is recommended to set environment variables on Linux with the help of BigDL-LLM:

pip install bigdl-llm
source bigdl-llm-init