ipex-llm/python/llm/example/CPU/PyTorch-Models/Model
Jin Qiao cc8b3be11c
Add GPU and CPU example for stablelm-zephyr-3b (#10643)
* Add example for StableLM

* fix

* add to readme
2024-04-03 16:28:31 +08:00
..
aquila2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
bark Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
bert Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
bluelm Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
chatglm Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
chatglm3 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
codellama Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
codeshell Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
deciLM-7b Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
deepseek Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
deepseek-moe Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
distil-whisper Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
flan-t5 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
fuyu Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
internlm-xcomposer Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
internlm2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
llama2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
llava Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
mamba Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
meta-llama Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
mistral Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
mixtral Replace ipex with ipex-llm (#10554) 2024-03-28 13:54:40 +08:00
openai-whisper Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
phi-1_5 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
phi-2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
phixtral Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
qwen-vl Fix Qwen-VL example problem (#10582) 2024-04-02 12:17:30 -07:00
qwen1.5 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
skywork Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
solar Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
stablelm Add GPU and CPU example for stablelm-zephyr-3b (#10643) 2024-04-03 16:28:31 +08:00
wizardcoder-python Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
yi Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
yuan2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
ziya Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
README.md Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00

IPEX-LLM INT4 Optimization for Large Language Model

You can use optimize_model API to accelerate general PyTorch models on Intel servers and PCs. This directory contains example scripts to help you quickly get started using IPEX-LLM to run some popular open-source models in the community. Each model has its own dedicated folder, where you can find detailed instructions on how to install and run it.

To run the examples, we recommend using Intel® Xeon® processors (server), or >= 12th Gen Intel® Core™ processor (client).

For OS, IPEX-LLM supports Ubuntu 20.04 or later, CentOS 7 or later, and Windows 10/11.

Best Known Configuration on Linux

For better performance, it is recommended to set environment variables on Linux with the help of IPEX-LLM:

pip install ipex-llm
source ipex-llm-init