ipex-llm/python/llm/example/CPU/HF-Transformers-AutoModels/Model
hxsz1997 245c7348bc
Add codegemma example (#10884)
* add codegemma example in GPU/HF-Transformers-AutoModels/

* add README of codegemma example in GPU/HF-Transformers-AutoModels/

* add codegemma example in GPU/PyTorch-Models/

* add readme of codegemma example in GPU/PyTorch-Models/

* add codegemma example in CPU/HF-Transformers-AutoModels/

* add readme of codegemma example in CPU/HF-Transformers-AutoModels/

* add codegemma example in CPU/PyTorch-Models/

* add readme of codegemma example in CPU/PyTorch-Models/

* fix typos

* fix filename typo

* add codegemma in tables

* add comments of lm_head

* remove comments of use_cache
2024-05-07 13:35:42 +08:00
..
aquila Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
aquila2 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
baichuan Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
baichuan2 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
bluelm Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
chatglm update chatglm readme (#10659) 2024-04-09 14:24:46 -07:00
chatglm2 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
chatglm3 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
codegemma Add codegemma example (#10884) 2024-05-07 13:35:42 +08:00
codellama Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
codeshell Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
deciLM-7b Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
deepseek Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
deepseek-moe Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
distil-whisper Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
dolly_v1 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
dolly_v2 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
falcon Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
flan-t5 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
fuyu Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
gemma Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
internlm fix_internlm-chat-7b-8k repo name in examples (#10747) 2024-04-12 10:15:48 -07:00
internlm-xcomposer Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
internlm2 fix_internlm-chat-7b-8k repo name in examples (#10747) 2024-04-12 10:15:48 -07:00
llama2 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
llama3 Fix the not stop issue of llama3 examples (#10860) 2024-04-23 19:10:09 +08:00
mistral Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
mixtral Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
moss Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
mpt Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
phi-1_5 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
phi-2 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
phi-3 Add example for phi-3 (#10881) 2024-04-29 16:43:55 +08:00
phixtral Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
phoenix Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
qwen Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
qwen-vl Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
qwen1.5 LLM:Add qwen moe example libs md (#10828) 2024-04-22 10:03:19 +08:00
redpajama Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
replit Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
skywork Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
solar Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
stablelm Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
starcoder Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
vicuna Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
whisper Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
wizardcoder-python Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
yi Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
yuan2 Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
ziya Upgrade to python 3.11 (#10711) 2024-04-09 17:41:17 +08:00
README.md Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00

IPEX-LLM Transformers INT4 Optimization for Large Language Model

You can use IPEX-LLM to run any Huggingface Transformer models with INT4 optimizations on either servers or laptops. This directory contains example scripts to help you quickly get started using IPEX-LLM to run some popular open-source models in the community. Each model has its own dedicated folder, where you can find detailed instructions on how to install and run it.

To run the examples, we recommend using Intel® Xeon® processors (server), or >= 12th Gen Intel® Core™ processor (client).

For OS, IPEX-LLM supports Ubuntu 20.04 or later (glibc>=2.17), CentOS 7 or later (glibc>=2.17), and Windows 10/11.

Best Known Configuration on Linux

For better performance, it is recommended to set environment variables on Linux with the help of IPEX-LLM:

pip install ipex-llm
source ipex-llm-init