ipex-llm/python/llm/example/GPU/HuggingFace/LLM
Guoqiong Song 380717f50d
fix gemma for 4.41 (#11531)
* fix gemma for 4.41
2024-07-18 15:02:50 -07:00
..
aquila Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
aquila2 Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
baichuan
baichuan2
bluelm Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
chatglm2
chatglm3
chinese-llama2
codegeex2
codegemma fix gemma for 4.41 (#11531) 2024-07-18 15:02:50 -07:00
codellama
codeshell
cohere Fix cohere model on transformers>=4.41 (#11575) 2024-07-17 17:18:59 -07:00
deciLM-7b Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
deepseek Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
dolly-v1
dolly-v2 Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
falcon
flan-t5
gemma fix gemma for 4.41 (#11531) 2024-07-18 15:02:50 -07:00
glm4 Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
gpt-j
internlm
internlm2
llama2
llama3
minicpm fix minicpm for transformers>=4.39 (#11533) 2024-07-18 15:01:57 -07:00
mistral Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
mixtral
mpt Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
phi-1_5 phi model readme (#11595) 2024-07-17 17:18:34 -07:00
phi-2 phi model readme (#11595) 2024-07-17 17:18:34 -07:00
phi-3 phi-3 on "transformers>=4.37.0,<=4.42.3" (#11534) 2024-07-17 17:19:57 -07:00
phixtral
qwen Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
qwen1.5 Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
qwen2
redpajama
replit
rwkv4 Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
rwkv5 Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
solar
stablelm
starcoder Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
vicuna Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
yi Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
yuan2 Update GPU HF-Transformers example structure (#11526) 2024-07-08 17:58:06 +08:00
README.md

IPEX-LLM Transformers INT4 Optimization for Large Language Model on Intel GPUs

You can use IPEX-LLM to run almost every Huggingface Transformer models with INT4 optimizations on your laptops with Intel GPUs. This directory contains example scripts to help you quickly get started using IPEX-LLM to run some popular open-source models in the community. Each model has its own dedicated folder, where you can find detailed instructions on how to install and run it.