ipex-llm/python/llm/example/GPU/HF-Transformers-AutoModels/Model
2024-03-25 10:06:02 +08:00
..
aquila Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
aquila2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
baichuan Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
baichuan2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
bluelm Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
chatglm2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
chatglm3 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
chinese-llama2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
codellama Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
codeshell Refactor bigdl.llm to ipex_llm (#24) 2024-03-22 15:41:21 +08:00
deciLM-7b Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
deepseek Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
distil-whisper Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
dolly-v1 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
dolly-v2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
falcon Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
flan-t5 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
gemma Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
gpt-j Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
internlm Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
internlm2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
llama2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
mistral Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
mixtral Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
mpt Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
phi-1_5 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
phi-2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
phixtral Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
qwen Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
qwen-vl Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
qwen1.5 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
redpajama Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
replit Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
rwkv4 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
rwkv5 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
solar Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
starcoder Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
vicuna Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
voiceassistant Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
whisper Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
yi Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
yuan2 Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00
README.md Update_document by heyang (#30) 2024-03-25 10:06:02 +08:00

IPEX-LLM Transformers INT4 Optimization for Large Language Model on Intel GPUs

You can use IPEX-LLM to run almost every Huggingface Transformer models with INT4 optimizations on your laptops with Intel GPUs. This directory contains example scripts to help you quickly get started using IPEX-LLM to run some popular open-source models in the community. Each model has its own dedicated folder, where you can find detailed instructions on how to install and run it.