ipex-llm/docs/mddocs/Overview/examples_cpu.md
2024-06-20 13:47:49 +08:00

6 KiB

IPEX-LLM Examples: CPU

Here, we provide some examples on how you could apply IPEX-LLM INT4 optimizations on popular open-source models in the community.

To run these examples, please first refer to here for more information about how to install ipex-llm, requirements and best practices for setting up your environment.

The following models have been verified on either servers or laptops with Intel CPUs.

Example of PyTorch API

Model Example of PyTorch API
LLaMA 2 link
ChatGLM link
Mistral link
Bark link
BERT link
Openai Whisper link
.. important::

   In addition to INT4 optimization, IPEX-LLM also provides other low bit optimizations (such as INT8, INT5, NF4, etc.). You may apply other low bit optimizations through PyTorch API as `example <https://github.com/intel-analytics/ipex-llm/tree/main/python/llm/example/CPU/PyTorch-Models/More-Data-Types>`_.

Example of transformers-style API

Model Example of transformers-style API
LLaMA (such as Vicuna, Guanaco, Koala, Baize, WizardLM, etc.) link1, link2
LLaMA 2 link
ChatGLM link
ChatGLM2 link
Mistral link
Falcon link
MPT link
Dolly-v1 link
Dolly-v2 link
Replit Code link
RedPajama link1, link2
Phoenix link1, link2
StarCoder link1, link2
Baichuan link
Baichuan2 link
InternLM link
Qwen link
Aquila link
MOSS link
Whisper link
.. important::

   In addition to INT4 optimization, IPEX-LLM also provides other low bit optimizations (such as INT8, INT5, NF4, etc.). You may apply other low bit optimizations through ``transformers``-style API as `example <https://github.com/intel-analytics/ipex-llm/tree/main/python/llm/example/CPU/HF-Transformers-AutoModels/More-Data-Types>`_.
.. seealso::

   See the complete examples `here <https://github.com/intel-analytics/ipex-llm/tree/main/python/llm/example/CPU>`_.