5.9 KiB
		
	
	
	
	
	
	
	
			
		
		
	
	
			5.9 KiB
		
	
	
	
	
	
	
	
BigDL-LLM Examples: CPU
Here, we provide some examples on how you could apply BigDL-LLM INT4 optimizations on popular open-source models in the community.
To run these examples, please first refer to here for more information about how to install bigdl-llm, requirements and best practices for setting up your environment.
The following models have been verified on either servers or laptops with Intel CPUs.
Example of PyTorch API
| Model | Example of PyTorch API | 
|---|---|
| LLaMA 2 | link | 
| ChatGLM | link | 
| Mistral | link | 
| Bark | link | 
| BERT | link | 
| Openai Whisper | link | 
.. important::
   In addition to INT4 optimization, BigDL-LLM also provides other low bit optimizations (such as INT8, INT5, NF4, etc.). You may apply other low bit optimizations through PyTorch API as `example <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/PyTorch-Models/More-Data-Types>`_.
Example of transformers-style API
| Model | Example of transformers-style API | 
|---|---|
| LLaMA (such as Vicuna, Guanaco, Koala, Baize, WizardLM, etc.) | link1, link2 | 
| LLaMA 2 | link | 
| ChatGLM | link | 
| ChatGLM2 | link | 
| Mistral | link | 
| Falcon | link | 
| MPT | link | 
| Dolly-v1 | link | 
| Dolly-v2 | link | 
| Replit Code | link | 
| RedPajama | link1, link2 | 
| Phoenix | link1, link2 | 
| StarCoder | link1, link2 | 
| Baichuan | link | 
| Baichuan2 | link | 
| InternLM | link | 
| Qwen | link | 
| Aquila | link | 
| MOSS | link | 
| Whisper | link | 
.. important::
   In addition to INT4 optimization, BigDL-LLM also provides other low bit optimizations (such as INT8, INT5, NF4, etc.). You may apply other low bit optimizations through ``transformers``-style API as `example <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU/HF-Transformers-AutoModels/More-Data-Types>`_.
.. seealso::
   See the complete examples `here <https://github.com/intel-analytics/BigDL/tree/main/python/llm/example/CPU>`_.