# BigDL-LLM INT4 Optimization for Large Language Model You can use `optimize_model` API to accelerate general PyTorch models on Intel servers and PCs. This directory contains example scripts to help you quickly get started using BigDL-LLM to run some popular open-source models in the community. Each model has its own dedicated folder, where you can find detailed instructions on how to install and run it. # Verified models | Model | Example | |----------------|----------------------------------------------------------| | LLaMA 2 | [link](llama2) | | ChatGLM | [link](chatglm) | | Openai Whisper | [link](openai-whisper) | | BERT | [link](bert) | | Bark | [link](bark) | | Mistral | [link](mistral) | | Flan-t5 | [link](flan-t5) | | Phi-1_5 | [link](phi-1_5) | | Qwen-VL | [link](qwen-vl) | | LLaVA | [link](llava) | ## Recommended Requirements To run the examples, we recommend using Intel® Xeon® processors (server), or >= 12th Gen Intel® Core™ processor (client). For OS, BigDL-LLM supports Ubuntu 20.04 or later, CentOS 7 or later, and Windows 10/11. ## Best Known Configuration on Linux For better performance, it is recommended to set environment variables on Linux with the help of BigDL-Nano: ```bash pip install bigdl-nano source bigdl-nano-init ```