* [ADD] rewrite new vllm docker quick start * [ADD] lora adapter doc finished * [ADD] mulit lora adapter test successfully * [ADD] add ipex-llm quantization doc * [UPDATE] update mmdocs vllm_docker_quickstart content * [REMOVE] rm tmp file * [UPDATE] tp and pp explaination and readthedoc link change * [FIX] fix the error description of tp+pp and quantization part * [FIX] fix the table of verifed model * [UPDATE] add full low bit para list * [UPDATE] update the load_in_low_bit params to verifed dtype |
||
|---|---|---|
| .. | ||
| docker_cpp_xpu_quickstart.md | ||
| docker_pytorch_inference_gpu.md | ||
| docker_run_pytorch_inference_in_vscode.md | ||
| docker_windows_gpu.md | ||
| fastchat_docker_quickstart.md | ||
| README.md | ||
| vllm_cpu_docker_quickstart.md | ||
| vllm_docker_quickstart.md | ||
IPEX-LLM Docker Container User Guides
In this section, you will find guides related to using IPEX-LLM with Docker, covering how to:
-
Inference in Python/C++
-
Serving