Update llama_cpp_quickstart.md (#13145)

Signed-off-by: Pranav Singh <pranav.singh@intel.com>
This commit is contained in:
Pranav Singh 2025-05-15 06:10:53 +05:30 committed by GitHub
parent bd71739e64
commit bd45bf7584
No known key found for this signature in database
GPG key ID: B5690EEEBB952194

View file

@ -3,7 +3,7 @@
<b>< English</b> | <a href='./llama_cpp_quickstart.zh-CN.md'>中文</a> >
</p>
[ggerganov/llama.cpp](https://github.com/ggerganov/llama.cpp) prvoides fast LLM inference in pure C++ across a variety of hardware; you can now use the C++ interface of [`ipex-llm`](https://github.com/intel-analytics/ipex-llm) as an accelerated backend for `llama.cpp` running on Intel **GPU** *(e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max)*.
[ggerganov/llama.cpp](https://github.com/ggerganov/llama.cpp) provides fast LLM inference in pure C++ across a variety of hardware; you can now use the C++ interface of [`ipex-llm`](https://github.com/intel-analytics/ipex-llm) as an accelerated backend for `llama.cpp` running on Intel **GPU** *(e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max)*.
> [!Important]
> You may use [llama.cpp Portable Zip](./llamacpp_portable_zip_gpu_quickstart.md) to directly run llama.cpp on Intel GPU with ipex-llm (***without the need of manual installations***).