Update llama_cpp_quickstart.md (#10663)
This commit is contained in:
parent
6699d86192
commit
ac65ab65c6
1 changed files with 3 additions and 0 deletions
|
|
@ -20,6 +20,8 @@ Visit the [Install IPEX-LLM on Linux with Intel GPU](https://ipex-llm.readthedoc
|
|||
#### Windows
|
||||
Visit the [Install IPEX-LLM on Windows with Intel GPU Guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html#install-prerequisites) to install [Visual Studio 2022](https://visualstudio.microsoft.com/downloads/) Community Edition, latest [GPU driver](https://www.intel.com/content/www/us/en/download/785597/intel-arc-iris-xe-graphics-windows.html) and Intel® oneAPI Base Toolkit 2024.0.
|
||||
|
||||
**Note**: IPEX-LLM backend only supports the more recent GPU drivers. Please make sure your GPU driver version is equal or newer than `31.0.101.5333`, otherwise you might find gibberish output.
|
||||
|
||||
### 1 Install IPEX-LLM for llama.cpp
|
||||
|
||||
To use `llama.cpp` with IPEX-LLM, first ensure that `ipex-llm[cpp]` is installed.
|
||||
|
|
@ -259,3 +261,4 @@ Log end
|
|||
|
||||
#### Fail to quantize model
|
||||
If you encounter `main: failed to quantize model from xxx`, please make sure you have created related output directory.
|
||||
|
||||
|
|
|
|||
Loading…
Reference in a new issue