Update llama_cpp_quickstart.md (#10663)
This commit is contained in:
		
							parent
							
								
									6699d86192
								
							
						
					
					
						commit
						ac65ab65c6
					
				
					 1 changed files with 3 additions and 0 deletions
				
			
		| 
						 | 
					@ -20,6 +20,8 @@ Visit the [Install IPEX-LLM on Linux with Intel GPU](https://ipex-llm.readthedoc
 | 
				
			||||||
#### Windows
 | 
					#### Windows
 | 
				
			||||||
Visit the [Install IPEX-LLM on Windows with Intel GPU Guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html#install-prerequisites) to install [Visual Studio 2022](https://visualstudio.microsoft.com/downloads/) Community Edition, latest [GPU driver](https://www.intel.com/content/www/us/en/download/785597/intel-arc-iris-xe-graphics-windows.html) and Intel® oneAPI Base Toolkit 2024.0.
 | 
					Visit the [Install IPEX-LLM on Windows with Intel GPU Guide](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html), and follow [Install Prerequisites](https://ipex-llm.readthedocs.io/en/latest/doc/LLM/Quickstart/install_windows_gpu.html#install-prerequisites) to install [Visual Studio 2022](https://visualstudio.microsoft.com/downloads/) Community Edition, latest [GPU driver](https://www.intel.com/content/www/us/en/download/785597/intel-arc-iris-xe-graphics-windows.html) and Intel® oneAPI Base Toolkit 2024.0.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					**Note**: IPEX-LLM backend only supports the more recent GPU drivers. Please make sure your GPU driver version is equal or newer than `31.0.101.5333`, otherwise you might find gibberish output.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### 1 Install IPEX-LLM for llama.cpp
 | 
					### 1 Install IPEX-LLM for llama.cpp
 | 
				
			||||||
 | 
					
 | 
				
			||||||
To use `llama.cpp` with IPEX-LLM, first ensure that `ipex-llm[cpp]` is installed.
 | 
					To use `llama.cpp` with IPEX-LLM, first ensure that `ipex-llm[cpp]` is installed.
 | 
				
			||||||
| 
						 | 
					@ -259,3 +261,4 @@ Log end
 | 
				
			||||||
 | 
					
 | 
				
			||||||
#### Fail to quantize model
 | 
					#### Fail to quantize model
 | 
				
			||||||
If you encounter `main: failed to quantize model from xxx`, please make sure you have created related output directory.
 | 
					If you encounter `main: failed to quantize model from xxx`, please make sure you have created related output directory.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue