Update llama_cpp_quickstart.md (#13145)
Signed-off-by: Pranav Singh <pranav.singh@intel.com>
This commit is contained in:
		
							parent
							
								
									bd71739e64
								
							
						
					
					
						commit
						bd45bf7584
					
				
					 1 changed files with 1 additions and 1 deletions
				
			
		| 
						 | 
				
			
			@ -3,7 +3,7 @@
 | 
			
		|||
  <b>< English</b> | <a href='./llama_cpp_quickstart.zh-CN.md'>中文</a> >
 | 
			
		||||
</p>
 | 
			
		||||
 | 
			
		||||
[ggerganov/llama.cpp](https://github.com/ggerganov/llama.cpp) prvoides fast LLM inference in pure C++ across a variety of hardware; you can now use the C++ interface of [`ipex-llm`](https://github.com/intel-analytics/ipex-llm) as an accelerated backend for `llama.cpp` running on Intel **GPU** *(e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max)*.
 | 
			
		||||
[ggerganov/llama.cpp](https://github.com/ggerganov/llama.cpp) provides fast LLM inference in pure C++ across a variety of hardware; you can now use the C++ interface of [`ipex-llm`](https://github.com/intel-analytics/ipex-llm) as an accelerated backend for `llama.cpp` running on Intel **GPU** *(e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max)*.
 | 
			
		||||
 | 
			
		||||
> [!Important]
 | 
			
		||||
> You may use [llama.cpp Portable Zip](./llamacpp_portable_zip_gpu_quickstart.md) to directly run llama.cpp on Intel GPU with ipex-llm (***without the need of manual installations***).
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue