[NPU] Add troubleshooting in portable zip doc (#12924)
This commit is contained in:
		
							parent
							
								
									b2d676f1c6
								
							
						
					
					
						commit
						091ab2bd59
					
				
					 2 changed files with 17 additions and 2 deletions
				
			
		| 
						 | 
				
			
			@ -17,6 +17,7 @@ IPEX-LLM provides llama.cpp support for running GGUF models on Intel NPU. This g
 | 
			
		|||
- [Step 2: Setup](#step-2-setup)
 | 
			
		||||
- [Step 3: Run GGUF Model](#step-3-run-gguf-model)
 | 
			
		||||
- [More details](npu_quickstart.md)
 | 
			
		||||
- [Troubleshooting](#troubleshooting)
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
## Prerequisites
 | 
			
		||||
| 
						 | 
				
			
			@ -34,7 +35,7 @@ Then, extract the zip file to a folder.
 | 
			
		|||
 | 
			
		||||
## Step 2: Setup
 | 
			
		||||
 | 
			
		||||
- Open "Command Prompt" (cmd), and enter the extracted folder through `cd /d PATH\TO\EXTRACTED\FOLDER`
 | 
			
		||||
- Open **"Command Prompt" (cmd)**, and enter the extracted folder through `cd /d PATH\TO\EXTRACTED\FOLDER`
 | 
			
		||||
- Runtime configuration based on your device:
 | 
			
		||||
  - For **Intel Core™ Ultra Processors (Series 2) with processor number 2xxV (code name Lunar Lake)**:
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -63,3 +64,9 @@ You could then use cli tool to run GGUF models on Intel NPU through running `lla
 | 
			
		|||
```cmd
 | 
			
		||||
llama-cli-npu.exe -m DeepSeek-R1-Distill-Qwen-7B-Q6_K.gguf -n 32 --prompt "What is AI?"
 | 
			
		||||
```
 | 
			
		||||
 | 
			
		||||
## Troubleshooting
 | 
			
		||||
 | 
			
		||||
### `L0 pfnCreate2 result: ZE_RESULT_ERROR_INVALID_ARGUMENT, code 0x78000004` error
 | 
			
		||||
 | 
			
		||||
First, verify that your NPU driver version meets the requirement. Then, check the runtime configuration based on your device. And please attention the difference between **Command Prompt** and **Windows PowerShell**. Take Arrow Lake for example, you need to use `set IPEX_LLM_NPU_ARL=1` in **Command Prompt** while `$env:IPEX_LLM_NPU_ARL = "1"` in **Windows PowerShell**.
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -17,6 +17,7 @@ IPEX-LLM 提供了 llama.cpp 的相关支持以在 Intel NPU 上运行 GGUF 模
 | 
			
		|||
- [步骤 2:启动](#步骤-2启动)
 | 
			
		||||
- [步骤 3:运行 GGUF 模型](#步骤-3运行-gguf-模型)
 | 
			
		||||
- [更多信息](npu_quickstart.md)
 | 
			
		||||
- [故障排除](#故障排除)
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
## 系统环境准备
 | 
			
		||||
| 
						 | 
				
			
			@ -34,7 +35,7 @@ IPEX-LLM 提供了 llama.cpp 的相关支持以在 Intel NPU 上运行 GGUF 模
 | 
			
		|||
 | 
			
		||||
## 步骤 2:启动
 | 
			
		||||
 | 
			
		||||
- 打开命令提示符(cmd),并通过在命令行输入指令 "cd /d PATH\TO\EXTRACTED\FOLDER" 进入解压缩后的文件夹
 | 
			
		||||
- 打开**命令提示符(cmd)**,并通过在命令行输入指令 "cd /d PATH\TO\EXTRACTED\FOLDER" 进入解压缩后的文件夹
 | 
			
		||||
- 根据你的设备完成运行配置:
 | 
			
		||||
  - 对于 **处理器为 2xxV 的 Intel Core™ Ultra Processors (Series 2) (代号 Lunar Lake)**:
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			@ -63,3 +64,10 @@ IPEX-LLM 提供了 llama.cpp 的相关支持以在 Intel NPU 上运行 GGUF 模
 | 
			
		|||
```cmd
 | 
			
		||||
llama-cli-npu.exe -m DeepSeek-R1-Distill-Qwen-7B-Q6_K.gguf -n 32 --prompt "What is AI?"
 | 
			
		||||
```
 | 
			
		||||
 | 
			
		||||
## 故障排除
 | 
			
		||||
 | 
			
		||||
### `L0 pfnCreate2 result: ZE_RESULT_ERROR_INVALID_ARGUMENT, code 0x78000004` 报错
 | 
			
		||||
 | 
			
		||||
首先确认你的 NPU 驱动版本是否符合要求,然后根据你的设备检查运行时配置,请注意 **命令提示符** 和 **Windows PowerShell** 的区别。
 | 
			
		||||
以 Arrow Lake 为例,在 **命令提示符** 中需要设置 `set IPEX_LLM_NPU_ARL=1`,而在 **Windows PowerShell** 中是 `$env:IPEX_LLM_NPU_ARL = "1"`。
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue