diff --git a/docs/mddocs/Quickstart/README.md b/docs/mddocs/Quickstart/README.md index 88bb1899..9f094652 100644 --- a/docs/mddocs/Quickstart/README.md +++ b/docs/mddocs/Quickstart/README.md @@ -24,6 +24,7 @@ This section includes efficient guide to show you how to: - [Run llama.cpp with IPEX-LLM on Intel GPU](./llama_cpp_quickstart.md) - [Run Ollama with IPEX-LLM on Intel GPU](./ollama_quickstart.md) - [Run Ollama Portable Zip on Intel GPU with IPEX-LLM](./ollama_portable_zip_quickstart.md) +- [Run llama.cpp Portable Zip on Intel NPU with IPEX-LLM](./llama_cpp_npu_portable_zip_quickstart.md) - [Run Llama 3 on Intel GPU using llama.cpp and ollama with IPEX-LLM](./llama3_llamacpp_ollama_quickstart.md) - [Run RAGFlow with IPEX-LLM on Intel GPU](./ragflow_quickstart.md) - [Run GraphRAG with IPEX-LLM on Intel GPU](./graphrag_quickstart.md) diff --git a/docs/mddocs/Quickstart/llama_cpp_npu_portable_zip_quickstart.md b/docs/mddocs/Quickstart/llama_cpp_npu_portable_zip_quickstart.md new file mode 100644 index 00000000..c5e225dc --- /dev/null +++ b/docs/mddocs/Quickstart/llama_cpp_npu_portable_zip_quickstart.md @@ -0,0 +1,65 @@ +# Run llama.cpp Portable Zip on Intel NPU with IPEX-LLM +
+ < English | 中文 > +
+ +IPEX-LLM provides llama.cpp support for running GGUF models on Intel NPU. This guide demonstrates how to use [llama.cpp NPU portable zip](https://github.com/intel/ipex-llm/releases/tag/v2.2.0-nightly) to directly run on Intel NPU (without the need of manual installations). + +> [!IMPORTANT] +> +> - IPEX-LLM currently only supports Windows on Intel NPU. +> - Only `meta-llama/Llama-3.2-3B-Instruct`, `deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B` and `deepseek-ai/DeepSeek-R1-Distill-Qwen-7B` are supported. + + +## Table of Contents +- [Prerequisites](#prerequisites) +- [Step 1: Download and Unzip](#step-1-download-and-unzip) +- [Step 2: Setup](#step-2-setup) +- [Step 3: Run GGUF Model](#step-3-run-gguf-model) +- [More details](npu_quickstart.md) + + +## Prerequisites + +Check your NPU driver version, and update it if needed: + +- Please use NPU driver version [32.0.100.3104](https://www.intel.com/content/www/us/en/download/794734/838895/intel-npu-driver-windows.html). +- And you could refer to [here](npu_quickstart.md#update-npu-driver) for details about NPU driver update. + +## Step 1: Download and Unzip + +Download IPEX-LLM llama.cpp NPU portable zip for Windows users from the [link](https://github.com/intel/ipex-llm/releases/tag/v2.2.0-nightly). + +Then, extract the zip file to a folder. + +## Step 2: Setup + +- Open "Command Prompt" (cmd), and enter the extracted folder through `cd /d PATH\TO\EXTRACTED\FOLDER` +- Runtime configuration based on your device: + - For **Intel Core™ Ultra Processors (Series 2) with processor number 2xxV (code name Lunar Lake)**: + + - For Intel Core™ Ultra 7 Processor 258V: + No runtime configuration required. + + - For Intel Core™ Ultra 5 Processor 228V & 226V: + ```cmd + set IPEX_LLM_NPU_DISABLE_COMPILE_OPT=1 + ``` + + - For **Intel Core™ Ultra Processors (Series 2) with processor number 2xxK or 2xxH (code name Arrow Lake)**: + ```cmd + set IPEX_LLM_NPU_ARL=1 + ``` + + - For **Intel Core™ Ultra Processors (Series 1) with processor number 1xxH (code name Meteor Lake)**: + ```cmd + set IPEX_LLM_NPU_MTL=1 + ``` + +## Step 3: Run GGUF Model + +You could then use cli tool to run GGUF models on Intel NPU through running `llama-cli-npu.exe` in the "Command Prompt" as following: + +```cmd +llama-cli-npu.exe -m DeepSeek-R1-Distill-Qwen-7B-Q6_K.gguf -n 32 --prompt "What is AI?" +``` diff --git a/docs/mddocs/Quickstart/llama_cpp_npu_portable_zip_quickstart.zh-CN.md b/docs/mddocs/Quickstart/llama_cpp_npu_portable_zip_quickstart.zh-CN.md new file mode 100644 index 00000000..1e093569 --- /dev/null +++ b/docs/mddocs/Quickstart/llama_cpp_npu_portable_zip_quickstart.zh-CN.md @@ -0,0 +1,65 @@ +# 使用 IPEX-LLM 在 Intel NPU 上运行 llama.cpp Portable Zip ++ < English | 中文 > +
+ +IPEX-LLM 提供了 llama.cpp 的相关支持以在 Intel NPU 上运行 GGUF 模型。本指南演示如何使用 [llama.cpp NPU portable zip](https://github.com/intel/ipex-llm/releases/tag/v2.2.0-nightly) 在 Intel NPU 上直接免安装运行。 + +> [!IMPORTANT] +> +> - IPEX-LLM 在 Intel NPU 上暂时只支持 Windows。 +> - 目前支持的模型有 `meta-llama/Llama-3.2-3B-Instruct`, `deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B` 和 `deepseek-ai/DeepSeek-R1-Distill-Qwen-7B`。 + + +## 目录 +- [系统环境安装](#系统环境准备) +- [步骤 1:下载和解压](#步骤-1下载和解压) +- [步骤 2:启动](#步骤-2启动) +- [步骤 3:运行 GGUF 模型](#步骤-3运行-gguf-模型) +- [更多信息](npu_quickstart.md) + + +## 系统环境准备 + +检查你的 NPU 驱动程序版本,并根据需要进行更新: + +- 请使用 NPU 驱动版本 [32.0.100.3104](https://www.intel.com/content/www/us/en/download/794734/838895/intel-npu-driver-windows.html) +- 你也可以参考这里 (https://github.com/intel/ipex-llm/blob/main/docs/mddocs/Quickstart/npu_quickstart.md#update-npu-driver) 了解更多关于 NPU 驱动程序更新的细节 + +## 步骤 1:下载和解压 + +从此[链接](https://github.com/intel/ipex-llm/releases/tag/v2.2.0-nightly)下载 IPEX-LLM llama.cpp NPU portable zip。 + +然后,将 zip 文件解压到一个文件夹中。 + +## 步骤 2:启动 + +- 打开命令提示符(cmd),并通过在命令行输入指令 "cd /d PATH\TO\EXTRACTED\FOLDER" 进入解压缩后的文件夹 +- 根据你的设备完成运行配置: + - 对于 **处理器为 2xxV 的 Intel Core™ Ultra Processors (Series 2) (代号 Lunar Lake)**: + + - 对于 Intel Core™ Ultra 7 Processor 258V: + 不需要额外的配置 + + - 对于 Intel Core™ Ultra 5 Processor 228V & 226V: + ```cmd + set IPEX_LLM_NPU_DISABLE_COMPILE_OPT=1 + ``` + + - 对于 **处理器为 2xxK 或者 2xxH 的 Intel Core™ Ultra Processors (Series 2) (代号 Arrow Lake)**: + ```cmd + set IPEX_LLM_NPU_ARL=1 + ``` + + - 对于 **处理器为 1xxH 的 Intel Core™ Ultra Processors (Series 1) (代号 Meteor Lake)**: + ```cmd + set IPEX_LLM_NPU_MTL=1 + ``` + +## 步骤 3:运行 GGUF 模型 + +你可以在命令行中使用 cli 工具 `llama-cli-npu.exe` 以在 Intel NPU 上运行 GGUF 模型: + +```cmd +llama-cli-npu.exe -m DeepSeek-R1-Distill-Qwen-7B-Q6_K.gguf -n 32 --prompt "What is AI?" +``` diff --git a/docs/mddocs/Quickstart/npu_quickstart.md b/docs/mddocs/Quickstart/npu_quickstart.md index b694969f..376afa8e 100644 --- a/docs/mddocs/Quickstart/npu_quickstart.md +++ b/docs/mddocs/Quickstart/npu_quickstart.md @@ -193,6 +193,10 @@ Refer to the following table for verified models: | LLaMA 3.2 | [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) | Meteor Lake, Lunar Lake, Arrow Lake | | DeepSeek-R1 | [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B), [deepseek-ai/DeepSeek-R1-Distill-Qwen-7B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B) | Meteor Lake, Lunar Lake, Arrow Lake | +> [!Important] +> +> You may use [llama.cpp portable zip](./llama_cpp_npu_portable_zip_quickstart.md) to directly run GGUF models on Intel NPU with ipex-llm (***without the need of manual installations***). + ### Run GGUF model using CLI tool #### Setup for running llama.cpp