Update Ollama portable zip QuickStart to fit new version (#12871)
* Update ollama portable zip quickstart * Update demo images
This commit is contained in:
parent
3ea5389a99
commit
21d6a78be0
2 changed files with 32 additions and 17 deletions
|
|
@ -5,6 +5,13 @@
|
||||||
|
|
||||||
This guide demonstrates how to use [Ollama portable zip](https://github.com/intel/ipex-llm/releases/tag/v2.2.0-nightly) to directly run Ollama on Intel GPU with `ipex-llm` (without the need of manual installations).
|
This guide demonstrates how to use [Ollama portable zip](https://github.com/intel/ipex-llm/releases/tag/v2.2.0-nightly) to directly run Ollama on Intel GPU with `ipex-llm` (without the need of manual installations).
|
||||||
|
|
||||||
|
> [!NOTE]
|
||||||
|
> Ollama portable zip has been verified on:
|
||||||
|
> - Intel Core Ultra processors
|
||||||
|
> - Intel Core 11th - 14th gen processors
|
||||||
|
> - Intel Arc A-Series GPU
|
||||||
|
> - Intel Arc B-Series GPU
|
||||||
|
|
||||||
## Table of Contents
|
## Table of Contents
|
||||||
- [Windows Quickstart](#windows-quickstart)
|
- [Windows Quickstart](#windows-quickstart)
|
||||||
- [Prerequisites](#prerequisites)
|
- [Prerequisites](#prerequisites)
|
||||||
|
|
@ -41,21 +48,21 @@ Then, extract the zip file to a folder.
|
||||||
|
|
||||||
### Step 2: Start Ollama Serve
|
### Step 2: Start Ollama Serve
|
||||||
|
|
||||||
Double-click `start-ollama.bat` in the extracted folder to start the Ollama service. A window will then pop up as shown below:
|
Start Ollama serve as follows:
|
||||||
|
|
||||||
|
- Open "Command Prompt" (cmd), and enter the extracted folder through `cd /d PATH\TO\EXTRACTED\FOLDER`
|
||||||
|
- Run `start-ollama.bat` in the "Command Prompt. A window will then pop up as shown below:
|
||||||
|
|
||||||
<div align="center">
|
<div align="center">
|
||||||
<img src="https://llm-assets.readthedocs.io/en/latest/_images/ollama_portable_start_ollama.png" width=80%/>
|
<img src="https://llm-assets.readthedocs.io/en/latest/_images/ollama_portable_start_ollama_new.png" width=80%/>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
### Step 3: Run Ollama
|
### Step 3: Run Ollama
|
||||||
|
|
||||||
You could then use Ollama to run LLMs on Intel GPUs as follows:
|
You could then use Ollama to run LLMs on Intel GPUs through running `ollama run deepseek-r1:7b` in the same "Command Prompt" (not the pop-up window). You may use any other model.
|
||||||
|
|
||||||
- Open "Command Prompt" (cmd), and enter the extracted folder through `cd /d PATH\TO\EXTRACTED\FOLDER`
|
|
||||||
- Run `ollama run deepseek-r1:7b` in the "Command Prompt" (you may use any other model)
|
|
||||||
|
|
||||||
<div align="center">
|
<div align="center">
|
||||||
<img src="https://llm-assets.readthedocs.io/en/latest/_images/ollama_portable_run_ollama.png" width=80%/>
|
<img src="https://llm-assets.readthedocs.io/en/latest/_images/ollama_portable_run_ollama_new.png" width=80%/>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
## Linux Quickstart
|
## Linux Quickstart
|
||||||
|
|
@ -110,7 +117,7 @@ For example, if you would like to run `deepseek-r1:7b` but the download speed fr
|
||||||
|
|
||||||
- For **Windows** users:
|
- For **Windows** users:
|
||||||
|
|
||||||
- Open "Command Prompt", and navigate to the extracted folder by `cd /d PATH\TO\EXTRACTED\FOLDER`
|
- In the "Command Prompt", navigate to the extracted folder by `cd /d PATH\TO\EXTRACTED\FOLDER`
|
||||||
- Run `set IPEX_LLM_MODEL_SOURCE=modelscope` in "Command Prompt"
|
- Run `set IPEX_LLM_MODEL_SOURCE=modelscope` in "Command Prompt"
|
||||||
- Run `ollama run deepseek-r1:7b`
|
- Run `ollama run deepseek-r1:7b`
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -5,6 +5,13 @@
|
||||||
|
|
||||||
本指南演示如何使用 [Ollama portable zip](https://github.com/intel/ipex-llm/releases/tag/v2.2.0-nightly) 通过 `ipex-llm` 在 Intel GPU 上直接免安装运行 Ollama。
|
本指南演示如何使用 [Ollama portable zip](https://github.com/intel/ipex-llm/releases/tag/v2.2.0-nightly) 通过 `ipex-llm` 在 Intel GPU 上直接免安装运行 Ollama。
|
||||||
|
|
||||||
|
> [!NOTE]
|
||||||
|
> Ollama portable zip 在如下设备上进行了验证:
|
||||||
|
> - Intel Core Ultra processors
|
||||||
|
> - Intel Core 11th - 14th gen processors
|
||||||
|
> - Intel Arc A-Series GPU
|
||||||
|
> - Intel Arc B-Series GPU
|
||||||
|
|
||||||
## 目录
|
## 目录
|
||||||
- [Windows用户指南](#windows用户指南)
|
- [Windows用户指南](#windows用户指南)
|
||||||
- [系统环境安装](#系统环境准备)
|
- [系统环境安装](#系统环境准备)
|
||||||
|
|
@ -41,21 +48,22 @@
|
||||||
|
|
||||||
### 步骤 2:启动 Ollama Serve
|
### 步骤 2:启动 Ollama Serve
|
||||||
|
|
||||||
在解压后的文件夹中双击 `start-ollama.bat` 即可启动 Ollama Serve。随后会弹出一个窗口,如下所示:
|
根据如下步骤启动 Ollama serve:
|
||||||
|
|
||||||
|
- 打开命令提示符(cmd),并通过在命令行输入指令 "cd /d PATH\TO\EXTRACTED\FOLDER" 进入解压缩后的文件夹
|
||||||
|
- 在命令提示符中运行 "start-ollama.bat" 即可启动 Ollama Serve。随后会弹出一个窗口,如下所示:
|
||||||
|
|
||||||
<div align="center">
|
<div align="center">
|
||||||
<img src="https://llm-assets.readthedocs.io/en/latest/_images/ollama_portable_start_ollama.png" width=80%/>
|
<img src="https://llm-assets.readthedocs.io/en/latest/_images/ollama_portable_start_ollama_new.png" width=80%/>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
### 步骤 3:运行 Ollama
|
### 步骤 3:运行 Ollama
|
||||||
|
|
||||||
在 Intel GPUs 上使用 Ollama 运行 LLMs,如下所示:
|
接下来通过在相同的命令提示符(非弹出的窗口)中运行 `ollama run deepseek-r1:7b`(可以将当前模型替换为你需要的模型),即可在 Intel GPUs 上使用 Ollama 运行 LLMs:
|
||||||
|
|
||||||
- 打开命令提示符(cmd),并通过在命令行输入指令 `cd /d PATH\TO\EXTRACTED\FOLDER` 进入解压后的文件夹
|
|
||||||
- 在命令提示符中运行 `ollama run deepseek-r1:7b`(可以将当前模型替换为你需要的模型)
|
|
||||||
|
|
||||||
<div align="center">
|
<div align="center">
|
||||||
<img src="https://llm-assets.readthedocs.io/en/latest/_images/ollama_portable_run_ollama.png" width=80%/>
|
<img src="https://llm-assets.readthedocs.io/en/latest/_images/ollama_portable_run_ollama_new.png" width=80%/>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
## Linux用户指南
|
## Linux用户指南
|
||||||
|
|
@ -108,7 +116,7 @@ Ollama 默认从 Ollama 库下载模型。通过在**运行 Ollama 之前**设
|
||||||
|
|
||||||
- 对于 **Windows** 用户:
|
- 对于 **Windows** 用户:
|
||||||
|
|
||||||
- 打开命令提示符(cmd),并通过 `cd /d PATH\TO\EXTRACTED\FOLDER` 命令进入解压后的文件夹
|
- 打开命令提示符通过 `cd /d PATH\TO\EXTRACTED\FOLDER` 命令进入解压后的文件夹
|
||||||
- 在命令提示符中运行 `set IPEX_LLM_MODEL_SOURCE=modelscope`
|
- 在命令提示符中运行 `set IPEX_LLM_MODEL_SOURCE=modelscope`
|
||||||
- 运行 `ollama run deepseek-r1:7b`
|
- 运行 `ollama run deepseek-r1:7b`
|
||||||
|
|
||||||
|
|
@ -134,7 +142,7 @@ Ollama 默认从 Ollama 库下载模型。通过在**运行 Ollama 之前**设
|
||||||
|
|
||||||
- 对于 **Windows** 用户:
|
- 对于 **Windows** 用户:
|
||||||
|
|
||||||
- 打开命令提示符(cmd),并通过 `cd /d PATH\TO\EXTRACTED\FOLDER` 命令进入解压后的文件夹
|
- 打开命令提示符,并通过 `cd /d PATH\TO\EXTRACTED\FOLDER` 命令进入解压后的文件夹
|
||||||
- 在命令提示符中将 `IPEX_LLM_NUM_CTX` 设置为所需长度,例如:`set IPEX_LLM_NUM_CTX=16384`
|
- 在命令提示符中将 `IPEX_LLM_NUM_CTX` 设置为所需长度,例如:`set IPEX_LLM_NUM_CTX=16384`
|
||||||
- 通过运行 `start-ollama.bat` 启动 Ollama serve
|
- 通过运行 `start-ollama.bat` 启动 Ollama serve
|
||||||
|
|
||||||
|
|
@ -161,7 +169,7 @@ Ollama 默认从 Ollama 库下载模型。通过在**运行 Ollama 之前**设
|
||||||
|
|
||||||
- 对于 **Windows** 用户:
|
- 对于 **Windows** 用户:
|
||||||
|
|
||||||
- 打开命令提示符(cmd),并通过 `cd /d PATH\TO\EXTRACTED\FOLDER` 命令进入解压后的文件夹
|
- 打开命令提示符,并通过 `cd /d PATH\TO\EXTRACTED\FOLDER` 命令进入解压后的文件夹
|
||||||
- 在命令提示符中设置 `ONEAPI_DEVICE_SELECTOR` 来定义你想使用的 Intel GPU,例如 `set ONEAPI_DEVICE_SELECTOR=level_zero:0`,其中`0`应该替换成你期望的 GPU id
|
- 在命令提示符中设置 `ONEAPI_DEVICE_SELECTOR` 来定义你想使用的 Intel GPU,例如 `set ONEAPI_DEVICE_SELECTOR=level_zero:0`,其中`0`应该替换成你期望的 GPU id
|
||||||
- 通过运行 `start-ollama.bat` 启动 Ollama serve
|
- 通过运行 `start-ollama.bat` 启动 Ollama serve
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue