Update ollama-release doc into v0.6.2 (#13094)
* Update ollama-release doc into v0.6.2 * update * revert signature changes
This commit is contained in:
		
							parent
							
								
									51b41faad7
								
							
						
					
					
						commit
						685a749adb
					
				
					 4 changed files with 13 additions and 10 deletions
				
			
		| 
						 | 
					@ -28,7 +28,7 @@ This guide demonstrates how to use [Ollama portable zip](https://github.com/ipex
 | 
				
			||||||
  - [Increase context length in Ollama](#increase-context-length-in-ollama)
 | 
					  - [Increase context length in Ollama](#increase-context-length-in-ollama)
 | 
				
			||||||
  - [Select specific GPU(s) to run Ollama when multiple ones are available](#select-specific-gpus-to-run-ollama-when-multiple-ones-are-available)
 | 
					  - [Select specific GPU(s) to run Ollama when multiple ones are available](#select-specific-gpus-to-run-ollama-when-multiple-ones-are-available)
 | 
				
			||||||
  - [Tune performance](#tune-performance)
 | 
					  - [Tune performance](#tune-performance)
 | 
				
			||||||
  - [Additional models supported after Ollama v0.5.4](#additional-models-supported-after-ollama-v054)
 | 
					  - [Additional models supported after Ollama v0.6.2](#additional-models-supported-after-ollama-v062)
 | 
				
			||||||
  - [Signature Verification](#signature-verification)
 | 
					  - [Signature Verification](#signature-verification)
 | 
				
			||||||
- [More details](ollama_quickstart.md)
 | 
					- [More details](ollama_quickstart.md)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -74,8 +74,10 @@ Check your GPU driver version, and update it if needed; we recommend following [
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### Step 1: Download and Extract
 | 
					### Step 1: Download and Extract
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
Download IPEX-LLM Ollama portable tgz for Ubuntu users from the [link](https://github.com/ipex-llm/ipex-llm/releases/tag/v2.3.0-nightly).
 | 
					Download IPEX-LLM Ollama portable tgz for Ubuntu users from the [link](https://github.com/ipex-llm/ipex-llm/releases/tag/v2.3.0-nightly).
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
Then open a terminal, extract the tgz file to a folder.
 | 
					Then open a terminal, extract the tgz file to a folder.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
```bash
 | 
					```bash
 | 
				
			||||||
| 
						 | 
					@ -206,9 +208,9 @@ To enable `SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS`, set it **before start
 | 
				
			||||||
> [!TIP]
 | 
					> [!TIP]
 | 
				
			||||||
> You could refer to [here](https://www.intel.com/content/www/us/en/developer/articles/guide/level-zero-immediate-command-lists.html) regarding more information about Level Zero Immediate Command Lists.
 | 
					> You could refer to [here](https://www.intel.com/content/www/us/en/developer/articles/guide/level-zero-immediate-command-lists.html) regarding more information about Level Zero Immediate Command Lists.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### Additional models supported after Ollama v0.5.4
 | 
					### Additional models supported after Ollama v0.6.2
 | 
				
			||||||
 | 
					
 | 
				
			||||||
The currently Ollama Portable Zip is based on Ollama v0.5.4; in addition, the following new models have also been supported in the Ollama Portable Zip:
 | 
					The currently Ollama Portable Zip is based on Ollama v0.6.2; in addition, the following new models have also been supported in the Ollama Portable Zip:
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| Model  | Download (Windows) | Download (Linux) | Model Link |
 | 
					| Model  | Download (Windows) | Download (Linux) | Model Link |
 | 
				
			||||||
| - | - | - | - |
 | 
					| - | - | - | - |
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -3,6 +3,7 @@
 | 
				
			||||||
   < <a href='./ollama_portable_zip_quickstart.md'>English</a> | <b>中文</b> >
 | 
					   < <a href='./ollama_portable_zip_quickstart.md'>English</a> | <b>中文</b> >
 | 
				
			||||||
</p>
 | 
					</p>
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					
 | 
				
			||||||
本指南演示如何使用 [Ollama portable zip](https://github.com/ipex-llm/ipex-llm/releases/tag/v2.3.0-nightly) 通过 `ipex-llm` 在 Intel GPU 上直接免安装运行 Ollama。
 | 
					本指南演示如何使用 [Ollama portable zip](https://github.com/ipex-llm/ipex-llm/releases/tag/v2.3.0-nightly) 通过 `ipex-llm` 在 Intel GPU 上直接免安装运行 Ollama。
 | 
				
			||||||
 | 
					
 | 
				
			||||||
> [!NOTE]
 | 
					> [!NOTE]
 | 
				
			||||||
| 
						 | 
					@ -28,7 +29,7 @@
 | 
				
			||||||
  - [在 Ollama 中增加上下文长度](#在-ollama-中增加上下文长度)
 | 
					  - [在 Ollama 中增加上下文长度](#在-ollama-中增加上下文长度)
 | 
				
			||||||
  - [在多块 GPU 可用时选择特定的 GPU 来运行 Ollama](#在多块-gpu-可用时选择特定的-gpu-来运行-ollama)
 | 
					  - [在多块 GPU 可用时选择特定的 GPU 来运行 Ollama](#在多块-gpu-可用时选择特定的-gpu-来运行-ollama)
 | 
				
			||||||
  - [性能调优](#性能调优)
 | 
					  - [性能调优](#性能调优)
 | 
				
			||||||
  - [Ollama v0.5.4 之后新增模型支持](#ollama-v054-之后新增模型支持)
 | 
					  - [Ollama v0.6.2 之后新增模型支持](#ollama-v062-之后新增模型支持)
 | 
				
			||||||
  - [签名验证](#签名验证)
 | 
					  - [签名验证](#签名验证)
 | 
				
			||||||
- [更多信息](ollama_quickstart.zh-CN.md)
 | 
					- [更多信息](ollama_quickstart.zh-CN.md)
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					@ -205,9 +206,9 @@ Ollama 默认从 Ollama 库下载模型。通过在**运行 Ollama 之前**设
 | 
				
			||||||
> [!TIP]
 | 
					> [!TIP]
 | 
				
			||||||
> 参考[此处文档](https://www.intel.com/content/www/us/en/developer/articles/guide/level-zero-immediate-command-lists.html)以获取更多 Level Zero Immediate Command Lists 相关信息。
 | 
					> 参考[此处文档](https://www.intel.com/content/www/us/en/developer/articles/guide/level-zero-immediate-command-lists.html)以获取更多 Level Zero Immediate Command Lists 相关信息。
 | 
				
			||||||
 | 
					
 | 
				
			||||||
### Ollama v0.5.4 之后新增模型支持
 | 
					### Ollama v0.6.2 之后新增模型支持
 | 
				
			||||||
 | 
					
 | 
				
			||||||
当前的 Ollama Portable Zip 基于 Ollama v0.5.4;此外,以下新模型也已在 Ollama Portable Zip 中得到支持:
 | 
					当前的 Ollama Portable Zip 基于 Ollama v0.6.2;此外,以下新模型也已在 Ollama Portable Zip 中得到支持:
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 模型  | 下载(Windows)| 下载(Linux)| 模型链接 |
 | 
					| 模型  | 下载(Windows)| 下载(Linux)| 模型链接 |
 | 
				
			||||||
| - | - | - | - |
 | 
					| - | - | - | - |
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -12,9 +12,9 @@
 | 
				
			||||||
> For installation on Intel Arc B-Series GPU (such as **B580**), please refer to this [guide](./bmg_quickstart.md).
 | 
					> For installation on Intel Arc B-Series GPU (such as **B580**), please refer to this [guide](./bmg_quickstart.md).
 | 
				
			||||||
 | 
					
 | 
				
			||||||
> [!NOTE]
 | 
					> [!NOTE]
 | 
				
			||||||
> Our current version is consistent with [v0.5.4](https://github.com/ollama/ollama/releases/tag/v0.5.4) of ollama.
 | 
					> Our current version is consistent with [v0.6.2](https://github.com/ollama/ollama/releases/tag/v0.6.2) of ollama.
 | 
				
			||||||
>
 | 
					>
 | 
				
			||||||
> `ipex-llm[cpp]==2.2.0b20250123` is consistent with [v0.5.1](https://github.com/ollama/ollama/releases/tag/v0.5.1) of ollama.
 | 
					> `ipex-llm[cpp]==2.2.0b20250413` is consistent with [v0.5.4](https://github.com/ollama/ollama/releases/tag/v0.5.4) of ollama.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
See the demo of running LLaMA2-7B on Intel Arc GPU below.
 | 
					See the demo of running LLaMA2-7B on Intel Arc GPU below.
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -12,9 +12,9 @@
 | 
				
			||||||
> 如果是在 Intel Arc B 系列 GPU 上安装(例如 **B580**),请参阅本[指南](./bmg_quickstart.md)。
 | 
					> 如果是在 Intel Arc B 系列 GPU 上安装(例如 **B580**),请参阅本[指南](./bmg_quickstart.md)。
 | 
				
			||||||
 | 
					
 | 
				
			||||||
> [!NOTE]
 | 
					> [!NOTE]
 | 
				
			||||||
>  `ipex-llm[cpp]` 的最新版本与官方 ollama 的 [v0.5.4](https://github.com/ollama/ollama/releases/tag/v0.5.4) 版本保持一致。
 | 
					>  `ipex-llm[cpp]` 的最新版本与官方 ollama 的 [v0.6.2](https://github.com/ollama/ollama/releases/tag/v0.6.2) 版本保持一致。
 | 
				
			||||||
>
 | 
					>
 | 
				
			||||||
> `ipex-llm[cpp]==2.2.0b20250123` 与官方 ollama 的 [v0.5.1](https://github.com/ollama/ollama/releases/tag/v0.5.1) 版本保持一致。
 | 
					> `ipex-llm[cpp]==2.2.0b20250413` 与官方 ollama 的 [v0.5.4](https://github.com/ollama/ollama/releases/tag/v0.5.4) 版本保持一致。
 | 
				
			||||||
 | 
					
 | 
				
			||||||
以下是在 Intel Arc GPU 上运行 LLaMA2-7B 的 DEMO 演示。
 | 
					以下是在 Intel Arc GPU 上运行 LLaMA2-7B 的 DEMO 演示。
 | 
				
			||||||
 | 
					
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue