[Ollama] Update ipex-llm ollama readme to v0.4.6 (#12542)

* Update ipex-llm ollama readme to v0.4.6
This commit is contained in:
SONG Ge 2024-12-13 16:26:12 +08:00 committed by GitHub
parent d20a968ce2
commit 5402fc65c8
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
2 changed files with 10 additions and 4 deletions

View file

@ -17,9 +17,9 @@ See the demo of running LLaMA2-7B on Intel Arc GPU below.
</table> </table>
> [!NOTE] > [!NOTE]
> `ipex-llm[cpp]==2.2.0b20240826` is consistent with [v0.1.39](https://github.com/ollama/ollama/releases/tag/v0.1.39) of ollama. > `ipex-llm[cpp]==2.2.0b20241204` is consistent with [v0.3.6](https://github.com/ollama/ollama/releases/tag/v0.3.6) of ollama.
> >
> Our current version is consistent with [v0.3.6](https://github.com/ollama/ollama/releases/tag/v0.3.6) of ollama. > Our current version is consistent with [v0.4.6](https://github.com/ollama/ollama/releases/tag/v0.4.6) of ollama.
> [!NOTE] > [!NOTE]
> Starting from `ipex-llm[cpp]==2.2.0b20240912`, oneAPI dependency of `ipex-llm[cpp]` on Windows will switch from `2024.0.0` to `2024.2.1` . > Starting from `ipex-llm[cpp]==2.2.0b20240912`, oneAPI dependency of `ipex-llm[cpp]` on Windows will switch from `2024.0.0` to `2024.2.1` .
@ -80,6 +80,7 @@ You may launch the Ollama service as below:
export ZES_ENABLE_SYSMAN=1 export ZES_ENABLE_SYSMAN=1
source /opt/intel/oneapi/setvars.sh source /opt/intel/oneapi/setvars.sh
export SYCL_CACHE_PERSISTENT=1 export SYCL_CACHE_PERSISTENT=1
export LD_LIBRARY_PATH=.:$LD_LIBRARY_PATH
# [optional] under most circumstances, the following environment variable may improve performance, but sometimes this may also cause performance degradation # [optional] under most circumstances, the following environment variable may improve performance, but sometimes this may also cause performance degradation
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1 export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1
# [optional] if you want to run on single GPU, use below command to limit GPU may improve performance # [optional] if you want to run on single GPU, use below command to limit GPU may improve performance
@ -177,6 +178,8 @@ Then you can create the model in Ollama by `ollama create example -f Modelfile`
- For **Linux users**: - For **Linux users**:
```bash ```bash
source /opt/intel/oneapi/setvars.sh
export LD_LIBRARY_PATH=.:$LD_LIBRARY_PATH
export no_proxy=localhost,127.0.0.1 export no_proxy=localhost,127.0.0.1
./ollama create example -f Modelfile ./ollama create example -f Modelfile
./ollama run example ./ollama run example

View file

@ -17,9 +17,9 @@
</table> </table>
> [!NOTE] > [!NOTE]
> `ipex-llm[cpp]==2.2.0b20240826` 版本与官方 ollama 版本 [v0.1.39](https://github.com/ollama/ollama/releases/tag/v0.1.39) 一致。 > `ipex-llm[cpp]==2.2.0b20241204` 版本与官方 ollama 版本 [v0.3.6](https://github.com/ollama/ollama/releases/tag/v0.3.6) 一致。
> >
> `ipex-llm[cpp]` 的最新版本与官方 ollama 版本 [v0.3.6](https://github.com/ollama/ollama/releases/tag/v0.3.6) 一致。 > `ipex-llm[cpp]` 的最新版本与官方 ollama 版本 [v0.4.6](https://github.com/ollama/ollama/releases/tag/v0.4.6) 一致。
> [!NOTE] > [!NOTE]
> 从 `ipex-llm[cpp]==2.2.0b20240912` 版本开始Windows 上 `ipex-llm[cpp]` 依赖的 oneAPI 版本已从 `2024.0.0` 更新到 `2024.2.1` > 从 `ipex-llm[cpp]==2.2.0b20240912` 版本开始Windows 上 `ipex-llm[cpp]` 依赖的 oneAPI 版本已从 `2024.0.0` 更新到 `2024.2.1`
@ -80,6 +80,7 @@ IPEX-LLM 现在已支持在 Linux 和 Windows 系统上运行 `Ollama`。
export ZES_ENABLE_SYSMAN=1 export ZES_ENABLE_SYSMAN=1
source /opt/intel/oneapi/setvars.sh source /opt/intel/oneapi/setvars.sh
export SYCL_CACHE_PERSISTENT=1 export SYCL_CACHE_PERSISTENT=1
export LD_LIBRARY_PATH=.:$LD_LIBRARY_PATH
# [optional] under most circumstances, the following environment variable may improve performance, but sometimes this may also cause performance degradation # [optional] under most circumstances, the following environment variable may improve performance, but sometimes this may also cause performance degradation
export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1 export SYCL_PI_LEVEL_ZERO_USE_IMMEDIATE_COMMANDLISTS=1
# [optional] if you want to run on single GPU, use below command to limit GPU may improve performance # [optional] if you want to run on single GPU, use below command to limit GPU may improve performance
@ -174,6 +175,8 @@ PARAMETER num_predict 64
```bash ```bash
export no_proxy=localhost,127.0.0.1 export no_proxy=localhost,127.0.0.1
source /opt/intel/oneapi/setvars.sh
export LD_LIBRARY_PATH=.:$LD_LIBRARY_PATH
./ollama create example -f Modelfile ./ollama create example -f Modelfile
./ollama run example ./ollama run example
``` ```