diff --git a/README.md b/README.md index 5dac42f5..13f09c75 100644 --- a/README.md +++ b/README.md @@ -92,8 +92,7 @@ output = tokenizer.batch_decode(output_ids) You may install **`bigdl-llm`** on Intel GPU as follows: > Note: See the [GPU installation guide](https://bigdl.readthedocs.io/en/latest/doc/LLM/Overview/install_gpu.html) for more details. ```bash -# below command will install intel_extension_for_pytorch==2.0.110+xpu as default -# you can install specific ipex/torch version for your need +# below command will install intel_extension_for_pytorch==2.1.10+xpu as default pip install --pre --upgrade bigdl-llm[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu ``` > Note: `bigdl-llm` has been tested on Python 3.9 diff --git a/docs/readthedocs/source/index.rst b/docs/readthedocs/source/index.rst index 508fa8eb..2f841b16 100644 --- a/docs/readthedocs/source/index.rst +++ b/docs/readthedocs/source/index.rst @@ -120,8 +120,7 @@ You may install ``bigdl-llm`` on Intel GPU as follows as follows: .. code-block:: console - # below command will install intel_extension_for_pytorch==2.0.110+xpu as default - # you can install specific ipex/torch version for your need + # below command will install intel_extension_for_pytorch==2.1.10+xpu as default pip install --pre --upgrade bigdl-llm[xpu] -f https://developer.intel.com/ipex-whl-stable-xpu .. note::