Miniconda/Anaconda -> Miniforge update in examples (#11194)
* Change installation address Change former address: "https://docs.conda.io/en/latest/miniconda.html#" to new address: "https://conda-forge.org/download/" for 63 occurrences under python\llm\example * Change Prompt Change "Anaconda Prompt" to "Miniforge Prompt" for 1 occurrence
This commit is contained in:
parent
5f13700c9f
commit
a644e9409b
64 changed files with 64 additions and 64 deletions
|
|
@ -21,7 +21,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
In the example [generate.py](./generate.py), we show a basic use case to load a GGUF LLaMA2 model into `ipex-llm` using `from_gguf()` API, with IPEX-LLM optimizations.
|
||||
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Aquila model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Aquila2 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a ChatGLM model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a CodeGemma model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a CodeShell model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a DeepSeek-MoE model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Recognize Tokens using `generate()` API
|
||||
In the example [recognize.py](./recognize.py), we show a basic use case for a Distil-Whisper model to conduct transcription using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Flan-t5 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for an Fuyu model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Multi-turn chat centered around an image using `chat()` API
|
||||
In the example [chat.py](./chat.py), we show a basic use case for an InternLM_XComposer model to start a multi-turn chat centered around an image using `chat()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -9,7 +9,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Mistral model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -9,7 +9,7 @@ To run these examples with IPEX-LLM on Intel CPUs, we have some recommended requ
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Mixtral model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations on Intel CPUs.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a phi-1_5 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a phi-2 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a phi-3 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a phixtral model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Multimodal chat using `chat()` API
|
||||
In the example [chat.py](./chat.py), we show a basic use case for a Qwen-VL model to start a multimodal chat using `chat()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for an Replit model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a StableLM model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for an Yi model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -9,7 +9,7 @@ In addition, you need to modify some files in Yuan2-2B-hf folder, since Flash at
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for an Yuan2 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Ziya model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Aquila2 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Synthesize speech with the given input text
|
||||
In the example [synthesize_speech.py](./synthesize_speech.py), we show a basic use case for Bark model to synthesize speech based on the given text, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Extract the feature of given text
|
||||
In the example [extract_feature.py](./extract_feature.py), we show a basic use case for a BERT model to extract the feature of given text, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a BlueLM model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a ChatGLM model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a ChatGLM3 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a CodeGemma model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a CodeLlama model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a CodeShell model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a DeciLM-7B model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a deepseek-moe model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Deepseek model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Recognize Tokens using `generate()` API
|
||||
In the example [recognize.py](./recognize.py), we show a basic use case for a Distil-Whisper model to conduct transcription using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Flan-t5 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for an Fuyu model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Multi-turn chat centered around an image using `chat()` API
|
||||
In the example [chat.py](./chat.py), we show a basic use case for an InternLM_XComposer model to start a multi-turn chat centered around an image using `chat()` API, with IPEX-LLM 'optimize_model' API.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a InternLM2 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Llama2 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Llama3 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Multi-turn chat centered around an image using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a LLaVA model to start a multi-turn chat centered around an image using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Mamba model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -9,7 +9,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Mistral model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -9,7 +9,7 @@ To run these examples with IPEX-LLM on Intel CPUs, we have some recommended requ
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Mixtral model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations on Intel CPUs.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a phi-1_5 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a phi-2 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -12,7 +12,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a phi-3 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a phixtral model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Multimodal chat using `chat()` API
|
||||
In the example [chat.py](./chat.py), we show a basic use case for a Qwen-VL model to start a multimodal chat using `chat()` API, with IPEX-LLM 'optimize_model' API.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Qwen1.5 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Skywork model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a SOLAR model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a StableLM model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a WizardCoder-Python model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Yi model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -9,7 +9,7 @@ In addition, you need to modify some files in Yuan2-2B-hf folder, since Flash at
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for an Yuan2 model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -7,7 +7,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
## Example: Predict Tokens using `generate()` API
|
||||
In the example [generate.py](./generate.py), we show a basic use case for a Ziya model to predict the next N tokens using `generate()` API, with IPEX-LLM 'optimize_model' API.
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
|
||||
|
|
|
|||
|
|
@ -8,7 +8,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
In this example, we run inference for a Llama2 model to showcase the speed of EAGLE with IPEX-LLM on MT-bench data on Intel CPUs.
|
||||
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
```bash
|
||||
|
|
|
|||
|
|
@ -19,7 +19,7 @@ To run these examples with IPEX-LLM, we have some recommended requirements for y
|
|||
In the example [generate.py](./generate.py), we show a basic use case to load a GGUF LLaMA2 model into `ipex-llm` using `from_gguf()` API, with IPEX-LLM optimizations.
|
||||
|
||||
### 1. Install
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
```bash
|
||||
|
|
|
|||
|
|
@ -10,7 +10,7 @@ To run these examples with IPEX-LLM on Intel GPUs, we have some recommended requ
|
|||
In the example [generate.py](./generate.py), we show a basic use case for a CodeGemma model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations on Intel GPUs.
|
||||
### 1. Install
|
||||
#### 1.1 Installation on Linux
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
```bash
|
||||
|
|
|
|||
|
|
@ -10,7 +10,7 @@ To run these examples with IPEX-LLM on Intel GPUs, we have some recommended requ
|
|||
In the example [generate.py](./generate.py), we show a basic use case for a CodeGemma model to predict the next N tokens using `generate()` API, with IPEX-LLM INT4 optimizations on Intel GPUs.
|
||||
### 1. Install
|
||||
#### 1.1 Installation on Linux
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://docs.conda.io/en/latest/miniconda.html#).
|
||||
We suggest using conda to manage the Python environment. For more information about conda installation, please refer to [here](https://conda-forge.org/download/).
|
||||
|
||||
After installing conda, create a Python environment for IPEX-LLM:
|
||||
```bash
|
||||
|
|
|
|||
|
|
@ -17,7 +17,7 @@ sudo apt install xpu-smi
|
|||
|
||||
### Usage
|
||||
|
||||
* After installing `ipex-llm`, open a terminal (on Linux) or **Anaconda Prompt** (on Windows), and activate the conda environment you have created for running `ipex-llm`:
|
||||
* After installing `ipex-llm`, open a terminal (on Linux) or **Miniforge Prompt** (on Windows), and activate the conda environment you have created for running `ipex-llm`:
|
||||
```
|
||||
conda activate llm
|
||||
```
|
||||
|
|
|
|||
Loading…
Reference in a new issue