Ollama quickstart update (#10806)
* add ollama doc for OLLAMA_NUM_GPU * remove useless params * revert unexpected changes back * move env setting to server part * update
This commit is contained in:
parent
08458b4f74
commit
fbd1743b5e
1 changed files with 10 additions and 19 deletions
|
|
@ -44,8 +44,7 @@ Activate the `llm-cpp` conda environment and initialize Ollama by executing the
|
|||
|
||||
### 3 Run Ollama Serve
|
||||
|
||||
|
||||
Launch the Ollama service:
|
||||
You may launch the Ollama service as below:
|
||||
|
||||
```eval_rst
|
||||
.. tabs::
|
||||
|
|
@ -53,6 +52,7 @@ Launch the Ollama service:
|
|||
|
||||
.. code-block:: bash
|
||||
|
||||
export OLLAMA_NUM_GPU=999
|
||||
export no_proxy=localhost,127.0.0.1
|
||||
export ZES_ENABLE_SYSMAN=1
|
||||
source /opt/intel/oneapi/setvars.sh
|
||||
|
|
@ -65,6 +65,7 @@ Launch the Ollama service:
|
|||
|
||||
.. code-block:: bash
|
||||
|
||||
set OLLAMA_NUM_GPU=999
|
||||
set no_proxy=localhost,127.0.0.1
|
||||
set ZES_ENABLE_SYSMAN=1
|
||||
call "C:\Program Files (x86)\Intel\oneAPI\setvars.bat"
|
||||
|
|
@ -74,6 +75,11 @@ Launch the Ollama service:
|
|||
```
|
||||
|
||||
```eval_rst
|
||||
.. note::
|
||||
|
||||
Please set environment variable ``OLLAMA_NUM_GPU`` to ``999`` to make sure all layers of your model are running on Intel GPU, otherwise, some layers may run on CPU.
|
||||
```
|
||||
|
||||
.. note::
|
||||
|
||||
To allow the service to accept connections from all IP addresses, use `OLLAMA_HOST=0.0.0.0 ./ollama serve` instead of just `./ollama serve`.
|
||||
|
|
@ -111,8 +117,7 @@ model**, e.g. `dolphin-phi`.
|
|||
{
|
||||
"model": "<model_name>",
|
||||
"prompt": "Why is the sky blue?",
|
||||
"stream": false,
|
||||
"options":{"num_gpu": 999}
|
||||
"stream": false
|
||||
}'
|
||||
|
||||
.. tab:: Windows
|
||||
|
|
@ -125,19 +130,12 @@ model**, e.g. `dolphin-phi`.
|
|||
{
|
||||
\"model\": \"<model_name>\",
|
||||
\"prompt\": \"Why is the sky blue?\",
|
||||
\"stream\": false,
|
||||
\"options\":{\"num_gpu\": 999}
|
||||
\"stream\": false
|
||||
}"
|
||||
|
||||
```
|
||||
|
||||
|
||||
```eval_rst
|
||||
.. note::
|
||||
|
||||
Please don't forget to set ``"options":{"num_gpu": 999}`` to make sure all layers of your model are running on Intel GPU, otherwise, some layers may run on CPU.
|
||||
```
|
||||
|
||||
#### Using Ollama Run GGUF models
|
||||
|
||||
Ollama supports importing GGUF models in the Modelfile, for example, suppose you have downloaded a `mistral-7b-instruct-v0.1.Q4_K_M.gguf` from [Mistral-7B-Instruct-v0.1-GGUF](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.1-GGUF/tree/main), then you can create a file named `Modelfile`:
|
||||
|
|
@ -145,16 +143,9 @@ Ollama supports importing GGUF models in the Modelfile, for example, suppose you
|
|||
```bash
|
||||
FROM ./mistral-7b-instruct-v0.1.Q4_K_M.gguf
|
||||
TEMPLATE [INST] {{ .Prompt }} [/INST]
|
||||
PARAMETER num_gpu 999
|
||||
PARAMETER num_predict 64
|
||||
```
|
||||
|
||||
```eval_rst
|
||||
.. note::
|
||||
|
||||
Please don't forget to set ``PARAMETER num_gpu 999`` to make sure all layers of your model are running on Intel GPU, otherwise, some layers may run on CPU.
|
||||
```
|
||||
|
||||
Then you can create the model in Ollama by `ollama create example -f Modelfile` and use `ollama run` to run the model directly on console.
|
||||
|
||||
```eval_rst
|
||||
|
|
|
|||
Loading…
Reference in a new issue