Tiny doc fix (#12405)
This commit is contained in:
parent
548dec5185
commit
d1cde7fac4
2 changed files with 2 additions and 2 deletions
|
|
@ -125,7 +125,7 @@ clinfo | grep "Device Name"
|
||||||
whose output should contain `Intel(R) Arc(TM) Graphics` or `Intel(R) Graphics` based on your GPU model.
|
whose output should contain `Intel(R) Arc(TM) Graphics` or `Intel(R) Graphics` based on your GPU model.
|
||||||
|
|
||||||
> [!TIP]
|
> [!TIP]
|
||||||
> You could refer to the [official driver guide for client GPUS](https://dgpu-docs.intel.com/driver/client/overview.html#installing-client-gpus-on-ubuntu-desktop-22-04-lts) for more information.
|
> You could refer to the [official driver guide for client GPUs](https://dgpu-docs.intel.com/driver/client/overview.html#installing-client-gpus-on-ubuntu-desktop-22-04-lts) for more information.
|
||||||
|
|
||||||
|
|
||||||
#### For other Intel iGPU and dGPU
|
#### For other Intel iGPU and dGPU
|
||||||
|
|
|
||||||
|
|
@ -155,7 +155,7 @@ To monitor your GPU's performance and status (e.g. memory consumption, utilizati
|
||||||
|
|
||||||
## A Quick Example
|
## A Quick Example
|
||||||
|
|
||||||
Now let's play with a real LLM. We'll be using the [Qwen2-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct) model, a 1.8 billion parameter LLM for this demonstration. Follow the steps below to setup and run the model, and observe how it responds to a prompt "What is AI?".
|
Now let's play with a real LLM. We'll be using the [Qwen2-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2-1.5B-Instruct) model, a 1.5 billion parameter LLM for this demonstration. Follow the steps below to setup and run the model, and observe how it responds to a prompt "What is AI?".
|
||||||
|
|
||||||
- Step 1: Follow [Runtime Configurations Section](#step-1-runtime-configurations) above to prepare your runtime environment.
|
- Step 1: Follow [Runtime Configurations Section](#step-1-runtime-configurations) above to prepare your runtime environment.
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue