[Nano] How-to Guide: Inference via Openvino on Intel GPUs (#7212)
* add openvino gpu inference guide * enable online doc for how-to * fix * fix layout error * update w.r.t. comments * fix * fix * fix error
This commit is contained in:
parent
cc6f9b4dd2
commit
3543a58723
3 changed files with 5 additions and 0 deletions
|
|
@ -118,6 +118,7 @@ subtrees:
|
||||||
- file: doc/Nano/Howto/Training/General/choose_num_processes_training
|
- file: doc/Nano/Howto/Training/General/choose_num_processes_training
|
||||||
- file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference
|
- file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference
|
||||||
- file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference_async
|
- file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference_async
|
||||||
|
- file: doc/Nano/Howto/Inference/OpenVINO/accelerate_inference_openvino_gpu
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
|
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
|
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
|
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,3 @@
|
||||||
|
{
|
||||||
|
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/openvino/accelerate_inference_openvino_gpu.ipynb"
|
||||||
|
}
|
||||||
|
|
@ -58,6 +58,7 @@ OpenVINO
|
||||||
|
|
||||||
* `How to run inference on OpenVINO model <Inference/OpenVINO/openvino_inference.html>`_
|
* `How to run inference on OpenVINO model <Inference/OpenVINO/openvino_inference.html>`_
|
||||||
* `How to run asynchronous inference on OpenVINO model <Inference/OpenVINO/openvino_inference_async.html>`_
|
* `How to run asynchronous inference on OpenVINO model <Inference/OpenVINO/openvino_inference_async.html>`_
|
||||||
|
* `How to accelerate a PyTorch / TensorFlow inference pipeline on Intel GPUs through OpenVINO <Inference/OpenVINO/accelerate_inference_openvino_gpu.html>`_
|
||||||
|
|
||||||
PyTorch
|
PyTorch
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue