[Nano] add pt dgpu inference how-to-guide (#8026)

* docs for arc dgpu how-to-guide

* minor adjustment + system info

* minor adjustment for appearance

* fix bugs

* add system info

* fix syntax errors

* adjust docs according to comments

* final adjustment

* delete gpu workflow testing
This commit is contained in:
Pingchuan Ma (Henry) 2023-04-12 19:18:16 +08:00 committed by GitHub
parent 2daaa6f7de
commit c6eccbfcc2
4 changed files with 7 additions and 1 deletions

View file

@ -169,6 +169,7 @@ subtrees:
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
- file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference - file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_gpu
- file: doc/Nano/Howto/Inference/TensorFlow/index - file: doc/Nano/Howto/Inference/TensorFlow/index
title: "TensorFlow" title: "TensorFlow"
subtrees: subtrees:

View file

@ -0,0 +1,3 @@
{
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/pytorch/accelerate_pytorch_inference_gpu.ipynb"
}

View file

@ -12,4 +12,5 @@ Inference Optimization: For PyTorch Users
* `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_ * `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_
* `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_ * `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_ * `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
* `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_ * `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_
* `How to accelerate a PyTorch inference pipeline using Intel ARC series dGPU <accelerate_pytorch_inference_gpu.html>`_

View file

@ -76,6 +76,7 @@ PyTorch
* `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_ * `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_ * `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
* `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_ * `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_
* `How to accelerate a PyTorch inference pipeline using Intel ARC series dGPU <Inference/PyTorch/accelerate_pytorch_inference_gpu.html>`_
TensorFlow TensorFlow
~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~