[Nano] Revise outdated info in PyTorch Inference how-to guides (#7453)

* Change order of PyTorch Inference how-to guides order

* Revise accelerate_pytorch_inference_jit_ipex.ipynb

* Revise accelerat_pytorch_inference_onnx.ipynb

* Revise accelerate_pytorch_inference_openvino.ipynb

* Revise inference_optimizer_optimize.ipynb

* Revise multi_instance_pytorch_inference.ipynb

* Revise pytorch_context_manager.ipynb

* Revise pytorch_save_and_load_ipex.ipynb and other small fixes

* Revise pytorch_save_and_load_jit.ipynb and other small fixes

* Revise pytorch_save_and_load_onnx.ipynb and other small fixese

* Revise pytorch_save_and_load_openvino.ipynb

* Revise quantize_pytorch_inference_inc.ipynb and other small fixes

* Revise quantize_pytorch_inference_pot.ipynb and other small fixes

* Fix failed links in the index page

* Small fixes

* Update based on comments
This commit is contained in:
Yuwen Hu 2023-02-09 17:50:50 +08:00 committed by GitHub
parent 56d699d935
commit 6960c87b54
3 changed files with 16 additions and 22 deletions

View file

@ -156,10 +156,10 @@ subtrees:
title: "PyTorch" title: "PyTorch"
subtrees: subtrees:
- entries: - entries:
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
- file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager - file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager
@ -167,7 +167,7 @@ subtrees:
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize - file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
- file: doc/Nano/Howto/Inference/TensorFlow/index - file: doc/Nano/Howto/Inference/TensorFlow/index
title: "TensorFlow" title: "TensorFlow"
subtrees: subtrees:

View file

@ -1,18 +1,15 @@
Inference Optimization: For PyTorch Users Inference Optimization: For PyTorch Users
============================================= =============================================
* `How to find accelerated method with minimal latency using InferenceOptimizer <inference_optimizer_optimize.html>`_
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <accelerate_pytorch_inference_onnx.html>`_ * `How to accelerate a PyTorch inference pipeline through ONNXRuntime <accelerate_pytorch_inference_onnx.html>`_
* `How to accelerate a PyTorch inference pipeline through OpenVINO <accelerate_pytorch_inference_openvino.html>`_ * `How to accelerate a PyTorch inference pipeline through OpenVINO <accelerate_pytorch_inference_openvino.html>`_
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <accelerate_pytorch_inference_jit_ipex.html>`_ * `How to accelerate a PyTorch inference pipeline through JIT/IPEX <accelerate_pytorch_inference_jit_ipex.html>`_
* `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_ * `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor <quantize_pytorch_inference_inc.html>`_
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <quantize_pytorch_inference_inc.html>`_ * `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools <quantize_pytorch_inference_pot.html>`_
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <quantize_pytorch_inference_pot.html>`_ * `How to enable automatic context management for PyTorch inference on Nano optimized models <pytorch_context_manager.html>`_
* |pytorch_inference_context_manager_link|_
* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
* `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
* `How to save and load optimized ONNXRuntime model <pytorch_save_and_load_onnx.html>`_ * `How to save and load optimized ONNXRuntime model <pytorch_save_and_load_onnx.html>`_
* `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_ * `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_
* `How to find accelerated method with minimal latency using InferenceOptimizer <inference_optimizer_optimize.html>`_ * `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context`` * `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_
.. _pytorch_inference_context_manager_link: pytorch_context_manager.html

View file

@ -63,21 +63,18 @@ OpenVINO
PyTorch PyTorch
~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~
* `How to find accelerated method with minimal latency using InferenceOptimizer <Inference/PyTorch/inference_optimizer_optimize.html>`_
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <Inference/PyTorch/accelerate_pytorch_inference_onnx.html>`_ * `How to accelerate a PyTorch inference pipeline through ONNXRuntime <Inference/PyTorch/accelerate_pytorch_inference_onnx.html>`_
* `How to accelerate a PyTorch inference pipeline through OpenVINO <Inference/PyTorch/accelerate_pytorch_inference_openvino.html>`_ * `How to accelerate a PyTorch inference pipeline through OpenVINO <Inference/PyTorch/accelerate_pytorch_inference_openvino.html>`_
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <Inference/PyTorch/accelerate_pytorch_inference_jit_ipex.html>`_ * `How to accelerate a PyTorch inference pipeline through JIT/IPEX <Inference/PyTorch/accelerate_pytorch_inference_jit_ipex.html>`_
* `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_ * `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor <Inference/PyTorch/quantize_pytorch_inference_inc.html>`_
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <Inference/PyTorch/quantize_pytorch_inference_inc.html>`_ * `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools <Inference/PyTorch/quantize_pytorch_inference_pot.html>`_
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <Inference/PyTorch/quantize_pytorch_inference_pot.html>`_ * `How to enable automatic context management for PyTorch inference on Nano optimized models <Inference/PyTorch/pytorch_context_manager.html>`_
* |pytorch_inference_context_manager_link|_
* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
* `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
* `How to save and load optimized ONNXRuntime model <Inference/PyTorch/pytorch_save_and_load_onnx.html>`_ * `How to save and load optimized ONNXRuntime model <Inference/PyTorch/pytorch_save_and_load_onnx.html>`_
* `How to save and load optimized OpenVINO model <Inference/PyTorch/pytorch_save_and_load_openvino.html>`_ * `How to save and load optimized OpenVINO model <Inference/PyTorch/pytorch_save_and_load_openvino.html>`_
* `How to find accelerated method with minimal latency using InferenceOptimizer <Inference/PyTorch/inference_optimizer_optimize.html>`_ * `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context`` * `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_
.. _pytorch_inference_context_manager_link: Inference/PyTorch/pytorch_context_manager.html
TensorFlow TensorFlow
~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~