[Nano] Revise outdated info in PyTorch Inference how-to guides (#7453)
* Change order of PyTorch Inference how-to guides order * Revise accelerate_pytorch_inference_jit_ipex.ipynb * Revise accelerat_pytorch_inference_onnx.ipynb * Revise accelerate_pytorch_inference_openvino.ipynb * Revise inference_optimizer_optimize.ipynb * Revise multi_instance_pytorch_inference.ipynb * Revise pytorch_context_manager.ipynb * Revise pytorch_save_and_load_ipex.ipynb and other small fixes * Revise pytorch_save_and_load_jit.ipynb and other small fixes * Revise pytorch_save_and_load_onnx.ipynb and other small fixese * Revise pytorch_save_and_load_openvino.ipynb * Revise quantize_pytorch_inference_inc.ipynb and other small fixes * Revise quantize_pytorch_inference_pot.ipynb and other small fixes * Fix failed links in the index page * Small fixes * Update based on comments
This commit is contained in:
parent
56d699d935
commit
6960c87b54
3 changed files with 16 additions and 22 deletions
|
|
@ -156,10 +156,10 @@ subtrees:
|
||||||
title: "PyTorch"
|
title: "PyTorch"
|
||||||
subtrees:
|
subtrees:
|
||||||
- entries:
|
- entries:
|
||||||
|
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
|
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
|
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
|
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
|
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc
|
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot
|
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager
|
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager
|
||||||
|
|
@ -167,7 +167,7 @@ subtrees:
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit
|
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
|
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
|
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
- file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
|
||||||
- file: doc/Nano/Howto/Inference/TensorFlow/index
|
- file: doc/Nano/Howto/Inference/TensorFlow/index
|
||||||
title: "TensorFlow"
|
title: "TensorFlow"
|
||||||
subtrees:
|
subtrees:
|
||||||
|
|
|
||||||
|
|
@ -1,18 +1,15 @@
|
||||||
Inference Optimization: For PyTorch Users
|
Inference Optimization: For PyTorch Users
|
||||||
=============================================
|
=============================================
|
||||||
|
|
||||||
|
* `How to find accelerated method with minimal latency using InferenceOptimizer <inference_optimizer_optimize.html>`_
|
||||||
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <accelerate_pytorch_inference_onnx.html>`_
|
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <accelerate_pytorch_inference_onnx.html>`_
|
||||||
* `How to accelerate a PyTorch inference pipeline through OpenVINO <accelerate_pytorch_inference_openvino.html>`_
|
* `How to accelerate a PyTorch inference pipeline through OpenVINO <accelerate_pytorch_inference_openvino.html>`_
|
||||||
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <accelerate_pytorch_inference_jit_ipex.html>`_
|
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <accelerate_pytorch_inference_jit_ipex.html>`_
|
||||||
* `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_
|
* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor <quantize_pytorch_inference_inc.html>`_
|
||||||
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <quantize_pytorch_inference_inc.html>`_
|
* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools <quantize_pytorch_inference_pot.html>`_
|
||||||
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <quantize_pytorch_inference_pot.html>`_
|
* `How to enable automatic context management for PyTorch inference on Nano optimized models <pytorch_context_manager.html>`_
|
||||||
* |pytorch_inference_context_manager_link|_
|
|
||||||
* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
|
|
||||||
* `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
|
|
||||||
* `How to save and load optimized ONNXRuntime model <pytorch_save_and_load_onnx.html>`_
|
* `How to save and load optimized ONNXRuntime model <pytorch_save_and_load_onnx.html>`_
|
||||||
* `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_
|
* `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_
|
||||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <inference_optimizer_optimize.html>`_
|
* `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
|
||||||
|
* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
|
||||||
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context``
|
* `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_
|
||||||
.. _pytorch_inference_context_manager_link: pytorch_context_manager.html
|
|
||||||
|
|
@ -63,21 +63,18 @@ OpenVINO
|
||||||
PyTorch
|
PyTorch
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
|
* `How to find accelerated method with minimal latency using InferenceOptimizer <Inference/PyTorch/inference_optimizer_optimize.html>`_
|
||||||
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <Inference/PyTorch/accelerate_pytorch_inference_onnx.html>`_
|
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <Inference/PyTorch/accelerate_pytorch_inference_onnx.html>`_
|
||||||
* `How to accelerate a PyTorch inference pipeline through OpenVINO <Inference/PyTorch/accelerate_pytorch_inference_openvino.html>`_
|
* `How to accelerate a PyTorch inference pipeline through OpenVINO <Inference/PyTorch/accelerate_pytorch_inference_openvino.html>`_
|
||||||
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <Inference/PyTorch/accelerate_pytorch_inference_jit_ipex.html>`_
|
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <Inference/PyTorch/accelerate_pytorch_inference_jit_ipex.html>`_
|
||||||
* `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_
|
* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor <Inference/PyTorch/quantize_pytorch_inference_inc.html>`_
|
||||||
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <Inference/PyTorch/quantize_pytorch_inference_inc.html>`_
|
* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools <Inference/PyTorch/quantize_pytorch_inference_pot.html>`_
|
||||||
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <Inference/PyTorch/quantize_pytorch_inference_pot.html>`_
|
* `How to enable automatic context management for PyTorch inference on Nano optimized models <Inference/PyTorch/pytorch_context_manager.html>`_
|
||||||
* |pytorch_inference_context_manager_link|_
|
|
||||||
* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
|
|
||||||
* `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
|
|
||||||
* `How to save and load optimized ONNXRuntime model <Inference/PyTorch/pytorch_save_and_load_onnx.html>`_
|
* `How to save and load optimized ONNXRuntime model <Inference/PyTorch/pytorch_save_and_load_onnx.html>`_
|
||||||
* `How to save and load optimized OpenVINO model <Inference/PyTorch/pytorch_save_and_load_openvino.html>`_
|
* `How to save and load optimized OpenVINO model <Inference/PyTorch/pytorch_save_and_load_openvino.html>`_
|
||||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <Inference/PyTorch/inference_optimizer_optimize.html>`_
|
* `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
|
||||||
|
* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
|
||||||
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context``
|
* `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_
|
||||||
.. _pytorch_inference_context_manager_link: Inference/PyTorch/pytorch_context_manager.html
|
|
||||||
|
|
||||||
TensorFlow
|
TensorFlow
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue