[Nano] Revise outdated info in PyTorch Inference how-to guides (#7453)
* Change order of PyTorch Inference how-to guides order * Revise accelerate_pytorch_inference_jit_ipex.ipynb * Revise accelerat_pytorch_inference_onnx.ipynb * Revise accelerate_pytorch_inference_openvino.ipynb * Revise inference_optimizer_optimize.ipynb * Revise multi_instance_pytorch_inference.ipynb * Revise pytorch_context_manager.ipynb * Revise pytorch_save_and_load_ipex.ipynb and other small fixes * Revise pytorch_save_and_load_jit.ipynb and other small fixes * Revise pytorch_save_and_load_onnx.ipynb and other small fixese * Revise pytorch_save_and_load_openvino.ipynb * Revise quantize_pytorch_inference_inc.ipynb and other small fixes * Revise quantize_pytorch_inference_pot.ipynb and other small fixes * Fix failed links in the index page * Small fixes * Update based on comments
This commit is contained in:
parent
56d699d935
commit
6960c87b54
3 changed files with 16 additions and 22 deletions
|
|
@ -156,10 +156,10 @@ subtrees:
|
|||
title: "PyTorch"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager
|
||||
|
|
@ -167,7 +167,7 @@ subtrees:
|
|||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/index
|
||||
title: "TensorFlow"
|
||||
subtrees:
|
||||
|
|
|
|||
|
|
@ -1,18 +1,15 @@
|
|||
Inference Optimization: For PyTorch Users
|
||||
=============================================
|
||||
|
||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <inference_optimizer_optimize.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <accelerate_pytorch_inference_onnx.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through OpenVINO <accelerate_pytorch_inference_openvino.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <accelerate_pytorch_inference_jit_ipex.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_
|
||||
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <quantize_pytorch_inference_inc.html>`_
|
||||
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <quantize_pytorch_inference_pot.html>`_
|
||||
* |pytorch_inference_context_manager_link|_
|
||||
* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
|
||||
* `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
|
||||
* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor <quantize_pytorch_inference_inc.html>`_
|
||||
* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools <quantize_pytorch_inference_pot.html>`_
|
||||
* `How to enable automatic context management for PyTorch inference on Nano optimized models <pytorch_context_manager.html>`_
|
||||
* `How to save and load optimized ONNXRuntime model <pytorch_save_and_load_onnx.html>`_
|
||||
* `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_
|
||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <inference_optimizer_optimize.html>`_
|
||||
|
||||
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context``
|
||||
.. _pytorch_inference_context_manager_link: pytorch_context_manager.html
|
||||
* `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
|
||||
* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_
|
||||
|
|
@ -63,21 +63,18 @@ OpenVINO
|
|||
PyTorch
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <Inference/PyTorch/inference_optimizer_optimize.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <Inference/PyTorch/accelerate_pytorch_inference_onnx.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through OpenVINO <Inference/PyTorch/accelerate_pytorch_inference_openvino.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <Inference/PyTorch/accelerate_pytorch_inference_jit_ipex.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_
|
||||
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <Inference/PyTorch/quantize_pytorch_inference_inc.html>`_
|
||||
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <Inference/PyTorch/quantize_pytorch_inference_pot.html>`_
|
||||
* |pytorch_inference_context_manager_link|_
|
||||
* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
|
||||
* `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
|
||||
* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor <Inference/PyTorch/quantize_pytorch_inference_inc.html>`_
|
||||
* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools <Inference/PyTorch/quantize_pytorch_inference_pot.html>`_
|
||||
* `How to enable automatic context management for PyTorch inference on Nano optimized models <Inference/PyTorch/pytorch_context_manager.html>`_
|
||||
* `How to save and load optimized ONNXRuntime model <Inference/PyTorch/pytorch_save_and_load_onnx.html>`_
|
||||
* `How to save and load optimized OpenVINO model <Inference/PyTorch/pytorch_save_and_load_openvino.html>`_
|
||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <Inference/PyTorch/inference_optimizer_optimize.html>`_
|
||||
|
||||
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context``
|
||||
.. _pytorch_inference_context_manager_link: Inference/PyTorch/pytorch_context_manager.html
|
||||
* `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
|
||||
* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_
|
||||
|
||||
TensorFlow
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
|
|
|||
Loading…
Reference in a new issue