From 6960c87b54eccdb9d4262f7d5062cae96d23e15b Mon Sep 17 00:00:00 2001 From: Yuwen Hu <54161268+Oscilloscope98@users.noreply.github.com> Date: Thu, 9 Feb 2023 17:50:50 +0800 Subject: [PATCH] [Nano] Revise outdated info in PyTorch Inference how-to guides (#7453) * Change order of PyTorch Inference how-to guides order * Revise accelerate_pytorch_inference_jit_ipex.ipynb * Revise accelerat_pytorch_inference_onnx.ipynb * Revise accelerate_pytorch_inference_openvino.ipynb * Revise inference_optimizer_optimize.ipynb * Revise multi_instance_pytorch_inference.ipynb * Revise pytorch_context_manager.ipynb * Revise pytorch_save_and_load_ipex.ipynb and other small fixes * Revise pytorch_save_and_load_jit.ipynb and other small fixes * Revise pytorch_save_and_load_onnx.ipynb and other small fixese * Revise pytorch_save_and_load_openvino.ipynb * Revise quantize_pytorch_inference_inc.ipynb and other small fixes * Revise quantize_pytorch_inference_pot.ipynb and other small fixes * Fix failed links in the index page * Small fixes * Update based on comments --- docs/readthedocs/source/_toc.yml | 4 ++-- .../doc/Nano/Howto/Inference/PyTorch/index.rst | 17 +++++++---------- .../readthedocs/source/doc/Nano/Howto/index.rst | 17 +++++++---------- 3 files changed, 16 insertions(+), 22 deletions(-) diff --git a/docs/readthedocs/source/_toc.yml b/docs/readthedocs/source/_toc.yml index 797407ea..af0d8864 100644 --- a/docs/readthedocs/source/_toc.yml +++ b/docs/readthedocs/source/_toc.yml @@ -156,10 +156,10 @@ subtrees: title: "PyTorch" subtrees: - entries: + - file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex - - file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot - file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager @@ -167,7 +167,7 @@ subtrees: - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino - - file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize + - file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference - file: doc/Nano/Howto/Inference/TensorFlow/index title: "TensorFlow" subtrees: diff --git a/docs/readthedocs/source/doc/Nano/Howto/Inference/PyTorch/index.rst b/docs/readthedocs/source/doc/Nano/Howto/Inference/PyTorch/index.rst index 6241866c..58c5b661 100644 --- a/docs/readthedocs/source/doc/Nano/Howto/Inference/PyTorch/index.rst +++ b/docs/readthedocs/source/doc/Nano/Howto/Inference/PyTorch/index.rst @@ -1,18 +1,15 @@ Inference Optimization: For PyTorch Users ============================================= +* `How to find accelerated method with minimal latency using InferenceOptimizer `_ * `How to accelerate a PyTorch inference pipeline through ONNXRuntime `_ * `How to accelerate a PyTorch inference pipeline through OpenVINO `_ * `How to accelerate a PyTorch inference pipeline through JIT/IPEX `_ -* `How to accelerate a PyTorch inference pipeline through multiple instances `_ -* `How to quantize your PyTorch model for inference using Intel Neural Compressor `_ -* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools `_ -* |pytorch_inference_context_manager_link|_ -* `How to save and load optimized IPEX model `_ -* `How to save and load optimized JIT model `_ +* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor `_ +* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools `_ +* `How to enable automatic context management for PyTorch inference on Nano optimized models `_ * `How to save and load optimized ONNXRuntime model `_ * `How to save and load optimized OpenVINO model `_ -* `How to find accelerated method with minimal latency using InferenceOptimizer `_ - -.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context`` -.. _pytorch_inference_context_manager_link: pytorch_context_manager.html \ No newline at end of file +* `How to save and load optimized JIT model `_ +* `How to save and load optimized IPEX model `_ +* `How to accelerate a PyTorch inference pipeline through multiple instances `_ \ No newline at end of file diff --git a/docs/readthedocs/source/doc/Nano/Howto/index.rst b/docs/readthedocs/source/doc/Nano/Howto/index.rst index 4e760eda..e6ea0e39 100644 --- a/docs/readthedocs/source/doc/Nano/Howto/index.rst +++ b/docs/readthedocs/source/doc/Nano/Howto/index.rst @@ -63,21 +63,18 @@ OpenVINO PyTorch ~~~~~~~~~~~~~~~~~~~~~~~~~ +* `How to find accelerated method with minimal latency using InferenceOptimizer `_ * `How to accelerate a PyTorch inference pipeline through ONNXRuntime `_ * `How to accelerate a PyTorch inference pipeline through OpenVINO `_ * `How to accelerate a PyTorch inference pipeline through JIT/IPEX `_ -* `How to accelerate a PyTorch inference pipeline through multiple instances `_ -* `How to quantize your PyTorch model for inference using Intel Neural Compressor `_ -* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools `_ -* |pytorch_inference_context_manager_link|_ -* `How to save and load optimized IPEX model `_ -* `How to save and load optimized JIT model `_ +* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor `_ +* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools `_ +* `How to enable automatic context management for PyTorch inference on Nano optimized models `_ * `How to save and load optimized ONNXRuntime model `_ * `How to save and load optimized OpenVINO model `_ -* `How to find accelerated method with minimal latency using InferenceOptimizer `_ - -.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context`` -.. _pytorch_inference_context_manager_link: Inference/PyTorch/pytorch_context_manager.html +* `How to save and load optimized JIT model `_ +* `How to save and load optimized IPEX model `_ +* `How to accelerate a PyTorch inference pipeline through multiple instances `_ TensorFlow ~~~~~~~~~~~~~~~~~~~~~~~~~