diff --git a/docs/readthedocs/source/_toc.yml b/docs/readthedocs/source/_toc.yml index 797407ea..af0d8864 100644 --- a/docs/readthedocs/source/_toc.yml +++ b/docs/readthedocs/source/_toc.yml @@ -156,10 +156,10 @@ subtrees: title: "PyTorch" subtrees: - entries: + - file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex - - file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot - file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager @@ -167,7 +167,7 @@ subtrees: - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino - - file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize + - file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference - file: doc/Nano/Howto/Inference/TensorFlow/index title: "TensorFlow" subtrees: diff --git a/docs/readthedocs/source/doc/Nano/Howto/Inference/PyTorch/index.rst b/docs/readthedocs/source/doc/Nano/Howto/Inference/PyTorch/index.rst index 6241866c..58c5b661 100644 --- a/docs/readthedocs/source/doc/Nano/Howto/Inference/PyTorch/index.rst +++ b/docs/readthedocs/source/doc/Nano/Howto/Inference/PyTorch/index.rst @@ -1,18 +1,15 @@ Inference Optimization: For PyTorch Users ============================================= +* `How to find accelerated method with minimal latency using InferenceOptimizer `_ * `How to accelerate a PyTorch inference pipeline through ONNXRuntime `_ * `How to accelerate a PyTorch inference pipeline through OpenVINO `_ * `How to accelerate a PyTorch inference pipeline through JIT/IPEX `_ -* `How to accelerate a PyTorch inference pipeline through multiple instances `_ -* `How to quantize your PyTorch model for inference using Intel Neural Compressor `_ -* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools `_ -* |pytorch_inference_context_manager_link|_ -* `How to save and load optimized IPEX model `_ -* `How to save and load optimized JIT model `_ +* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor `_ +* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools `_ +* `How to enable automatic context management for PyTorch inference on Nano optimized models `_ * `How to save and load optimized ONNXRuntime model `_ * `How to save and load optimized OpenVINO model `_ -* `How to find accelerated method with minimal latency using InferenceOptimizer `_ - -.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context`` -.. _pytorch_inference_context_manager_link: pytorch_context_manager.html \ No newline at end of file +* `How to save and load optimized JIT model `_ +* `How to save and load optimized IPEX model `_ +* `How to accelerate a PyTorch inference pipeline through multiple instances `_ \ No newline at end of file diff --git a/docs/readthedocs/source/doc/Nano/Howto/index.rst b/docs/readthedocs/source/doc/Nano/Howto/index.rst index 4e760eda..e6ea0e39 100644 --- a/docs/readthedocs/source/doc/Nano/Howto/index.rst +++ b/docs/readthedocs/source/doc/Nano/Howto/index.rst @@ -63,21 +63,18 @@ OpenVINO PyTorch ~~~~~~~~~~~~~~~~~~~~~~~~~ +* `How to find accelerated method with minimal latency using InferenceOptimizer `_ * `How to accelerate a PyTorch inference pipeline through ONNXRuntime `_ * `How to accelerate a PyTorch inference pipeline through OpenVINO `_ * `How to accelerate a PyTorch inference pipeline through JIT/IPEX `_ -* `How to accelerate a PyTorch inference pipeline through multiple instances `_ -* `How to quantize your PyTorch model for inference using Intel Neural Compressor `_ -* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools `_ -* |pytorch_inference_context_manager_link|_ -* `How to save and load optimized IPEX model `_ -* `How to save and load optimized JIT model `_ +* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor `_ +* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools `_ +* `How to enable automatic context management for PyTorch inference on Nano optimized models `_ * `How to save and load optimized ONNXRuntime model `_ * `How to save and load optimized OpenVINO model `_ -* `How to find accelerated method with minimal latency using InferenceOptimizer `_ - -.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context`` -.. _pytorch_inference_context_manager_link: Inference/PyTorch/pytorch_context_manager.html +* `How to save and load optimized JIT model `_ +* `How to save and load optimized IPEX model `_ +* `How to accelerate a PyTorch inference pipeline through multiple instances `_ TensorFlow ~~~~~~~~~~~~~~~~~~~~~~~~~