[Nano] Revise outdated info in PyTorch Inference how-to guides (#7453)
* Change order of PyTorch Inference how-to guides order * Revise accelerate_pytorch_inference_jit_ipex.ipynb * Revise accelerat_pytorch_inference_onnx.ipynb * Revise accelerate_pytorch_inference_openvino.ipynb * Revise inference_optimizer_optimize.ipynb * Revise multi_instance_pytorch_inference.ipynb * Revise pytorch_context_manager.ipynb * Revise pytorch_save_and_load_ipex.ipynb and other small fixes * Revise pytorch_save_and_load_jit.ipynb and other small fixes * Revise pytorch_save_and_load_onnx.ipynb and other small fixese * Revise pytorch_save_and_load_openvino.ipynb * Revise quantize_pytorch_inference_inc.ipynb and other small fixes * Revise quantize_pytorch_inference_pot.ipynb and other small fixes * Fix failed links in the index page * Small fixes * Update based on comments
This commit is contained in:
		
							parent
							
								
									56d699d935
								
							
						
					
					
						commit
						6960c87b54
					
				
					 3 changed files with 16 additions and 22 deletions
				
			
		| 
						 | 
					@ -156,10 +156,10 @@ subtrees:
 | 
				
			||||||
                          title: "PyTorch"
 | 
					                          title: "PyTorch"
 | 
				
			||||||
                          subtrees:
 | 
					                          subtrees:
 | 
				
			||||||
                            - entries: 
 | 
					                            - entries: 
 | 
				
			||||||
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
 | 
					 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager
 | 
				
			||||||
| 
						 | 
					@ -167,7 +167,7 @@ subtrees:
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
 | 
				
			||||||
                              - file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
 | 
					                              - file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
 | 
				
			||||||
                        - file: doc/Nano/Howto/Inference/TensorFlow/index
 | 
					                        - file: doc/Nano/Howto/Inference/TensorFlow/index
 | 
				
			||||||
                          title: "TensorFlow"
 | 
					                          title: "TensorFlow"
 | 
				
			||||||
                          subtrees:
 | 
					                          subtrees:
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
| 
						 | 
					@ -1,18 +1,15 @@
 | 
				
			||||||
Inference Optimization: For PyTorch Users
 | 
					Inference Optimization: For PyTorch Users
 | 
				
			||||||
=============================================
 | 
					=============================================
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					* `How to find accelerated method with minimal latency using InferenceOptimizer <inference_optimizer_optimize.html>`_
 | 
				
			||||||
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <accelerate_pytorch_inference_onnx.html>`_
 | 
					* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <accelerate_pytorch_inference_onnx.html>`_
 | 
				
			||||||
* `How to accelerate a PyTorch inference pipeline through OpenVINO <accelerate_pytorch_inference_openvino.html>`_
 | 
					* `How to accelerate a PyTorch inference pipeline through OpenVINO <accelerate_pytorch_inference_openvino.html>`_
 | 
				
			||||||
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <accelerate_pytorch_inference_jit_ipex.html>`_
 | 
					* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <accelerate_pytorch_inference_jit_ipex.html>`_
 | 
				
			||||||
* `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_
 | 
					* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor <quantize_pytorch_inference_inc.html>`_
 | 
				
			||||||
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <quantize_pytorch_inference_inc.html>`_
 | 
					* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools <quantize_pytorch_inference_pot.html>`_
 | 
				
			||||||
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <quantize_pytorch_inference_pot.html>`_
 | 
					* `How to enable automatic context management for PyTorch inference on Nano optimized models <pytorch_context_manager.html>`_
 | 
				
			||||||
* |pytorch_inference_context_manager_link|_
 | 
					 | 
				
			||||||
* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
 | 
					 | 
				
			||||||
* `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
 | 
					 | 
				
			||||||
* `How to save and load optimized ONNXRuntime model <pytorch_save_and_load_onnx.html>`_
 | 
					* `How to save and load optimized ONNXRuntime model <pytorch_save_and_load_onnx.html>`_
 | 
				
			||||||
* `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_
 | 
					* `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_
 | 
				
			||||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <inference_optimizer_optimize.html>`_
 | 
					* `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
 | 
				
			||||||
 | 
					* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
 | 
				
			||||||
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context``
 | 
					* `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_
 | 
				
			||||||
.. _pytorch_inference_context_manager_link: pytorch_context_manager.html
 | 
					 | 
				
			||||||
| 
						 | 
					@ -63,21 +63,18 @@ OpenVINO
 | 
				
			||||||
PyTorch
 | 
					PyTorch
 | 
				
			||||||
~~~~~~~~~~~~~~~~~~~~~~~~~
 | 
					~~~~~~~~~~~~~~~~~~~~~~~~~
 | 
				
			||||||
 | 
					
 | 
				
			||||||
 | 
					* `How to find accelerated method with minimal latency using InferenceOptimizer <Inference/PyTorch/inference_optimizer_optimize.html>`_
 | 
				
			||||||
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <Inference/PyTorch/accelerate_pytorch_inference_onnx.html>`_
 | 
					* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <Inference/PyTorch/accelerate_pytorch_inference_onnx.html>`_
 | 
				
			||||||
* `How to accelerate a PyTorch inference pipeline through OpenVINO <Inference/PyTorch/accelerate_pytorch_inference_openvino.html>`_
 | 
					* `How to accelerate a PyTorch inference pipeline through OpenVINO <Inference/PyTorch/accelerate_pytorch_inference_openvino.html>`_
 | 
				
			||||||
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <Inference/PyTorch/accelerate_pytorch_inference_jit_ipex.html>`_
 | 
					* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <Inference/PyTorch/accelerate_pytorch_inference_jit_ipex.html>`_
 | 
				
			||||||
* `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_
 | 
					* `How to quantize your PyTorch model in INT8 for inference using Intel Neural Compressor <Inference/PyTorch/quantize_pytorch_inference_inc.html>`_
 | 
				
			||||||
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <Inference/PyTorch/quantize_pytorch_inference_inc.html>`_
 | 
					* `How to quantize your PyTorch model in INT8 for inference using OpenVINO Post-training Optimization Tools <Inference/PyTorch/quantize_pytorch_inference_pot.html>`_
 | 
				
			||||||
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <Inference/PyTorch/quantize_pytorch_inference_pot.html>`_
 | 
					* `How to enable automatic context management for PyTorch inference on Nano optimized models <Inference/PyTorch/pytorch_context_manager.html>`_
 | 
				
			||||||
* |pytorch_inference_context_manager_link|_
 | 
					 | 
				
			||||||
* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
 | 
					 | 
				
			||||||
* `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
 | 
					 | 
				
			||||||
* `How to save and load optimized ONNXRuntime model <Inference/PyTorch/pytorch_save_and_load_onnx.html>`_
 | 
					* `How to save and load optimized ONNXRuntime model <Inference/PyTorch/pytorch_save_and_load_onnx.html>`_
 | 
				
			||||||
* `How to save and load optimized OpenVINO model <Inference/PyTorch/pytorch_save_and_load_openvino.html>`_
 | 
					* `How to save and load optimized OpenVINO model <Inference/PyTorch/pytorch_save_and_load_openvino.html>`_
 | 
				
			||||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <Inference/PyTorch/inference_optimizer_optimize.html>`_
 | 
					* `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
 | 
				
			||||||
 | 
					* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
 | 
				
			||||||
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context``
 | 
					* `How to accelerate a PyTorch inference pipeline through multiple instances <Inference/PyTorch/multi_instance_pytorch_inference.html>`_
 | 
				
			||||||
.. _pytorch_inference_context_manager_link: Inference/PyTorch/pytorch_context_manager.html
 | 
					 | 
				
			||||||
 | 
					
 | 
				
			||||||
TensorFlow
 | 
					TensorFlow
 | 
				
			||||||
~~~~~~~~~~~~~~~~~~~~~~~~~
 | 
					~~~~~~~~~~~~~~~~~~~~~~~~~
 | 
				
			||||||
| 
						 | 
					
 | 
				
			||||||
		Loading…
	
		Reference in a new issue