[Nano] add how to guide save and load jit, ipex, onnx, openvino (#6659)
* add how to guide: * acclerate with jit_ipex * save and load jit, ipex, onnx, openvino * add these five above .nblink files ; * add index of sl files * clear all the output & fix the bug of title * remove extra blank indent * format the jupter with prettier * fix the bug of error words * add blank line before unorder list * * remove the normal inference in accelerate using jit/ipex; * add note to example why we should pass in the orginal model to get the optimized one in sl ipex * fix:new pip install shell cmd & indent improve
This commit is contained in:
parent
cd04f7dbdc
commit
bf5ccae4ef
7 changed files with 25 additions and 0 deletions
|
|
@ -114,8 +114,13 @@ subtrees:
|
|||
- file: doc/Nano/Howto/Training/General/choose_num_processes_training
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_ipex
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
||||
- file: doc/Nano/Howto/install_in_colab
|
||||
- file: doc/Nano/Howto/windows_guide
|
||||
|
|
|
|||
|
|
@ -0,0 +1,3 @@
|
|||
{
|
||||
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/pytorch/accelerate_pytorch_inference_jit_ipex.ipynb"
|
||||
}
|
||||
|
|
@ -0,0 +1,3 @@
|
|||
{
|
||||
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/pytorch/pytorch_save_and_load_ipex.ipynb"
|
||||
}
|
||||
|
|
@ -0,0 +1,3 @@
|
|||
{
|
||||
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/pytorch/pytorch_save_and_load_jit.ipynb"
|
||||
}
|
||||
|
|
@ -0,0 +1,3 @@
|
|||
{
|
||||
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/pytorch/pytorch_save_and_load_onnx.ipynb"
|
||||
}
|
||||
|
|
@ -0,0 +1,3 @@
|
|||
{
|
||||
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/pytorch/pytorch_save_and_load_openvino.ipynb"
|
||||
}
|
||||
|
|
@ -49,8 +49,13 @@ PyTorch
|
|||
|
||||
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <Inference/PyTorch/accelerate_pytorch_inference_onnx.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through OpenVINO <Inference/PyTorch/accelerate_pytorch_inference_openvino.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <Inference/PyTorch/accelerate_pytorch_inference_jit_ipex.html>`_
|
||||
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <Inference/PyTorch/quantize_pytorch_inference_inc.html>`_
|
||||
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <Inference/PyTorch/quantize_pytorch_inference_pot.html>`_
|
||||
* `How to save and load optimized IPEX model <Inference/PyTorch/pytorch_save_and_load_ipex.html>`_
|
||||
* `How to save and load optimized JIT model <Inference/PyTorch/pytorch_save_and_load_jit.html>`_
|
||||
* `How to save and load optimized ONNXRuntime model <Inference/PyTorch/pytorch_save_and_load_onnx.html>`_
|
||||
* `How to save and load optimized OpenVINO model <Inference/PyTorch/pytorch_save_and_load_openvino.html>`_
|
||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <Inference/PyTorch/inference_optimizer_optimize.html>`_
|
||||
|
||||
Install
|
||||
|
|
|
|||
Loading…
Reference in a new issue