[Nano] Add how-to-guide of load/save API for tensorflow inference (#7180)
* feat(docs): add load/save onnx and opnevino model for tensorflow * fix bugs after previewing * fix order issues of insertion for toc.yml * change link title for tensorflow
This commit is contained in:
parent
d950992b91
commit
2858a1b5bf
4 changed files with 10 additions and 0 deletions
|
|
@ -132,6 +132,8 @@ subtrees:
|
||||||
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
||||||
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_onnx
|
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_onnx
|
||||||
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_openvino
|
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_openvino
|
||||||
|
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_onnx
|
||||||
|
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_openvino
|
||||||
- file: doc/Nano/Howto/install_in_colab
|
- file: doc/Nano/Howto/install_in_colab
|
||||||
- file: doc/Nano/Howto/windows_guide
|
- file: doc/Nano/Howto/windows_guide
|
||||||
- file: doc/Nano/Overview/known_issues
|
- file: doc/Nano/Overview/known_issues
|
||||||
|
|
|
||||||
|
|
@ -0,0 +1,3 @@
|
||||||
|
{
|
||||||
|
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/tensorflow/tensorflow_save_and_load_onnx.ipynb"
|
||||||
|
}
|
||||||
|
|
@ -0,0 +1,3 @@
|
||||||
|
{
|
||||||
|
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/tensorflow/tensorflow_save_and_load_openvino.ipynb"
|
||||||
|
}
|
||||||
|
|
@ -82,6 +82,8 @@ TensorFlow
|
||||||
~~~~~~~~~~~~~~~~~~~~~~~~~
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
* `How to accelerate a TensorFlow inference pipeline through ONNXRuntime <Inference/TensorFlow/accelerate_tensorflow_inference_onnx.html>`_
|
* `How to accelerate a TensorFlow inference pipeline through ONNXRuntime <Inference/TensorFlow/accelerate_tensorflow_inference_onnx.html>`_
|
||||||
* `How to accelerate a TensorFlow inference pipeline through OpenVINO <Inference/TensorFlow/accelerate_tensorflow_inference_openvino.html>`_
|
* `How to accelerate a TensorFlow inference pipeline through OpenVINO <Inference/TensorFlow/accelerate_tensorflow_inference_openvino.html>`_
|
||||||
|
* `How to save and load optimized ONNXRuntime model in TensorFlow <Inference/TensorFlow/tensorflow_save_and_load_onnx.html>`_
|
||||||
|
* `How to save and load optimized OpenVINO model in TensorFlow <Inference/TensorFlow/tensorflow_save_and_load_openvino.html>`_
|
||||||
|
|
||||||
Install
|
Install
|
||||||
-------------------------
|
-------------------------
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue