[Nano] add how-to-guide for tensorflow inference by onnxruntime and openvino (#7149)
* Feat(docs): add how-to-guide for tensorflow inference by onnxruntime and openvino * fix bugs for index.rst * revise according to PR comments * revise minor parts according to PR comments * revise bugs according to PR comments
This commit is contained in:
		
							parent
							
								
									ab9a3b69fb
								
							
						
					
					
						commit
						18ae409b7c
					
				
					 4 changed files with 13 additions and 0 deletions
				
			
		| 
						 | 
				
			
			@ -110,6 +110,8 @@ subtrees:
 | 
			
		|||
                  - file: doc/Nano/Howto/Training/PyTorch/use_nano_decorator_pytorch_training
 | 
			
		||||
                  - file: doc/Nano/Howto/Training/TensorFlow/accelerate_tensorflow_training_multi_instance
 | 
			
		||||
                  - file: doc/Nano/Howto/Training/TensorFlow/tensorflow_training_embedding_sparseadam
 | 
			
		||||
                  - file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_onnx
 | 
			
		||||
                  - file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_openvino
 | 
			
		||||
                  - file: doc/Nano/Howto/Training/General/choose_num_processes_training
 | 
			
		||||
                  - file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference
 | 
			
		||||
                  - file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference_async
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -0,0 +1,3 @@
 | 
			
		|||
{
 | 
			
		||||
    "path": "../../../../../../../../python/nano/tutorial/notebook/inference/tensorflow/accelerate_tensorflow_inference_onnx.ipynb"
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			@ -0,0 +1,3 @@
 | 
			
		|||
{
 | 
			
		||||
    "path": "../../../../../../../../python/nano/tutorial/notebook/inference/tensorflow/accelerate_tensorflow_inference_openvino.ipynb"
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			@ -66,6 +66,11 @@ PyTorch
 | 
			
		|||
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context``
 | 
			
		||||
.. _pytorch_inference_context_manager_link: Inference/PyTorch/pytorch_context_manager.html
 | 
			
		||||
 | 
			
		||||
TensorFlow
 | 
			
		||||
~~~~~~~~~~~~~~~~~~~~~~~~~
 | 
			
		||||
* `How to accelerate a TensorFlow inference pipeline through ONNXRuntime <Inference/TensorFlow/accelerate_tensorflow_inference_onnx.html>`_
 | 
			
		||||
* `How to accelerate a TensorFlow inference pipeline through OpenVINO <Inference/TensorFlow/accelerate_tensorflow_inference_openvino.html>`_
 | 
			
		||||
 | 
			
		||||
Install
 | 
			
		||||
-------------------------
 | 
			
		||||
* `How to install BigDL-Nano in Google Colab <install_in_colab.html>`_
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue