[Nano] Add how-to guide for TensorFlow Keras inference using bf16 mixed precision (#7337)
* Add basic structure for tf bf16 infer how-to guide * Add how-to use BFloat16 mixed precision for TensorFlow Keras inference * Small fixes * Add instruction requires box * Add outputs to show inference time diff * Small fixes regarding hardware requirements and others * Update based on comments * Small fixes * Small fixes
This commit is contained in:
		
							parent
							
								
									08a556239c
								
							
						
					
					
						commit
						c7674f52b2
					
				
					 3 changed files with 6 additions and 1 deletions
				
			
		| 
						 | 
				
			
			@ -133,6 +133,7 @@ subtrees:
 | 
			
		|||
                  - file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
 | 
			
		||||
                  - file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_onnx
 | 
			
		||||
                  - file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_openvino
 | 
			
		||||
                  - file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_inference_bf16
 | 
			
		||||
                  - file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_onnx
 | 
			
		||||
                  - file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_openvino
 | 
			
		||||
                  - file: doc/Nano/Howto/install_in_colab
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
| 
						 | 
				
			
			@ -0,0 +1,3 @@
 | 
			
		|||
{
 | 
			
		||||
    "path": "../../../../../../../../python/nano/tutorial/notebook/inference/tensorflow/tensorflow_inference_bf16.ipynb"
 | 
			
		||||
}
 | 
			
		||||
| 
						 | 
				
			
			@ -41,7 +41,7 @@ TensorFlow
 | 
			
		|||
~~~~~~~~~~~~~~~~~~~~~~~~~
 | 
			
		||||
* `How to accelerate a TensorFlow Keras application on training workloads through multiple instances <Training/TensorFlow/accelerate_tensorflow_training_multi_instance.html>`_
 | 
			
		||||
* |tensorflow_training_embedding_sparseadam_link|_
 | 
			
		||||
* `How to conduct BFloat16 Mixed Precision training in your TensorFlow application <Training/TensorFlow/tensorflow_training_bf16.html>`_
 | 
			
		||||
* `How to conduct BFloat16 Mixed Precision training in your TensorFlow Keras application <Training/TensorFlow/tensorflow_training_bf16.html>`_
 | 
			
		||||
 | 
			
		||||
.. |tensorflow_training_embedding_sparseadam_link| replace:: How to optimize your model with a sparse ``Embedding`` layer and ``SparseAdam`` optimizer
 | 
			
		||||
.. _tensorflow_training_embedding_sparseadam_link: Training/TensorFlow/tensorflow_training_embedding_sparseadam.html
 | 
			
		||||
| 
						 | 
				
			
			@ -83,6 +83,7 @@ TensorFlow
 | 
			
		|||
~~~~~~~~~~~~~~~~~~~~~~~~~
 | 
			
		||||
* `How to accelerate a TensorFlow inference pipeline through ONNXRuntime <Inference/TensorFlow/accelerate_tensorflow_inference_onnx.html>`_
 | 
			
		||||
* `How to accelerate a TensorFlow inference pipeline through OpenVINO <Inference/TensorFlow/accelerate_tensorflow_inference_openvino.html>`_
 | 
			
		||||
* `How to conduct BFloat16 Mixed Precision inference in a TensorFlow Keras application <Inference/TensorFlow/tensorflow_inference_bf16.html>`_
 | 
			
		||||
* `How to save and load optimized ONNXRuntime model in TensorFlow <Inference/TensorFlow/tensorflow_save_and_load_onnx.html>`_
 | 
			
		||||
* `How to save and load optimized OpenVINO model in TensorFlow <Inference/TensorFlow/tensorflow_save_and_load_openvino.html>`_
 | 
			
		||||
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
		Reference in a new issue