[Nano] Add how-to guide for TensorFlow Keras inference using bf16 mixed precision (#7337)
* Add basic structure for tf bf16 infer how-to guide * Add how-to use BFloat16 mixed precision for TensorFlow Keras inference * Small fixes * Add instruction requires box * Add outputs to show inference time diff * Small fixes regarding hardware requirements and others * Update based on comments * Small fixes * Small fixes
This commit is contained in:
parent
08a556239c
commit
c7674f52b2
3 changed files with 6 additions and 1 deletions
|
|
@ -133,6 +133,7 @@ subtrees:
|
|||
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_onnx
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_openvino
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_inference_bf16
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_onnx
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_openvino
|
||||
- file: doc/Nano/Howto/install_in_colab
|
||||
|
|
|
|||
|
|
@ -0,0 +1,3 @@
|
|||
{
|
||||
"path": "../../../../../../../../python/nano/tutorial/notebook/inference/tensorflow/tensorflow_inference_bf16.ipynb"
|
||||
}
|
||||
|
|
@ -41,7 +41,7 @@ TensorFlow
|
|||
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
* `How to accelerate a TensorFlow Keras application on training workloads through multiple instances <Training/TensorFlow/accelerate_tensorflow_training_multi_instance.html>`_
|
||||
* |tensorflow_training_embedding_sparseadam_link|_
|
||||
* `How to conduct BFloat16 Mixed Precision training in your TensorFlow application <Training/TensorFlow/tensorflow_training_bf16.html>`_
|
||||
* `How to conduct BFloat16 Mixed Precision training in your TensorFlow Keras application <Training/TensorFlow/tensorflow_training_bf16.html>`_
|
||||
|
||||
.. |tensorflow_training_embedding_sparseadam_link| replace:: How to optimize your model with a sparse ``Embedding`` layer and ``SparseAdam`` optimizer
|
||||
.. _tensorflow_training_embedding_sparseadam_link: Training/TensorFlow/tensorflow_training_embedding_sparseadam.html
|
||||
|
|
@ -83,6 +83,7 @@ TensorFlow
|
|||
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
* `How to accelerate a TensorFlow inference pipeline through ONNXRuntime <Inference/TensorFlow/accelerate_tensorflow_inference_onnx.html>`_
|
||||
* `How to accelerate a TensorFlow inference pipeline through OpenVINO <Inference/TensorFlow/accelerate_tensorflow_inference_openvino.html>`_
|
||||
* `How to conduct BFloat16 Mixed Precision inference in a TensorFlow Keras application <Inference/TensorFlow/tensorflow_inference_bf16.html>`_
|
||||
* `How to save and load optimized ONNXRuntime model in TensorFlow <Inference/TensorFlow/tensorflow_save_and_load_onnx.html>`_
|
||||
* `How to save and load optimized OpenVINO model in TensorFlow <Inference/TensorFlow/tensorflow_save_and_load_openvino.html>`_
|
||||
|
||||
|
|
|
|||
Loading…
Reference in a new issue