[Nano] Improve How-to Guides Navigations (#7396)
* Remove deprecated option enable_auto_doc_ref for recommonmark * Add first level navigation structure for Nano how-to guides * Update navigation for How-to Training part * Update navigation for How-to Inference part * Update navigation for How-to Preprocessing/Install part and other small fixes * Fix wrong link path caused by position changes of how-to install related guides * Small fix
This commit is contained in:
parent
bd29010854
commit
c31136df0b
18 changed files with 252 additions and 42 deletions
|
|
@ -101,43 +101,87 @@ subtrees:
|
|||
title: "How-to Guides"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Preprocessing/PyTorch/accelerate_pytorch_cv_data_pipeline
|
||||
- file: doc/Nano/Howto/Training/PyTorchLightning/accelerate_pytorch_lightning_training_ipex
|
||||
- file: doc/Nano/Howto/Training/PyTorchLightning/accelerate_pytorch_lightning_training_multi_instance
|
||||
- file: doc/Nano/Howto/Training/PyTorchLightning/pytorch_lightning_training_channels_last
|
||||
- file: doc/Nano/Howto/Training/PyTorchLightning/pytorch_lightning_training_bf16
|
||||
- file: doc/Nano/Howto/Training/PyTorch/convert_pytorch_training_torchnano
|
||||
- file: doc/Nano/Howto/Training/PyTorch/use_nano_decorator_pytorch_training
|
||||
- file: doc/Nano/Howto/Training/PyTorch/accelerate_pytorch_training_ipex
|
||||
- file: doc/Nano/Howto/Training/PyTorch/accelerate_pytorch_training_multi_instance
|
||||
- file: doc/Nano/Howto/Training/PyTorch/pytorch_training_channels_last
|
||||
- file: doc/Nano/Howto/Training/PyTorch/accelerate_pytorch_training_bf16
|
||||
- file: doc/Nano/Howto/Training/TensorFlow/accelerate_tensorflow_training_multi_instance
|
||||
- file: doc/Nano/Howto/Training/TensorFlow/tensorflow_training_embedding_sparseadam
|
||||
- file: doc/Nano/Howto/Training/TensorFlow/tensorflow_training_bf16
|
||||
- file: doc/Nano/Howto/Training/General/choose_num_processes_training
|
||||
- file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference
|
||||
- file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference_async
|
||||
- file: doc/Nano/Howto/Inference/OpenVINO/accelerate_inference_openvino_gpu
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_ipex
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_onnx
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_openvino
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_inference_bf16
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_onnx
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_openvino
|
||||
- file: doc/Nano/Howto/install_in_colab
|
||||
- file: doc/Nano/Howto/windows_guide
|
||||
- file: doc/Nano/Howto/Preprocessing/index
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Preprocessing/PyTorch/index
|
||||
title: "PyTorch"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Preprocessing/PyTorch/accelerate_pytorch_cv_data_pipeline
|
||||
- file: doc/Nano/Howto/Training/index
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Training/PyTorchLightning/index
|
||||
title: "PyTorch Lightning"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Training/PyTorchLightning/accelerate_pytorch_lightning_training_ipex
|
||||
- file: doc/Nano/Howto/Training/PyTorchLightning/accelerate_pytorch_lightning_training_multi_instance
|
||||
- file: doc/Nano/Howto/Training/PyTorchLightning/pytorch_lightning_training_channels_last
|
||||
- file: doc/Nano/Howto/Training/PyTorchLightning/pytorch_lightning_training_bf16
|
||||
- file: doc/Nano/Howto/Training/PyTorch/index
|
||||
title: "PyTorch"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Training/PyTorch/convert_pytorch_training_torchnano
|
||||
- file: doc/Nano/Howto/Training/PyTorch/use_nano_decorator_pytorch_training
|
||||
- file: doc/Nano/Howto/Training/PyTorch/accelerate_pytorch_training_ipex
|
||||
- file: doc/Nano/Howto/Training/PyTorch/accelerate_pytorch_training_multi_instance
|
||||
- file: doc/Nano/Howto/Training/PyTorch/pytorch_training_channels_last
|
||||
- file: doc/Nano/Howto/Training/PyTorch/accelerate_pytorch_training_bf16
|
||||
- file: doc/Nano/Howto/Training/TensorFlow/index
|
||||
title: "TensorFlow"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Training/TensorFlow/accelerate_tensorflow_training_multi_instance
|
||||
- file: doc/Nano/Howto/Training/TensorFlow/tensorflow_training_embedding_sparseadam
|
||||
- file: doc/Nano/Howto/Training/TensorFlow/tensorflow_training_bf16
|
||||
- file: doc/Nano/Howto/Training/General/index
|
||||
title: "General"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Training/General/choose_num_processes_training
|
||||
- file: doc/Nano/Howto/Inference/index
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Inference/OpenVINO/index
|
||||
title: "OpenVINO"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference
|
||||
- file: doc/Nano/Howto/Inference/OpenVINO/openvino_inference_async
|
||||
- file: doc/Nano/Howto/Inference/OpenVINO/accelerate_inference_openvino_gpu
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/index
|
||||
title: "PyTorch"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_onnx
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_openvino
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/accelerate_pytorch_inference_jit_ipex
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/multi_instance_pytorch_inference
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_inc
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/quantize_pytorch_inference_pot
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_context_manager
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_ipex
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_jit
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_onnx
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/pytorch_save_and_load_openvino
|
||||
- file: doc/Nano/Howto/Inference/PyTorch/inference_optimizer_optimize
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/index
|
||||
title: "TensorFlow"
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_onnx
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/accelerate_tensorflow_inference_openvino
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_inference_bf16
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_onnx
|
||||
- file: doc/Nano/Howto/Inference/TensorFlow/tensorflow_save_and_load_openvino
|
||||
- file: doc/Nano/Howto/Install/index
|
||||
subtrees:
|
||||
- entries:
|
||||
- file: doc/Nano/Howto/Install/install_in_colab
|
||||
- file: doc/Nano/Howto/Install/windows_guide
|
||||
- file: doc/Nano/Overview/known_issues
|
||||
title: "Tips and Known Issues"
|
||||
- file: doc/Nano/Overview/troubshooting
|
||||
|
|
|
|||
|
|
@ -269,8 +269,7 @@ def setup(app):
|
|||
'auto_toc_tree_section': 'Contents',
|
||||
'enable_math': False,
|
||||
'enable_inline_math': False,
|
||||
'enable_eval_rst': True,
|
||||
'enable_auto_doc_ref': True,
|
||||
'enable_eval_rst': True
|
||||
}, True)
|
||||
app.add_transform(AutoStructify)
|
||||
|
||||
|
|
|
|||
|
|
@ -0,0 +1,6 @@
|
|||
Inference Optimization: For OpenVINO Users
|
||||
=============================================
|
||||
|
||||
* `How to run inference on OpenVINO model <openvino_inference.html>`_
|
||||
* `How to run asynchronous inference on OpenVINO model <openvino_inference_async.html>`_
|
||||
* `How to accelerate a PyTorch / TensorFlow inference pipeline on Intel GPUs through OpenVINO <accelerate_inference_openvino_gpu.html>`_
|
||||
|
|
@ -0,0 +1,18 @@
|
|||
Inference Optimization: For PyTorch Users
|
||||
=============================================
|
||||
|
||||
* `How to accelerate a PyTorch inference pipeline through ONNXRuntime <accelerate_pytorch_inference_onnx.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through OpenVINO <accelerate_pytorch_inference_openvino.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through JIT/IPEX <accelerate_pytorch_inference_jit_ipex.html>`_
|
||||
* `How to accelerate a PyTorch inference pipeline through multiple instances <multi_instance_pytorch_inference.html>`_
|
||||
* `How to quantize your PyTorch model for inference using Intel Neural Compressor <quantize_pytorch_inference_inc.html>`_
|
||||
* `How to quantize your PyTorch model for inference using OpenVINO Post-training Optimization Tools <quantize_pytorch_inference_pot.html>`_
|
||||
* |pytorch_inference_context_manager_link|_
|
||||
* `How to save and load optimized IPEX model <pytorch_save_and_load_ipex.html>`_
|
||||
* `How to save and load optimized JIT model <pytorch_save_and_load_jit.html>`_
|
||||
* `How to save and load optimized ONNXRuntime model <pytorch_save_and_load_onnx.html>`_
|
||||
* `How to save and load optimized OpenVINO model <pytorch_save_and_load_openvino.html>`_
|
||||
* `How to find accelerated method with minimal latency using InferenceOptimizer <inference_optimizer_optimize.html>`_
|
||||
|
||||
.. |pytorch_inference_context_manager_link| replace:: How to use context manager through ``get_context``
|
||||
.. _pytorch_inference_context_manager_link: pytorch_context_manager.html
|
||||
|
|
@ -0,0 +1,8 @@
|
|||
Inference Optimization: For TensorFlow Users
|
||||
=============================================
|
||||
|
||||
* `How to accelerate a TensorFlow inference pipeline through ONNXRuntime <accelerate_tensorflow_inference_onnx.html>`_
|
||||
* `How to accelerate a TensorFlow inference pipeline through OpenVINO <accelerate_tensorflow_inference_openvino.html>`_
|
||||
* `How to conduct BFloat16 Mixed Precision inference in a TensorFlow Keras application <tensorflow_inference_bf16.html>`_
|
||||
* `How to save and load optimized ONNXRuntime model in TensorFlow <tensorflow_save_and_load_onnx.html>`_
|
||||
* `How to save and load optimized OpenVINO model in TensorFlow <tensorflow_save_and_load_openvino.html>`_
|
||||
33
docs/readthedocs/source/doc/Nano/Howto/Inference/index.rst
Normal file
33
docs/readthedocs/source/doc/Nano/Howto/Inference/index.rst
Normal file
|
|
@ -0,0 +1,33 @@
|
|||
Inference Optimization
|
||||
=========================
|
||||
|
||||
Here you could find detailed guides on how to apply BigDL-Nano to optimize your inference workloads. Select your desired use case below for further navigation:
|
||||
|
||||
.. grid:: 1 2 2 2
|
||||
|
||||
.. grid-item::
|
||||
|
||||
.. button-link:: OpenVINO/index.html
|
||||
:color: primary
|
||||
:expand:
|
||||
:outline:
|
||||
|
||||
I use **OpenVINO** toolkit.
|
||||
|
||||
.. grid-item::
|
||||
|
||||
.. button-link:: PyTorch/index.html
|
||||
:color: primary
|
||||
:expand:
|
||||
:outline:
|
||||
|
||||
I am a **PyTorch** user.
|
||||
|
||||
.. grid-item::
|
||||
|
||||
.. button-link:: TensorFlow/index.html
|
||||
:color: primary
|
||||
:expand:
|
||||
:outline:
|
||||
|
||||
I am a **TensorFlow** user.
|
||||
7
docs/readthedocs/source/doc/Nano/Howto/Install/index.rst
Normal file
7
docs/readthedocs/source/doc/Nano/Howto/Install/index.rst
Normal file
|
|
@ -0,0 +1,7 @@
|
|||
Install
|
||||
=========================
|
||||
|
||||
Here you could find detailed guides on how to install BigDL-Nano for different use cases:
|
||||
|
||||
* `How to install BigDL-Nano in Google Colab <install_in_colab.html>`_
|
||||
* `How to install BigDL-Nano on Windows <windows_guide.html>`_
|
||||
|
|
@ -0,0 +1,4 @@
|
|||
Preprocessing Optimization: For PyTorch Users
|
||||
==============================================
|
||||
|
||||
* `How to accelerate a computer vision data processing pipeline <accelerate_pytorch_cv_data_pipeline.html>`_
|
||||
|
|
@ -0,0 +1,15 @@
|
|||
Preprocessing Optimization
|
||||
===========================
|
||||
|
||||
Here you could find detailed guides on how to apply BigDL-Nano to accelerate your data preprocess pipeline. Select your desired use case below for further navigation:
|
||||
|
||||
.. grid:: 1 2 2 2
|
||||
|
||||
.. grid-item::
|
||||
|
||||
.. button-link:: PyTorch/index.html
|
||||
:color: primary
|
||||
:expand:
|
||||
:outline:
|
||||
|
||||
I am a **PyTorch** user.
|
||||
|
|
@ -0,0 +1,4 @@
|
|||
Training Optimization: General Tips
|
||||
====================================
|
||||
|
||||
* `How to choose the number of processes for multi-instance training <choose_num_processes_training.html>`_
|
||||
|
|
@ -0,0 +1,14 @@
|
|||
Training Optimization: For PyTorch Users
|
||||
=========================================
|
||||
|
||||
* |convert_pytorch_training_torchnano|_
|
||||
* |use_nano_decorator_pytorch_training|_
|
||||
* `How to accelerate a PyTorch application on training workloads through Intel® Extension for PyTorch* <accelerate_pytorch_training_ipex.html>`_
|
||||
* `How to accelerate a PyTorch application on training workloads through multiple instances <accelerate_pytorch_training_multi_instance.html>`_
|
||||
* `How to use the channels last memory format in your PyTorch application for training <pytorch_training_channels_last.html>`_
|
||||
* `How to conduct BFloat16 Mixed Precision training in your PyTorch application <accelerate_pytorch_training_bf16.html>`_
|
||||
|
||||
.. |use_nano_decorator_pytorch_training| replace:: How to accelerate your PyTorch training loop with ``@nano`` decorator
|
||||
.. _use_nano_decorator_pytorch_training: use_nano_decorator_pytorch_training.html
|
||||
.. |convert_pytorch_training_torchnano| replace:: How to convert your PyTorch training loop to use ``TorchNano`` for acceleration
|
||||
.. _convert_pytorch_training_torchnano: convert_pytorch_training_torchnano.html
|
||||
|
|
@ -0,0 +1,7 @@
|
|||
Training Optimization: For PyTorch Lightning Users
|
||||
===================================================
|
||||
|
||||
* `How to accelerate a PyTorch Lightning application on training workloads through Intel® Extension for PyTorch* <accelerate_pytorch_lightning_training_ipex.html>`_
|
||||
* `How to accelerate a PyTorch Lightning application on training workloads through multiple instances <accelerate_pytorch_lightning_training_multi_instance.html>`_
|
||||
* `How to use the channels last memory format in your PyTorch Lightning application for training <pytorch_lightning_training_channels_last.html>`_
|
||||
* `How to conduct BFloat16 Mixed Precision training in your PyTorch Lightning application <pytorch_lightning_training_bf16.html>`_
|
||||
|
|
@ -0,0 +1,9 @@
|
|||
Training Optimization: For TensorFlow Users
|
||||
============================================
|
||||
|
||||
* `How to accelerate a TensorFlow Keras application on training workloads through multiple instances <accelerate_tensorflow_training_multi_instance.html>`_
|
||||
* |tensorflow_training_embedding_sparseadam_link|_
|
||||
* `How to conduct BFloat16 Mixed Precision training in your TensorFlow application <tensorflow_training_bf16.html>`_
|
||||
|
||||
.. |tensorflow_training_embedding_sparseadam_link| replace:: How to optimize your model with a sparse ``Embedding`` layer and ``SparseAdam`` optimizer
|
||||
.. _tensorflow_training_embedding_sparseadam_link: tensorflow_training_embedding_sparseadam.html
|
||||
42
docs/readthedocs/source/doc/Nano/Howto/Training/index.rst
Normal file
42
docs/readthedocs/source/doc/Nano/Howto/Training/index.rst
Normal file
|
|
@ -0,0 +1,42 @@
|
|||
Training Optimization
|
||||
=========================
|
||||
|
||||
Here you could find detailed guides on how to apply BigDL-Nano to optimize your training workloads. Select your desired use case below for further navigation:
|
||||
|
||||
.. grid:: 1 2 2 2
|
||||
|
||||
.. grid-item::
|
||||
|
||||
.. button-link:: PyTorchLightning/index.html
|
||||
:color: primary
|
||||
:expand:
|
||||
:outline:
|
||||
|
||||
I am a **PyTorch Lightning** user.
|
||||
|
||||
.. grid-item::
|
||||
|
||||
.. button-link:: PyTorch/index.html
|
||||
:color: primary
|
||||
:expand:
|
||||
:outline:
|
||||
|
||||
I am a **PyTorch** user.
|
||||
|
||||
.. grid-item::
|
||||
|
||||
.. button-link:: TensorFlow/index.html
|
||||
:color: primary
|
||||
:expand:
|
||||
:outline:
|
||||
|
||||
I am a **TensorFlow** user.
|
||||
|
||||
.. grid-item::
|
||||
|
||||
.. button-link:: General/index.html
|
||||
:color: primary
|
||||
:expand:
|
||||
:outline:
|
||||
|
||||
I want to know general optimization tips.
|
||||
|
|
@ -89,5 +89,5 @@ TensorFlow
|
|||
|
||||
Install
|
||||
-------------------------
|
||||
* `How to install BigDL-Nano in Google Colab <install_in_colab.html>`_
|
||||
* `How to install BigDL-Nano on Windows <windows_guide.html>`_
|
||||
* `How to install BigDL-Nano in Google Colab <Install/install_in_colab.html>`_
|
||||
* `How to install BigDL-Nano on Windows <Install/windows_guide.html>`_
|
||||
|
|
@ -91,7 +91,7 @@ For Linux, Ubuntu (22.04/20.04/18.04) is recommended.
|
|||
|
||||
For Windows OS, users could only run `bigdl-nano-init` every time they open a new cmd terminal.
|
||||
|
||||
We recommend using Windows Subsystem for Linux 2 (WSL2) to run BigDL-Nano. Please refer to [Nano Windows install guide](../Howto/windows_guide.md) for instructions.
|
||||
We recommend using Windows Subsystem for Linux 2 (WSL2) to run BigDL-Nano. Please refer to [Nano Windows install guide](../Howto/Install/windows_guide.md) for instructions.
|
||||
|
||||
### Install on MacOS
|
||||
#### MacOS with Intel Chip
|
||||
|
|
|
|||
Loading…
Reference in a new issue