[BigDL 2.0][PPML] PPML readthedocs refine (#5024)
* PPML Overview * Benchmark doc
This commit is contained in:
parent
18674dde18
commit
07f6f3e5dd
2 changed files with 9 additions and 9 deletions
|
|
@ -4,7 +4,7 @@
|
||||||
Protecting privacy and confidentiality is critical for large-scale data analysis and machine learning. Analytics Zoo ***PPML*** combines various low level hardware and software security technologies (e.g., Intel SGX, LibOS such as Graphene and Occlum, Federated Learning, etc.), so that users can continue to apply standard Big Data and AI technologies (such as Apache Spark, Apache Flink, Tensorflow, PyTorch, etc.) without sacrificing privacy.
|
Protecting privacy and confidentiality is critical for large-scale data analysis and machine learning. Analytics Zoo ***PPML*** combines various low level hardware and software security technologies (e.g., Intel SGX, LibOS such as Graphene and Occlum, Federated Learning, etc.), so that users can continue to apply standard Big Data and AI technologies (such as Apache Spark, Apache Flink, Tensorflow, PyTorch, etc.) without sacrificing privacy.
|
||||||
|
|
||||||
## 1.1 PPML for Big Data AI
|
## 1.1 PPML for Big Data AI
|
||||||
Analytics Zoo provides a distributed PPML platform for protecting the *end-to-end Big Data AI pipeline* (from data ingestion, data analysis, all the way to machine learning and deep learning). In particular, it extends the single-node [Trusted Execution Environment](https://en.wikipedia.org/wiki/Trusted_execution_environment) to provide a *Trusted Cluster Environment*, so as to run unmodified Big Data analysis and ML/DL programs in a secure fashion on (private or public) cloud:
|
Analytics Zoo/BigDL provides a distributed PPML platform for protecting the *end-to-end Big Data AI pipeline* (from data ingestion, data analysis, all the way to machine learning and deep learning). In particular, it extends the single-node [Trusted Execution Environment](https://en.wikipedia.org/wiki/Trusted_execution_environment) to provide a *Trusted Cluster Environment*, so as to run unmodified Big Data analysis and ML/DL programs in a secure fashion on (private or public) cloud:
|
||||||
|
|
||||||
* Compute and memory protected by SGX Enclaves
|
* Compute and memory protected by SGX Enclaves
|
||||||
* Network communication protected by remote attestation and TLS
|
* Network communication protected by remote attestation and TLS
|
||||||
|
|
@ -16,7 +16,7 @@ That is, even when the program runs in an untrusted cloud environment, all the d
|
||||||
In the current release, two types of trusted Big Data AI applications are supported:
|
In the current release, two types of trusted Big Data AI applications are supported:
|
||||||
|
|
||||||
1. Big Data analytics and ML/DL (supporting [Apache Spark](https://spark.apache.org/) and [BigDL](https://github.com/intel-analytics/BigDL))
|
1. Big Data analytics and ML/DL (supporting [Apache Spark](https://spark.apache.org/) and [BigDL](https://github.com/intel-analytics/BigDL))
|
||||||
2. Realtime compute and ML/DL (supporting [Apache Flink](https://flink.apache.org/) and Analytics Zoo [Cluster Serving](https://www.usenix.org/conference/opml20/presentation/song))
|
2. Realtime compute and ML/DL (supporting [Apache Flink](https://flink.apache.org/) and BigDL [Cluster Serving](https://www.usenix.org/conference/opml20/presentation/song))
|
||||||
|
|
||||||
## 2. Trusted Big Data Analytics and ML
|
## 2. Trusted Big Data Analytics and ML
|
||||||
With the trusted Big Data analytics and ML/DL support, users can run standard Spark data analysis (such as Spark SQL, Dataframe, MLlib, etc.) and distributed deep learning (using BigDL) in a secure and trusted fashion.
|
With the trusted Big Data analytics and ML/DL support, users can run standard Spark data analysis (such as Spark SQL, Dataframe, MLlib, etc.) and distributed deep learning (using BigDL) in a secure and trusted fashion.
|
||||||
|
|
@ -87,7 +87,7 @@ cd analytics-zoo/ppml/
|
||||||
|
|
||||||
Pull docker image from Dockerhub
|
Pull docker image from Dockerhub
|
||||||
```bash
|
```bash
|
||||||
docker pull intelanalytics/analytics-zoo-ppml-trusted-big-data-ml-scala-graphene:0.12.0-SNAPSHOT
|
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-scala-graphene:0.14.0-SNAPSHOT
|
||||||
```
|
```
|
||||||
|
|
||||||
Alternatively, you can build docker image from Dockerfile (this will take some time):
|
Alternatively, you can build docker image from Dockerfile (this will take some time):
|
||||||
|
|
@ -263,7 +263,7 @@ Then stop the service:
|
||||||
Pull docker image from Dockerhub
|
Pull docker image from Dockerhub
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
docker pull intelanalytics/analytics-zoo-ppml-trusted-big-data-ml-python-graphene:0.11-SNAPSHOT
|
docker pull intelanalytics/bigdl-ppml-trusted-big-data-ml-python-graphene:0.14-SNAPSHOT
|
||||||
```
|
```
|
||||||
|
|
||||||
Alternatively, you can build docker image from Dockerfile (this will take some time):
|
Alternatively, you can build docker image from Dockerfile (this will take some time):
|
||||||
|
|
@ -697,12 +697,12 @@ Pull docker image from Dockerhub
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# For Graphene
|
# For Graphene
|
||||||
docker pull intelanalytics/analytics-zoo-ppml-trusted-realtime-ml-scala-graphene:0.12.0-SNAPSHOT
|
docker pull intelanalytics/bigdl-ppml-trusted-realtime-ml-scala-graphene:0.14.0-SNAPSHOT
|
||||||
```
|
```
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
# For Occlum
|
# For Occlum
|
||||||
docker pull intelanalytics/analytics-zoo-ppml-trusted-realtime-ml-scala-occlum:0.12.0-SNAPSHOT
|
docker pull intelanalytics/bigdl-ppml-trusted-realtime-ml-scala-occlum:0.14.0-SNAPSHOT
|
||||||
```
|
```
|
||||||
|
|
||||||
Also, you can build docker image from Dockerfile (this will take some time).
|
Also, you can build docker image from Dockerfile (this will take some time).
|
||||||
|
|
@ -782,7 +782,7 @@ After all services are ready, you can directly push inference requests int queue
|
||||||
```python
|
```python
|
||||||
from zoo.serving.client import InputQueue
|
from zoo.serving.client import InputQueue
|
||||||
input_api = InputQueue()
|
input_api = InputQueue()
|
||||||
input_api.enqueue('my-image1', user_define_key={"path: 'path/to/image1'})
|
input_api.enqueue('my-image1', user_define_key={"path": 'path/to/image1'})
|
||||||
```
|
```
|
||||||
|
|
||||||
Cluster Serving service is a long running service in container, you can stop it as follows:
|
Cluster Serving service is a long running service in container, you can stop it as follows:
|
||||||
|
|
|
||||||
|
|
@ -10,9 +10,9 @@ Prior to deploying PPML Cluster Serving, please make sure the following is setup
|
||||||
## Deploy Trusted Realtime ML for Kubernetes ##
|
## Deploy Trusted Realtime ML for Kubernetes ##
|
||||||
1. Pull docker image from dockerhub
|
1. Pull docker image from dockerhub
|
||||||
```
|
```
|
||||||
$ docker pull intelanalytics/analytics-zoo-ppml-trusted-realtime-ml-scala-graphene:0.12.0-SNAPSHOT
|
$ docker pull intelanalytics/bigdl-ppml-trusted-realtime-ml-scala-graphene:0.14.0-SNAPSHOT
|
||||||
```
|
```
|
||||||
2. Pull the source code of Analytics Zoo and enter PPML graphene k8s directory
|
2. Pull the source code of Analytics Zoo/BigDL and enter PPML graphene k8s directory
|
||||||
```
|
```
|
||||||
$ git clone https://github.com/intel-analytics/analytics-zoo.git
|
$ git clone https://github.com/intel-analytics/analytics-zoo.git
|
||||||
$ cd analytics-zoo/ppml/trusted-realtime-ml/scala/docker-graphene/kubernetes
|
$ cd analytics-zoo/ppml/trusted-realtime-ml/scala/docker-graphene/kubernetes
|
||||||
|
|
|
||||||
Loading…
Reference in a new issue