diff --git a/example/README.md b/example/README.md index f9c2950b..1fc56593 100644 --- a/example/README.md +++ b/example/README.md @@ -55,9 +55,9 @@ Now users could send inference request with server certificates (`server.crt`). There are prebuilt docker images could be used for the examples, either in the following docker way or [`kubernates`](./kubernetes/) way. Users could pull them directly and try the example. ``` -docker pull occlum/init_ra_server:0.29.1 -docker pull occlum/tf_demo:0.29.1 -docker pull occlum/tf_demo_client:0.29.1 +docker pull occlum/init_ra_server:0.29.2-ubuntu20.04 +docker pull occlum/tf_demo:0.29.2-ubuntu20.04 +docker pull occlum/tf_demo_client:0.29.2-ubuntu20.04 ``` If users want to build or customize the images, please check below part. @@ -66,11 +66,11 @@ If users want to build or customize the images, please check below part. Our target is to deploy the demo in separated container images, so docker build is necessary steps. Thanks to the `docker run in docker` method, this example build could be done in Occlum development container image. -First, please make sure `docker` is installed successfully in your host. Then start the Occlum container (use version `0.29.1-ubuntu20.04` for example) as below. +First, please make sure `docker` is installed successfully in your host. Then start the Occlum container (use version `0.29.2-ubuntu20.04` for example) as below. ``` $ sudo docker run --rm -itd --network host \ -v $(which docker):/usr/bin/docker -v /var/run/docker.sock:/var/run/docker.sock \ - occlum/occlum:0.29.1-ubuntu20.04 + occlum/occlum:0.29.2-ubuntu20.04 ``` All the following are running in the above container. diff --git a/example/kubernetes/README.md b/example/kubernetes/README.md index dc2cd806..920085a5 100644 --- a/example/kubernetes/README.md +++ b/example/kubernetes/README.md @@ -56,12 +56,12 @@ usage: build.sh [OPTION]... For example, below command generates three container images. ``` -# ./build.sh -r demo -g 0.29.1 +# ./build.sh -r demo -g 0.29.2 ``` -* **`demo/init_ra_server:0.29.1`** acts as key broker pod. -* **`demo/tf_demo:0.29.1`** acts as tensorflow serving pod. -* **`demo/tf_demo_client:0.29.1`** acts as client. +* **`demo/init_ra_server:0.29.2`** acts as key broker pod. +* **`demo/tf_demo:0.29.2`** acts as tensorflow serving pod. +* **`demo/tf_demo_client:0.29.2`** acts as client. ## How to test @@ -110,7 +110,7 @@ In default, only one replica for the tensorflow serving pod. ### Try the inference request ``` -$ docker run --rm --network host demo/tf_demo_client:0.29.1 python3 inception_client.py --server=localhost:31001 --crt server.crt --image cat.jpg +$ docker run --rm --network host demo/tf_demo_client:0.29.2 python3 inception_client.py --server=localhost:31001 --crt server.crt --image cat.jpg ``` If successful, it prints the classification results. @@ -120,7 +120,7 @@ If successful, it prints the classification results. Below command can do benchmark test for the tensorflow serving service running in Occlum. ``` -$ docker run --rm --network host demo/tf_demo_client:0.29.1 python3 benchmark.py --server localhost:31001 --crt server.crt --cnum 4 --loop 10 --image cat.jpg +$ docker run --rm --network host demo/tf_demo_client:0.29.2 python3 benchmark.py --server localhost:31001 --crt server.crt --cnum 4 --loop 10 --image cat.jpg ``` Try scale up the tensorflow serving pods number, better `tps` can be achieved.