From 4d2158031efcbe60b9379645c52ee13404381c7e Mon Sep 17 00:00:00 2001 From: MRXLT Date: Fri, 3 Apr 2020 11:31:37 +0800 Subject: [PATCH 1/4] fix doc --- README.md | 2 +- README_CN.md | 2 +- doc/SERVER_DAG.md | 2 +- doc/TRAIN_TO_SERVICE.md | 4 ++-- doc/TRAIN_TO_SERVICE_CN.md | 4 ++-- 5 files changed, 7 insertions(+), 7 deletions(-) diff --git a/README.md b/README.md index f20012d0b..3e42a11fb 100644 --- a/README.md +++ b/README.md @@ -249,7 +249,7 @@ curl -H "Content-Type:application/json" -X POST -d '{"url": "https://paddle-serv - [Compile from source code](doc/COMPILE.md) ### About Efficiency -- [How to profile Paddle Serving latency?](https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/util) +- [How to profile Paddle Serving latency?](python/examples/util) - [CPU Benchmarks(Chinese)](doc/BENCHMARKING.md) - [GPU Benchmarks(Chinese)](doc/GPU_BENCHMARKING.md) diff --git a/README_CN.md b/README_CN.md index f5ca91ef5..f379dec0b 100644 --- a/README_CN.md +++ b/README_CN.md @@ -255,7 +255,7 @@ curl -H "Content-Type:application/json" -X POST -d '{"url": "https://paddle-serv - [如何编译PaddleServing?](doc/COMPILE_CN.md) ### 关于Paddle Serving性能 -- [如何测试Paddle Serving性能?](https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/util/) +- [如何测试Paddle Serving性能?](python/examples/util/) - [CPU版Benchmarks](doc/BENCHMARKING.md) - [GPU版Benchmarks](doc/GPU_BENCHMARKING.md) diff --git a/doc/SERVER_DAG.md b/doc/SERVER_DAG.md index fdfcec948..5a5c851ef 100644 --- a/doc/SERVER_DAG.md +++ b/doc/SERVER_DAG.md @@ -14,7 +14,7 @@ Deep neural nets often have some preprocessing steps on input data, and postproc ## How to define Node -PaddleServing has some predefined Computation Node in the framework. A very commonly used Computation Graph is the simple reader-inference-response mode that can cover most of the single model inference scenarios. A example graph and the corresponding DAG defination code is as follows. +PaddleServing has some predefined Computation Node in the framework. A very commonly used Computation Graph is the simple reader-inference-response mode that can cover most of the single model inference scenarios. A example graph and the corresponding DAG definition code is as follows.
diff --git a/doc/TRAIN_TO_SERVICE.md b/doc/TRAIN_TO_SERVICE.md index 4219e6694..795c329e8 100644 --- a/doc/TRAIN_TO_SERVICE.md +++ b/doc/TRAIN_TO_SERVICE.md @@ -5,9 +5,9 @@ Paddle Serving is Paddle's high-performance online inference service framework, which can flexibly support the deployment of most models. In this article, the IMDB review sentiment analysis task is used as an example to show the entire process from model training to deployment of inference service through 9 steps. ## Step1:Prepare for Running Environment -Paddle Serving can be deployed on Linux environments such as Centos and Ubuntu. On other systems or in environments where you do not want to install the serving module, you can still access the server-side prediction service through the http service. +Paddle Serving can be deployed on Linux environments.Currently the server supports deployment on Centos7. [Docker deployment is recommended](RUN_IN_DOCKER.md). The rpc client supports deploymen on Centos7 and Ubuntu 18.On other systems or in environments where you do not want to install the serving module, you can still access the server-side prediction service through the http service. -You can choose to install the cpu or gpu version of the server module according to the requirements and machine environment, and install the client module on the client machine. When you want to access the server with http +You can choose to install the cpu or gpu version of the server module according to the requirements and machine environment, and install the client module on the client machine. When you want to access the server with http, there is not need to install client module. ```shell pip install paddle_serving_server #cpu version server side diff --git a/doc/TRAIN_TO_SERVICE_CN.md b/doc/TRAIN_TO_SERVICE_CN.md index 8349723fb..ad2a43c30 100644 --- a/doc/TRAIN_TO_SERVICE_CN.md +++ b/doc/TRAIN_TO_SERVICE_CN.md @@ -6,9 +6,9 @@ Paddle Serving是Paddle的高性能在线预测服务框架,可以灵活支持 ## Step1:准备环境 -Paddle Serving可以部署在Centos和Ubuntu等Linux环境上,在其他系统上或者不希望安装serving模块的环境中仍然可以通过http服务来访问server端的预测服务。 +Paddle Serving可以部署在Linux环境上,目前server端支持在Centos7上部署,推荐使用[Docker部署](RUN_IN_DOCKER_CN.md)。rpc client端可以在Centos7和Ubuntu18上部署,在其他系统上或者不希望安装serving模块的环境中仍然可以通过http服务来访问server端的预测服务。 -可以根据需求和机器环境来选择安装cpu或gpu版本的server模块,在client端机器上安装client模块。当希望同http来访问server端 +可以根据需求和机器环境来选择安装cpu或gpu版本的server模块,在client端机器上安装client模块。使用http请求的方式来访问server时,client端机器不需要安装client模块。 ```shell pip install paddle_serving_server #cpu版本server端 From e1e87a2a60b68447c32f8dff37e19c5b153b3320 Mon Sep 17 00:00:00 2001 From: MRXLT Date: Fri, 3 Apr 2020 11:44:49 +0800 Subject: [PATCH 2/4] add mem_optim argument --- README.md | 1 + README_CN.md | 1 + 2 files changed, 2 insertions(+) diff --git a/README.md b/README.md index 3e42a11fb..9a9036c7e 100644 --- a/README.md +++ b/README.md @@ -65,6 +65,7 @@ python -m paddle_serving_server.serve --model uci_housing_model --thread 10 --po | `port` | int | `9292` | Exposed port of current service to users| | `name` | str | `""` | Service name, can be used to generate HTTP request url | | `model` | str | `""` | Path of paddle model directory to be served | +| `mem_optim` | bool | `False` | Enable memory optimization | Here, we use `curl` to send a HTTP POST request to the service we just started. Users can use any python library to send HTTP POST as well, e.g, [requests](https://requests.readthedocs.io/en/master/). diff --git a/README_CN.md b/README_CN.md index f379dec0b..07c9e1f23 100644 --- a/README_CN.md +++ b/README_CN.md @@ -69,6 +69,7 @@ python -m paddle_serving_server.serve --model uci_housing_model --thread 10 --po | `port` | int | `9292` | Exposed port of current service to users| | `name` | str | `""` | Service name, can be used to generate HTTP request url | | `model` | str | `""` | Path of paddle model directory to be served | +| `mem_optim` | bool | `False` | Enable memory optimization | 我们使用 `curl` 命令来发送HTTP POST请求给刚刚启动的服务。用户也可以调用python库来发送HTTP POST请求,请参考英文文档 [requests](https://requests.readthedocs.io/en/master/)。 From 07da8f38863cbf9ac3244faf99621e8b16d9679e Mon Sep 17 00:00:00 2001 From: MRXLT Date: Fri, 3 Apr 2020 15:37:55 +0800 Subject: [PATCH 3/4] fix demos list --- README.md | 2 +- README_CN.md | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 9a9036c7e..5aac66a24 100644 --- a/README.md +++ b/README.md @@ -232,7 +232,7 @@ curl -H "Content-Type:application/json" -X POST -d '{"url": "https://paddle-serv | Key | Value | | :----------------- | :----------------------------------------------------------- | | Model Name | DNN-CTR(with cube) | -| URL | None(Get model by [local_train.py](python/examples/criteo_ctr_with_cube/local_train.py)) | +| URL | https://paddle-serving.bj.bcebos.com/unittest/ctr_cube_unittest.tar.gz | | Client/Server Code | https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/criteo_ctr_with_cube | | Description | Get click probability from a feature vector of item | diff --git a/README_CN.md b/README_CN.md index 07c9e1f23..8901fa417 100644 --- a/README_CN.md +++ b/README_CN.md @@ -237,7 +237,7 @@ curl -H "Content-Type:application/json" -X POST -d '{"url": "https://paddle-serv | Key | Value | | :----------------- | :----------------------------------------------------------- | | 模型名 | DNN-CTR(with cube) | -| 下载链接 | None(Get model by [local_train.py](python/examples/criteo_ctr_with_cube/local_train.py)) | +| 下载链接 | https://paddle-serving.bj.bcebos.com/unittest/ctr_cube_unittest.tar.gz | | 客户端/服务端代码 | https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/criteo_ctr_with_cube | | 介绍 | 从项目的特征向量中获得点击概率 | From f2318ded5f4e6948ab471b4d00ee1e348a042b15 Mon Sep 17 00:00:00 2001 From: MRXLT Date: Fri, 3 Apr 2020 19:38:15 +0800 Subject: [PATCH 4/4] update demo list --- README.md | 2 +- README_CN.md | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 5aac66a24..b31f72af5 100644 --- a/README.md +++ b/README.md @@ -223,7 +223,7 @@ curl -H "Content-Type:application/json" -X POST -d '{"url": "https://paddle-serv | Key | Value | | :----------------- | :----------------------------------------------------------- | | Model Name | DNN-CTR | -| URL | None(Get model by [local_train.py](./python/examples/criteo_ctr/local_train.py)) | +| URL | https://paddle-serving.bj.bcebos.com/criteo_ctr_example/criteo_ctr_demo_model.tar.gz | | Client/Server Code | https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/criteo_ctr | | Description | Get click probability from a feature vector of item | diff --git a/README_CN.md b/README_CN.md index 8901fa417..477dc4d55 100644 --- a/README_CN.md +++ b/README_CN.md @@ -228,7 +228,7 @@ curl -H "Content-Type:application/json" -X POST -d '{"url": "https://paddle-serv | Key | Value | | :----------------- | :----------------------------------------------------------- | | 模型名 | DNN-CTR | -| 下载链接 | None(Get model by [local_train.py](./python/examples/criteo_ctr/local_train.py)) | +| 下载链接 | https://paddle-serving.bj.bcebos.com/criteo_ctr_example/criteo_ctr_demo_model.tar.gz | | 客户端/服务端代码 | https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/criteo_ctr | | 介绍 | 从项目的特征向量中获得点击概率 |