Skip to content

Commit 6090e6a

Browse files
committed
Correct some spellings & links.
1 parent c910bf8 commit 6090e6a

35 files changed

+209
-200
lines changed

README.md

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -33,17 +33,17 @@ PaddleOCR aims to create multilingual, awesome, leading, and practical OCR tools
3333
- [more](./doc/doc_en/update_en.md)
3434

3535
## Features
36-
- PP-OCR series of high-quality pre-trained models, comparable to commercial effects
36+
- PP-OCR - A series of high-quality pre-trained models, comparable to commercial products
3737
- Ultra lightweight PP-OCRv2 series models: detection (3.1M) + direction classifier (1.4M) + recognition 8.5M) = 13.0M
3838
- Ultra lightweight PP-OCR mobile series models: detection (3.0M) + direction classifier (1.4M) + recognition (5.0M) = 9.4M
3939
- General PP-OCR server series models: detection (47.1M) + direction classifier (1.4M) + recognition (94.9M) = 143.4M
4040
- Support Chinese, English, and digit recognition, vertical text recognition, and long text recognition
41-
- Support multi-language recognition: about 80 languages like Korean, Japanese, German, French, etc
41+
- Support multi-lingual recognition: about 80 languages like Korean, Japanese, German, French, etc
4242
- PP-Structure: a document structurize system
43-
- support layout analysis and table recognition (support export to Excel)
44-
- support key information extraction
45-
- support DocVQA
46-
- Rich toolkits related to the OCR areas
43+
- Support layout analysis and table recognition (support export to Excel)
44+
- Support key information extraction
45+
- Support DocVQA
46+
- Rich OCR toolkit
4747
- Semi-automatic data annotation tool, i.e., PPOCRLabel: support fast and efficient data annotation
4848
- Data synthesis tool, i.e., Style-Text: easy to synthesize a large number of images which are similar to the target scene image
4949
- Support user-defined training, provides rich predictive inference deployment solutions
@@ -62,7 +62,7 @@ The above pictures are the visualizations of the general ppocr_server model. For
6262

6363
<a name="Community"></a>
6464
## Community
65-
- Scan the QR code below with your Wechat, you can access to official technical exchange group. Look forward to your participation.
65+
- Scan the QR code below with your Wechat, you can join the official technical discussion group. Looking forward to your participation.
6666

6767
<div align="center">
6868
<img src="https://raw.githubusercontent.com/PaddlePaddle/PaddleOCR/dygraph/doc/joinus.PNG" width = "200" height = "200" />
@@ -120,8 +120,8 @@ For a new language request, please refer to [Guideline for new language_requests
120120
- [PP-Structure: Information Extraction](./ppstructure/README.md)
121121
- [Layout Parser](./ppstructure/layout/README.md)
122122
- [Table Recognition](./ppstructure/table/README.md)
123-
- [DocVQA](https://github.com/PaddlePaddle/PaddleOCR/tree/release/2.4/ppstructure/vqa)
124-
- [Key Information Extraction](https://github.com/PaddlePaddle/PaddleOCR/blob/release/2.4/ppstructure/docs/kie.md)
123+
- [DocVQA](./ppstructure/vqa/README.md)
124+
- [Key Information Extraction](./ppstructure/docs/kie.md)
125125
- Academic Circles
126126
- [Two-stage Algorithm](./doc/doc_en/algorithm_overview_en.md)
127127
- [PGNet Algorithm](./doc/doc_en/pgnet_en.md)

README_ch.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -99,8 +99,8 @@ PaddleOCR旨在打造一套丰富、领先、且实用的OCR工具库,助力
9999
- [PP-Structure信息提取](./ppstructure/README_ch.md)
100100
- [版面分析](./ppstructure/layout/README_ch.md)
101101
- [表格识别](./ppstructure/table/README_ch.md)
102-
- [DocVQA](https://github.com/PaddlePaddle/PaddleOCR/tree/release/2.4/ppstructure/vqa)
103-
- [关键信息提取](https://github.com/PaddlePaddle/PaddleOCR/blob/release/2.4/ppstructure/docs/kie.md)
102+
- [DocVQA](./ppstructure/vqa/README_ch.md)
103+
- [关键信息提取](./ppstructure/docs/kie.md)
104104
- OCR学术圈
105105
- [两阶段模型介绍与下载](./doc/doc_ch/algorithm_overview.md)
106106
- [端到端PGNet算法](./doc/doc_ch/pgnet.md)

deploy/cpp_infer/readme_en.md

Lines changed: 35 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,8 @@
11
# Server-side C++ Inference
22

3-
This chapter introduces the C++ deployment method of the PaddleOCR model, and the corresponding python predictive deployment method refers to [document](../../doc/doc_ch/inference.md).
4-
C++ is better than python in terms of performance calculation. Therefore, in most CPU and GPU deployment scenarios, C++ deployment is mostly used.
5-
This section will introduce how to configure the C++ environment and complete it in the Linux\Windows (CPU\GPU) environment
6-
PaddleOCR model deployment.
3+
This chapter introduces the C++ deployment steps of the PaddleOCR model. The corresponding Python predictive deployment method refers to [document](../../doc/doc_ch/inference.md).
4+
C++ is better than python in terms of performance. Therefore, in CPU and GPU deployment scenarios, C++ deployment is mostly used.
5+
This section will introduce how to configure the C++ environment and deploy PaddleOCR in Linux (CPU\GPU) environment. For Windows deployment please refer to [Windows](./docs/windows_vs2019_build.md) compilation guidelines.
76

87

98
## 1. Prepare the Environment
@@ -15,17 +14,17 @@ PaddleOCR model deployment.
1514

1615
### 1.1 Compile OpenCV
1716

18-
* First of all, you need to download the source code compiled package in the Linux environment from the opencv official website. Taking opencv3.4.7 as an example, the download command is as follows.
17+
* First of all, you need to download the source code compiled package in the Linux environment from the OpenCV official website. Taking OpenCV 3.4.7 as an example, the download command is as follows.
1918

2019
```bash
2120
cd deploy/cpp_infer
2221
wget https://paddleocr.bj.bcebos.com/libs/opencv/opencv-3.4.7.tar.gz
2322
tar -xf opencv-3.4.7.tar.gz
2423
```
2524

26-
Finally, you can see the folder of `opencv-3.4.7/` in the current directory.
25+
Finally, you will see the folder of `opencv-3.4.7/` in the current directory.
2726

28-
* Compile opencv, the opencv source path (`root_path`) and installation path (`install_path`) should be set by yourself. Enter the opencv source code path and compile it in the following way.
27+
* Compile OpenCV, the OpenCV source path (`root_path`) and installation path (`install_path`) should be set by yourself. Enter the OpenCV source code path and compile it in the following way.
2928

3029

3130
```shell
@@ -58,11 +57,11 @@ make -j
5857
make install
5958
```
6059

61-
Among them, `root_path` is the downloaded opencv source code path, and `install_path` is the installation path of opencv. After `make install` is completed, the opencv header file and library file will be generated in this folder for later OCR source code compilation.
60+
In the above commands, `root_path` is the downloaded OpenCV source code path, and `install_path` is the installation path of OpenCV. After `make install` is completed, the OpenCV header file and library file will be generated in this folder for later OCR source code compilation.
6261

6362

6463

65-
The final file structure under the opencv installation path is as follows.
64+
The final file structure under the OpenCV installation path is as follows.
6665

6766
```
6867
opencv3/
@@ -79,28 +78,28 @@ opencv3/
7978

8079
#### 1.2.1 Direct download and installation
8180

82-
[Paddle inference library official website](https://paddle-inference.readthedocs.io/en/latest/user_guides/download_lib.html). You can view and select the appropriate version of the inference library on the official website.
81+
[Paddle inference library official website](https://paddle-inference.readthedocs.io/en/latest/user_guides/download_lib.html). You can review and select the appropriate version of the inference library on the official website.
8382

8483

85-
* After downloading, use the following method to uncompress.
84+
* After downloading, use the following command to extract files.
8685

8786
```
8887
tar -xf paddle_inference.tgz
8988
```
9089

91-
Finally you can see the following files in the folder of `paddle_inference/`.
90+
Finally you will see the the folder of `paddle_inference/` in the current path.
9291

93-
#### 1.2.2 Compile from the source code
94-
* If you want to get the latest Paddle inference library features, you can download the latest code from Paddle github repository and compile the inference library from the source code. It is recommended to download the inference library with paddle version greater than or equal to 2.0.1.
95-
* You can refer to [Paddle inference library] (https://www.paddlepaddle.org.cn/documentation/docs/en/advanced_guide/inference_deployment/inference/build_and_install_lib_en.html) to get the Paddle source code from github, and then compile To generate the latest inference library. The method of using git to access the code is as follows.
92+
#### 1.2.2 Compile the inference source code
93+
* If you want to get the latest Paddle inference library features, you can download the latest code from Paddle GitHub repository and compile the inference library from the source code. It is recommended to download the inference library with paddle version greater than or equal to 2.0.1.
94+
* You can refer to [Paddle inference library] (https://www.paddlepaddle.org.cn/documentation/docs/en/advanced_guide/inference_deployment/inference/build_and_install_lib_en.html) to get the Paddle source code from GitHub, and then compile To generate the latest inference library. The method of using git to access the code is as follows.
9695

9796

9897
```shell
9998
git clone https://github.com/PaddlePaddle/Paddle.git
10099
git checkout develop
101100
```
102101

103-
* After entering the Paddle directory, the commands to compile the paddle inference library are as follows.
102+
* Enter the Paddle directory and run the following commands to compile the paddle inference library.
104103

105104
```shell
106105
rm -rf build
@@ -133,14 +132,14 @@ build/paddle_inference_install_dir/
133132
|-- version.txt
134133
```
135134

136-
Among them, `paddle` is the Paddle library required for C++ prediction later, and `version.txt` contains the version information of the current inference library.
135+
`paddle` is the Paddle library required for C++ prediction later, and `version.txt` contains the version information of the current inference library.
137136

138137

139138
## 2. Compile and Run the Demo
140139

141140
### 2.1 Export the inference model
142141

143-
* You can refer to [Model inference](../../doc/doc_ch/inference.md)export the inference model. After the model is exported, assuming it is placed in the `inference` directory, the directory structure is as follows.
142+
* You can refer to [Model inference](../../doc/doc_ch/inference.md) and export the inference model. After the model is exported, assuming it is placed in the `inference` directory, the directory structure is as follows.
144143

145144
```
146145
inference/
@@ -171,20 +170,28 @@ CUDA_LIB_DIR=your_cuda_lib_dir
171170
CUDNN_LIB_DIR=your_cudnn_lib_dir
172171
```
173172

174-
`OPENCV_DIR` is the opencv installation path; `LIB_DIR` is the download (`paddle_inference` folder)
173+
`OPENCV_DIR` is the OpenCV installation path; `LIB_DIR` is the download (`paddle_inference` folder)
175174
or the generated Paddle inference library path (`build/paddle_inference_install_dir` folder);
176-
`CUDA_LIB_DIR` is the cuda library file path, in docker; it is `/usr/local/cuda/lib64`; `CUDNN_LIB_DIR` is the cudnn library file path, in docker it is `/usr/lib/x86_64-linux-gnu/`.
175+
`CUDA_LIB_DIR` is the CUDA library file path, in docker; it is `/usr/local/cuda/lib64`; `CUDNN_LIB_DIR` is the cuDNN library file path, in docker it is `/usr/lib/x86_64-linux-gnu/`.
177176

178177

179178
* After the compilation is completed, an executable file named `ppocr` will be generated in the `build` folder.
180179

181180

182181
### Run the demo
183-
Execute the built executable file:
182+
Execute the built executable file:
184183
```shell
185184
./build/ppocr <mode> [--param1] [--param2] [...]
186185
```
187-
Here, `mode` is a required parameter,and the value range is ['det', 'rec', 'system'], representing using detection only, using recognition only and using the end-to-end system respectively. Specifically,
186+
`mode` is a required parameter,and the valid values are
187+
188+
mode value | Model used
189+
-----|------
190+
det | Detection only
191+
rec | Recognition only
192+
system | End-to-end system
193+
194+
Specifically,
188195

189196
##### 1. run det demo:
190197
```shell
@@ -214,9 +221,9 @@ Here, `mode` is a required parameter,and the value range is ['det', 'rec', 'sy
214221
--image_dir=../../doc/imgs/12.jpg
215222
```
216223

217-
More parameters are as follows,
224+
More parameters are as follows,
218225

219-
- common parameters
226+
- Common parameters
220227

221228
|parameter|data type|default|meaning|
222229
| --- | --- | --- | --- |
@@ -226,7 +233,7 @@ More parameters are as follows,
226233
|cpu_math_library_num_threads|int|10|Number of threads when using CPU inference. When machine cores is enough, the large the value, the faster the inference speed|
227234
|use_mkldnn|bool|true|Whether to use mkdlnn library|
228235

229-
- detection related parameters
236+
- Detection related parameters
230237

231238
|parameter|data type|default|meaning|
232239
| --- | --- | --- | --- |
@@ -238,15 +245,15 @@ More parameters are as follows,
238245
|use_polygon_score|bool|false|Whether to use polygon box to calculate bbox score, false means to use rectangle box to calculate. Use rectangular box to calculate faster, and polygonal box more accurate for curved text area.|
239246
|visualize|bool|true|Whether to visualize the results,when it is set as true, The prediction result will be save in the image file `./ocr_vis.png`.|
240247

241-
- classifier related parameters
248+
- Classifier related parameters
242249

243250
|parameter|data type|default|meaning|
244251
| --- | --- | --- | --- |
245252
|use_angle_cls|bool|false|Whether to use the direction classifier|
246253
|cls_model_dir|string|-|Address of direction classifier inference model|
247254
|cls_thresh|float|0.9|Score threshold of the direction classifier|
248255

249-
- recogniton related parameters
256+
- Recognition related parameters
250257

251258
|parameter|data type|default|meaning|
252259
| --- | --- | --- | --- |
@@ -265,4 +272,4 @@ The detection results will be shown on the screen, which is as follows.
265272

266273
### 2.3 Notes
267274

268-
* Paddle2.0.0 inference model library is recommended for this toturial.
275+
* Paddle 2.0.0 inference model library is recommended for this tutorial.

deploy/docker/hubserving/README.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,9 @@
11
English | [简体中文](README_cn.md)
22

33
## Introduction
4-
Many users hope package the PaddleOCR service into a docker image, so that it can be quickly released and used in the docker or k8s environment.
4+
Many users hope package the PaddleOCR service into a docker image, so that it can be quickly released and used in the docker or K8s environment.
55

6-
This page provides some standardized code to achieve this goal. You can quickly publish the PaddleOCR project into a callable Restful API service through the following steps. (At present, the deployment based on the HubServing mode is implemented first, and author plans to increase the deployment of the PaddleServing mode in the futrue)
6+
This page provides some standardized code to achieve this goal. You can quickly publish the PaddleOCR project into a callable Restful API service through the following steps. (At present, the deployment based on the HubServing mode is implemented first, and author plans to increase the deployment of the PaddleServing mode in the future)
77

88
## 1. Prerequisites
99

@@ -14,7 +14,7 @@ c. NVIDIA Container Toolkit(GPU,Docker 19.03+ can skip this)
1414
d. cuDNN 7.6+(GPU)
1515

1616
## 2. Build Image
17-
a. Goto Dockerfile directory(ps:Need to distinguish between cpu and gpu version, the following takes cpu as an example, gpu version needs to replace the keyword)
17+
a. Go to Dockerfile directory(PS: Need to distinguish between CPU and GPU version, the following takes CPU as an example, GPU version needs to replace the keyword)
1818
```
1919
cd deploy/docker/hubserving/cpu
2020
```
@@ -42,13 +42,13 @@ docker logs -f paddle_ocr
4242
```
4343

4444
## 4. Test
45-
a. Calculate the Base64 encoding of the picture to be recognized (if you just test, you can use a free online tool, like:https://freeonlinetools24.com/base64-image/
45+
a. Calculate the Base64 encoding of the picture to be recognized (For test purpose, you can use a free online tool such as https://freeonlinetools24.com/base64-image/ )
4646
b. Post a service request(sample request in sample_request.txt)
4747

4848
```
4949
curl -H "Content-Type:application/json" -X POST --data "{\"images\": [\"Input image Base64 encode(need to delete the code 'data:image/jpg;base64,')\"]}" http://localhost:8868/predict/ocr_system
5050
```
51-
c. Get resposne(If the call is successful, the following result will be returned)
51+
c. Get response(If the call is successful, the following result will be returned)
5252
```
5353
{"msg":"","results":[[{"confidence":0.8403433561325073,"text":"约定","text_region":[[345,377],[641,390],[634,540],[339,528]]},{"confidence":0.8131805658340454,"text":"最终相遇","text_region":[[356,532],[624,530],[624,596],[356,598]]}]],"status":"0"}
5454
```

deploy/lite/readme_en.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,8 @@
11
# Tutorial of PaddleOCR Mobile deployment
22

3-
This tutorial will introduce how to use [Paddle Lite](https://github.com/PaddlePaddle/Paddle-Lite) to deploy paddleOCR ultra-lightweight Chinese and English detection models on mobile phones.
3+
This tutorial will introduce how to use [Paddle Lite](https://github.com/PaddlePaddle/Paddle-Lite) to deploy PaddleOCR ultra-lightweight Chinese and English detection models on mobile phones.
44

5-
paddle-lite is a lightweight inference engine for PaddlePaddle. It provides efficient inference capabilities for mobile phones and IoTs, and extensively integrates cross-platform hardware to provide lightweight deployment solutions for end-side deployment issues.
5+
paddle-lite is a lightweight inference engine for PaddlePaddle. It provides efficient inference capabilities for mobile phones and IoT, and extensively integrates cross-platform hardware to provide lightweight deployment solutions for end-side deployment issues.
66

77
## 1. Preparation
88

deploy/pdserving/README_CN.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,6 +22,7 @@ PaddleOCR提供2种服务部署方式:
2222
- [环境准备](#环境准备)
2323
- [模型转换](#模型转换)
2424
- [Paddle Serving pipeline部署](#部署)
25+
- [Windows用户](#Windows用户)
2526
- [FAQ](#FAQ)
2627

2728
<a name="环境准备"></a>
@@ -187,9 +188,10 @@ python3 -m paddle_serving_client.convert --dirname ./ch_PP-OCRv2_rec_infer/ \
187188
2021-05-13 03:42:36,979 chl2(In: ['rec'], Out: ['@DAGExecutor']) size[0/0]
188189
```
189190
190-
## WINDOWS用户
191+
<a name="Windows用户"></a>
192+
## Windows用户
191193
192-
Windows用户不能使用上述的启动方式,需要使用Web Service,详情参见[Windows平台使用Paddle Serving指导](https://github.com/PaddlePaddle/Serving/blob/develop/doc/WINDOWS_TUTORIAL_CN.md)
194+
Windows用户不能使用上述的启动方式,需要使用Web Service,详情参见[Windows平台使用Paddle Serving指导](https://github.com/PaddlePaddle/Serving/blob/develop/doc/Windows_Tutorial_CN.md)
193195
194196
**WINDOWS只能使用0.5.0版本的CPU模式**
195197

0 commit comments

Comments
 (0)