-## MEASUREMENT
+## Measurement
Explanation:
diff --git a/doc/doc_en/config_en.md b/doc/doc_en/config_en.md
index 4ac6758ff6..aa78263e4b 100644
--- a/doc/doc_en/config_en.md
+++ b/doc/doc_en/config_en.md
@@ -1,4 +1,12 @@
-## Optional parameter list
+# Configuration
+
+- [1. Optional Parameter List](#1-optional-parameter-list)
+- [2. Intorduction to Global Parameters of Configuration File](#2-intorduction-to-global-parameters-of-configuration-file)
+- [3. Multilingual Config File Generation](#3-multilingual-config-file-generation)
+
+
+
+## 1. Optional Parameter List
The following list can be viewed through `--help`
@@ -7,7 +15,9 @@ The following list can be viewed through `--help`
| -c | ALL | Specify configuration file to use | None | **Please refer to the parameter introduction for configuration file usage** |
| -o | ALL | set configuration options | None | Configuration using -o has higher priority than the configuration file selected with -c. E.g: -o Global.use_gpu=false |
-## INTRODUCTION TO GLOBAL PARAMETERS OF CONFIGURATION FILE
+
+
+## 2. Intorduction to Global Parameters of Configuration File
Take rec_chinese_lite_train_v2.0.yml as an example
### Global
@@ -121,8 +131,9 @@ In PaddleOCR, the network is divided into four stages: Transform, Backbone, Neck
| drop_last | Whether to discard the last incomplete mini-batch because the number of samples in the data set cannot be divisible by batch_size | True | \ |
| num_workers | The number of sub-processes used to load data, if it is 0, the sub-process is not started, and the data is loaded in the main process | 8 | \ |
+
-## 3. MULTILINGUAL CONFIG FILE GENERATION
+## 3. Multilingual Config File Generation
PaddleOCR currently supports 80 (except Chinese) language recognition. A multi-language configuration file template is
provided under the path `configs/rec/multi_languages`: [rec_multi_language_lite_train.yml](../../configs/rec/multi_language/rec_multi_language_lite_train.yml)。
@@ -187,21 +198,21 @@ Italian is made up of Latin letters, so after executing the command, you will ge
...
character_type: it # language
character_dict_path: {path/of/dict} # path of dict
-
+
Train:
dataset:
name: SimpleDataSet
data_dir: train_data/ # root directory of training data
label_file_list: ["./train_data/train_list.txt"] # train label path
...
-
+
Eval:
dataset:
name: SimpleDataSet
data_dir: train_data/ # root directory of val data
label_file_list: ["./train_data/val_list.txt"] # val label path
...
-
+
```
diff --git a/doc/doc_en/detection_en.md b/doc/doc_en/detection_en.md
index d3f6f3da10..14180c6faa 100644
--- a/doc/doc_en/detection_en.md
+++ b/doc/doc_en/detection_en.md
@@ -1,23 +1,23 @@
-# CONTENT
-
-- [Paste Your Document In Here](#paste-your-document-in-here)
-- [1. TEXT DETECTION](#1-text-detection)
- * [1.1 DATA PREPARATION](#11-data-preparation)
- * [1.2 DOWNLOAD PRETRAINED MODEL](#12-download-pretrained-model)
- * [1.3 START TRAINING](#13-start-training)
- * [1.4 LOAD TRAINED MODEL AND CONTINUE TRAINING](#14-load-trained-model-and-continue-training)
- * [1.5 TRAINING WITH NEW BACKBONE](#15-training-with-new-backbone)
- * [1.6 EVALUATION](#16-evaluation)
- * [1.7 TEST](#17-test)
- * [1.8 INFERENCE MODEL PREDICTION](#18-inference-model-prediction)
-- [2. FAQ](#2-faq)
-
-
-# 1. TEXT DETECTION
+# Text Detection
This section uses the icdar2015 dataset as an example to introduce the training, evaluation, and testing of the detection model in PaddleOCR.
-## 1.1 DATA PREPARATION
+- [1. Data and Weights Preparation](#1-data-and-weights-preparatio)
+ * [1.1 Data Preparation](#11-data-preparation)
+ * [1.2 Download Pretrained Model](#12-download-pretrained-model)
+- [2. Training](#2-training)
+ * [2.1 Start Training](#21-start-training)
+ * [2.2 Load Trained Model and Continue Training](#22-load-trained-model-and-continue-training)
+ * [2.3 Training with New Backbone](#23-training-with-new-backbone)
+- [3. Evaluation and Test](#3-evaluation-and-test)
+ * [3.1 Evaluation](#31-evaluation)
+ * [3.2 Test](#32-test)
+- [4. Inference](#4-inference)
+- [5. FAQ](#2-faq)
+
+## 1. Data and Weights Preparation
+
+### 1.1 Data Preparation
The icdar2015 dataset contains train set which has 1000 images obtained with wearable cameras and test set which has 500 images obtained with wearable cameras. The icdar2015 can be obtained from [official website](https://rrc.cvc.uab.es/?ch=4&com=downloads). Registration is required for downloading.
@@ -59,7 +59,7 @@ The `points` in the dictionary represent the coordinates (x, y) of the four poin
If you want to train PaddleOCR on other datasets, please build the annotation file according to the above format.
-## 1.2 DOWNLOAD PRETRAINED MODEL
+### 1.2 Download Pretrained Model
First download the pretrained model. The detection model of PaddleOCR currently supports 3 backbones, namely MobileNetV3, ResNet18_vd and ResNet50_vd. You can use the model in [PaddleClas](https://github.com/PaddlePaddle/PaddleClas/tree/release/2.0/ppcls/modeling/architectures) to replace backbone according to your needs.
And the responding download link of backbone pretrain weights can be found in (https://github.com/PaddlePaddle/PaddleClas/blob/release%2F2.0/README_cn.md#resnet%E5%8F%8A%E5%85%B6vd%E7%B3%BB%E5%88%97).
@@ -75,7 +75,10 @@ wget -P ./pretrain_models/ https://paddle-imagenet-models-name.bj.bcebos.com/dyg
```
-## 1.3 START TRAINING
+# 2. TRAINING
+
+### 2.1 Start Training
+
*If CPU version installed, please set the parameter `use_gpu` to `false` in the configuration.*
```shell
python3 tools/train.py -c configs/det/det_mv3_db.yml \
@@ -98,7 +101,7 @@ python3 -m paddle.distributed.launch --gpus '0,1,2,3' tools/train.py -c configs
```
-## 1.4 LOAD TRAINED MODEL AND CONTINUE TRAINING
+### 2.2 Load Trained Model and Continue Training
If you expect to load trained model and continue the training again, you can specify the parameter `Global.checkpoints` as the model path to be loaded.
For example:
@@ -109,7 +112,7 @@ python3 tools/train.py -c configs/det/det_mv3_db.yml -o Global.checkpoints=./you
**Note**: The priority of `Global.checkpoints` is higher than that of `Global.pretrain_weights`, that is, when two parameters are specified at the same time, the model specified by `Global.checkpoints` will be loaded first. If the model path specified by `Global.checkpoints` is wrong, the one specified by `Global.pretrain_weights` will be loaded.
-## 1.5 TRAINING WITH NEW BACKBONE
+### 2.3 Training with New Backbone
The network part completes the construction of the network, and PaddleOCR divides the network into four parts, which are under [ppocr/modeling](../../ppocr/modeling). The data entering the network will pass through these four parts in sequence(transforms->backbones->
necks->heads).
@@ -159,7 +162,9 @@ After adding the four-part modules of the network, you only need to configure th
**NOTE**: More details about replace Backbone and other mudule can be found in [doc](add_new_algorithm_en.md).
-## 1.6 EVALUATION
+## 3. Evaluation and Test
+
+### 3.1 Evaluation
PaddleOCR calculates three indicators for evaluating performance of OCR detection task: Precision, Recall, and Hmean(F-Score).
@@ -174,7 +179,7 @@ python3 tools/eval.py -c configs/det/det_mv3_db.yml -o Global.checkpoints="{pat
* Note: `box_thresh` and `unclip_ratio` are parameters required for DB post-processing, and not need to be set when evaluating the EAST and SAST model.
-## 1.7 TEST
+### 3.2 Test
Test the detection result on a single image:
```shell
@@ -192,7 +197,7 @@ Test the detection result on all images in the folder:
python3 tools/infer_det.py -c configs/det/det_mv3_db.yml -o Global.infer_img="./doc/imgs_en/" Global.pretrained_model="./output/det_db/best_accuracy"
```
-## 1.8 INFERENCE MODEL PREDICTION
+## 4. Inference
The inference model (the model saved by `paddle.jit.save`) is generally a solidified model saved after the model training is completed, and is mostly used to give prediction in deployment.
@@ -215,7 +220,7 @@ If it is other detection algorithms, such as the EAST, the det_algorithm paramet
python3 tools/infer/predict_det.py --det_algorithm="EAST" --det_model_dir="./output/det_db_inference/" --image_dir="./doc/imgs/" --use_gpu=True
```
-# 2. FAQ
+## 5. FAQ
Q1: The prediction results of trained model and inference model are inconsistent?
**A**: Most of the problems are caused by the inconsistency of the pre-processing and post-processing parameters during the prediction of the trained model and the pre-processing and post-processing parameters during the prediction of the inference model. Taking the model trained by the det_mv3_db.yml configuration file as an example, the solution to the problem of inconsistent prediction results between the training model and the inference model is as follows:
diff --git a/doc/doc_en/inference_en.md b/doc/doc_en/inference_en.md
index e30355fb8e..b445232fee 100755
--- a/doc/doc_en/inference_en.md
+++ b/doc/doc_en/inference_en.md
@@ -1,5 +1,5 @@
-# Reasoning based on Python prediction engine
+# Inference Based on Python Prediction Engine
The inference model (the model saved by `paddle.jit.save`) is generally a solidified model saved after the model training is completed, and is mostly used to give prediction in deployment.
@@ -10,37 +10,36 @@ For more details, please refer to the document [Classification Framework](https:
Next, we first introduce how to convert a trained model into an inference model, and then we will introduce text detection, text recognition, angle class, and the concatenation of them based on inference model.
-- [CONVERT TRAINING MODEL TO INFERENCE MODEL](#CONVERT)
- - [Convert detection model to inference model](#Convert_detection_model)
- - [Convert recognition model to inference model](#Convert_recognition_model)
- - [Convert angle classification model to inference model](#Convert_angle_class_model)
+- [1. Convert Training Model to Inference Model](#CONVERT)
+ - [1.1 Convert Detection Model to Inference Model](#Convert_detection_model)
+ - [1.2 Convert Recognition Model to Inference Model](#Convert_recognition_model)
+ - [1.3 Convert Angle Classification Model to Inference Model](#Convert_angle_class_model)
-- [TEXT DETECTION MODEL INFERENCE](#DETECTION_MODEL_INFERENCE)
- - [1. LIGHTWEIGHT CHINESE DETECTION MODEL INFERENCE](#LIGHTWEIGHT_DETECTION)
- - [2. DB TEXT DETECTION MODEL INFERENCE](#DB_DETECTION)
- - [3. EAST TEXT DETECTION MODEL INFERENCE](#EAST_DETECTION)
- - [4. SAST TEXT DETECTION MODEL INFERENCE](#SAST_DETECTION)
- - [5. Multilingual model inference](#Multilingual model inference)
+- [2. Text Detection Model Inference](#DETECTION_MODEL_INFERENCE)
+ - [2.1 Lightweight Chinese Detection Model Inference](#LIGHTWEIGHT_DETECTION)
+ - [2.2 DB Text Detection Model Inference](#DB_DETECTION)
+ - [2.3 East Text Detection Model Inference](#EAST_DETECTION)
+ - [2.4 Sast Text Detection Model Inference](#SAST_DETECTION)
+
+- [3. Text Recognition Model Inference](#RECOGNITION_MODEL_INFERENCE)
+ - [3.1 Lightweight Chinese Text Recognition Model Reference](#LIGHTWEIGHT_RECOGNITION)
+ - [3.2 CTC-Based Text Recognition Model Inference](#CTC-BASED_RECOGNITION)
+ - [3.3 SRN-Based Text Recognition Model Inference](#SRN-BASED_RECOGNITION)
+ - [3.4 Text Recognition Model Inference Using Custom Characters Dictionary](#USING_CUSTOM_CHARACTERS)
+ - [3.5 Multilingual Model Inference](#MULTILINGUAL_MODEL_INFERENCE)
-- [TEXT RECOGNITION MODEL INFERENCE](#RECOGNITION_MODEL_INFERENCE)
- - [1. LIGHTWEIGHT CHINESE MODEL](#LIGHTWEIGHT_RECOGNITION)
- - [2. CTC-BASED TEXT RECOGNITION MODEL INFERENCE](#CTC-BASED_RECOGNITION)
- - [3. SRN-BASED TEXT RECOGNITION MODEL INFERENCE](#SRN-BASED_RECOGNITION)
- - [3. TEXT RECOGNITION MODEL INFERENCE USING CUSTOM CHARACTERS DICTIONARY](#USING_CUSTOM_CHARACTERS)
- - [4. MULTILINGUAL MODEL INFERENCE](MULTILINGUAL_MODEL_INFERENCE)
+- [4. Angle Classification Model Inference](#ANGLE_CLASS_MODEL_INFERENCE)
-- [ANGLE CLASSIFICATION MODEL INFERENCE](#ANGLE_CLASS_MODEL_INFERENCE)
- - [1. ANGLE CLASSIFICATION MODEL INFERENCE](#ANGLE_CLASS_MODEL_INFERENCE)
-
-- [TEXT DETECTION ANGLE CLASSIFICATION AND RECOGNITION INFERENCE CONCATENATION](#CONCATENATION)
- - [1. LIGHTWEIGHT CHINESE MODEL](#LIGHTWEIGHT_CHINESE_MODEL)
- - [2. OTHER MODELS](#OTHER_MODELS)
+- [5. Text Detection Angle Classification And Recognition Inference Concatenation](#CONCATENATION)
+ - [5.1 Lightweight Chinese Model](#LIGHTWEIGHT_CHINESE_MODEL)
+ - [5.2 Other Models](#OTHER_MODELS)
-## CONVERT TRAINING MODEL TO INFERENCE MODEL
+## 1. Convert Training Model to Inference Model
-### Convert detection model to inference model
+
+### 1.1 Convert Detection Model to Inference Model
Download the lightweight Chinese detection model:
```
@@ -67,7 +66,7 @@ inference/det_db/
```
-### Convert recognition model to inference model
+### 1.2 Convert Recognition Model to Inference Model
Download the lightweight Chinese recognition model:
```
@@ -95,7 +94,7 @@ inference/det_db/
```
-### Convert angle classification model to inference model
+### 1.3 Convert Angle Classification Model to Inference Model
Download the angle classification model:
```
@@ -122,13 +121,13 @@ inference/det_db/
-## TEXT DETECTION MODEL INFERENCE
+## 2. Text Detection Model Inference
The following will introduce the lightweight Chinese detection model inference, DB text detection model inference and EAST text detection model inference. The default configuration is based on the inference setting of the DB text detection model.
Because EAST and DB algorithms are very different, when inference, it is necessary to **adapt the EAST text detection algorithm by passing in corresponding parameters**.
-### 1. LIGHTWEIGHT CHINESE DETECTION MODEL INFERENCE
+### 2.1 Lightweight Chinese Detection Model Inference
For lightweight Chinese detection model inference, you can execute the following commands:
@@ -163,7 +162,7 @@ python3 tools/infer/predict_det.py --image_dir="./doc/imgs/1.jpg" --det_model_di
```
-### 2. DB TEXT DETECTION MODEL INFERENCE
+### 2.2 DB Text Detection Model Inference
First, convert the model saved in the DB text detection training process into an inference model. Taking the model based on the Resnet50_vd backbone network and trained on the ICDAR2015 English dataset as an example ([model download link](https://paddleocr.bj.bcebos.com/dygraph_v2.0/en/det_r50_vd_db_v2.0_train.tar)), you can use the following command to convert:
@@ -184,7 +183,7 @@ The visualized text detection results are saved to the `./inference_results` fol
**Note**: Since the ICDAR2015 dataset has only 1,000 training images, mainly for English scenes, the above model has very poor detection result on Chinese text images.
-### 3. EAST TEXT DETECTION MODEL INFERENCE
+### 2.3 EAST TEXT DETECTION MODEL INFERENCE
First, convert the model saved in the EAST text detection training process into an inference model. Taking the model based on the Resnet50_vd backbone network and trained on the ICDAR2015 English dataset as an example ([model download link](https://paddleocr.bj.bcebos.com/dygraph_v2.0/en/det_r50_vd_east_v2.0_train.tar)), you can use the following command to convert:
@@ -205,7 +204,7 @@ The visualized text detection results are saved to the `./inference_results` fol
-### 4. SAST TEXT DETECTION MODEL INFERENCE
+### 2.4 Sast Text Detection Model Inference
#### (1). Quadrangle text detection model (ICDAR2015)
First, convert the model saved in the SAST text detection training process into an inference model. Taking the model based on the Resnet50_vd backbone network and trained on the ICDAR2015 English dataset as an example ([model download link](https://paddleocr.bj.bcebos.com/dygraph_v2.0/en/det_r50_vd_sast_icdar15_v2.0_train.tar)), you can use the following command to convert:
@@ -243,13 +242,13 @@ The visualized text detection results are saved to the `./inference_results` fol
**Note**: SAST post-processing locality aware NMS has two versions: Python and C++. The speed of C++ version is obviously faster than that of Python version. Due to the compilation version problem of NMS of C++ version, C++ version NMS will be called only in Python 3.5 environment, and python version NMS will be called in other cases.
-## TEXT RECOGNITION MODEL INFERENCE
+## 3. Text Recognition Model Inference
The following will introduce the lightweight Chinese recognition model inference, other CTC-based and Attention-based text recognition models inference. For Chinese text recognition, it is recommended to choose the recognition model based on CTC loss. In practice, it is also found that the result of the model based on Attention loss is not as good as the one based on CTC loss. In addition, if the characters dictionary is modified during training, make sure that you use the same characters set during inferencing. Please check below for details.
-### 1. LIGHTWEIGHT CHINESE TEXT RECOGNITION MODEL REFERENCE
+### 3.1 Lightweight Chinese Text Recognition Model Reference
For lightweight Chinese recognition model inference, you can execute the following commands:
@@ -269,7 +268,7 @@ Predicts of ./doc/imgs_words_en/word_10.png:('PAIN', 0.9897658)
```
-### 2. CTC-BASED TEXT RECOGNITION MODEL INFERENCE
+### 3.2 CTC-Based Text Recognition Model Inference
Taking CRNN as an example, we introduce the recognition model inference based on CTC loss. Rosetta and Star-Net are used in a similar way, No need to set the recognition algorithm parameter rec_algorithm.
@@ -292,6 +291,7 @@ After executing the command, the recognition result of the above image is as fol
```bash
Predicts of ./doc/imgs_words_en/word_336.png:('super', 0.9999073)
```
+
**Note**:Since the above model refers to [DTRB](https://arxiv.org/abs/1904.01906) text recognition training and evaluation process, it is different from the training of lightweight Chinese recognition model in two aspects:
- The image resolution used in training is different: the image resolution used in training the above model is [3,32,100], while during our Chinese model training, in order to ensure the recognition effect of long text, the image resolution used in training is [3, 32, 320]. The default shape parameter of the inference stage is the image resolution used in training phase, that is [3, 32, 320]. Therefore, when running inference of the above English model here, you need to set the shape of the recognition image through the parameter `rec_image_shape`.
@@ -304,7 +304,7 @@ dict_character = list(self.character_str)
```
-### 3. SRN-BASED TEXT RECOGNITION MODEL INFERENCE
+### 3.3 SRN-Based Text Recognition Model Inference
The recognition model based on SRN requires additional setting of the recognition algorithm parameter
--rec_algorithm="SRN". At the same time, it is necessary to ensure that the predicted shape is consistent
@@ -319,7 +319,7 @@ python3 tools/infer/predict_rec.py --image_dir="./doc/imgs_words_en/word_336.png
```
-### 4. TEXT RECOGNITION MODEL INFERENCE USING CUSTOM CHARACTERS DICTIONARY
+### 3.4 Text Recognition Model Inference Using Custom Characters Dictionary
If the text dictionary is modified during training, when using the inference model to predict, you need to specify the dictionary path used by `--rec_char_dict_path`, and set `rec_char_type=ch`
```
@@ -327,7 +327,8 @@ python3 tools/infer/predict_rec.py --image_dir="./doc/imgs_words_en/word_336.png
```
-### 5. MULTILINGAUL MODEL INFERENCE
+
+### 3.5 Multilingual Model Inference
If you need to predict other language models, when using inference model prediction, you need to specify the dictionary path used by `--rec_char_dict_path`. At the same time, in order to get the correct visualization results,
You need to specify the visual font path through `--vis_font_path`. There are small language fonts provided by default under the `doc/fonts` path, such as Korean recognition:
@@ -343,13 +344,7 @@ Predicts of ./doc/imgs_words/korean/1.jpg:('바탕으로', 0.9948904)
```
-## ANGLE CLASSIFICATION MODEL INFERENCE
-
-The following will introduce the angle classification model inference.
-
-
-
-### 1.ANGLE CLASSIFICATION MODEL INFERENCE
+## 4. Angle Classification Model Inference
For angle classification model inference, you can execute the following commands:
@@ -371,10 +366,10 @@ After executing the command, the prediction results (classification angle and sc
```
-## TEXT DETECTION ANGLE CLASSIFICATION AND RECOGNITION INFERENCE CONCATENATION
+## 5. Text Detection Angle Classification and Recognition Inference Concatenation
-### 1. LIGHTWEIGHT CHINESE MODEL
+### 5.1 Lightweight Chinese Model
When performing prediction, you need to specify the path of a single image or a folder of images through the parameter `image_dir`, the parameter `det_model_dir` specifies the path to detect the inference model, the parameter `cls_model_dir` specifies the path to angle classification inference model and the parameter `rec_model_dir` specifies the path to identify the inference model. The parameter `use_angle_cls` is used to control whether to enable the angle classification model. The parameter `use_mp` specifies whether to use multi-process to infer `total_process_num` specifies process number when using multi-process. The parameter . The visualized recognition results are saved to the `./inference_results` folder by default.
@@ -388,14 +383,14 @@ python3 tools/infer/predict_system.py --image_dir="./doc/imgs/00018069.jpg" --de
# use multi-process
python3 tools/infer/predict_system.py --image_dir="./doc/imgs/00018069.jpg" --det_model_dir="./inference/det_db/" --rec_model_dir="./inference/rec_crnn/" --use_angle_cls=false --use_mp=True --total_process_num=6
```
-```
+
After executing the command, the recognition result image is as follows:

-### 2. OTHER MODELS
+### 5.2 Other Models
If you want to try other detection algorithms or recognition algorithms, please refer to the above text detection model inference and text recognition model inference, update the corresponding configuration and model.
diff --git a/doc/doc_en/models_en.md b/doc/doc_en/models_en.md
new file mode 100644
index 0000000000..37c4a17456
--- /dev/null
+++ b/doc/doc_en/models_en.md
@@ -0,0 +1,46 @@
+# PP-OCR Model Zoo
+The PP-OCR model zoo section explains some basic concepts of the OCR model and how to quickly use the models in the PP-OCR model library.
+
+This section contains two parts. Firstly, [PP-OCR Model Download](./models_list_en.md) explains the concept of PP-OCR model types and provides links to download all models. The next [Python Inference for PP-OCR Model Zoo](./inference_ppocr_en.md) is an introduction to the use of the PP-OCR model library, which can quickly utilize the rich model library models to obtain test results through the Python inference engine.
+
+------
+
+Let's first understand some basic concepts.
+
+- [Introduction about OCR](#introduction-about-ocr)
+ * [Basic Concepts of OCR Detection Model](#basic-concepts-of-ocr-detection-model)
+ * [Basic Concepts of OCR Recognition Model](#basic-concepts-of-ocr-recognition-model)
+ * [PP-OCR Model](#pp-ocr-model)
+
+
+## 1. Introduction about OCR
+
+This section briefly introduces the basic concepts of OCR detection model and recognition model, and introduces PaddleOCR's PP-OCR model.
+
+OCR (Optical Character Recognition, Optical Character Recognition) is currently the general term for text recognition. It is not limited to document or book text recognition, but also includes recognizing text in natural scenes. It can also be called STR (Scene Text Recognition).
+
+OCR text recognition generally includes two parts, text detection and text recognition. The text detection module first uses detection algorithms to detect text lines in the image. And then the recognition algorithm to identify the specific text in the text line.
+
+
+### 1.1 Basic Concepts of OCR Detection Model
+
+Text detection can locate the text area in the image, and then usually mark the word or text line in the form of a bounding box. Traditional text detection algorithms mostly extract features manually, which are characterized by fast speed and good effect in simple scenes, but the effect will be greatly reduced when faced with natural scenes. Currently, deep learning methods are mostly used.
+
+Text detection algorithms based on deep learning can be roughly divided into the following categories:
+1. Method based on target detection. Generally, after the text box is predicted, the final text box is filtered through NMS, which is mostly four-point text box, which is not ideal for curved text scenes. Typical algorithms are methods such as EAST and Text Box.
+2. Method based on text segmentation. The text line is regarded as the segmentation target, and then the external text box is constructed through the segmentation result, which can handle curved text, and the effect is not ideal for the text cross scene problem. Typical algorithms are DB, PSENet and other methods.
+3. Hybrid target detection and segmentation method.
+
+
+### 1.2 Basic Concepts of OCR Recognition Model
+
+The input of the OCR recognition algorithm is generally text lines images which has less background information, and the text information occupies the main part. The recognition algorithm can be divided into two types of algorithms:
+1. CTC-based method. The text prediction module of the recognition algorithm is based on CTC, and the commonly used algorithm combination is CNN+RNN+CTC. There are also some algorithms that try to add transformer modules to the network and so on.
+2. Attention-based method. The text prediction module of the recognition algorithm is based on Attention, and the commonly used algorithm combination is CNN+RNN+Attention.
+
+
+### 1.3 PP-OCR Model
+
+PaddleOCR integrates many OCR algorithms, text detection algorithms include DB, EAST, SAST, etc., text recognition algorithms include CRNN, RARE, StarNet, Rosetta, SRN and other algorithms.
+
+Among them, PaddleOCR has released the PP-OCR series model for the general OCR in Chinese and English natural scenes. The PP-OCR model is composed of the DB+CRNN algorithm. It uses massive Chinese data training and model tuning methods to have high text detection and recognition capabilities in Chinese scenes. And PaddleOCR has launched a high-precision and ultra-lightweight PP-OCRv2 model. The detection model is only 3M, and the recognition model is only 8.5M. Using [PaddleSlim](https://github.com/PaddlePaddle/PaddleSlim)'s model quantification method, the detection model can be compressed to 0.8M without reducing the accuracy. The recognition is compressed to 3M, which is more suitable for mobile deployment scenarios.
diff --git a/doc/doc_en/paddleOCR_overview_en.md b/doc/doc_en/paddleOCR_overview_en.md
index 403cd99415..073c3ec889 100644
--- a/doc/doc_en/paddleOCR_overview_en.md
+++ b/doc/doc_en/paddleOCR_overview_en.md
@@ -36,4 +36,4 @@ If you getting this error `OSError: [WinError 126] The specified module could no
Please try to download Shapely whl file using [http://www.lfd.uci.edu/~gohlke/pythonlibs/#shapely](http://www.lfd.uci.edu/~gohlke/pythonlibs/#shapely).
-Reference: [Solve shapely installation on windows](
\ No newline at end of file
+Reference: [Solve shapely installation on windows](https://stackoverflow.com/questions/44398265/install-shapely-oserror-winerror-126-the-specified-module-could-not-be-found)
\ No newline at end of file
diff --git a/doc/doc_en/quickstart_en.md b/doc/doc_en/quickstart_en.md
index c4fb506819..0055d8f7a8 100644
--- a/doc/doc_en/quickstart_en.md
+++ b/doc/doc_en/quickstart_en.md
@@ -5,7 +5,7 @@
+ [1. Install PaddleOCR Whl Package](#1-install-paddleocr-whl-package)
* [2. Easy-to-Use](#2-easy-to-use)
- + [2.1 Use by command line](#21-use-by-command-line)
+ + [2.1 Use by Command Line](#21-use-by-command-line)
- [2.1.1 English and Chinese Model](#211-english-and-chinese-model)
- [2.1.2 Multi-language Model](#212-multi-language-model)
- [2.1.3 Layout Analysis](#213-layoutAnalysis)
@@ -39,7 +39,7 @@ pip install "paddleocr>=2.0.1" # Recommend to use version 2.0.1+
-### 2.1 Use by command line
+### 2.1 Use by Command Line
PaddleOCR provides a series of test images, click [here](https://paddleocr.bj.bcebos.com/dygraph_v2.1/ppocr_img.zip) to download, and then switch to the corresponding directory in the terminal
@@ -95,7 +95,7 @@ If you do not use the provided test image, you can replace the following `--imag
['PAIN', 0.990372]
```
-If you need to use the 2.0 model, please specify the parameter `--version 2.0`, paddleocr uses the 2.1 model by default. More whl package usage can be found in [whl package](./whl_en.md)
+If you need to use the 2.0 model, please specify the parameter `--version PP-OCR`, paddleocr uses the 2.1 model by default(`--versioin PP-OCRv2`). More whl package usage can be found in [whl package](./whl_en.md)
#### 2.1.2 Multi-language Model
diff --git a/doc/doc_en/recognition_en.md b/doc/doc_en/recognition_en.md
index 347c000789..84f5541562 100644
--- a/doc/doc_en/recognition_en.md
+++ b/doc/doc_en/recognition_en.md
@@ -1,24 +1,23 @@
-# TEXT RECOGNITION
+# Text Recognition
-- [1 DATA PREPARATION](#DATA_PREPARATION)
+- [1. Data Preparation](#DATA_PREPARATION)
- [1.1 Costom Dataset](#Costom_Dataset)
- [1.2 Dataset Download](#Dataset_download)
- [1.3 Dictionary](#Dictionary)
- [1.4 Add Space Category](#Add_space_category)
-- [2 TRAINING](#TRAINING)
+- [2. Training](#TRAINING)
- [2.1 Data Augmentation](#Data_Augmentation)
- [2.2 General Training](#Training)
- [2.3 Multi-language Training](#Multi_language)
-- [3 EVALUATION](#EVALUATION)
+- [3. Evaluation](#EVALUATION)
-- [4 PREDICTION](#PREDICTION)
- - [4.1 Training engine prediction](#Training_engine_prediction)
-- [5 CONVERT TO INFERENCE MODEL](#Inference)
+- [4. Prediction](#PREDICTION)
+- [5. Convert to Inference Model](#Inference)
-## 1 DATA PREPARATION
+## 1. Data Preparation
PaddleOCR supports two data formats:
@@ -37,7 +36,7 @@ mklink /d rs`vXm00Y!IoL~7D7~7$8oEs|$fZss>V~8
Hwl(Gq>0V0*%wckJ1P+h~xvfalmO;)$T
znH#y)T>lO`HQavZtmYz-ETR6cwHr%{R)f#m9^L=W0?Z&wU*x!wD6J1n#XxMxyYK_y
z5@7}^o8i~6Bql;6#odEv`NFM^$z_r
HX138Xh#i7PAXaD97h{3FMa@?+~1*%8;!=VWp+DX!gh!uT1eO@pTy5_
zn$;7@_nM^l8NMwm%&dWHWk{V6Ip`}zb=Ul3EH@GUe8jG5RMu1TA0!!Au6?^y(iT_W
zUXbzuXj@Q;^fP^9r}w{7m+zE
zh(I%ao4R<6Uj=XrQTWjBIoZiC?=VC>G`XFZ9A$|I@zs#1fOz9VylF`SzfbAFnuIMF
ziTJu-aA$Qhp*xF^M~Ayn|Ku0>cN3;|8?Dc(NX)ci@zBc(GF+aUw$1H1NK&_C!w$}8
zj0*{MkHonLKqQj~@sO6c<{~yrbPTBzyoe(Xq=a_V30Q6JEt;o@5E_i4S?$?`nF
z+K=HSUI46r@I}ikS6+n5BHQt;&rLS}TX?>b;;RNO7-`eX0f#*oGh8{6d{4c}f(sAJ
zi^^E=%%ga>D>lt5b9C{9)c{wKC^FStYpjS#6??yRFQG2kDO2fsXV=y@yEC4GA)Z&n
z)NC!cv;+
zGaNzJdwFIkTn!^>##<*rzegu+Z!M>GWWF`WAAX`(OUFbm*%6P
@5wyMof-t*g24{oBMs^F8>%FVBarHOA1paP(?rb(|>8X
zyVxD%cbW34_epUM0#!V`>gI)_XW}{38k9YJCv!JcXuGV$T41rf-gz7mL||7g@pk+U
zCF6YL#2k(w+Ol*nBL6llizJTP_Y2wW-u0%Pa