Merge pull request #3930 from Evezerest/2.3
Divide PP-OCR model inference and other model inference
This commit is contained in:
commit
c00b005f59
|
@ -103,13 +103,12 @@ For a new language request, please refer to [Guideline for new language_requests
|
|||
- [PP-OCR Model and Configuration](./doc/doc_en/models_and_config_en.md)
|
||||
- [PP-OCR Model Download](./doc/doc_en/models_list_en.md)
|
||||
- [Yml Configuration](./doc/doc_en/config_en.md)
|
||||
- [Python Inference](./doc/doc_en/inference_en.md)
|
||||
- [Python Inference for PP-OCR Model Library](./doc/doc_en/inference_ppocr_en.md)
|
||||
- [PP-OCR Training](./doc/doc_en/training_en.md)
|
||||
- [Text Detection](./doc/doc_en/detection_en.md)
|
||||
- [Text Recognition](./doc/doc_en/recognition_en.md)
|
||||
- [Direction Classification](./doc/doc_en/angle_class_en.md)
|
||||
- Inference and Deployment
|
||||
- [Python Inference](./doc/doc_en/inference_en.md)
|
||||
- [C++ Inference](./deploy/cpp_infer/readme_en.md)
|
||||
- [Serving](./deploy/pdserving/README.md)
|
||||
- [Mobile](./deploy/lite/readme_en.md)
|
||||
|
@ -120,6 +119,7 @@ For a new language request, please refer to [Guideline for new language_requests
|
|||
- Academic Circles
|
||||
- [Two-stage Algorithm](./doc/doc_en/algorithm_overview_en.md)
|
||||
- [PGNet Algorithm](./doc/doc_en/algorithm_overview_en.md)
|
||||
- [Python Inference](./doc/doc_en/inference_en.md)
|
||||
- Data Annotation and Synthesis
|
||||
- [Semi-automatic Annotation Tool: PPOCRLabel](./PPOCRLabel/README.md)
|
||||
- [Data Synthesis Tool: Style-Text](./StyleText/README.md)
|
||||
|
|
|
@ -95,13 +95,12 @@ PaddleOCR旨在打造一套丰富、领先、且实用的OCR工具库,助力
|
|||
- [PP-OCR模型与配置文件](./doc/doc_ch/models_and_config.md)
|
||||
- [PP-OCR模型下载](./doc/doc_ch/models_list.md)
|
||||
- [配置文件内容与生成](./doc/doc_ch/config.md)
|
||||
- [模型库快速使用](./doc/doc_ch/inference.md)
|
||||
- [PP-OCR模型库快速推理](./doc/doc_ch/inference_ppocr.md)
|
||||
- [PP-OCR模型训练](./doc/doc_ch/training.md)
|
||||
- [文本检测](./doc/doc_ch/detection.md)
|
||||
- [文本识别](./doc/doc_ch/recognition.md)
|
||||
- [方向分类器](./doc/doc_ch/angle_class.md)
|
||||
- PP-OCR模型推理部署
|
||||
- [基于Python脚本预测引擎推理](./doc/doc_ch/inference.md)
|
||||
- [基于C++预测引擎推理](./deploy/cpp_infer/readme.md)
|
||||
- [服务化部署](./deploy/pdserving/README_CN.md)
|
||||
- [端侧部署](./deploy/lite/readme.md)
|
||||
|
@ -117,6 +116,7 @@ PaddleOCR旨在打造一套丰富、领先、且实用的OCR工具库,助力
|
|||
- OCR学术圈
|
||||
- [两阶段模型介绍与下载](./doc/doc_ch/algorithm_overview.md)
|
||||
- [端到端PGNet算法](./doc/doc_ch/pgnet.md)
|
||||
- [基于Python脚本预测引擎推理](./doc/doc_ch/inference.md)
|
||||
- 数据集
|
||||
- [通用中英文OCR数据集](./doc/doc_ch/datasets.md)
|
||||
- [手写中文OCR数据集](./doc/doc_ch/handwritten_datasets.md)
|
||||
|
|
|
@ -2,7 +2,7 @@
|
|||
# PP-OCR模型与配置文件
|
||||
PP-OCR模型与配置文件一章主要补充一些OCR模型的基本概念、配置文件的内容与作用以便对模型后续的参数调整和训练中拥有更好的体验。
|
||||
|
||||
本节包含三个部分,首先在[PP-OCR模型下载](./models_list.md)中解释PP-OCR模型的类型概念,并提供所有模型的下载链接。然后在[配置文件内容与生成](./config.md)中详细说明调整PP-OCR模型所需的参数。最后的[模型库快速使用](./inference.md)是对第一节PP-OCR模型库使用方法的介绍,可以通过Python推理引擎快速利用丰富的模型库模型获得测试结果。
|
||||
本章包含三个部分,首先在[PP-OCR模型下载](./models_list.md)中解释PP-OCR模型的类型概念,并提供所有模型的下载链接。然后在[配置文件内容与生成](./config.md)中详细说明调整PP-OCR模型所需的参数。最后的[模型库快速使用](./inference_ppocr.md)是对第一节PP-OCR模型库使用方法的介绍,可以通过Python推理引擎快速利用丰富的模型库模型获得测试结果。
|
||||
|
||||
------
|
||||
|
||||
|
|
|
@ -90,7 +90,7 @@ cd /path/to/ppocr_img
|
|||
```
|
||||
|
||||
|
||||
更多whl包使用可参考[whl包文档](./whl.md)
|
||||
如需使用2.0模型,请指定参数`--version 2.0`,paddleocr默认使用2.1模型。更多whl包使用可参考[whl包文档](./whl.md)
|
||||
|
||||
|
||||
<a name="212"></a>
|
||||
|
|
|
@ -7,15 +7,13 @@
|
|||
- [1.2 数据下载](#数据下载)
|
||||
- [1.3 字典](#字典)
|
||||
- [1.4 支持空格](#支持空格)
|
||||
|
||||
- [2 启动训练](#启动训练)
|
||||
- [2.1 数据增强](#数据增强)
|
||||
- [2.2 通用模型训练](#通用模型训练)
|
||||
- [2.3 多语言模型训练](#多语言模型训练)
|
||||
|
||||
- [3 评估](#评估)
|
||||
|
||||
- [4 预测](#预测)
|
||||
- [5 转Inference模型测试](#Inference)
|
||||
|
||||
|
||||
<a name="数据准备"></a>
|
||||
|
@ -424,3 +422,39 @@ python3 tools/infer_rec.py -c configs/rec/ch_ppocr_v2.0/rec_chinese_lite_train_v
|
|||
infer_img: doc/imgs_words/ch/word_1.jpg
|
||||
result: ('韩国小馆', 0.997218)
|
||||
```
|
||||
|
||||
<a name="Inference"></a>
|
||||
|
||||
## 5. 转Inference模型测试
|
||||
|
||||
识别模型转inference模型与检测的方式相同,如下:
|
||||
|
||||
```
|
||||
# -c 后面设置训练算法的yml配置文件
|
||||
# -o 配置可选参数
|
||||
# Global.pretrained_model 参数设置待转换的训练模型地址,不用添加文件后缀 .pdmodel,.pdopt或.pdparams。
|
||||
# Global.save_inference_dir参数设置转换的模型将保存的地址。
|
||||
|
||||
python3 tools/export_model.py -c configs/rec/ch_ppocr_v2.0/rec_chinese_lite_train_v2.0.yml -o Global.pretrained_model=./ch_lite/ch_ppocr_mobile_v2.0_rec_train/best_accuracy Global.save_inference_dir=./inference/rec_crnn/
|
||||
```
|
||||
|
||||
**注意:**如果您是在自己的数据集上训练的模型,并且调整了中文字符的字典文件,请注意修改配置文件中的`character_dict_path`是否是所需要的字典文件。
|
||||
|
||||
转换成功后,在目录下有三个文件:
|
||||
|
||||
```
|
||||
/inference/rec_crnn/
|
||||
├── inference.pdiparams # 识别inference模型的参数文件
|
||||
├── inference.pdiparams.info # 识别inference模型的参数信息,可忽略
|
||||
└── inference.pdmodel # 识别inference模型的program文件
|
||||
```
|
||||
|
||||
- 自定义模型推理
|
||||
|
||||
如果训练时修改了文本的字典,在使用inference模型预测时,需要通过`--rec_char_dict_path`指定使用的字典路径,并且设置 `rec_char_type=ch`
|
||||
|
||||
```
|
||||
python3 tools/infer/predict_rec.py --image_dir="./doc/imgs_words_en/word_336.png" --rec_model_dir="./your inference model" --rec_image_shape="3, 32, 100" --rec_char_type="ch" --rec_char_dict_path="your text dict path"
|
||||
```
|
||||
|
||||
|
||||
|
|
|
@ -1,22 +1,22 @@
|
|||
|
||||
# Reasoning based on Python prediction engine
|
||||
# Python Inference for PP-OCR Model Library
|
||||
|
||||
This article introduces the use of the Python inference engine for the PP-OCR model library. The content is in order of text detection, text recognition, direction classifier and the prediction method of the three in series on the CPU and GPU.
|
||||
|
||||
|
||||
- [TEXT DETECTION MODEL INFERENCE](#DETECTION_MODEL_INFERENCE)
|
||||
- [Text Detection Model Inference](#DETECTION_MODEL_INFERENCE)
|
||||
|
||||
- [TEXT RECOGNITION MODEL INFERENCE](#RECOGNITION_MODEL_INFERENCE)
|
||||
- [1. LIGHTWEIGHT CHINESE MODEL](#LIGHTWEIGHT_RECOGNITION)
|
||||
- [2. MULTILINGUAL MODEL INFERENCE](MULTILINGUAL_MODEL_INFERENCE)
|
||||
- [Text Recognition Model Inference](#RECOGNITION_MODEL_INFERENCE)
|
||||
- [1. Lightweight Chinese Recognition Model Inference](#LIGHTWEIGHT_RECOGNITION)
|
||||
- [2. Multilingaul Model Inference](#MULTILINGUAL_MODEL_INFERENCE)
|
||||
|
||||
- [ANGLE CLASSIFICATION MODEL INFERENCE](#ANGLE_CLASS_MODEL_INFERENCE)
|
||||
- [Angle Classification Model Inference](#ANGLE_CLASS_MODEL_INFERENCE)
|
||||
|
||||
- [TEXT DETECTION ANGLE CLASSIFICATION AND RECOGNITION INFERENCE CONCATENATION](#CONCATENATION)
|
||||
- [Text Detection Angle Classification and Recognition Inference Concatenation](#CONCATENATION)
|
||||
|
||||
<a name="DETECTION_MODEL_INFERENCE"></a>
|
||||
|
||||
## TEXT DETECTION MODEL INFERENCE
|
||||
## Text Detection Model Inference
|
||||
|
||||
The default configuration is based on the inference setting of the DB text detection model. For lightweight Chinese detection model inference, you can execute the following commands:
|
||||
|
||||
|
@ -52,11 +52,11 @@ python3 tools/infer/predict_det.py --image_dir="./doc/imgs/1.jpg" --det_model_di
|
|||
|
||||
<a name="RECOGNITION_MODEL_INFERENCE"></a>
|
||||
|
||||
## TEXT RECOGNITION MODEL INFERENCE
|
||||
## Text Recognition Model Inference
|
||||
|
||||
|
||||
<a name="LIGHTWEIGHT_RECOGNITION"></a>
|
||||
### 1. LIGHTWEIGHT CHINESE TEXT RECOGNITION MODEL REFERENCE
|
||||
### 1. Lightweight Chinese Recognition Model Inference
|
||||
|
||||
For lightweight Chinese recognition model inference, you can execute the following commands:
|
||||
|
||||
|
@ -77,7 +77,7 @@ Predicts of ./doc/imgs_words_en/word_10.png:('PAIN', 0.9897658)
|
|||
|
||||
<a name="MULTILINGUAL_MODEL_INFERENCE"></a>
|
||||
|
||||
### 2. MULTILINGAUL MODEL INFERENCE
|
||||
### 2. Multilingaul Model Inference
|
||||
If you need to predict other language models, when using inference model prediction, you need to specify the dictionary path used by `--rec_char_dict_path`. At the same time, in order to get the correct visualization results,
|
||||
You need to specify the visual font path through `--vis_font_path`. There are small language fonts provided by default under the `doc/fonts` path, such as Korean recognition:
|
||||
|
||||
|
@ -94,7 +94,7 @@ Predicts of ./doc/imgs_words/korean/1.jpg:('바탕으로', 0.9948904)
|
|||
|
||||
<a name="ANGLE_CLASS_MODEL_INFERENCE"></a>
|
||||
|
||||
## ANGLE CLASSIFICATION MODEL INFERENCE
|
||||
## Angle Classification Model Inference
|
||||
|
||||
For angle classification model inference, you can execute the following commands:
|
||||
|
||||
|
@ -114,7 +114,7 @@ After executing the command, the prediction results (classification angle and sc
|
|||
```
|
||||
|
||||
<a name="CONCATENATION"></a>
|
||||
## TEXT DETECTION ANGLE CLASSIFICATION AND RECOGNITION INFERENCE CONCATENATION
|
||||
## Text Detection Angle Classification and Recognition Inference Concatenation
|
||||
|
||||
When performing prediction, you need to specify the path of a single image or a folder of images through the parameter `image_dir`, the parameter `det_model_dir` specifies the path to detect the inference model, the parameter `cls_model_dir` specifies the path to angle classification inference model and the parameter `rec_model_dir` specifies the path to identify the inference model. The parameter `use_angle_cls` is used to control whether to enable the angle classification model. The parameter `use_mp` specifies whether to use multi-process to infer `total_process_num` specifies process number when using multi-process. The parameter . The visualized recognition results are saved to the `./inference_results` folder by default.
|
||||
|
||||
|
|
|
@ -1,4 +1,12 @@
|
|||
# PP-OCR Model and Configuration
|
||||
The chapter on PP-OCR model and configuration file mainly adds some basic concepts of OCR model and the content and role of configuration file to have a better experience in the subsequent parameter adjustment and training of the model.
|
||||
|
||||
This chapter contains three parts. Firstly, [PP-OCR Model Download](. /models_list_en.md) explains the concept of PP-OCR model types and provides links to download all models. Then in [Yml Configuration](. /config_en.md) details the parameters needed to fine-tune the PP-OCR models. The final [Python Inference for PP-OCR Model Library](. /inference_ppocr_en.md) is an introduction to the use of the PP-OCR model library in the first section, which can quickly utilize the rich model library models to obtain test results through the Python inference engine.
|
||||
|
||||
------
|
||||
|
||||
Let's first understand some basic concepts.
|
||||
|
||||
- [INTRODUCTION ABOUT OCR](#introduction-about-ocr)
|
||||
* [BASIC CONCEPTS OF OCR DETECTION MODEL](#basic-concepts-of-ocr-detection-model)
|
||||
* [Basic concepts of OCR recognition model](#basic-concepts-of-ocr-recognition-model)
|
||||
|
|
|
@ -95,7 +95,7 @@ If you do not use the provided test image, you can replace the following `--imag
|
|||
['PAIN', 0.990372]
|
||||
```
|
||||
|
||||
More whl package usage can be found in [whl package](./whl_en.md)
|
||||
If you need to use the 2.0 model, please specify the parameter `--version 2.0`, paddleocr uses the 2.1 model by default. More whl package usage can be found in [whl package](./whl_en.md)
|
||||
<a name="212-multi-language-model"></a>
|
||||
|
||||
#### 2.1.2 Multi-language Model
|
||||
|
|
|
@ -15,6 +15,7 @@
|
|||
|
||||
- [4 PREDICTION](#PREDICTION)
|
||||
- [4.1 Training engine prediction](#Training_engine_prediction)
|
||||
- [5 CONVERT TO INFERENCE MODEL](#Inference)
|
||||
|
||||
<a name="DATA_PREPARATION"></a>
|
||||
## 1 DATA PREPARATION
|
||||
|
@ -361,6 +362,7 @@ Eval:
|
|||
```
|
||||
|
||||
<a name="EVALUATION"></a>
|
||||
|
||||
## 3 EVALUATION
|
||||
|
||||
The evaluation dataset can be set by modifying the `Eval.dataset.label_file_list` field in the `configs/rec/rec_icdar15_train.yml` file.
|
||||
|
@ -432,3 +434,40 @@ Get the prediction result of the input image:
|
|||
infer_img: doc/imgs_words/ch/word_1.jpg
|
||||
result: ('韩国小馆', 0.997218)
|
||||
```
|
||||
|
||||
<a name="Inference"></a>
|
||||
|
||||
## 5 CONVERT TO INFERENCE MODEL
|
||||
|
||||
The recognition model is converted to the inference model in the same way as the detection, as follows:
|
||||
|
||||
```
|
||||
# -c Set the training algorithm yml configuration file
|
||||
# -o Set optional parameters
|
||||
# Global.pretrained_model parameter Set the training model address to be converted without adding the file suffix .pdmodel, .pdopt or .pdparams.
|
||||
# Global.save_inference_dir Set the address where the converted model will be saved.
|
||||
|
||||
python3 tools/export_model.py -c configs/rec/ch_ppocr_v2.0/rec_chinese_lite_train_v2.0.yml -o Global.pretrained_model=./ch_lite/ch_ppocr_mobile_v2.0_rec_train/best_accuracy Global.save_inference_dir=./inference/rec_crnn/
|
||||
```
|
||||
|
||||
If you have a model trained on your own dataset with a different dictionary file, please make sure that you modify the `character_dict_path` in the configuration file to your dictionary file path.
|
||||
|
||||
After the conversion is successful, there are three files in the model save directory:
|
||||
|
||||
```
|
||||
inference/det_db/
|
||||
├── inference.pdiparams # The parameter file of recognition inference model
|
||||
├── inference.pdiparams.info # The parameter information of recognition inference model, which can be ignored
|
||||
└── inference.pdmodel # The program file of recognition model
|
||||
```
|
||||
|
||||
- Text recognition model Inference using custom characters dictionary
|
||||
|
||||
If the text dictionary is modified during training, when using the inference model to predict, you need to specify the dictionary path used by `--rec_char_dict_path`, and set `rec_char_type=ch`
|
||||
|
||||
```
|
||||
python3 tools/infer/predict_rec.py --image_dir="./doc/imgs_words_en/word_336.png" --rec_model_dir="./your inference model" --rec_image_shape="3, 32, 100" --rec_char_type="ch" --rec_char_dict_path="your text dict path"
|
||||
```
|
||||
|
||||
|
||||
|
||||
|
|
Loading…
Reference in New Issue