简体中文 | English
Seal text recognition is a technology that automatically extracts and recognizes seal content from documents or images. The recognition of seal text is part of document processing and has various applications in many scenarios, such as contract comparison, inventory access approval, and invoice reimbursement approval.
The Seal Text Recognition pipeline includes a layout area analysis module, a seal text detection module, and a text recognition module.
If you prioritize model accuracy, please choose a model with higher accuracy. If you prioritize inference speed, please choose a model with faster inference. If you prioritize model storage size, please choose a model with a smaller storage footprint.
The pre trained model production line provided by PaddleX can quickly experience the effect. You can experience the effect of the seal text recognition production line online, or use the command line or Python locally to experience the effect of the seal text recognition production line.
You can experience online the effect of seal text recognition in the v3 production line for extracting document scene information, using official demo images for recognition, for example:
If you are satisfied with the performance of the production line, you can directly integrate and deploy the production line. If you are not satisfied, you can also use private data to fine tune the models in the production line online.
Before using the seal text recognition production line locally, please ensure that you have completed the wheel package installation of PaddleX according to the PaddleX Local Installation Guide.
One command can quickly experience the effect of seal text recognition production line, use test file, and replace --input with the local path for prediction
paddlex --pipeline seal_recognition --input seal_text_det.png --device gpu:0 --save_path output
Parameter description:
--Pipeline: Production line name, here is the seal text recognition production line
--Input: The local path or URL of the input image to be processed
--The GPU serial number used by the device (e.g. GPU: 0 indicates the use of the 0th GPU, GPU: 1,2 indicates the use of the 1st and 2nd GPUs), or the CPU (-- device CPU) can be selected for use
When executing the above Python script, the default seal text recognition production line configuration file is loaded. If you need to customize the configuration file, you can execute the following command to obtain it:
After running, the result obtained is:
The visualized image not saved by default. You can customize the save path through --save_path, and then all results will be saved in the specified path.
A few lines of code can complete the fast inference of the production line. Taking the seal text recognition production line as an example:
from paddlex import create_pipeline
pipeline = create_pipeline(pipeline="seal_recognition")
output = pipeline.predict("seal_text_det.png")
for res in output:
res.print()
res.save_to_img("./output/") # Save the results in img
The result obtained is the same as the command line method.
In the above Python script, the following steps were executed:
(1)Instantiate the production line object using create_pipeline: Specific parameter descriptions are as follows:
| Parameter | Description | Type | Default |
|-|-|-|-|
|pipeline| The name of the production line or the path to the production line configuration file. If it is the name of the production line, it must be supported by PaddleX. |str|None|
|device| The device for production line model inference. Supports: "gpu", "cpu". |str|gpu|
|use_hpip| Whether to enable high-performance inference, only available if the production line supports it. |bool|False|
(2)Invoke the predict method of the production line object for inference prediction: The predict method parameter is x, which is used to input data to be predicted, supporting multiple input methods, as shown in the following examples:
| Parameter Type | Parameter Description |
|---|---|
| Python Var | Supports directly passing in Python variables, such as numpy.ndarray representing image data. |
| str | Supports passing in the path of the file to be predicted, such as the local path of an image file: /root/data/img.jpg. |
| str | Supports passing in the URL of the file to be predicted, such as the network URL of an image file: Example. |
| str | Supports passing in a local directory, which should contain files to be predicted, such as the local path: /root/data/. |
| dict | Supports passing in a dictionary type, where the key needs to correspond to a specific task, such as "img" for image classification tasks. The value of the dictionary supports the above types of data, for example: {"img": "/root/data1"}. |
| list | Supports passing in a list, where the list elements need to be of the above types of data, such as [numpy.ndarray, numpy.ndarray], ["/root/data/img1.jpg", "/root/data/img2.jpg"], ["/root/data1", "/root/data2"], [{"img": "/root/data1"}, {"img": "/root/data2/img.jpg"}]. |
(3)Obtain the prediction results by calling the predict method: The predict method is a generator, so prediction results need to be obtained through iteration. The predict method predicts data in batches, so the prediction results are in the form of a list.
(4)Process the prediction results: The prediction result for each sample is of dict type and supports printing or saving to files, with the supported file types depending on the specific pipeline. For example:
| Method | Description | Method Parameters |
|---|---|---|
| save_to_img | Save the results as an img format file | - save_path: str, the path to save the file. When it's a directory, the saved file name will be consistent with the input file type; |
Where save_to_img can save visualization results (including OCR result images, layout analysis result images).
If you have a configuration file, you can customize the configurations of the seal recognition pipeline by simply modifying the pipeline parameter in the create_pipeline method to the path of the pipeline configuration file.
For example, if your configuration file is saved in / my_path/seal_recognition.yaml , Then only need to execute:
from paddlex import create_pipeline
pipeline = create_pipeline(pipeline="./my_path/seal_recognition.yaml")
output = pipeline.predict("seal_text_det.png")
for res in output:
res.print() ## 打印预测的结构化输出
res.save_to_img("./output/") ## 保存可视化结果
If the production line can meet your requirements for inference speed and accuracy, you can directly develop integration/deployment.
If you need to directly apply the production line to your Python project, you can refer to the example code in 2.2.2 Python scripting.
In addition, PaddleX also offers three other deployment methods, detailed as follows:
🚀 ** High performance deployment: In actual production environments, many applications have strict standards for the performance indicators of deployment strategies, especially response speed, to ensure efficient system operation and smooth user experience. To this end, PaddleX provides a high-performance inference plugin aimed at deep performance optimization of model inference and pre-processing, achieving significant acceleration of end-to-end processes. For a detailed high-performance deployment process, please refer to the PaddleX High Performance Deployment Guide.
☁️ ** Service deployment * *: Service deployment is a common form of deployment in actual production environments. By encapsulating inference functions as services, clients can access these services through network requests to obtain inference results. PaddleX supports users to achieve service-oriented deployment of production lines at low cost. For detailed service-oriented deployment processes, please refer to the PaddleX Service Deployment Guide (../../../ipeline_deploy/service_deploy. md).
Here are API references and examples of calling multilingual services:
If the default model weights provided by the seal text recognition production line are not satisfactory in terms of accuracy or speed in your scenario, you can try using your own specific domain or application scenario data to further fine tune the existing model to improve the recognition performance of the seal text recognition production line in your scenario.
Due to the fact that the seal text recognition production line consists of three modules, the performance of the model production line may not be as expected due to any of these modules.
You can analyze images with poor recognition performance and refer to the following rules for analysis and model fine-tuning:
If seal texts are undetected (i.e., text miss detection), the text detection model may be insufficient. You need to refer to the Customization section in the Text Detection Module Development Tutorial and use your private dataset to fine-tune the text detection model.
If many detected texts contain recognition errors (i.e., the recognized text content does not match the actual text content), the text recognition model requires further improvement. You need to refer to the Customization section.
After completing fine-tuning training using a private dataset, you can obtain a local model weight file.
If you need to use the fine tuned model weights, simply modify the production line configuration file and replace the local path of the fine tuned model weights with the corresponding position in the production line configuration file
......
Pipeline:
layout_model: RT-DETR-H_layout_3cls #can be modified to the local path of the fine tuned model
text_det_model: PP-OCRv4_server_seal_det #can be modified to the local path of the fine tuned model
text_rec_model: PP-OCRv4_server_rec #can be modified to the local path of the fine tuned model
layout_batch_size: 1
text_rec_batch_size: 1
device: "gpu:0"
......
Subsequently, refer to the command line or Python script in the local experience to load the modified production line configuration file.
PaddleX supports various mainstream hardware devices such as Nvidia GPU, Kunlun Core XPU, Ascend NPU, and Cambrian MLU, and can seamlessly switch between different hardware devices by simply modifying the --device parameter.
For example, if you use Nvidia GPU for inference on a seal text recognition production line, the Python command you use is:
paddlex --pipeline seal_recognition --input seal_text_det.png --device gpu:0 --save_path output
At this point, if you want to switch the hardware to Ascend NPU, simply modify the --device in the Python command to NPU:
paddlex --pipeline seal_recognition --input seal_text_det.png --device npu:0 --save_path output
If you want to use the seal text recognition production line on a wider range of hardware, please refer to the PaddleX Multi Hardware Usage Guide。