[go: up one dir, main page]

CN118799890B - A method for express information extraction and distributed representation based on deep learning - Google Patents

A method for express information extraction and distributed representation based on deep learning Download PDF

Info

Publication number
CN118799890B
CN118799890B CN202411280328.3A CN202411280328A CN118799890B CN 118799890 B CN118799890 B CN 118799890B CN 202411280328 A CN202411280328 A CN 202411280328A CN 118799890 B CN118799890 B CN 118799890B
Authority
CN
China
Prior art keywords
information
express
digital
text
address
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202411280328.3A
Other languages
Chinese (zh)
Other versions
CN118799890A (en
Inventor
杨洋
李经纬
李志洪
崔凯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mopai Space Tianjin Co ltd
Original Assignee
Mopai Space Tianjin Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mopai Space Tianjin Co ltd filed Critical Mopai Space Tianjin Co ltd
Priority to CN202411280328.3A priority Critical patent/CN118799890B/en
Publication of CN118799890A publication Critical patent/CN118799890A/en
Application granted granted Critical
Publication of CN118799890B publication Critical patent/CN118799890B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/191Design or setup of recognition systems or techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
    • G06V30/19173Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/08Logistics, e.g. warehousing, loading or distribution; Inventory or stock management
    • G06Q10/083Shipping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/18Extraction of features or characteristics of the image
    • G06V30/1801Detecting partial patterns, e.g. edges or contours, or configurations, e.g. loops, corners, strokes or intersections
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/19Recognition using electronic means
    • G06V30/19007Matching; Proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • G06V30/42Document-oriented image-based pattern recognition based on the type of document

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Business, Economics & Management (AREA)
  • Molecular Biology (AREA)
  • Economics (AREA)
  • Medical Informatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Marketing (AREA)
  • Human Resources & Organizations (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Entrepreneurship & Innovation (AREA)
  • General Business, Economics & Management (AREA)
  • Strategic Management (AREA)
  • Tourism & Hospitality (AREA)
  • Development Economics (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a deep learning-based express information extraction and distributed characterization method, which comprises the following steps: data preparation, namely acquiring an express delivery face list data set required by training through data enhancement and preprocessing; model construction, namely training an express delivery face list area and a text area identification model by using YOLOv target detection network; identifying an express delivery face list, and cutting out an express delivery face list area; acquiring the effective information of the face list and acquiring the coordinates of a target text area; paddleOCR identifying texts, screening and identifying text information; text screening, namely classifying and digitizing the identified text information; the distributed representation is used for respectively converting address, telephone and name information into digital codes; express matching, namely matching express characteristic information with characteristic information in a client database by comparing digital codes, and binding corresponding clients. The method improves the efficiency and accuracy of express information processing and delivery, and reduces the error rate of manual operation.

Description

Express information extraction and distributed characterization method based on deep learning
Technical Field
The invention relates to the field of logistics storage, computer vision and natural language processing, in particular to an express information extraction and distributed characterization method based on deep learning.
Background
With the rapid development of electronic commerce, the business volume of the express industry is rapidly increased, and how to efficiently and accurately manage and distribute express information becomes a problem to be solved urgently. The traditional express information processing method generally depends on manual operation, and is low in efficiency and easy to make mistakes. The rapid development of target detection and Optical Character Recognition (OCR) technology based on deep learning provides a new technical means for solving the problem.
Disclosure of Invention
The invention aims to provide an express information extraction and distributed characterization method based on deep learning, which realizes the efficient identification of an express bill and a target text region by using YOLOv target detection network and PaddleOCR optical character identification technology, digitizes the identified text information and forms uniform digital codes so as to realize the automatic management and efficient distribution of the express information. The traditional express information processing method generally depends on manual operation, and is low in efficiency and easy to make mistakes.
An express information extraction and distributed characterization method based on deep learning, the method comprises the following steps:
S1, acquiring an express delivery face list data set required by training, enriching the training set by a data enhancement method, preprocessing the data set and finishing data labeling;
S2, training a network aiming at express bill area identification by using YOLOv target detection network, and predicting the bill position and rotation angle; carrying out detection model reasoning, carrying out rotation target detection on the face sheet according to the detected position and angle, outputting a face sheet image and a face sheet rotation angle, and righting the cut face sheet in a single direction to obtain a complete face sheet area;
S3, training a model for detecting the position of the target text by using YOLOv target detection network, and predicting the position of the target text; performing detection model reasoning, and calculating text region coordinates according to the detected position to obtain a YOLO target information frame;
S4, identifying express bill information by using PaddleOCR optical character identification technology to obtain PaddleOCR text frames, and calculating PaddleOCR text frame coordinates and YOLO target information frame coordinates on the basis of YOLOv target identification results and PaddleOCR optical character identification technology text identification results;
S5, establishing an information digital distributed characterization method to classify the obtained text information, and storing the classified text information in an Excel table;
S6, establishing a complete customer database according to the information digital distributed characterization method, wherein the characteristic information of each customer in the database comprises a name, an address and a telephone number and is stored in a digital mode;
S7, according to an information digital distributed characterization method, the identified express characteristic information containing characters such as literal, name, address, telephone number and the like is converted into digital characteristic information, and then the express characteristic information is coded into a series of ordered codes; when the characteristic information of a certain express delivery face list is not identified, the information is marked by a single byte character 0 through coding;
S8, the express characteristic information obtained according to the information digital distributed characterization method is a string of digital codes which are arranged according to a set sequence, the digital codes are compared with the digital codes corresponding to the customer characteristic information in the database, the digital address codes, the digital telephone codes and the digital name codes are sequentially matched, and if the digital address codes, the digital telephone codes and the digital name codes can be matched correctly, the express is bound to the customer; if the matching is not correct, the information is input into a customer database.
As a further technical scheme of the invention, the information digital distributed characterization method comprises the following steps:
a) Identifying an address, and converting each section of address (including communities, buildings and family numbers) in the literal address into a digital address, wherein the address is an 8-bit digital string;
b) Identifying a telephone number of the customer, the telephone number being an 11-digit string;
c) And identifying the customer name, wherein the customer name is a 12-bit number string, and one Chinese character corresponds to 4 digits according to the comparison of the four-corner number dictionary to form a code string.
As a further technical scheme of the invention, the data labeling of the data set in the S1 comprises labeling an express delivery face sheet area and an effective information area in the face sheet, wherein the express delivery face sheet area data set collects multiple groups of express delivery with face sheets shot at different angles, RGB images are converted into 640 multiplied by 480 resolution, the effective information data set in the face sheet collects the cut express delivery face sheet, gray images are converted into 640 multiplied by 480 resolution, and data enhancement is carried out by using a data enhancement method, which comprises rotation, cutting, overturning and noise addition; marking the express delivery face list area is to use Labelme software to carry out rectangular label marking on the express delivery face list area in RGB images of the express delivery face list shot at different angles, and the label storage format is selected from the YOLO format; the effective information area in the labeling menu is to use Labelme software to label the number of the express menu and the recipient information in the gray level image of the cut express menu by rectangular labels, and the label storage format is selected from the Yolo format.
As a further technical scheme of the invention, a YOLOv target detection network is used for training in the S2 to obtain a model, and the model can effectively detect and cut and extract overlapped surface sheets, multi-surface sheets, fuzzy surface sheets, surface sheets with incomplete information and fold surface sheets; the method comprises the steps of firstly rotating a face sheet angle, carrying out preliminary correction by adopting affine transformation of OpenCV, then judging the face sheet orientation by using an orientation classification model, carrying out final correction on the face sheet in the reverse direction and the left and right sides, and finally outputting the face sheet which is a forward angle with characters upwards; the method comprises the steps that the face sheet information extraction is based on YOLOv rotation target detection, the face sheet is subjected to rotation target detection, a complete face sheet area is cut according to a rotation rectangular frame predicted by a model, and other non-face sheet areas are not included; training the YOLOv target detection network trained in the step S3, and dividing a training set, a verification set and a test set by a sample according to the ratio of 3:1:1 to obtain a model for detecting the position of a target text; the target position detection is to select effective text area information in the picture, including the name, address, telephone number and express bill number of the customer in the bill.
As a further technical scheme of the present invention, the PaddleOCR optical character recognition technology used in S4 is an optical character recognition technology based on a convolutional neural network, characters are converted into image information through an optical input mode, the optical input mode includes scanning, the image information is converted into editable computer text by utilizing the character recognition technology, and text information is screened by calculating the coincidence degree of PaddleOCR frame coordinates and YOLO frame coordinates; and calculating the coincidence degree of PaddleOCR text boxes and YOLOv text boxes, and reserving PaddleOCR text boxes with the coincidence degree reaching 0.5, so as to achieve the aim of screening text information.
As a further technical solution of the present invention, the preset sequence of the information digital distributed characterization method in S5 is as follows: sequentially comprises communities, buildings, family numbers, telephone numbers and names from left to right; a code is formed by 31-bit single-byte characters to represent specific characteristic information, the code consists of 5 sections, namely a community code section, a building code section, a household code section, a telephone code section and a name code section, wherein the community code section consists of 2-bit single-byte characters, the building code section consists of 2-bit single-byte characters, the household code section consists of 4-bit single-byte characters, the telephone code section consists of 11-bit single-byte characters, and the name code section consists of 12-bit single-byte characters.
As a further technical solution of the present invention, the address code segment of the information digital distributed characterization method in S5 includes: a tenth address code segment, a hundredth address code segment, a thousandth address code segment and a thousandth address code segment, wherein the tenth address code segment is a first second bit character of the address code segment, the hundredth address code segment is a third fourth bit character of the address code segment, the thousandth address code segment is a fifth sixth bit character of the segment code segment, and the thousandth address code segment is a seventh eighth bit character of the address code segment.
As a further technical scheme of the invention, the complete customer database in S6 is the digital characteristic information of the community customers, the community customers enter personal information, and the community customers' characteristic information is converted into digital coded ordered codes according to an information digital distributed characterization method to obtain the characteristic digital information of the community customers.
As a further technical scheme of the invention, the express characteristic information in S7 is that when express delivery is carried out, an express bill is shot through a camera, information on the express bill is acquired by using an optical character recognition technology based on deep learning, the identified express bill information is converted into digital characteristic information according to an information digitizing method, digital information distribution carries out a principle of step-by-step distribution from front to back, a series of ordered codes are formed, and the express bill information is not existed due to external factors, and the non-existent information is converted into digital 0 when the express bill information is converted according to the information digitizing method, wherein the external factors comprise blurring, overlapping, stains and folds of the express bill.
As a further technical scheme of the invention, the express characteristic information obtained by the information digital distributed characterization method in the S8 is a string of digital codes arranged according to a set sequence, and if the express characteristic information and the client characteristic information are matched correctly, the express is bound to corresponding clients; if the matching is not right, judging that the client information of the client to which the express belongs is not in the client database, and storing the express characteristic information in the client database as the client characteristic information in the database.
Compared with the prior art, the invention has the beneficial effects that:
According to the invention, by combining YOLOv target detection network and PaddleOCR optical character recognition technology, the high-efficiency recognition of the express bill and the target text area is realized, and the recognized text information is converted into uniform digital codes by an information digitizing method, so that the automatic management and high-efficiency distribution of the express information are realized, the efficiency and accuracy of the express information processing are improved, and the error rate of manual operation is reduced.
Drawings
Fig. 1 is a flowchart of a method for extracting express information and performing distributed characterization according to an embodiment of the present invention.
Fig. 2 is a schematic structural diagram of a method for digitally representing information in an embodiment of the present invention.
Detailed Description
The technical scheme of the invention is further described in detail below with reference to the specific embodiments.
Referring to fig. 1 to 2, an embodiment of the present invention provides a deep learning-based express information extraction and distributed characterization method thereof, the method includes the following steps:
S1, acquiring an express delivery face list data set required by training, enriching the training set by a data enhancement method, preprocessing the data set and finishing data labeling;
the method comprises the steps of obtaining a large number of express delivery face list data sets required by training from different express delivery companies, wherein data samples comprise face list pictures shot under various angles and illumination conditions, and the face list data sets comprise multiple groups of express delivery with face lists shot at different angles. The RGB image is converted to 640 x 480 resolution size and data enhancement operations such as rotation, cropping, flipping, and noise addition are performed. Preprocessing a dataset, marking rectangular labels on express delivery face single areas shot at different angles by using Labelme software, selecting a Yolo format as a label storage format, and marking the marked express delivery face single areas with a sample size of about 4800. And processing the cut express bill, converting the cut express bill into a gray image with the resolution of 640 multiplied by 480, marking the effective information (such as express bill number and recipient information) in the express bill by using Labelme software with rectangular labels, selecting a Yolo format for a label storage format, and marking the marked express bill with the sample size of about 4800.
The data labeling of the data set in the S1 comprises labeling an express delivery face list area and an effective information area in the face list, wherein the labeling of the express delivery face list area is to use Labelme software to label rectangular labels of the express delivery face list area in RGB images of the express delivery face list shot at different angles, and a label storage format is selected from a YOLO format; the effective information area in the labeling menu is to use Labelme software to label the number of the express menu and the recipient information in the gray level image of the cut express menu by rectangular labels, and the label storage format is selected from the Yolo format.
S2, training a network aiming at express bill area identification by using YOLOv target detection network, and predicting the bill position and rotation angle; carrying out detection model reasoning, carrying out rotation target detection on the face sheet according to the detected position and angle, outputting a face sheet image and a face sheet rotation angle, and righting the cut face sheet in a single direction to obtain a complete face sheet area;
Training the S2 by using YOLOv target detection network to obtain a model, wherein the model can effectively detect and cut and extract overlapped surface sheets, multi-surface sheets, fuzzy surface sheets and surface sheets with incomplete information; the method comprises the steps of firstly rotating a face sheet angle, carrying out preliminary correction by adopting affine transformation of OpenCV, then judging the face sheet orientation by using an orientation classification model, carrying out final correction on the face sheet in the reverse direction and the left and right sides, and finally outputting the face sheet which is a forward angle with characters upwards; the method comprises the steps that the face sheet information extraction is based on YOLOv rotation target detection, the face sheet is subjected to rotation target detection, a complete face sheet area is cut according to a rotation rectangular frame predicted by a model, and other non-face sheet areas are not included; training the YOLOv target detection network trained in the step S3, dividing a training set, a verification set and a test set according to the ratio of 3:1:1 by using about 1440 samples for training the model, about 480 samples for verifying the model, about 480 samples for testing the model, and training the marked data set to obtain a model for detecting the position of the target text; the target position detection is to select effective text area information in the picture, including the name, address, telephone number and express bill number of the customer in the bill. The model calculates the text region coordinates to realize effective text region information frame selection.
S3, training a model for detecting the position of the target text by using YOLOv target detection network, and predicting the position of the target text; performing detection model reasoning, and calculating text region coordinates according to the detected position to obtain a YOLO target information frame;
Inputting an original express image into a trained model for detecting the position of a target text, carrying out detection model reasoning on a test set, outputting a face list image and a rotation angle, carrying out preliminary correction by using affine transformation of OpenCV, carrying out final correction on the face list in the reverse direction, the left and right directions, obtaining a forward face list image with characters upwards, identifying and cutting out an express face list area, and obtaining an independent express face list image.
S4, identifying express bill information by using PaddleOCR optical character identification technology to obtain PaddleOCR text frames, and calculating PaddleOCR text frame coordinates and YOLO target information frame coordinates on the basis of YOLOv target identification results and PaddleOCR optical character identification technology text identification results;
The PaddleOCR optical character recognition technology used in the S4 is an optical character recognition technology based on a convolutional neural network, characters are converted into image information through an optical input mode, the optical input mode comprises scanning, the image information is converted into editable computer texts through the character recognition technology, and text information is screened through calculating the coincidence degree of PaddleOCR frame coordinates and YOLO frame coordinates; and calculating the coincidence degree of PaddleOCR text boxes and YOLOv text boxes, and reserving PaddleOCR text boxes with the coincidence degree reaching 0.5, so as to achieve the aim of screening text information.
S5, establishing an information digital distributed characterization method to classify the obtained text information, and storing the classified text information in an Excel table;
the information digital distributed characterization method comprises the following steps:
a) Identifying an address, and converting each section of address (including communities, buildings and family numbers) in the literal address into a digital address, wherein the address is an 8-bit digital string;
b) Identifying a telephone number of the customer, the telephone number being an 11-digit string;
c) And identifying the customer name, wherein the customer name is a 12-bit number string, and one Chinese character corresponds to 4 digits according to the comparison of the four-corner number dictionary to form a code string.
The preset sequence of the information digital distributed characterization method in the S5 is as follows: sequentially comprises communities, buildings, family numbers, telephone numbers and names from left to right; a code is formed by 31-bit single-byte characters to represent specific characteristic information, the code consists of 5 sections, namely a community code section, a building code section, a household code section, a telephone code section and a name code section, wherein the community code section consists of 2-bit single-byte characters, the building code section consists of 2-bit single-byte characters, the household code section consists of 4-bit single-byte characters, the telephone code section consists of 11-bit single-byte characters, and the name code section consists of 12-bit single-byte characters.
The address code segment of the information digital distributed characterization method in S5 includes: a tenth address code segment, a hundredth address code segment, a thousandth address code segment and a thousandth address code segment, wherein the tenth address code segment is a first second bit character of the address code segment, the hundredth address code segment is a third fourth bit character of the address code segment, the thousandth address code segment is a fifth sixth bit character of the segment code segment, and the thousandth address code segment is a seventh eighth bit character of the address code segment.
S6, establishing a complete customer database according to the information digital distributed characterization method, wherein the characteristic information of each customer in the database comprises a name, an address and a telephone number and is stored in a digital mode;
And S6, the complete customer database is the digital characteristic information of the community customer, the community customer inputs personal information, and the community customer characteristic information is converted into a digital coded ordered code according to an information digital distributed characterization method to obtain the characteristic digital information of the community customer.
S7, according to an information digital distributed characterization method, the identified express characteristic information containing characters such as literal, name, address, telephone number and the like is converted into digital characteristic information, and then the express characteristic information is coded into a series of ordered codes; when the characteristic information of a certain express delivery face list is not identified, the information is marked by a single byte character 0 through coding;
The express characteristic information in the S7 is that when express delivery is carried out, an express bill is shot through a camera, information on the express bill is acquired by using an optical character recognition technology based on deep learning, the identified express bill information is converted into digital characteristic information according to an information digitizing method, a principle of gradual distribution from front to back is carried out on digital information distribution, a string of ordered codes is formed, the information of the express bill is not existed due to external factors, the information which is not existed is converted into digital 0 when the information is converted according to the information digitizing method, and the external factors comprise blurring, overlapping, stains and folds of the express bill.
S8, the express characteristic information obtained according to the information digital distributed characterization method is a string of digital codes which are arranged according to a set sequence, the digital codes are compared with the digital codes corresponding to the customer characteristic information in the database, the digital address codes, the digital telephone codes and the digital name codes are sequentially matched, and if the digital address codes, the digital telephone codes and the digital name codes can be matched correctly, the express is bound to the customer; if the matching is not correct, the information is input into a customer database.
The express characteristic information obtained by the information digital distributed characterization method in the S8 is a string of digital codes arranged according to a set sequence, and if the express characteristic information and the client characteristic information are matched correctly, the express is bound to corresponding clients; if the matching is not right, judging that the client information of the client to which the express belongs is not in the client database, and storing the express characteristic information in the client database as the client characteristic information in the database.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The foregoing description is only of the preferred embodiments of the present invention, and is not intended to limit the scope of the invention, but rather is intended to cover any equivalents of the structures or equivalent processes disclosed herein or in the alternative, which may be employed directly or indirectly in other related arts.

Claims (10)

1. The express information extraction and distributed characterization method based on deep learning is characterized by comprising the following steps of:
S1, acquiring an express delivery face list data set required by training, enriching the training set by a data enhancement method, preprocessing the data set and finishing data labeling;
S2, training a network aiming at express bill area identification by using YOLOv target detection network, and predicting the bill position and rotation angle; carrying out detection model reasoning, carrying out rotation target detection on the face sheet according to the detected position and angle, outputting a face sheet image and a face sheet rotation angle, and righting the cut face sheet in a single direction to obtain a complete face sheet area;
S3, training a model for detecting the position of the target text by using YOLOv target detection network, and predicting the position of the target text; performing detection model reasoning, and calculating text region coordinates according to the detected position to obtain a YOLO target information frame;
S4, identifying express bill information by using PaddleOCR optical character identification technology to obtain PaddleOCR text frames, and calculating PaddleOCR text frame coordinates and YOLO target information frame coordinates on the basis of YOLOv target identification results and PaddleOCR optical character identification technology text identification results;
S5, establishing an information digital distributed characterization method to classify the obtained text information, and storing the classified text information in an Excel table;
S6, establishing a complete customer database according to the information digital distributed characterization method, wherein the characteristic information of each customer in the database comprises a name, an address and a telephone number and is stored in a digital mode;
S7, according to an information digital distributed characterization method, the identified express characteristic information containing the literal, name, address and telephone number is converted into digital characteristic information, and then the express characteristic information is encoded into a string of ordered codes; when the characteristic information of a certain express delivery face list is not identified, the information is marked by a single byte character 0 through coding;
S8, the express characteristic information obtained according to the information digital distributed characterization method is a string of digital codes which are arranged according to a set sequence, the digital codes are compared with the digital codes corresponding to the customer characteristic information in the database, the digital address codes, the digital telephone codes and the digital name codes are sequentially matched, and if the digital address codes, the digital telephone codes and the digital name codes can be matched correctly, the express is bound to the customer; if the matching is not correct, the information is input into a customer database.
2. The deep learning-based express information extraction and distributed characterization method according to claim 1, wherein the information digital distributed characterization method comprises the following steps:
a) Identifying an address, converting each segment of address in the literal address into a digital address, wherein the address is an 8-bit digital string;
b) Identifying a telephone number of the customer, the telephone number being an 11-digit string;
c) And identifying the customer name, wherein the customer name is a 12-bit number string, and one Chinese character corresponds to 4 digits according to the comparison of the four-corner number dictionary to form a code string.
3. The method for extracting and distributing type characterization of express information based on deep learning according to claim 1, wherein the data labeling of the data set in S1 includes labeling two kinds of express bill area and effective information area in the express bill, the express bill area data set collects multiple groups of express with bills shot at different angles, RGB images are converted into 640 x 480 resolution, the effective information data set in the express bill collects the cut express bill, gray images are converted into 640 x 480 resolution, and data enhancement is performed by using a data enhancement method, wherein the data enhancement method includes rotation, cutting, overturning and noise addition; marking the express delivery face list area is to use Labelme software to carry out rectangular label marking on the express delivery face list area in RGB images of the express delivery face list shot at different angles, and the label storage format is selected from the YOLO format; the effective information area in the labeling menu is to use Labelme software to label the number of the express menu and the recipient information in the gray level image of the cut express menu by rectangular labels, and the label storage format is selected from the Yolo format.
4. The method for extracting express information and carrying out distributed characterization on the basis of deep learning according to claim 1, wherein a YOLOv target detection network is used for training in the step S2 to obtain a model, and the model is used for effectively detecting and cutting and extracting overlapped surface sheets, multi-surface sheets, fuzzy surface sheets and fold surface sheets with incomplete information; the method comprises the steps of firstly rotating a face sheet angle, carrying out preliminary correction by adopting affine transformation of OpenCV, then judging the face sheet orientation by using an orientation classification model, carrying out final correction on the face sheet in the reverse direction and the left and right sides, and finally outputting the face sheet which is a forward angle with characters upwards; the method comprises the steps that the face sheet information extraction is based on YOLOv rotation target detection, the face sheet is subjected to rotation target detection, a complete face sheet area is cut according to a rotation rectangular frame predicted by a model, and other non-face sheet areas are not included; training the YOLOv target detection network trained in the step S3, and dividing a training set, a verification set and a test set by a sample according to the ratio of 3:1:1 to obtain a model for detecting the position of a target text; the target position detection is to select effective text area information in the picture, including the name, address, telephone number and express bill number of the customer in the bill.
5. The method for extracting and distributing the express information based on the deep learning according to claim 1, wherein PaddleOCR optical character recognition technology used in the step S4 is optical character recognition technology based on convolutional neural network, characters are converted into image information through an optical input mode, the optical input mode comprises scanning, the image information is converted into editable computer text through the character recognition technology, and text information is screened through calculating the coincidence degree of PaddleOCR frame coordinates and YOLO frame coordinates; and calculating the coincidence degree of PaddleOCR text boxes and YOLOv text boxes, and reserving PaddleOCR text boxes with the coincidence degree reaching 0.5, so as to achieve the aim of screening text information.
6. The deep learning-based express information extraction and distributed characterization method according to claim 2, wherein the preset sequence of the information digital distributed characterization method in S5 is as follows: sequentially comprises communities, buildings, family numbers, telephone numbers and names from left to right; a code is formed by 31-bit single-byte characters to represent specific characteristic information, the code consists of 5 sections, namely a community code section, a building code section, a household code section, a telephone code section and a name code section, wherein the community code section consists of 2-bit single-byte characters, the building code section consists of 2-bit single-byte characters, the household code section consists of 4-bit single-byte characters, the telephone code section consists of 11-bit single-byte characters, and the name code section consists of 12-bit single-byte characters.
7. The deep learning-based express delivery information extraction and distributed characterization method according to claim 6, wherein the address code section of the information digital distributed characterization method in S5 includes: a tenth address code segment, a hundredth address code segment, a thousandth address code segment and a thousandth address code segment, wherein the tenth address code segment is a first second bit character of the address code segment, the hundredth address code segment is a third fourth bit character of the address code segment, the thousandth address code segment is a fifth sixth bit character of the segment code segment, and the thousandth address code segment is a seventh eighth bit character of the address code segment.
8. The deep learning-based express information extraction and distributed characterization method according to claim 1, wherein the complete customer database in S6 is digital characteristic information of the community customers, the community customers enter principal information, and the community customers' characteristic information is converted into digital coded ordered codes according to the information digital distributed characterization method, so as to obtain characteristic digital information of the community customers.
9. The method for extracting and distributing the express delivery information based on the deep learning according to claim 1, wherein the express delivery characteristic information in the S7 is that when the express delivery is delivered, the express delivery bill is shot by a camera, the information on the express delivery bill is acquired by using an optical character recognition technology based on the deep learning, the recognized express delivery bill information is converted into digital characteristic information according to an information digitizing method, a principle of gradual distribution from front to back is carried out by digital information distribution, a series of ordered codes are formed, the information of the express delivery bill is not existed due to external factors, the information which does not exist is converted into digital 0 when the information is converted according to the information digitizing method, and the external factors comprise blurring, overlapping, stains and folds of the express delivery bill.
10. The method for extracting and distributing the express information based on the deep learning according to claim 1, wherein the express characteristic information obtained by the information digital distributed characterization method in the step S8 is a string of digital codes arranged according to a set sequence, and if the express characteristic information and the client characteristic information are matched correctly, the express is bound to the corresponding client; if the matching is not right, judging that the client information of the client to which the express belongs is not in the client database, and storing the express characteristic information in the client database as the client characteristic information in the database.
CN202411280328.3A 2024-09-13 2024-09-13 A method for express information extraction and distributed representation based on deep learning Active CN118799890B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202411280328.3A CN118799890B (en) 2024-09-13 2024-09-13 A method for express information extraction and distributed representation based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202411280328.3A CN118799890B (en) 2024-09-13 2024-09-13 A method for express information extraction and distributed representation based on deep learning

Publications (2)

Publication Number Publication Date
CN118799890A CN118799890A (en) 2024-10-18
CN118799890B true CN118799890B (en) 2024-11-12

Family

ID=93035266

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202411280328.3A Active CN118799890B (en) 2024-09-13 2024-09-13 A method for express information extraction and distributed representation based on deep learning

Country Status (1)

Country Link
CN (1) CN118799890B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113919345A (en) * 2021-10-15 2022-01-11 北京同城必应科技有限公司 Intelligent information extraction method for express orders in same city
CN116129416A (en) * 2023-01-31 2023-05-16 广东海洋大学 AI algorithm-based vehicle management system with double systems and double modes

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110377686B (en) * 2019-07-04 2021-09-17 浙江大学 Address information feature extraction method based on deep neural network model
US20230134651A1 (en) * 2021-10-28 2023-05-04 Akporefe Agbamu Synchronized Identity, Document, and Transaction Management
CN118155226A (en) * 2024-03-13 2024-06-07 国网安徽省电力有限公司涡阳县供电公司 Power consumption information input method and system based on image recognition and computer equipment

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113919345A (en) * 2021-10-15 2022-01-11 北京同城必应科技有限公司 Intelligent information extraction method for express orders in same city
CN116129416A (en) * 2023-01-31 2023-05-16 广东海洋大学 AI algorithm-based vehicle management system with double systems and double modes

Also Published As

Publication number Publication date
CN118799890A (en) 2024-10-18

Similar Documents

Publication Publication Date Title
CN1103087C (en) Recognition and Correction Method of Optical Scanning Form
JP5522408B2 (en) Pattern recognition device
JP5105561B2 (en) Method for processing mail pieces using client code associated with digital imprint
CN112508011A (en) OCR (optical character recognition) method and device based on neural network
CN108960223A (en) The method for automatically generating voucher based on bill intelligent recognition
US20140185933A1 (en) Document image compression method and its application in document authentication
US20140268250A1 (en) Systems and methods for receipt-based mobile image capture
CN114550158B (en) Scene character recognition method and system
CN108364037A (en) Method, system and the equipment of Handwritten Chinese Character Recognition
CN113158895A (en) Bill identification method and device, electronic equipment and storage medium
CN113887484B (en) Card type file image identification method and device
CN112149401A (en) Document comparison identification method and system based on ocr
CN112149654B (en) Invoice text information identification method based on deep learning
CN113657377A (en) Structured recognition method for airplane ticket printing data image
CN115116068B (en) Archive intelligent archiving system based on OCR
CN116311292A (en) Document image information extraction method, device, computer equipment and storage medium
CN111213157A (en) Express information input method and system based on intelligent terminal
CN118799890B (en) A method for express information extraction and distributed representation based on deep learning
CN108090728B (en) Express information input method and system based on intelligent terminal
US9679179B2 (en) Method for processing information from a hand-held scanning device
CN116756358A (en) Electronic management method for flight manifest
JP2009146245A (en) Image collation method, image collation apparatus, and image collation program
CN114241485A (en) Information identification method, device, equipment and storage medium of property certificate
JP3090070B2 (en) Form identification method and device
Kumar et al. Optical Character Recognition (OCR) Using Opencv and Python: Implementation and Performance Analysis

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant