CN115497114B - Structured information extraction method for cigarette logistics receiving bill - Google Patents

Structured information extraction method for cigarette logistics receiving bill Download PDF

Info

Publication number
CN115497114B
CN115497114B CN202211442689.4A CN202211442689A CN115497114B CN 115497114 B CN115497114 B CN 115497114B CN 202211442689 A CN202211442689 A CN 202211442689A CN 115497114 B CN115497114 B CN 115497114B
Authority
CN
China
Prior art keywords
picture
value
template
identified
key
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202211442689.4A
Other languages
Chinese (zh)
Other versions
CN115497114A (en
Inventor
曾华
徐伟
刘永海
朱小晓
胡晓峰
李涛
周幸
曾鹏程
李�瑞
廖健
王静雅
付雯
龙涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Aimo Technology Co ltd
China National Tobacco Corp Sichuan Branch
Original Assignee
Shenzhen Aimo Technology Co ltd
China National Tobacco Corp Sichuan Branch
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Aimo Technology Co ltd, China National Tobacco Corp Sichuan Branch filed Critical Shenzhen Aimo Technology Co ltd
Priority to CN202211442689.4A priority Critical patent/CN115497114B/en
Publication of CN115497114A publication Critical patent/CN115497114A/en
Application granted granted Critical
Publication of CN115497114B publication Critical patent/CN115497114B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • G06V30/41Analysis of document content
    • G06V30/414Extracting the geometrical structure, e.g. layout tree; Block segmentation, e.g. bounding boxes for graphics or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/12Detection or correction of errors, e.g. by rescanning the pattern
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • G06V30/41Analysis of document content
    • G06V30/418Document matching, e.g. of document images
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a structured information extraction method of a cigarette logistics receiving bill, which comprises a pre-labeling step and an identification step, wherein the pre-labeling step comprises the following steps: setting template picture standards of the bill, selecting standard template pictures, and labeling key and value on the template pictures, wherein the key is a fixed keyword in the bill, and the value is variable content in the bill; and (3) identification: determining a picture to be identified, matching a key of the picture to be identified with a key of a template picture, setting text boxes except the key in the picture to be identified as value candidate boxes, aligning the picture to be identified with the template picture according to the corresponding relation between the keys, correcting the value candidate boxes in a dislocation mode, and extracting structural information according to the content in the value text boxes of the template picture. The method is flexible and high in applicability, can solve the problem of printing dislocation, and is high in recognition accuracy.

Description

Structured information extraction method for cigarette logistics receiving bill
Technical Field
The invention relates to the field of logistics, in particular to a structured information extraction method of a cigarette logistics receipt.
Background
In the tobacco logistics scene, the receiver needs to confirm the information of the delivery party, and the information of the bill needs to be checked and checked with the information recorded by the system. The manual auditing needs a lot of time and is easy to make mistakes, and an alternative scheme is to automatically extract the structured information (date, number and the like) on the bill through an image recognition algorithm and compare the structured information with the structured information recorded by the system.
At present, two main methods are mainly used for extracting structural information of bills, one is to use related rules such as regular expression matching and the like to carry out post-processing on the result of ocr (optical character recognition) and optical character recognition, the method is flexible but the accuracy is not high, especially the situation of printing dislocation cannot be solved, the other is to use a deep learning method to detect the position of each field and then carry out ocr content recognition on each field, the method accuracy is high, but a large amount of data are required to be collected for labeling and training for each bill, and the method is not flexible and the applicability is not high.
Disclosure of Invention
Based on the above, the invention provides a structured information extraction method for cigarette logistics receiving bills, which is based on structured information extraction of template alignment and dislocation correction, only one template picture needs to be marked for each bill, is flexible and high in applicability, can solve the printing dislocation condition, and is high in recognition accuracy.
The technical scheme of the invention is as follows:
the method for extracting the structured information of the cigarette logistics receiving bill is characterized by comprising the following steps of:
pre-marking: setting template picture standards of the bill, selecting standard template pictures, and labeling key and value on the template pictures, wherein the key is a fixed keyword in the bill, and the value is variable content in the bill;
and (3) identification: determining a picture to be identified, matching a key of the picture to be identified with a key of a template picture, setting text boxes except the key in the picture to be identified as value candidate boxes, aligning the picture to be identified with the template picture according to the corresponding relation between the keys, correcting the value candidate boxes in a dislocation mode, and extracting structural information according to the content in the value text boxes of the template picture.
The thought of the technical scheme is as follows:
a ticket consists of two parts: a fixed key (such as name) and a variable value (such as Zhang San) of the content, each bill accords with a specific typesetting style, the key content is unchanged and the positions are completely aligned, the value content is variable and even the length and the number of lines are variable, but the positions basically only fluctuate near a preset position.
Based on the characteristics of the bills, the design scheme of the invention selects a standard template picture for each bill to carry out key and value marking, matches the key in the picture to be identified with the template key for correlation, carries out template alignment through perspective transformation, and then can extract corresponding structural information near the preset value frame position.
In the pre-marking step, the template picture standard of the bill is flat, non-inclined and printed picture without dislocation.
In the pre-labeling step, the step of labeling the key is as follows:
and performing rectangular frame labeling and text content labeling on the template picture, and setting a compact rectangular frame as a keyword area.
In the pre-labeling step, the labeling value step is as follows:
and marking rectangular frames and field names of fields needing to be identified except for marking keys in the template picture.
The identifying step further comprises the following steps:
all text boxes and text content in the picture to be identified are detected and identified at ocr.
The identifying step further comprises the following steps:
and judging whether the text boxes and text contents in the obtained picture to be identified belong to key text boxes of the template picture or not through keyword matching, if so, associating the keys in the picture to be identified with the keys of the template picture to form a group of key corresponding relations, if not, taking the text boxes and the text contents in the picture to be identified as value candidate boxes, and if no group of key corresponding relations exist, failing to identify the current picture.
The identifying step further comprises the following steps:
aligning the picture to be identified with the template picture according to the corresponding relation between the keys, extracting 4 vertexes of the text box, and establishing the corresponding relation of 4 groups of vertexes according to the corresponding relation between each group of picture keys to be identified and the template picture key; when N groups of key corresponding relations exist, a corresponding relation of N4 groups of vertex coordinates is established, a homography matrix is calculated according to the corresponding relation between the vertex coordinates, and the picture to be identified and the template picture are aligned through perspective transformation.
The identifying step further comprises the following steps:
and translating all the value candidate frames at least once according to a preset rule, calculating the alignment degree of the value candidate frames and the template value frame of each displacement, selecting the displacement with the highest alignment degree as final dislocation displacement, and carrying out dislocation correction on all the value candidate frames according to the displacement to obtain the dislocation corrected value frame of the picture to be identified and the content thereof.
The identifying step further comprises the following steps:
extracting structural information near the template picture value text boxes, finding one template picture value text box with the largest overlapping area for each value candidate box, if the overlapping degree of the two template picture value text boxes is larger than a set threshold value, associating the value candidate box with the corresponding template picture value text box, otherwise, neglecting the value candidate box;
after the association of all the value candidate boxes is completed, the value candidate box associated with each template picture value text box is the content corresponding to the template picture value field, the text content is connected in series to obtain the extracted structured information of the field, and if the template picture value text box is not associated with the value candidate box, the field cannot be identified.
The beneficial effects of the invention are as follows:
1. structured information extraction based on template alignment and dislocation correction only needs to mark one template picture for each bill, and the method is flexible and high in applicability, meanwhile, the problem of printing dislocation can be solved, and the recognition accuracy is high;
2. selecting a standard template picture for each bill to carry out key and value labeling, matching the key in the picture to be identified with the template key to carry out association, and carrying out template alignment through perspective transformation, so that corresponding structured information can be extracted near the preset value frame position;
3. in order to solve the interference caused by the value printing dislocation, after the key template is aligned, dislocation correction is carried out according to the alignment degree of the to-be-identified picture value and the template value, so that the accuracy of the extraction of the structured information is greatly improved.
Detailed Description
The following describes embodiments of the present invention in detail.
Examples:
a method for extracting structured information of a cigarette logistics receiving bill comprises the following steps:
pre-marking: selecting a standard template picture, and marking the template picture by key and value;
and (3) identification: ocr detection and recognition are carried out on the picture to be recognized, a key text box is matched with a template key from ocr, the rest text boxes are used as value candidate boxes, the picture to be recognized and the template picture are aligned according to the corresponding relation between the keys, dislocation correction is carried out on the value candidate boxes, and structural information is extracted near the value boxes preset by the template.
The idea of the above embodiment is as follows:
a ticket consists of two parts: a fixed key (such as name) and a variable value (such as Zhang San) of the content, each bill accords with a specific typesetting style, the key content is unchanged and the positions are completely aligned, the value content is variable and even the length and the number of lines are variable, but the positions basically only fluctuate near a preset position.
Based on the characteristics of the bills, the design scheme of the invention selects a standard template picture for each bill to carry out key and value marking, matches the key in the picture to be identified with the template key for correlation, carries out template alignment through perspective transformation, and then can extract corresponding structural information near the preset value frame position.
In the pre-marking step, the selected standard template picture is a flat picture which does not have inclination and is printed without dislocation.
In the pre-labeling step, keys are fixed keywords in the bill, rectangular frame labeling and text content labeling are carried out on the keys, compact rectangular frames only comprise keyword areas, value is variable content in the bill, all the contents are not labeled, only fields needing to be identified are labeled, and rectangular frame labeling and field name labeling are carried out on the keys.
In the identification step, all text boxes and text contents in the picture are detected and identified through ocr.
In the identification step, judging whether all ocr text boxes belong to key text boxes through keyword matching, if so, associating the text boxes with corresponding template keys to form a group of key corresponding relations, if not, taking the text boxes as value candidate boxes, and if any group of key corresponding relations do not exist, failing to identify the current picture.
In the identification step, the picture to be identified is aligned with the template picture according to the corresponding relation between the keys, 4 vertexes of the text box are extracted, the corresponding relation of 4 groups of vertexes is established according to the corresponding relation between each group of key text box and the template key, if N groups of key corresponding relations exist, the corresponding relation of N multiplied by 4 groups of vertex coordinates can be established, a homography matrix is calculated according to the corresponding relation between the vertex coordinates, the picture to be identified is aligned with the template picture through perspective transformation, and the positions of the value candidate boxes are also transformed onto the aligned picture through the same transformation matrix.
In the identification step, dislocation correction is carried out on the value candidate frame, after template alignment is carried out, if printing dislocation does not exist, the value candidate frame falls inside the template value frame, if printing dislocation exists, the value candidate frame has offset, and alignment with the template value frame falls on the edge of the template value frame or outside the template value frame.
And translating all the value candidate frames in a certain range for multiple times, translating all the directions up, down, left and right in a range of taking the original position of the value candidate frame as a center and taking the radius of the x and y directions as 50 pixels, wherein the total translation times are (50/10 x 2+1) 2=121 in units of 10 pixels, calculating the alignment degree of the value candidate frame and the template value frame of each displacement, selecting the displacement with the highest alignment degree as the final dislocation displacement, and carrying out dislocation correction on all the value candidate frames according to the displacement.
Marking the ith template value frame asNumber of framesN, j candidate value box +.>The number of frames is m, the intersection is an intersection operation, the area is an area operation, the bin is a binarization function (1 if the condition is satisfied, or 0 if the condition is satisfied), and the calculation formula of the alignment_ratio is as follows:
in the identifying step, structural information is extracted near a value frame preset by a template, one template value frame with the largest overlapping area is found for each value candidate frame, if the overlapping degree of the two value candidate frames is larger than a set threshold value, the value candidate frame is associated with the template value frame, otherwise, the value candidate frame is ignored, after all the value candidate frames are associated, all the value candidate frames associated with each template value frame are the content corresponding to the template value field, the text content of each value candidate frame is connected in series to obtain the structural information of the extracted field, if the template value frame is not associated with any value candidate frame, the field cannot be identified, the template value frame is recorded as tv, the value candidate frame is v, and the overlapping degree overlap-ratio is calculated as follows:
the method is flexible and high in applicability, can solve the problem of printing dislocation, and is high in recognition accuracy. And selecting a standard template picture for each bill to carry out key and value labeling, matching the key in the picture to be identified with the template key to carry out association, and carrying out template alignment through perspective transformation, so that corresponding structured information can be extracted near the preset value frame position. In order to solve the interference caused by the value printing dislocation, after the key template is aligned, dislocation correction is carried out according to the alignment degree of the to-be-identified picture value and the template value, so that the accuracy of the extraction of the structured information is greatly improved.
The steps in the pre-labeling stage are as follows:
1. a standard template picture is selected.
The template picture should be as flat as possible, without tilting and without misplacement of the printing.
2. And (5) performing key and value labeling on the template picture.
The key is a fixed keyword in the bill, and rectangular frame marking (a compact rectangular frame only comprises a keyword area) and text content marking are carried out on the key.
The value is variable content in the bill, all the content is not required to be marked, only the field to be identified is marked, and rectangular frame marking (a wide rectangular frame can cover all the position range where the field content appears) and field name marking (the key and the value are not in one-to-one correspondence, and some value have no key at all, so the field corresponding to the value is directly specified).
After combing in the above examples, the following is described in further detail.
The steps in the identification phase are as follows:
1. and ocr, detecting and identifying the picture to be identified.
All text boxes and text content in the picture are detected and identified by ocr.
2. From the ocr result, the matching key text box is associated with the template key, and the rest text boxes are used as value candidate boxes.
Judging whether all ocr text boxes belong to key text boxes through keyword matching, if so, associating the text boxes with corresponding template keys to form a group of key corresponding relations, and if not, taking the text boxes as value candidate boxes.
If any group of key corresponding relations do not exist, the current picture cannot be identified, otherwise, the step 3 is continuously executed.
3. And aligning the picture to be identified with the template picture according to the corresponding relation between the keys.
4 vertexes of the text box are extracted, a corresponding relation of 4 groups of vertexes can be established according to the corresponding relation of each group of key text boxes and the template keys, and if N groups of key corresponding relations exist, a corresponding relation of N multiplied by 4 groups of vertex coordinates can be established. According to the corresponding relation between the vertex coordinates, a homography matrix is calculated, the picture to be identified and the template picture are aligned through perspective transformation, and the position of the value candidate frame is also transformed to the aligned picture through the same transformation matrix.
The related description and principle of homography matrix and perspective transformation are the prior art and are not repeated
4. And performing dislocation correction on the value candidate frame.
After the template alignment is performed, if there is no print misalignment, the value candidate frame should fall as far as possible inside the template value frame, whereas if there is print misalignment, there is a certain degree of offset of the value candidate frame, the alignment with the template value frame is not good (falls on or outside the edge of the template value frame).
And translating all the value candidate frames in a certain range for a plurality of times (for example, translating all the values in the directions up, down, left and right in a range of taking the original position of the value candidate frame as the center and taking the radius of the x and y directions as 50 pixels, taking 10 pixels as a unit, wherein the total translation times are (50/10 x 2+1) 2=121), calculating the alignment degree of the value candidate frame and the template value frame of each displacement, selecting one displacement with the highest alignment degree as the final dislocation displacement, and carrying out dislocation correction on all the value candidate frames according to the displacement.
Marking the ith template value frame asThe number of frames is n, the j candidate value frame is +.>The number of frames is m, the intersection is an intersection operation, the area is an area operation, the bin is a binarization function (1 if the condition is satisfied, or 0 if the condition is satisfied), and the calculation formula of the alignment_ratio is as follows:
5. and extracting the structural information near a value frame preset by the template.
For each value candidate frame, finding a template value frame with the largest overlapping area, if the overlapping degree of the template value frame and the template value frame is larger than a set threshold value (such as 0.6), associating the value candidate frame to the template value frame, otherwise, ignoring the value candidate frame. After all the value candidate boxes are associated, all the value candidate boxes associated with each template value box are the contents corresponding to the template value field, the text contents of the template value boxes are connected in series to obtain the extracted structural information of the field, and if the template value box is not associated with any value candidate box, the field cannot be identified.
And (3) recording a template value frame as tv and a value candidate frame as v, and calculating the overlapping degree overlap ratio according to the following formula:
the homography matrix is explained as follows:
respectively acquiring coordinates of a text box of a key marked by a template picture and coordinates of a text box of a key matched by a picture to be identified, and establishing a corresponding homography matrix according to the acquired corresponding relation between the coordinates; and transforming the coordinates of the value of the picture to be identified into aligned coordinates according to the homography matrix of the marked key of the template picture and the matched key of the picture to be identified, so as to obtain the aligned value of the picture to be identified.
Specifically, for the text box corresponding to each key, respectively obtaining four vertex coordinates of the text box of the key marked by the template picture and four vertex coordinates of the text box of the key matched with the picture to be identified, and establishing the following homography matrix according to the obtained correspondence between the four groups of vertex coordinates;
(1);
wherein (x 1, y 1), (x 2, y 2), (x 3, y 3), (x 4, y 4) are four vertex coordinates of a text box corresponding to the current key identified by the picture to be identified, (x '1, y' 1), (x '2, y' 2), (x '3, y' 3), (x '4, y' 4) are four vertex coordinates of a text box corresponding to the current key marked by the template picture, and h11, h12, h13, h21, h22, h23, h31, h32 are unknown parameters to be solved. And 8 unknown parameters of the homography matrix can be obtained by bringing corresponding coordinate values into the formula (1), and the solved homography matrix is input into an image transformation model to obtain the aligned keys of the picture to be identified. The image transformation model of the present embodiment is a perspective transformation model. The alignment of the matched value for the picture to be identified is shown above and will not be described in detail here.
In a real application scenario, each set of coordinate points (e.g., (x 1, y 1) and (x '1, y' 1) calculated in the above embodiment form a set of coordinate points, which will be hereinafter referred to as a point pair) will all include noise. For example, the position of the coordinate point deviates by several pixels, even a phenomenon of mismatching of the characteristic point pair occurs, and if the homography matrix is calculated by using only four points, a large error occurs. So to make the computation more accurate, a homography matrix is typically computed using much more than four points.
In the above embodiment, four vertices of the text boxes of all keys are used to calculate a homography matrix, and the method used is RANSAC, which specifically includes the steps of:
(1) Randomly selecting 4 pairs of matching characteristic points from the initial matching point pair set S as an internal point set Si, and estimating an initial homography matrix Hi;
(2) The remaining matching point pairs in S are calculated with Hi. If the projection error of a certain characteristic point is smaller than a threshold t, adding the characteristic point into Si;
(3) Recording the number of matching point pairs in the Si set;
(4) Repeating the steps (2) - (3) until the iteration number is greater than K;
(5) The point pair number obtained by the iterative calculation is the most, and the estimation model with the most point pair number is the homography matrix to be solved.
The foregoing examples merely illustrate specific embodiments of the invention, which are described in greater detail and are not to be construed as limiting the scope of the invention. It should be noted that it will be apparent to those skilled in the art that several variations and modifications can be made without departing from the spirit of the invention, which are all within the scope of the invention.

Claims (6)

1. The method for extracting the structured information of the cigarette logistics receiving bill is characterized by comprising the following steps of:
pre-marking: setting template picture standards of the bill, selecting standard template pictures, and labeling key and value on the template pictures, wherein the key is a fixed keyword in the bill, and the value is variable content in the bill;
and (3) identification: determining a picture to be identified, matching a key of the picture to be identified with a key of a template picture, setting text boxes except the key in the picture to be identified as value candidate boxes, aligning the picture to be identified with the template picture according to the corresponding relation between the keys, correcting the value candidate boxes in a dislocation manner, and extracting structural information according to the content in the value text boxes of the template picture;
in the pre-labeling step, the step of labeling the key is as follows: rectangular frame labeling and text content labeling are carried out on the template picture, and compact rectangular frames are set as keyword areas; in the pre-labeling step, the labeling value step is as follows: performing rectangular frame labeling and field name labeling on fields needing to be identified except for labeling keys in the template picture;
in the identifying step, the step of correcting the value candidate frame by dislocation is as follows: and translating all the value candidate frames at least once according to a preset rule, calculating the alignment degree of the value candidate frames and the template value frame of each displacement, selecting the displacement with the highest alignment degree as final dislocation displacement, and carrying out dislocation correction on all the value candidate frames according to the displacement to obtain the dislocation corrected value frame of the picture to be identified and the content thereof.
2. The method for extracting structured information from a cigarette logistics and goods receipt according to claim 1, wherein in the pre-labeling step, the template picture standard of the receipt is a flat, non-inclined and printed picture without dislocation.
3. The method for extracting structured information from a cigarette logistics and goods receipt of claim 1 or 2, wherein the step of identifying further comprises the steps of:
all text boxes and text content in the picture to be identified are detected and identified at ocr.
4. The method for extracting structured information from a cigarette logistics receipt of claim 3, wherein the step of identifying further comprises the steps of:
and judging whether the text boxes and text contents in the obtained picture to be identified belong to key text boxes of the template picture or not through keyword matching, if so, associating the keys in the picture to be identified with the keys of the template picture to form a group of key corresponding relations, if not, taking the text boxes and the text contents in the picture to be identified as value candidate boxes, and if no group of key corresponding relations exist, failing to identify the current picture.
5. The method for extracting structured information from a cigarette logistics receipt of claim 4, wherein the step of identifying further comprises the steps of:
aligning the picture to be identified with the template picture according to the corresponding relation between the keys, extracting 4 vertexes of the text box, and establishing the corresponding relation of 4 groups of vertexes according to the corresponding relation between each group of picture keys to be identified and the template picture key;
when N groups of key corresponding relations exist, a corresponding relation of N4 groups of vertex coordinates is established, a homography matrix is calculated according to the corresponding relation between the vertex coordinates, and the picture to be identified and the template picture are aligned through perspective transformation.
6. The method for extracting structured information from a cigarette logistics receipt of claim 5, wherein the step of identifying further comprises the steps of:
extracting structural information near the template picture value text boxes, finding one template picture value text box with the largest overlapping area for each value candidate box, if the overlapping degree of the two template picture value text boxes is larger than a set threshold value, associating the value candidate box with the corresponding template picture value text box, otherwise, neglecting the value candidate box;
after the association of all the value candidate boxes is completed, the value candidate box associated with each template picture value text box is the content corresponding to the template picture value field, the text content is connected in series to obtain the extracted structured information of the field, and if the template picture value text box is not associated with the value candidate box, the field cannot be identified.
CN202211442689.4A 2022-11-18 2022-11-18 Structured information extraction method for cigarette logistics receiving bill Active CN115497114B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211442689.4A CN115497114B (en) 2022-11-18 2022-11-18 Structured information extraction method for cigarette logistics receiving bill

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211442689.4A CN115497114B (en) 2022-11-18 2022-11-18 Structured information extraction method for cigarette logistics receiving bill

Publications (2)

Publication Number Publication Date
CN115497114A CN115497114A (en) 2022-12-20
CN115497114B true CN115497114B (en) 2024-03-12

Family

ID=85116135

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211442689.4A Active CN115497114B (en) 2022-11-18 2022-11-18 Structured information extraction method for cigarette logistics receiving bill

Country Status (1)

Country Link
CN (1) CN115497114B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116382696B (en) * 2023-03-18 2024-06-07 宝钢工程技术集团有限公司 Engineering attribute dynamic analysis and submission method based on factory object position number

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6539112B1 (en) * 1999-02-26 2003-03-25 Raf Technology, Inc. Methods and system for identifying a reference region on an image of a dropped-out form
US8724907B1 (en) * 2012-03-28 2014-05-13 Emc Corporation Method and system for using OCR data for grouping and classifying documents
CN111241974A (en) * 2020-01-07 2020-06-05 深圳追一科技有限公司 Bill information acquisition method and device, computer equipment and storage medium
CN111931784A (en) * 2020-09-17 2020-11-13 深圳壹账通智能科技有限公司 Bill recognition method, system, computer device and computer-readable storage medium
CN112613367A (en) * 2020-12-14 2021-04-06 盈科票据服务(深圳)有限公司 Bill information text box acquisition method, system, equipment and storage medium
CN112699867A (en) * 2020-09-27 2021-04-23 民生科技有限责任公司 Fixed format target image element information extraction method and system
CN112861782A (en) * 2021-03-07 2021-05-28 上海大学 Bill photo key information extraction system and method
CN113158895A (en) * 2021-04-20 2021-07-23 北京中科江南信息技术股份有限公司 Bill identification method and device, electronic equipment and storage medium
CN113191348A (en) * 2021-05-31 2021-07-30 山东新一代信息产业技术研究院有限公司 Template-based text structured extraction method and tool
CN113657377A (en) * 2021-07-22 2021-11-16 西南财经大学 Structured recognition method for airplane ticket printing data image
CN113903024A (en) * 2021-09-28 2022-01-07 合肥高维数据技术有限公司 Handwritten bill numerical value information identification method, system, medium and device
CN115063784A (en) * 2022-06-08 2022-09-16 杭州未名信科科技有限公司 Bill image information extraction method and device, storage medium and electronic equipment
CN115147855A (en) * 2021-03-30 2022-10-04 上海聚均科技有限公司 Method and system for carrying out batch OCR (optical character recognition) on bills
CN115240178A (en) * 2022-06-24 2022-10-25 深源恒际科技有限公司 Structured information extraction method and system for bill image

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111709339B (en) * 2020-06-09 2023-09-19 北京百度网讯科技有限公司 Bill image recognition method, device, equipment and storage medium

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6539112B1 (en) * 1999-02-26 2003-03-25 Raf Technology, Inc. Methods and system for identifying a reference region on an image of a dropped-out form
US8724907B1 (en) * 2012-03-28 2014-05-13 Emc Corporation Method and system for using OCR data for grouping and classifying documents
CN111241974A (en) * 2020-01-07 2020-06-05 深圳追一科技有限公司 Bill information acquisition method and device, computer equipment and storage medium
WO2022057471A1 (en) * 2020-09-17 2022-03-24 深圳壹账通智能科技有限公司 Bill identification method, system, computer device, and computer-readable storage medium
CN111931784A (en) * 2020-09-17 2020-11-13 深圳壹账通智能科技有限公司 Bill recognition method, system, computer device and computer-readable storage medium
CN112699867A (en) * 2020-09-27 2021-04-23 民生科技有限责任公司 Fixed format target image element information extraction method and system
CN112613367A (en) * 2020-12-14 2021-04-06 盈科票据服务(深圳)有限公司 Bill information text box acquisition method, system, equipment and storage medium
CN112861782A (en) * 2021-03-07 2021-05-28 上海大学 Bill photo key information extraction system and method
CN115147855A (en) * 2021-03-30 2022-10-04 上海聚均科技有限公司 Method and system for carrying out batch OCR (optical character recognition) on bills
CN113158895A (en) * 2021-04-20 2021-07-23 北京中科江南信息技术股份有限公司 Bill identification method and device, electronic equipment and storage medium
CN113191348A (en) * 2021-05-31 2021-07-30 山东新一代信息产业技术研究院有限公司 Template-based text structured extraction method and tool
CN113657377A (en) * 2021-07-22 2021-11-16 西南财经大学 Structured recognition method for airplane ticket printing data image
CN113903024A (en) * 2021-09-28 2022-01-07 合肥高维数据技术有限公司 Handwritten bill numerical value information identification method, system, medium and device
CN115063784A (en) * 2022-06-08 2022-09-16 杭州未名信科科技有限公司 Bill image information extraction method and device, storage medium and electronic equipment
CN115240178A (en) * 2022-06-24 2022-10-25 深源恒际科技有限公司 Structured information extraction method and system for bill image

Also Published As

Publication number Publication date
CN115497114A (en) 2022-12-20

Similar Documents

Publication Publication Date Title
CN108961236B (en) Circuit board defect detection method and device
CN109658584B (en) Bill information identification method and device
JP6354589B2 (en) Object identification device, method and program
WO2020098250A1 (en) Character recognition method, server, and computer readable storage medium
US6778703B1 (en) Form recognition using reference areas
CN110546651B (en) Method, system and computer readable medium for identifying objects
CN104584071B (en) Object detector, object identification method
CN101339566B (en) Image processing method, image processing apparatus, image reading apparatus and image forming apparatus
CN104584073A (en) Object discrimination device, object discrimination method, and program
US7463772B1 (en) De-warping of scanned images
CN115497114B (en) Structured information extraction method for cigarette logistics receiving bill
CN107680112B (en) Image registration method
CN111160427B (en) Method for detecting mass flow data type based on neural network
CN107862319B (en) Heterogeneous high-light optical image matching error eliminating method based on neighborhood voting
CN114359553B (en) Signature positioning method and system based on Internet of things and storage medium
CN111860527A (en) Image correction method, image correction device, computer device, and storage medium
CN111667479A (en) Pattern verification method and device for target image, electronic device and storage medium
CN115375917B (en) Target edge feature extraction method, device, terminal and storage medium
CN111858977B (en) Bill information acquisition method, device, computer equipment and storage medium
CN110120013A (en) A kind of cloud method and device
JP2014164622A (en) Information system and server
CN113095187A (en) Examination paper correction method based on image feature matching alignment
JPH04306067A (en) Method and system for identifying image region
CN111414905A (en) Text detection method, text detection device, electronic equipment and storage medium
CN108992033B (en) Grading device, equipment and storage medium for vision test

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant