WO2019071660A1 - Bill information identification method, electronic device, and readable storage medium - Google Patents
Bill information identification method, electronic device, and readable storage medium Download PDFInfo
- Publication number
- WO2019071660A1 WO2019071660A1 PCT/CN2017/108735 CN2017108735W WO2019071660A1 WO 2019071660 A1 WO2019071660 A1 WO 2019071660A1 CN 2017108735 W CN2017108735 W CN 2017108735W WO 2019071660 A1 WO2019071660 A1 WO 2019071660A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- ticket
- identified
- picture
- model
- sample
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/40—Document-oriented image-based pattern recognition
- G06V30/41—Analysis of document content
- G06V30/414—Extracting the geometrical structure, e.g. layout tree; Block segmentation, e.g. bounding boxes for graphics or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
Definitions
- the present application relates to the field of computer technologies, and in particular, to a ticket information identification method, an electronic device, and a readable storage medium.
- the purpose of the present application is to provide a ticket information identification method, an electronic device, and a readable storage medium, which are intended to improve the efficiency of ticket information identification and reduce the error rate of ticket information recognition.
- a first aspect of the present application provides an electronic device, where the electronic device includes a memory and a processor, where the memory stores a ticket information recognition system operable on the processor, and the ticket information
- the identification system implements the following steps when executed by the processor:
- the area recognition model After receiving the picture of the ticket to be processed, determining a region recognition model corresponding to each field to be identified in the ticket image according to a predetermined mapping relationship between the field to be identified and the region identification model, and calling corresponding to each field to be identified
- the area recognition model performs area recognition on the line character area of the ticket picture, and identifies a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and the included character information is in the same line.
- the frames are spliced together in the order of recognition to form a target line character region containing character information;
- a second aspect of the present application provides a ticket information identification method, where the ticket information identification method includes:
- Step 1 After receiving the picture of the bill to be processed, determining the area recognition model corresponding to each field to be identified in the ticket picture according to the mapping relationship between the predetermined field to be identified and the area identification model, for each field to be identified, Invoking a corresponding area recognition model to perform area recognition on the line character area of the ticket picture, identifying a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and placing the included character information in the same
- the target boxes of the rows are stitched together in the order of recognition. Forming a target line character area containing character information;
- Step 2 Determine, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and call a corresponding character recognition model for each of the target line character regions of the to-be-identified field Character recognition is performed to respectively identify character information included in a target line character region of each of the to-be-identified fields.
- a third aspect of the present application provides a computer readable storage medium storing a ticket information identification system, the ticket information identification system being executable by at least one processor, Taking the at least one processor to perform the following steps:
- the area recognition model After receiving the picture of the ticket to be processed, determining a region recognition model corresponding to each field to be identified in the ticket image according to a predetermined mapping relationship between the field to be identified and the region identification model, and calling corresponding to each field to be identified
- the area recognition model performs area recognition on the line character area of the ticket picture, and identifies a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and the included character information is in the same line.
- the frames are spliced together in the order of recognition to form a target line character region containing character information;
- the area recognition model corresponding to each field to be identified in the ticket image is used to identify the area of each line to be recognized in the line character area of the ticket picture, and identify A small frame containing character information and a fixed width is a preset value, and the small boxes containing the character information in the same line are sequentially stitched to form a target line character area containing character information, and the character recognition model corresponding to the field to be identified is called. Character recognition is performed on the target line character area.
- the identified line character area containing the character information is the width of the unified fixed preset value, the character information can be specific to the smaller sub-area, and the sub-area containing the character information has a good approximation.
- the target line character area when character recognition is performed by the character recognition model there are much less interference factors than the character information, thereby reducing the error rate of the ticket information recognition.
- FIG. 1 is a schematic diagram of an operating environment of a preferred embodiment of the ticket information identification system 10 of the present application;
- FIG. 2 is a schematic flowchart diagram of an embodiment of a method for identifying a bill information according to the present application.
- FIG. 1 is a schematic diagram of an operating environment of a preferred embodiment of the ticket information identification system 10 of the present application.
- the ticket information identification system 10 is installed and operated in the electronic device 1.
- the electronic device 1 may include, but is not limited to, a memory 11, a processor 12, and a display 13.
- Figure 1 shows only the electronic device 1 with components 11-13, but it should be understood that not all illustrated components may be implemented, and more or fewer components may be implemented instead.
- the memory 11 comprises at least one type of readable storage medium, which in some embodiments may be an internal storage unit of the electronic device 1, such as a hard disk or memory of the electronic device 1.
- the memory 11 may also be an external storage device of the electronic device 1 in other embodiments, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (SMC), and a secure digital device. (Secure Digital, SD) card, flash card, etc.
- SMC smart memory card
- SD Secure Digital
- flash card etc.
- the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device.
- the memory 11 is used to store application software and various types of data installed in the electronic device 1, such as program codes of the ticket information recognition system 10, and the like.
- the memory 11 can also be used to temporarily store data that has been output or is about to be output.
- the processor 12 in some embodiments, may be a central processing unit (CPU), a microprocessor or other data processing chip for running program code or processing data stored in the memory 11, for example
- the ticket information recognition system 10 and the like are executed.
- the display 13 in some embodiments may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like.
- the display 13 is for displaying information processed in the electronic device 1 and a user interface for displaying visualization, such as a bill picture to be processed, recognized character information, and the like.
- the components 11-13 of the electronic device 1 communicate with one another via a system bus.
- the ticket information identification system 10 includes at least one computer readable instruction stored in the memory 11, the at least one computer readable instruction being executable by the processor 12 to implement various embodiments of the present application.
- Step S1 After receiving the picture of the bill to be processed, determining, according to a predetermined mapping relationship between the field to be identified and the area identification model, an area recognition model corresponding to each field to be identified in the ticket picture, for each field to be identified, Invoking a corresponding area recognition model to perform area recognition on the line character area of the ticket picture, identifying a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and placing the included character information in the same
- the target frames of the rows are stitched together in the order of recognition to form a target line character region containing character information.
- the ticket information identification system 10 receives a bill picture of the to-be-identified processing sent by the user through the terminal device 2, and the bill picture includes a bill picture related to insurance, medical, financial, and the like, such as an outpatient or hospital bill picture.
- a bill picture related to insurance, medical, financial, and the like such as an outpatient or hospital bill picture.
- receiving a picture of a ticket sent by a user on a client installed in a terminal device such as a mobile phone, a tablet computer, or a self-service terminal device
- receiving the user to send the message on a browser system in a terminal such as a mobile phone, a tablet computer, or a self-service terminal device.
- a region identification model corresponding to the type of the field to be identified is pre-configured, for example, a first recognition model is pre-set for the text class field, and a second recognition model is preset for the digital class field, for the date/ The time class field is pre-set with a third recognition model, the fourth recognition model is pre-set for the currency class field, and so on.
- the method may include:
- the pre-trained bill picture recognition model is used to identify the bill type in the received picture, and output the identification result of the bill category (for example, the category of the medical bill includes the outpatient bill, Hospitalization bills, as well as other types of notes).
- A2 Performing a tilt correction on the received ticket image by using a predetermined correction rule; in an optional implementation manner, the predetermined correction rule is: using a Hough probability algorithm to find the ticket As many small straight lines as possible in the image; all straight lines are determined from the found small straight lines, and the straight lines whose x coordinate values are not much different in the determined straight line are sequentially connected in the order of the corresponding y coordinate values.
- the size of the x coordinate value it is divided into several classes, or the straight lines whose y coordinate values are not much different in the determined straight line are sequentially connected in the order of the corresponding x coordinate values, and are classified into several classes according to the size of the y coordinate value; All horizontal lines belonging to a class are used as a target class line, and the longest line closest to each target class line is found by least squares method; the slope of each long line is calculated, and the median of the slopes of each long line is calculated. Mean, compare the median and mean of the calculated slope to determine the smaller one, and adjust the image tilt according to the smaller one to determine the received bill No inclination correction is normal piece of the picture.
- A3. Determine, according to a mapping relationship between the predetermined ticket category and the to-be-identified field, a field to be identified corresponding to the identified ticket category;
- A4. Determine, according to a predetermined mapping relationship between the to-be-identified field and the area recognition model, an area recognition model corresponding to each of the to-be-identified fields.
- the area recognition model is a convolutional neural network model
- the training process for the area recognition model corresponding to a field to be identified is as follows:
- the ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;
- C6 performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;
- verification pass rate is greater than or equal to a preset threshold (for example, 98%), the training is completed, or if the verification pass rate is less than the preset threshold, increase the number of ticket picture samples, and repeat steps C2, C3, and C4. , C5, C6.
- a preset threshold for example, 98%)
- Step S2 determining, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and target line characters for each of the to-be-identified fields
- the area is called by the corresponding character recognition model for character recognition to respectively identify the character information included in the target line character area of each of the to-be-identified fields.
- the character recognition corresponding to each of the to-be-identified fields may be determined according to a predetermined mapping relationship between the to-be-identified field and the character recognition model.
- a model in response to the identified target line character regions of each of the to-be-identified fields, calling a corresponding character recognition model for character recognition to respectively identify character information included in a target line character region of each of the to-be-identified fields, completing the entire The character information of the ticket picture is identified.
- the character recognition model is a Long-Short Term Memory (LSTM), and the training process for a character recognition model corresponding to a field to be identified is as follows:
- the ticket picture sample contains only one line of character information of the to-be-identified field, the font is black, the background is white, and each ticket is The name of the picture sample is named as the character information of the field to be identified contained therein;
- the bill picture samples are divided into a first data set and a second data set according to a ratio of X:Y (for example, 8:2), and the number of bill picture samples in the first data set is larger than the bill picture sample in the second data set.
- Quantity the first data set as a training set, and the second data set as a test set, where X is greater than 0 and Y is greater than 0;
- the sample of the bill image in the first data set is sent to the time recurrent neural network model for model training, and the second data is used for the trained model every certain period of time or a preset number of iterations (for example, every 1000 iterations).
- the set is tested to evaluate the effect of the currently trained model.
- the trained model is used to identify the character information of the ticket image sample in the second data set, and compares with the name of the tested ticket picture sample to calculate the error of the recognition result and the labeling result, and the error calculation uses the editing distance. As a calculation standard.
- the training model obtains divergence of the character information recognition error of the ticket picture sample during the test, the training parameters are adjusted and retrained, so that the error of the character information recognition of the ticket picture sample can be converged during the training. After the error converges, the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
- the area identification model corresponding to each to-be-identified field in the ticket picture performs area identification on each line character area in the ticket picture, and identifies the character information and the fixed width.
- a small frame of preset values, and the small boxes containing the character information in the same row are sequentially stitched to form a target line character region containing character information, and the character recognition model corresponding to the field to be recognized is called to the target line character region.
- the target line character area when character recognition is performed by the character recognition model, there are much less interference factors than the character information, thereby reducing the error rate of the ticket information recognition.
- the ticket picture recognition model is a deep convolutional neural network model (eg, the deep convolutional neural network model may be in a CaffeNet environment)
- the selected deep convolutional neural network SSD (Single Shot MultiBox Detector) algorithm model the deep convolutional neural network model consists of one input layer, 13 convolutional layers, 5 pooling layers, 2 fully connected layers, One classification layer is formed.
- Table 1 The detailed structure of the deep convolutional neural network model is shown in Table 1 below:
- Layer Name indicates the name of each layer
- Input indicates the input layer
- Conv indicates the convolution layer of the model
- Conv1 indicates the first convolution layer of the model
- MaxPool indicates the maximum pooling layer of the model
- MaxPool1 indicates the model.
- Fc represents the fully connected layer in the model
- Fc1 represents the first fully connected layer in the model
- Softmax represents the Softmax classifier
- Batch Size represents the number of input images of the current layer
- Kernel Size represents the current layer volume
- the scale of the kernel (for example, the Kernel Size can be equal to 3, indicating that the scale of the convolution kernel is 3x3)
- the Stride Size indicates the moving step size of the convolution kernel, that is, the distance moved to the next convolution position after one convolution is completed.
- Pad Size indicates the size of the image fill in the current network layer.
- pooling mode of the pooling layer in this embodiment includes, but is not limited to, Mean pooling, Max pooling, Overlapping, L2pooling, Local Contrast. Normalization, Stochasticpooling, Def-pooling, and more.
- the training process of the ticket picture recognition model is as follows:
- the transposition of the bill picture is determined, and the flip adjustment is made: when the aspect ratio is greater than 1, the height and width of the bill picture are reversed, and if the stamp position is on the left side of the bill picture, the bill is The image is rotated clockwise by ninety degrees. If the stamp position is on the right side of the bill image, the bill image is rotated counterclockwise by ninety degrees. When the aspect ratio is less than 1, the bill image height and width are not reversed. The position is on the lower side of the ticket picture, and the ticket image is rotated clockwise by one hundred and eighty degrees.
- the annotation data of each object refers to the position information of the rectangular frame of the object, and the coordinates of the upper left corner of the rectangle (xmin, ymin) and the coordinates of the lower right corner (xmax, Ymax) indicates four numbers. If xmax ⁇ xmin, reverse the position and do the same for the y coordinate to ensure max>min.
- the sample picture of the ticket for the model training is a picture of the ticket whose height and width are not reversed and marked accurately, so as to facilitate the subsequent model training to be more accurate and effective.
- FIG. 2 is a schematic flowchart of a method for identifying a ticket information according to an embodiment of the present invention.
- the method for identifying a ticket information includes the following steps:
- Step S10 After receiving the picture of the ticket to be processed, determining, according to the mapping relationship between the field to be identified and the area identification model, the area recognition model corresponding to each field to be identified in the ticket picture, for each field to be identified, Invoking a corresponding area recognition model to perform area recognition on the line character area of the ticket picture, identifying a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and placing the included character information in the same
- the target frames of the rows are stitched together in the order of recognition to form a target line character region containing character information.
- the ticket information identification system 10 receives a bill picture of the to-be-identified processing sent by the user through the terminal device 2, and the bill picture includes a bill picture related to insurance, medical, financial, and the like, such as an outpatient or hospital bill picture.
- a bill picture related to insurance, medical, financial, and the like such as an outpatient or hospital bill picture.
- receiving a picture of a ticket sent by a user on a client installed in a terminal device such as a mobile phone, a tablet computer, or a self-service terminal device
- receiving the user to send the message on a browser system in a terminal such as a mobile phone, a tablet computer, or a self-service terminal device.
- a region identification model corresponding to the type of the field to be identified is pre-configured, for example, a first recognition model is pre-set for the text class field, and a second recognition model is preset for the digital class field, for the date/ The time class field is pre-set with a third recognition model, the fourth recognition model is pre-set for the currency class field, and so on.
- a predetermined mapping relationship between the to-be-identified field such as a text class field, a numeric class field, a date/time class field, a currency class field, and the like
- An area recognition model corresponding to each of the to-be-identified fields For each of the to-be-identified fields, the corresponding area recognition model is called to perform area recognition on the line character area of the ticket picture, and the character information is recognized from the ticket picture and the fixed width is a preset value (for example, 16 pieces)
- the small frame of the pixel width is the target frame, and the small boxes containing the character information in the same line are stitched together in order to form a target line character region containing character information.
- the method may include:
- the pre-trained bill picture recognition model is used to identify the bill type in the received picture, and output the identification result of the bill category (for example, the category of the medical bill includes the outpatient bill, Hospitalization bills, as well as other types of notes).
- A2 Performing a tilt correction on the received ticket image by using a predetermined correction rule; in an optional implementation manner, the predetermined correction rule is: using a Hough probability algorithm to find the ticket As many small straight lines as possible in the image; all straight lines are determined from the found small straight lines, and the straight lines whose x coordinate values are not much different in the determined straight line are sequentially connected in the order of the corresponding y coordinate values.
- the size of the x coordinate value it is divided into several classes, or the straight lines whose y coordinate values are not much different in the determined straight line are sequentially connected in the order of the corresponding x coordinate values, and are classified into several classes according to the size of the y coordinate value; All horizontal lines belonging to a class are used as a target class line, and the longest line closest to each target class line is found by least squares method; the slope of each long line is calculated, and the median of the slopes of each long line is calculated. Mean, compare the median and mean of the calculated slope to determine the smaller one, and adjust the image tilt according to the smaller one to determine the received bill No inclination correction is normal piece of the picture.
- A3. Determine, according to a mapping relationship between the predetermined ticket category and the to-be-identified field, a field to be identified corresponding to the identified ticket category;
- A4. Determine, according to a predetermined mapping relationship between the to-be-identified field and the area recognition model, an area recognition model corresponding to each of the to-be-identified fields.
- the area recognition model is a convolutional neural network model
- the training process for the area recognition model corresponding to a field to be identified is as follows:
- the ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;
- C6 performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;
- verification pass rate is greater than or equal to a preset threshold (for example, 98%), the training is completed, or if the verification pass rate is less than the preset threshold, increase the number of ticket picture samples, and repeat steps C2, C3, and C4. , C5, C6.
- a preset threshold for example, 98%)
- Step S20 Determine, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and call a corresponding character recognition model for each of the target line character regions of the to-be-identified field.
- Character recognition to identify each of the to-be-identified words The character information contained in the target line character area of the segment.
- the character recognition corresponding to each of the to-be-identified fields may be determined according to a predetermined mapping relationship between the to-be-identified field and the character recognition model.
- a model in response to the identified target line character regions of each of the to-be-identified fields, calling a corresponding character recognition model for character recognition to respectively identify character information included in a target line character region of each of the to-be-identified fields, completing the entire The character information of the ticket picture is identified.
- the character recognition model is a Long-Short Term Memory (LSTM), and the training process for a character recognition model corresponding to a field to be identified is as follows:
- the ticket picture sample contains only one line of character information of the to-be-identified field, the font is black, the background is white, and each ticket is The name of the picture sample is named as the character information of the field to be identified contained therein;
- the bill picture samples are divided into a first data set and a second data set according to a ratio of X:Y (for example, 8:2), and the number of bill picture samples in the first data set is larger than the bill picture sample in the second data set.
- Quantity the first data set as a training set, and the second data set as a test set, where X is greater than 0 and Y is greater than 0;
- the sample of the bill image in the first data set is sent to the time recurrent neural network model for model training, and the second data is used for the trained model every certain period of time or a preset number of iterations (for example, every 1000 iterations).
- the set is tested to evaluate the effect of the currently trained model.
- the trained model is used to identify the character information of the ticket image sample in the second data set, and compares with the name of the tested ticket picture sample to calculate the error of the recognition result and the labeling result, and the error calculation uses the editing distance. As a calculation standard.
- the training model obtains divergence of the character information recognition error of the ticket picture sample during the test, the training parameters are adjusted and retrained, so that the error of the character information recognition of the ticket picture sample can be converged during the training. After the error converges, the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
- the area identification model corresponding to each to-be-identified field in the ticket picture performs area identification on each line character area in the ticket picture, and identifies the character information and the fixed width.
- a small frame of preset values, and the small boxes containing the character information in the same row are sequentially stitched to form a target line character region containing character information, and the character recognition model corresponding to the field to be recognized is called to the target line character region.
- the target line character area when character recognition is performed by the character recognition model, there are much less interference factors than the character information, thereby reducing the error rate of the ticket information recognition.
- the ticket picture recognition model is a deep convolutional neural network model (for example, the deep convolutional neural network model may be selected based on a CaffeNet environment) Deep Spool (Single Shot MultiBox Detector) algorithm model, the deep convolutional neural network model consists of 1 input layer, 13 convolutional layers, 5 pooling layers, 2 fully connected layers, 1 classification Layer composition.
- the detailed structure of the deep convolutional neural network model is shown in Table 1 below:
- Layer Name indicates the name of each layer
- Input indicates the input layer
- Conv indicates the convolution layer of the model
- Conv1 indicates the first convolution layer of the model
- MaxPool indicates the maximum pooling layer of the model
- MaxPool1 indicates the model.
- Fc represents the fully connected layer in the model
- Fc1 represents the first fully connected layer in the model
- Softmax represents the Softmax classifier
- Batch Size represents the number of input images of the current layer
- Kernel Size represents the current layer volume
- the scale of the kernel (for example, the Kernel Size can be equal to 3, indicating that the scale of the convolution kernel is 3x3)
- the Stride Size indicates the moving step size of the convolution kernel, that is, the distance moved to the next convolution position after one convolution is completed.
- Pad Size indicates the size of the image fill in the current network layer.
- pooling mode of the pooling layer in this embodiment includes, but is not limited to, Mean pooling, Max pooling, Overlapping, L2pooling, Local Contrast. Normalization, Stochasticpooling, Def-pooling, and more.
- the training process of the ticket picture recognition model is as follows:
- the preset ticket category may include two types of outpatient bills and hospital bills
- preparing a preset number for example, 1000 sheets
- the bill image sample is processed as follows:
- the transposition of the bill picture is determined, and the flip adjustment is made: when the aspect ratio is greater than 1, the height and width of the bill picture are reversed, and if the stamp position is on the left side of the bill picture, the bill is The image is rotated clockwise by ninety degrees. If the stamp position is on the right side of the bill image, the bill image is rotated counterclockwise by ninety degrees. When the aspect ratio is less than 1, the bill image height and width are not reversed. The position is on the lower side of the ticket picture, and the ticket image is rotated clockwise by one hundred and eighty degrees.
- the annotation data of each object refers to the position information of the rectangular frame of the object, and the coordinates of the upper left corner of the rectangle (xmin, ymin) and the coordinates of the lower right corner (xmax, Ymax) indicates four numbers. If xmax ⁇ xmin, reverse the position and do the same for the y coordinate to ensure max>min.
- the sample picture of the ticket for the model training is a picture of the ticket whose height and width are not reversed and marked accurately, so as to facilitate the subsequent model training to be more accurate and effective.
- the present application also provides a computer readable storage medium storing a ticket information identification system, the ticket information identification system being executable by at least one processor to cause the at least one processor.
- the foregoing embodiment method can be implemented by means of software plus a necessary general hardware platform, and can also be implemented by hardware, but in many cases, the former is A better implementation.
- the technical solution of the present application which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM, disk,
- the optical disc includes a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the methods described in various embodiments of the present application.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- Multimedia (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Geometry (AREA)
- Computer Graphics (AREA)
- Image Analysis (AREA)
- Character Discrimination (AREA)
- Character Input (AREA)
Abstract
The present application relates to a bill information identification method, an electronic device, and a readable storage medium. The method comprises: determining, according to a predetermined mapping relationship between fields to be identified and region identification models, a corresponding region identification model for each field to be identified in a bill image; calling the corresponding region identification models to perform region identification on character line regions of the bill image; identifying, from the bill image, target boxes containing character information and having a fixed width of a predetermined value, and joining, according to an identification order, the target boxes containing character information in the same line to form a target character line region containing the character information; determining, according to a predetermined mapping relationship between the fields to be identified and character identification models, a corresponding character identification model for each field to be identified; and calling the corresponding character identification model to perform character identification for the target character line region of the field to be identified. The present application reduces the error rate when identifying bill information.
Description
本申请基于巴黎公约申明享有2017年10月9日递交的申请号为CN201710930679.8、名称为“票据信息识别方法、电子装置及可读存储介质”中国专利申请的优先权,该中国专利申请的整体内容以参考的方式结合在本申请中。The present application is based on the priority of the Chinese Patent Application entitled "Payment Information Identification Method, Electronic Device and Readable Storage Medium", filed on October 9, 2017, with the application number of CN201710930679.8, which is filed on October 9, 2017. The entire content is incorporated herein by reference.
本申请涉及计算机技术领域,尤其涉及一种票据信息识别方法、电子装置及可读存储介质。The present application relates to the field of computer technologies, and in particular, to a ticket information identification method, an electronic device, and a readable storage medium.
如今随着经济的发展和人们生活水平的提高,越来越多的人选择购买医疗、商业、金融等保险。为了改善用户的保险理赔体验,提升保险理赔效率,目前,有些保险公司推出了自助理赔业务,比如用户在进行医疗保险理赔过程中,只需要将门诊或住院票据拍照上传到保险公司系统,保险公司业务员会将用户上传的票据图片上的信息录入到理赔系统中,以进行下一步操作,这种自助理赔方式大大方便了用户进行理赔的过程,然而,这种自助理赔方式在带来了便捷的理赔过程的同时,却增加了保险公司业务人员的工作压力,问题主要表现在需要花费大量的人力来处理用户上传的票据图像,效率低下,且数据录入的错误率居高不下。Nowadays, with the development of the economy and the improvement of people's living standards, more and more people choose to purchase medical, commercial, financial and other insurance. In order to improve the user's insurance claims experience and improve the efficiency of insurance claims, some insurance companies have launched self-service claims services. For example, in the process of medical insurance claims, users only need to upload photos of outpatient or hospital bills to the insurance company system, insurance companies. The salesperson will enter the information on the picture uploaded by the user into the claim system for the next step. This self-service settlement method greatly facilitates the user's process of claim settlement. However, this self-service settlement method brings convenience. At the same time of the claims process, it increases the work pressure of the insurance company's business personnel. The problem is mainly caused by the need to spend a lot of manpower to process the image uploaded by the user, which is inefficient and the error rate of data entry is high.
发明内容Summary of the invention
本申请的目的在于提供一种票据信息识别方法、电子装置及可读存储介质,旨在提高票据信息识别效率和降低票据信息识别的错误率。The purpose of the present application is to provide a ticket information identification method, an electronic device, and a readable storage medium, which are intended to improve the efficiency of ticket information identification and reduce the error rate of ticket information recognition.
为实现上述目的,本申请第一方面提供一种电子装置,所述电子装置包括存储器、处理器,所述存储器上存储有可在所述处理器上运行的票据信息识别系统,所述票据信息识别系统被所述处理器执行时实现如下步骤:In order to achieve the above object, a first aspect of the present application provides an electronic device, where the electronic device includes a memory and a processor, where the memory stores a ticket information recognition system operable on the processor, and the ticket information The identification system implements the following steps when executed by the processor:
在收到待处理的票据图片后,根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型,针对各个待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值的目标框,并将所包含的字符信息处于同一行的目标框按照识别的先后顺序拼接在一起形成包含字符信息的目标行字符区域;After receiving the picture of the ticket to be processed, determining a region recognition model corresponding to each field to be identified in the ticket image according to a predetermined mapping relationship between the field to be identified and the region identification model, and calling corresponding to each field to be identified The area recognition model performs area recognition on the line character area of the ticket picture, and identifies a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and the included character information is in the same line. The frames are spliced together in the order of recognition to form a target line character region containing character information;
根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对各个所述待识别字段的目标行字符区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字段的目标行字符区域包含的字符信息。Determining, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and calling a corresponding character recognition model for character recognition for each target character region of the to-be-identified field And respectively identifying character information included in a target line character region of each of the to-be-identified fields.
此外,为实现上述目的,本申请第二方面提供一种票据信息识别方法,所述票据信息识别方法包括:In addition, in order to achieve the above object, a second aspect of the present application provides a ticket information identification method, where the ticket information identification method includes:
步骤一、在收到待处理的票据图片后,根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型,针对各个待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值的目标框,并将所包含的字符信息处于同一行的目标框按照识别的先后顺序拼接在一起
形成包含字符信息的目标行字符区域;Step 1: After receiving the picture of the bill to be processed, determining the area recognition model corresponding to each field to be identified in the ticket picture according to the mapping relationship between the predetermined field to be identified and the area identification model, for each field to be identified, Invoking a corresponding area recognition model to perform area recognition on the line character area of the ticket picture, identifying a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and placing the included character information in the same The target boxes of the rows are stitched together in the order of recognition.
Forming a target line character area containing character information;
步骤二、根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对各个所述待识别字段的目标行字符区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字段的目标行字符区域包含的字符信息。Step 2: Determine, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and call a corresponding character recognition model for each of the target line character regions of the to-be-identified field Character recognition is performed to respectively identify character information included in a target line character region of each of the to-be-identified fields.
进一步地,为实现上述目的,本申请第三方面提供一种计算机可读存储介质,所述计算机可读存储介质存储有票据信息识别系统,所述票据信息识别系统可被至少一个处理器执行,以使所述至少一个处理器执行如下步骤:Further, in order to achieve the above object, a third aspect of the present application provides a computer readable storage medium storing a ticket information identification system, the ticket information identification system being executable by at least one processor, Taking the at least one processor to perform the following steps:
在收到待处理的票据图片后,根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型,针对各个待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值的目标框,并将所包含的字符信息处于同一行的目标框按照识别的先后顺序拼接在一起形成包含字符信息的目标行字符区域;After receiving the picture of the ticket to be processed, determining a region recognition model corresponding to each field to be identified in the ticket image according to a predetermined mapping relationship between the field to be identified and the region identification model, and calling corresponding to each field to be identified The area recognition model performs area recognition on the line character area of the ticket picture, and identifies a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and the included character information is in the same line. The frames are spliced together in the order of recognition to form a target line character region containing character information;
根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对各个所述待识别字段的目标行字符区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字段的目标行字符区域包含的字符信息。Determining, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and calling a corresponding character recognition model for character recognition for each target character region of the to-be-identified field And respectively identifying character information included in a target line character region of each of the to-be-identified fields.
本申请提出的票据信息识别方法、系统及可读存储介质,通过票据图片中各个待识别字段对应的区域识别模型对各个待识别字段在所述票据图片中的行字符区域进行区域识别,识别出包含字符信息且固定宽度为预设值的小框,并将所包含的字符信息处于同一行的小框按顺序拼接形成包含字符信息的目标行字符区域,调用与待识别字段对应的字符识别模型对该目标行字符区域进行字符识别。由于识别出的包含字符信息的行字符区域为统一固定预设值的宽度,这样,可以将字符信息具体到更小的子区域,并对包含字符信息的各个子区域有一个很好的逼近,在利用字符识别模型进行字符识别时的目标行字符区域中除字符信息之外的其它干扰因素会少很多,从而降低票据信息识别的错误率。The method for identifying the ticket information and the readable storage medium proposed by the present application, the area recognition model corresponding to each field to be identified in the ticket image is used to identify the area of each line to be recognized in the line character area of the ticket picture, and identify A small frame containing character information and a fixed width is a preset value, and the small boxes containing the character information in the same line are sequentially stitched to form a target line character area containing character information, and the character recognition model corresponding to the field to be identified is called. Character recognition is performed on the target line character area. Since the identified line character area containing the character information is the width of the unified fixed preset value, the character information can be specific to the smaller sub-area, and the sub-area containing the character information has a good approximation. In the target line character area when character recognition is performed by the character recognition model, there are much less interference factors than the character information, thereby reducing the error rate of the ticket information recognition.
图1为本申请票据信息识别系统10较佳实施例的运行环境示意图;1 is a schematic diagram of an operating environment of a preferred embodiment of the ticket information identification system 10 of the present application;
图2为本申请票据信息识别方法一实施例的流程示意图。FIG. 2 is a schematic flowchart diagram of an embodiment of a method for identifying a bill information according to the present application.
为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处所描述的具体实施例仅用以解释本申请,并不用于限定本申请。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。In order to make the objects, technical solutions, and advantages of the present application more comprehensible, the present application will be further described in detail below with reference to the accompanying drawings and embodiments. It is understood that the specific embodiments described herein are merely illustrative of the application and are not intended to be limiting. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of the present application without departing from the inventive scope are the scope of the present application.
需要说明的是,在本申请中涉及“第一”、“第二”等的描述仅用于描述目的,而不能理解为指示或暗示其相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”的特征可以明示或者隐含地包括至少一个该特征。另外,各个实施例之间的技术方案可以相互结合,但是必须是以本领域普通技术人员能够实现为基础,当技术方案的结合出现相互矛盾或无法实现时
应当认为这种技术方案的结合不存在,也不在本申请要求的保护范围之内。It should be noted that the descriptions of "first", "second" and the like in the present application are for the purpose of description only, and are not to be construed as indicating or implying their relative importance or implicitly indicating the number of technical features indicated. . Thus, features defining "first" and "second" may include at least one of the features, either explicitly or implicitly. In addition, the technical solutions between the various embodiments may be combined with each other, but must be based on the realization of those skilled in the art, when the combination of technical solutions is contradictory or impossible to achieve.
It should be considered that the combination of such technical solutions does not exist and is not within the scope of protection claimed herein.
本申请提供一种票据信息识别系统。请参阅图1,是本申请票据信息识别系统10较佳实施例的运行环境示意图。The application provides a ticket information identification system. Please refer to FIG. 1 , which is a schematic diagram of an operating environment of a preferred embodiment of the ticket information identification system 10 of the present application.
在本实施例中,所述的票据信息识别系统10安装并运行于电子装置1中。该电子装置1可包括,但不仅限于,存储器11、处理器12及显示器13。图1仅示出了具有组件11-13的电子装置1,但是应理解的是,并不要求实施所有示出的组件,可以替代的实施更多或者更少的组件。In the present embodiment, the ticket information identification system 10 is installed and operated in the electronic device 1. The electronic device 1 may include, but is not limited to, a memory 11, a processor 12, and a display 13. Figure 1 shows only the electronic device 1 with components 11-13, but it should be understood that not all illustrated components may be implemented, and more or fewer components may be implemented instead.
所述存储器11至少包括一种类型的可读存储介质,所述存储器11在一些实施例中可以是所述电子装置1的内部存储单元,例如该电子装置1的硬盘或内存。所述存储器11在另一些实施例中也可以是所述电子装置1的外部存储设备,例如所述电子装置1上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器11还可以既包括所述电子装置1的内部存储单元也包括外部存储设备。所述存储器11用于存储安装于所述电子装置1的应用软件及各类数据,例如所述票据信息识别系统10的程序代码等。所述存储器11还可以用于暂时地存储已经输出或者将要输出的数据。The memory 11 comprises at least one type of readable storage medium, which in some embodiments may be an internal storage unit of the electronic device 1, such as a hard disk or memory of the electronic device 1. The memory 11 may also be an external storage device of the electronic device 1 in other embodiments, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (SMC), and a secure digital device. (Secure Digital, SD) card, flash card, etc. Further, the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device. The memory 11 is used to store application software and various types of data installed in the electronic device 1, such as program codes of the ticket information recognition system 10, and the like. The memory 11 can also be used to temporarily store data that has been output or is about to be output.
所述处理器12在一些实施例中可以是一中央处理器(Central Processing Unit,CPU),微处理器或其他数据处理芯片,用于运行所述存储器11中存储的程序代码或处理数据,例如执行所述票据信息识别系统10等。The processor 12, in some embodiments, may be a central processing unit (CPU), a microprocessor or other data processing chip for running program code or processing data stored in the memory 11, for example The ticket information recognition system 10 and the like are executed.
所述显示器13在一些实施例中可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(Organic Light-Emitting Diode,有机发光二极管)触摸器等。所述显示器13用于显示在所述电子装置1中处理的信息以及用于显示可视化的用户界面,例如待处理的票据图片、识别出的字符信息等。所述电子装置1的部件11-13通过系统总线相互通信。The display 13 in some embodiments may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like. The display 13 is for displaying information processed in the electronic device 1 and a user interface for displaying visualization, such as a bill picture to be processed, recognized character information, and the like. The components 11-13 of the electronic device 1 communicate with one another via a system bus.
所述票据信息识别系统10包括至少一个存储在所述存储器11中的计算机可读指令,该至少一个计算机可读指令可被所述处理器12执行,以实现本申请各实施例。The ticket information identification system 10 includes at least one computer readable instruction stored in the memory 11, the at least one computer readable instruction being executable by the processor 12 to implement various embodiments of the present application.
其中,上述票据信息识别系统10被所述处理器12执行时实现如下步骤:Wherein, when the ticket information identification system 10 is executed by the processor 12, the following steps are implemented:
步骤S1,在收到待处理的票据图片后,根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型,针对各个待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值的目标框,并将所包含的字符信息处于同一行的目标框按照识别的先后顺序拼接在一起形成包含字符信息的目标行字符区域。Step S1: After receiving the picture of the bill to be processed, determining, according to a predetermined mapping relationship between the field to be identified and the area identification model, an area recognition model corresponding to each field to be identified in the ticket picture, for each field to be identified, Invoking a corresponding area recognition model to perform area recognition on the line character area of the ticket picture, identifying a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and placing the included character information in the same The target frames of the rows are stitched together in the order of recognition to form a target line character region containing character information.
本实施例中,票据信息识别系统10接收用户通过终端设备2发送的待识别处理的票据图片,该票据图片包括与医疗、商业、金融等保险相关的票据图片,如门诊或住院票据图片。例如,接收用户在手机、平板电脑、自助终端设备等终端设备中预先安装的客户端上发送来的票据图片,或接收用户在手机、平板电脑、自助终端设备等终端中的浏览器系统上发送来的票据图片。In this embodiment, the ticket information identification system 10 receives a bill picture of the to-be-identified processing sent by the user through the terminal device 2, and the bill picture includes a bill picture related to insurance, medical, financial, and the like, such as an outpatient or hospital bill picture. For example, receiving a picture of a ticket sent by a user on a client installed in a terminal device such as a mobile phone, a tablet computer, or a self-service terminal device, or receiving the user to send the message on a browser system in a terminal such as a mobile phone, a tablet computer, or a self-service terminal device. The picture of the bill coming.
预先根据待识别字段的不同类型预设有与之对应的区域识别模型,例如,针对文本类字段对应预设有第一识别模型,针对数字类字段对应预设有第二识别模型,针对日期/时间类字段对应预设有第三识别模型,针对货币类字段对应预设有第四识别模型,等等。这样,在收到待处理的票据图片后,根据预先确定的待识别字段(如文本类字段、数字类字段、日期/时间类字段、货币类字段等等)
与区域识别模型的映射关系,确定各个所述待识别字段对应的区域识别模型,针对各个所述待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值(例如,16个像素宽度)的小框即目标框,并将所包含的字符信息处于同一行的小框按照先后顺序拼接在一起形成包含字符信息的目标行字符区域。其中,在确定各个待识别字段对应的区域识别模型时可包括:A region identification model corresponding to the type of the field to be identified is pre-configured, for example, a first recognition model is pre-set for the text class field, and a second recognition model is preset for the digital class field, for the date/ The time class field is pre-set with a third recognition model, the fourth recognition model is pre-set for the currency class field, and so on. In this way, after receiving the picture of the ticket to be processed, according to a predetermined field to be identified (such as a text class field, a numeric class field, a date/time class field, a currency class field, etc.)
Determining, by the mapping relationship with the area identification model, an area identification model corresponding to each of the to-be-identified fields, and calling, for each of the to-be-identified fields, a corresponding area recognition model to perform area recognition on the line character area of the ticket picture. A small frame that contains character information and has a fixed width of a preset value (for example, 16 pixels width) is identified on the note image, and the small frame containing the character information in the same line is stitched together in order. A target line character area containing character information is formed. Wherein, when determining the area recognition model corresponding to each to-be-identified field, the method may include:
A1、在收到待处理的票据图片后,利用预先训练的票据图片识别模型对收到的图片中的票据类别进行识别,并输出票据类别的识别结果(例如,医疗票据的类别包括门诊票据,住院票据,以及其他类票据)。A1. After receiving the picture of the bill to be processed, the pre-trained bill picture recognition model is used to identify the bill type in the received picture, and output the identification result of the bill category (for example, the category of the medical bill includes the outpatient bill, Hospitalization bills, as well as other types of notes).
A2、利用预先确定的矫正规则对收到的票据图片进行倾斜矫正;在一种可选的实施方式中,所述预先确定的矫正规则为:用霍夫变换(Hough)的概率算法找出票据图像中尽可能多的小段直线;从找出的小段直线中确定出所有偏水平的直线,并将确定出的直线中x坐标值相差不大的直线按对应的y坐标值的大小顺序依次相连,按照x坐标值大小分为若干类,或者,将确定出的直线中y坐标值相差不大的直线按对应的x坐标值的大小顺序依次相连,按照y坐标值大小分为若干类;将属于一类的所有水平直线作为一个目标类直线,并通过最小二乘法找出最接近各个目标类直线的长直线;计算出各个长直线的斜率,计算出各个长直线的斜率的中位数和均值,比较计算出的斜率的中位数和均值的大小以确定出较小者,并根据确定出的较小者调整图像倾角,以将收到的票据图片矫正为正常无倾角的图片。A2: Performing a tilt correction on the received ticket image by using a predetermined correction rule; in an optional implementation manner, the predetermined correction rule is: using a Hough probability algorithm to find the ticket As many small straight lines as possible in the image; all straight lines are determined from the found small straight lines, and the straight lines whose x coordinate values are not much different in the determined straight line are sequentially connected in the order of the corresponding y coordinate values. According to the size of the x coordinate value, it is divided into several classes, or the straight lines whose y coordinate values are not much different in the determined straight line are sequentially connected in the order of the corresponding x coordinate values, and are classified into several classes according to the size of the y coordinate value; All horizontal lines belonging to a class are used as a target class line, and the longest line closest to each target class line is found by least squares method; the slope of each long line is calculated, and the median of the slopes of each long line is calculated. Mean, compare the median and mean of the calculated slope to determine the smaller one, and adjust the image tilt according to the smaller one to determine the received bill No inclination correction is normal piece of the picture.
A3、根据预先确定的票据类别与待识别字段的映射关系,确定识别的票据类别对应的待识别字段;A3. Determine, according to a mapping relationship between the predetermined ticket category and the to-be-identified field, a field to be identified corresponding to the identified ticket category;
A4、根据预先确定的待识别字段与区域识别模型的映射关系,确定各个所述待识别字段对应的区域识别模型。A4. Determine, according to a predetermined mapping relationship between the to-be-identified field and the area recognition model, an area recognition model corresponding to each of the to-be-identified fields.
在一种可选的实施方式中,所述区域识别模型为卷积神经网络模型,针对一个待识别字段对应的区域识别模型的训练过程如下:In an optional implementation manner, the area recognition model is a convolutional neural network model, and the training process for the area recognition model corresponding to a field to be identified is as follows:
C1、针对该待识别字段,获取预设数量(例如,10万)的票据图片样本;C1. Obtain a preset number (for example, 100,000) of bill picture samples for the to-be-identified field;
C2、在各个票据图片样本上每隔第一预设数量(例如,16个)的像素,设置第二预设数量(例如,10个)的不同高宽比的且固定宽度为预设值(例如,16个像素宽度)的小框;C2, setting a second preset number (for example, 10) of different aspect ratios and setting a fixed width to a preset value on each of the first bill number samples (for example, 16 pixels) For example, a small frame of 16 pixels wide);
C3、在各个票据图片样本上对包含该待识别字段的部分或者全部字符信息的小框进行标记;C3. Mark a small frame containing part or all of the character information of the to-be-identified field on each ticket picture sample;
C4、将包含该待识别字段的字符信息的票据图片样本归入第一训练集,并将不包含该待识别字段的字符信息的票据图片样本归入第二训练集;C4. The ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;
C5、分别从第一训练集和第二训练集中提取出第一预设比例(例如,80%)的票据图片样本作为待训练的样本图片,并将第一训练集和第二训练集中剩余的票据图片样本作为待验证的样本图片;C5. Extracting, from the first training set and the second training set, a first preset ratio (for example, 80%) of the ticket picture samples as the sample picture to be trained, and remaining the first training set and the second training set. a sample of the bill image as a sample image to be verified;
C6、利用提取的各个待训练的样本图片进行模型训练,以生成所述区域识别模型,并利用各个待验证的样本图片对生成的所述区域识别模型进行验证;C6: performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;
C7、若验证通过率大于等于预设阈值(例如,98%),则训练完成,或者,若验证通过率小于预设阈值,则增加票据图片样本的数量,并重复执行步骤C2、C3、C4、C5、C6。C7. If the verification pass rate is greater than or equal to a preset threshold (for example, 98%), the training is completed, or if the verification pass rate is less than the preset threshold, increase the number of ticket picture samples, and repeat steps C2, C3, and C4. , C5, C6.
步骤S2,根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对各个所述待识别字段的目标行字符
区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字段的目标行字符区域包含的字符信息。Step S2: determining, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and target line characters for each of the to-be-identified fields
The area is called by the corresponding character recognition model for character recognition to respectively identify the character information included in the target line character area of each of the to-be-identified fields.
本实施例中,在利用区域识别模型识别出各个待识别字段的目标行字符区域后,可根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对识别出的各个所述待识别字段的目标行字符区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字段的目标行字符区域包含的字符信息,完成整个票据图片的字符信息识别。In this embodiment, after the target line character area of each to-be-identified field is identified by using the area recognition model, the character recognition corresponding to each of the to-be-identified fields may be determined according to a predetermined mapping relationship between the to-be-identified field and the character recognition model. a model, in response to the identified target line character regions of each of the to-be-identified fields, calling a corresponding character recognition model for character recognition to respectively identify character information included in a target line character region of each of the to-be-identified fields, completing the entire The character information of the ticket picture is identified.
在一种可选的实施方式中,所述字符识别模型为时间递归神经网络模型(Long-Short Term Memory,LSTM),针对一个待识别字段对应的字符识别模型的训练过程如下:In an optional implementation manner, the character recognition model is a Long-Short Term Memory (LSTM), and the training process for a character recognition model corresponding to a field to be identified is as follows:
D1、针对该待识别字段,获取预设数量(例如,10万)的票据图片样本,票据图片样本中仅包含一行该待识别字段的字符信息,字体为黑色,背景为白色,并将各个票据图片样本的名称命名为其所包含的该待识别字段的字符信息;D1. For the to-be-identified field, obtain a preset number (for example, 100,000) of ticket picture samples, where the ticket picture sample contains only one line of character information of the to-be-identified field, the font is black, the background is white, and each ticket is The name of the picture sample is named as the character information of the field to be identified contained therein;
D2、将所述票据图片样本按照X:Y(例如,8:2)的比例分成第一数据集和第二数据集,第一数据集中的票据图片样本数量大于第二数据集中的票据图片样本数量,第一数据集作为训练集,第二数据集作为测试集,其中,X大于0,Y大于0;D2. The bill picture samples are divided into a first data set and a second data set according to a ratio of X:Y (for example, 8:2), and the number of bill picture samples in the first data set is larger than the bill picture sample in the second data set. Quantity, the first data set as a training set, and the second data set as a test set, where X is greater than 0 and Y is greater than 0;
D3、将第一数据集中的票据图片样本送入时间递归神经网络模型进行模型训练,每隔一段时间或预设次数的迭代(例如每进行1000次迭代),对训练得到的模型使用第二数据集进行测试,以评估当前训练的模型效果。测试时,使用训练得到的模型对第二数据集中的票据图片样本进行字符信息识别,并和测试的票据图片样本的名称做对比,以计算识别的结果和标注结果的误差,误差计算采用编辑距离作为计算标准。若训练得到的模型在测试时对票据图片样本的字符信息识别误差出现发散,则调整训练参数并重新训练,使训练时模型对票据图片样本的字符信息识别的误差能够收敛。当误差收敛后,结束模型训练,生成的模型作为最终的该待识别字段对应的字符识别模型。D3. The sample of the bill image in the first data set is sent to the time recurrent neural network model for model training, and the second data is used for the trained model every certain period of time or a preset number of iterations (for example, every 1000 iterations). The set is tested to evaluate the effect of the currently trained model. During the test, the trained model is used to identify the character information of the ticket image sample in the second data set, and compares with the name of the tested ticket picture sample to calculate the error of the recognition result and the labeling result, and the error calculation uses the editing distance. As a calculation standard. If the training model obtains divergence of the character information recognition error of the ticket picture sample during the test, the training parameters are adjusted and retrained, so that the error of the character information recognition of the ticket picture sample can be converged during the training. After the error converges, the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
与现有技术相比,本实施例通过票据图片中各个待识别字段对应的区域识别模型对各个待识别字段在所述票据图片中的行字符区域进行区域识别,识别出包含字符信息且固定宽度为预设值的小框,并将所包含的字符信息处于同一行的小框按顺序拼接形成包含字符信息的目标行字符区域,调用与待识别字段对应的字符识别模型对该目标行字符区域进行字符识别。由于识别出的包含字符信息的行字符区域为统一固定预设值的宽度,这样,可以将字符信息具体到更小的子区域,并对包含字符信息的各个子区域有一个很好的逼近,在利用字符识别模型进行字符识别时的目标行字符区域中除字符信息之外的其它干扰因素会少很多,从而降低票据信息识别的错误率。Compared with the prior art, in this embodiment, the area identification model corresponding to each to-be-identified field in the ticket picture performs area identification on each line character area in the ticket picture, and identifies the character information and the fixed width. a small frame of preset values, and the small boxes containing the character information in the same row are sequentially stitched to form a target line character region containing character information, and the character recognition model corresponding to the field to be recognized is called to the target line character region. Perform character recognition. Since the identified line character area containing the character information is the width of the unified fixed preset value, the character information can be specific to the smaller sub-area, and the sub-area containing the character information has a good approximation. In the target line character area when character recognition is performed by the character recognition model, there are much less interference factors than the character information, thereby reducing the error rate of the ticket information recognition.
在一可选的实施例中,在上述图1的实施例的基础上,所述票据图片识别模型为深度卷积神经网络模型(例如,该深度卷积神经网络模型可以为在CaffeNet的环境下选取的基于深度卷积神经网络SSD(Single Shot MultiBox Detector)算法模型),该深度卷积神经网络模型由1个输入层、13个卷积层、5个池化层、2个全连接层、1个分类层构成。所述深度卷积神经网络模型的详细结构如下表1所示:In an optional embodiment, based on the embodiment of FIG. 1 above, the ticket picture recognition model is a deep convolutional neural network model (eg, the deep convolutional neural network model may be in a CaffeNet environment) The selected deep convolutional neural network SSD (Single Shot MultiBox Detector) algorithm model, the deep convolutional neural network model consists of one input layer, 13 convolutional layers, 5 pooling layers, 2 fully connected layers, One classification layer is formed. The detailed structure of the deep convolutional neural network model is shown in Table 1 below:
表1Table 1
其中:Layer Name表示每一层的名称,Input表示输入层,Conv表示模型的卷积层,Conv1表示模型的第1个卷积层,MaxPool表示模型的最大值池化层,MaxPool1表示模型的第1个最大值池化层,Fc表示模型中的全连接层,Fc1表示模型中第1个全连接层,Softmax表示Softmax分类器;Batch Size表示当前层的输入图像数目;Kernel Size表示当前层卷积核的尺度(例如,Kernel Size可以等于3,表示卷积核的尺度为3x3);Stride Size表示卷积核的移动步长,即做完一次卷积之后移动到下一个卷积位置的距离;Pad Size表示对当前网络层之中的图像填充的大小。需要说明的是,本实施例中池化层的池化方式包括但不限于Mean pooling(均值采样)、Max pooling(最大值采样)、Overlapping(重叠采样)、L2pooling(均方采样)、Local Contrast Normalization(归一化采样)、Stochasticpooling(随即采样)、Def-pooling(形变约束采样)等等。Where: Layer Name indicates the name of each layer, Input indicates the input layer, Conv indicates the convolution layer of the model, Conv1 indicates the first convolution layer of the model, MaxPool indicates the maximum pooling layer of the model, and MaxPool1 indicates the model. 1 maximum pooling layer, Fc represents the fully connected layer in the model, Fc1 represents the first fully connected layer in the model, Softmax represents the Softmax classifier; Batch Size represents the number of input images of the current layer; Kernel Size represents the current layer volume The scale of the kernel (for example, the Kernel Size can be equal to 3, indicating that the scale of the convolution kernel is 3x3); the Stride Size indicates the moving step size of the convolution kernel, that is, the distance moved to the next convolution position after one convolution is completed. ; Pad Size indicates the size of the image fill in the current network layer. It should be noted that the pooling mode of the pooling layer in this embodiment includes, but is not limited to, Mean pooling, Max pooling, Overlapping, L2pooling, Local Contrast. Normalization, Stochasticpooling, Def-pooling, and more.
所述票据图片识别模型的训练过程如下:The training process of the ticket picture recognition model is as follows:
B1、为每一个预设票据类别(例如,预设票据类别可包括门诊票据和住院票据2种)准备预设数量(例如,1000张)的标注有对应的票据类别的票据图片
样本;本实施例中,在训练之前,针对票据图片样本还做如下处理:B1. Prepare a preset number (for example, 1000 sheets) of bill pictures marked with corresponding bill categories for each preset bill category (for example, the preset bill category may include two types of outpatient bills and hospital bills)
Sample; in this embodiment, before the training, the ticket picture sample is also processed as follows:
根据其高宽比信息以及印章的位置判断票据图片的转置情况,并做翻转调整:当高宽比大于1时,说明票据图片高宽颠倒,若印章位置在票据图片左侧,则对票据图像做顺时针旋转九十度处理,若印章位置在票据图片右侧,则对票据图像做逆时针旋转九十度处理;当高宽比小于1时,说明票据图片高宽未颠倒,若印章位置在票据图片下侧,则对票据图像做顺时针旋转一百八十度处理。According to the aspect ratio information and the position of the seal, the transposition of the bill picture is determined, and the flip adjustment is made: when the aspect ratio is greater than 1, the height and width of the bill picture are reversed, and if the stamp position is on the left side of the bill picture, the bill is The image is rotated clockwise by ninety degrees. If the stamp position is on the right side of the bill image, the bill image is rotated counterclockwise by ninety degrees. When the aspect ratio is less than 1, the bill image height and width are not reversed. The position is on the lower side of the ticket picture, and the ticket image is rotated clockwise by one hundred and eighty degrees.
找出标注存在严重问题的数据,比如关键位置信息缺失或超出整张图片范围,以及印章标注位置位于票据中央等明显标注错误的数据,对这些数据进行清理,确保数据标注准确无误。Find out the data with serious problems, such as the missing or beyond the entire image range, and the markedly located data in the center of the ticket, such as the wrongly marked data, to clean up the data to ensure that the data is accurate.
对经过翻转后的标注数据做修正,每个对象的标注数据指的是框出这个对象的矩形框的位置信息,用这个矩形框的左上角坐标(xmin,ymin)和右下角坐标(xmax,ymax)四个数来表示,如果xmax<xmin,则颠倒二者位置,对y坐标做同样的处理,以确保max>min。Correcting the inverted annotation data, the annotation data of each object refers to the position information of the rectangular frame of the object, and the coordinates of the upper left corner of the rectangle (xmin, ymin) and the coordinates of the lower right corner (xmax, Ymax) indicates four numbers. If xmax < xmin, reverse the position and do the same for the y coordinate to ensure max>min.
这样,能保证进行模型训练的票据图片样本均为高宽未颠倒且标注准确无误的票据图片,以利于后续的模型训练更加准确有效。In this way, it is ensured that the sample picture of the ticket for the model training is a picture of the ticket whose height and width are not reversed and marked accurately, so as to facilitate the subsequent model training to be more accurate and effective.
B2、将每一个预设票据类别对应的票据图片样本分为第一比例(例如,80%)的训练子集和第二比例(例如,20%)的验证子集,将各个训练子集中的票据图片样本进行混合以得到训练集,并将各个验证子集中的票据图片样本进行混合以得到验证集;B2, dividing the ticket picture sample corresponding to each preset ticket category into a training subset of a first ratio (for example, 80%) and a verification subset of a second ratio (for example, 20%), and respectively The ticket picture samples are mixed to obtain a training set, and the ticket picture samples in the respective verification subsets are mixed to obtain a verification set;
B3、利用所述训练集训练所述票据图片识别模型;B3. Training the ticket picture recognition model by using the training set;
B4、利用所述验证集验证训练的所述票据图片识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加每一个预设票据类别对应的票据图片样本的数量,并重新执行步骤B2、B3、B4。B4. Using the verification set to verify the accuracy of the ticket picture recognition model of the training, if the accuracy rate is greater than or equal to the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, then each one is added. The number of ticket picture samples corresponding to the preset ticket category is re-executed, and steps B2, B3, and B4 are re-executed.
如图2所示,图2为本申请票据信息识别方法一实施例的流程示意图,该票据信息识别方法包括以下步骤:As shown in FIG. 2, FIG. 2 is a schematic flowchart of a method for identifying a ticket information according to an embodiment of the present invention. The method for identifying a ticket information includes the following steps:
步骤S10,在收到待处理的票据图片后,根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型,针对各个待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值的目标框,并将所包含的字符信息处于同一行的目标框按照识别的先后顺序拼接在一起形成包含字符信息的目标行字符区域。Step S10: After receiving the picture of the ticket to be processed, determining, according to the mapping relationship between the field to be identified and the area identification model, the area recognition model corresponding to each field to be identified in the ticket picture, for each field to be identified, Invoking a corresponding area recognition model to perform area recognition on the line character area of the ticket picture, identifying a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and placing the included character information in the same The target frames of the rows are stitched together in the order of recognition to form a target line character region containing character information.
本实施例中,票据信息识别系统10接收用户通过终端设备2发送的待识别处理的票据图片,该票据图片包括与医疗、商业、金融等保险相关的票据图片,如门诊或住院票据图片。例如,接收用户在手机、平板电脑、自助终端设备等终端设备中预先安装的客户端上发送来的票据图片,或接收用户在手机、平板电脑、自助终端设备等终端中的浏览器系统上发送来的票据图片。In this embodiment, the ticket information identification system 10 receives a bill picture of the to-be-identified processing sent by the user through the terminal device 2, and the bill picture includes a bill picture related to insurance, medical, financial, and the like, such as an outpatient or hospital bill picture. For example, receiving a picture of a ticket sent by a user on a client installed in a terminal device such as a mobile phone, a tablet computer, or a self-service terminal device, or receiving the user to send the message on a browser system in a terminal such as a mobile phone, a tablet computer, or a self-service terminal device. The picture of the bill coming.
预先根据待识别字段的不同类型预设有与之对应的区域识别模型,例如,针对文本类字段对应预设有第一识别模型,针对数字类字段对应预设有第二识别模型,针对日期/时间类字段对应预设有第三识别模型,针对货币类字段对应预设有第四识别模型,等等。这样,在收到待处理的票据图片后,根据预先确定的待识别字段(如文本类字段、数字类字段、日期/时间类字段、货币类字段等等)与区域识别模型的映射关系,确定各个所述待识别字段对应的区域识别模型,针
对各个所述待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值(例如,16个像素宽度)的小框即目标框,并将所包含的字符信息处于同一行的小框按照先后顺序拼接在一起形成包含字符信息的目标行字符区域。其中,在确定各个待识别字段对应的区域识别模型时可包括:A region identification model corresponding to the type of the field to be identified is pre-configured, for example, a first recognition model is pre-set for the text class field, and a second recognition model is preset for the digital class field, for the date/ The time class field is pre-set with a third recognition model, the fourth recognition model is pre-set for the currency class field, and so on. In this way, after receiving the picture of the ticket to be processed, determining according to a predetermined mapping relationship between the to-be-identified field (such as a text class field, a numeric class field, a date/time class field, a currency class field, and the like) and the region recognition model. An area recognition model corresponding to each of the to-be-identified fields,
For each of the to-be-identified fields, the corresponding area recognition model is called to perform area recognition on the line character area of the ticket picture, and the character information is recognized from the ticket picture and the fixed width is a preset value (for example, 16 pieces) The small frame of the pixel width is the target frame, and the small boxes containing the character information in the same line are stitched together in order to form a target line character region containing character information. Wherein, when determining the area recognition model corresponding to each to-be-identified field, the method may include:
A1、在收到待处理的票据图片后,利用预先训练的票据图片识别模型对收到的图片中的票据类别进行识别,并输出票据类别的识别结果(例如,医疗票据的类别包括门诊票据,住院票据,以及其他类票据)。A1. After receiving the picture of the bill to be processed, the pre-trained bill picture recognition model is used to identify the bill type in the received picture, and output the identification result of the bill category (for example, the category of the medical bill includes the outpatient bill, Hospitalization bills, as well as other types of notes).
A2、利用预先确定的矫正规则对收到的票据图片进行倾斜矫正;在一种可选的实施方式中,所述预先确定的矫正规则为:用霍夫变换(Hough)的概率算法找出票据图像中尽可能多的小段直线;从找出的小段直线中确定出所有偏水平的直线,并将确定出的直线中x坐标值相差不大的直线按对应的y坐标值的大小顺序依次相连,按照x坐标值大小分为若干类,或者,将确定出的直线中y坐标值相差不大的直线按对应的x坐标值的大小顺序依次相连,按照y坐标值大小分为若干类;将属于一类的所有水平直线作为一个目标类直线,并通过最小二乘法找出最接近各个目标类直线的长直线;计算出各个长直线的斜率,计算出各个长直线的斜率的中位数和均值,比较计算出的斜率的中位数和均值的大小以确定出较小者,并根据确定出的较小者调整图像倾角,以将收到的票据图片矫正为正常无倾角的图片。A2: Performing a tilt correction on the received ticket image by using a predetermined correction rule; in an optional implementation manner, the predetermined correction rule is: using a Hough probability algorithm to find the ticket As many small straight lines as possible in the image; all straight lines are determined from the found small straight lines, and the straight lines whose x coordinate values are not much different in the determined straight line are sequentially connected in the order of the corresponding y coordinate values. According to the size of the x coordinate value, it is divided into several classes, or the straight lines whose y coordinate values are not much different in the determined straight line are sequentially connected in the order of the corresponding x coordinate values, and are classified into several classes according to the size of the y coordinate value; All horizontal lines belonging to a class are used as a target class line, and the longest line closest to each target class line is found by least squares method; the slope of each long line is calculated, and the median of the slopes of each long line is calculated. Mean, compare the median and mean of the calculated slope to determine the smaller one, and adjust the image tilt according to the smaller one to determine the received bill No inclination correction is normal piece of the picture.
A3、根据预先确定的票据类别与待识别字段的映射关系,确定识别的票据类别对应的待识别字段;A3. Determine, according to a mapping relationship between the predetermined ticket category and the to-be-identified field, a field to be identified corresponding to the identified ticket category;
A4、根据预先确定的待识别字段与区域识别模型的映射关系,确定各个所述待识别字段对应的区域识别模型。A4. Determine, according to a predetermined mapping relationship between the to-be-identified field and the area recognition model, an area recognition model corresponding to each of the to-be-identified fields.
在一种可选的实施方式中,所述区域识别模型为卷积神经网络模型,针对一个待识别字段对应的区域识别模型的训练过程如下:In an optional implementation manner, the area recognition model is a convolutional neural network model, and the training process for the area recognition model corresponding to a field to be identified is as follows:
C1、针对该待识别字段,获取预设数量(例如,10万)的票据图片样本;C1. Obtain a preset number (for example, 100,000) of bill picture samples for the to-be-identified field;
C2、在各个票据图片样本上每隔第一预设数量(例如,16个)的像素,设置第二预设数量(例如,10个)的不同高宽比的且固定宽度为预设值(例如,16个像素宽度)的小框;C2, setting a second preset number (for example, 10) of different aspect ratios and setting a fixed width to a preset value on each of the first bill number samples (for example, 16 pixels) For example, a small frame of 16 pixels wide);
C3、在各个票据图片样本上对包含该待识别字段的部分或者全部字符信息的小框进行标记;C3. Mark a small frame containing part or all of the character information of the to-be-identified field on each ticket picture sample;
C4、将包含该待识别字段的字符信息的票据图片样本归入第一训练集,并将不包含该待识别字段的字符信息的票据图片样本归入第二训练集;C4. The ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;
C5、分别从第一训练集和第二训练集中提取出第一预设比例(例如,80%)的票据图片样本作为待训练的样本图片,并将第一训练集和第二训练集中剩余的票据图片样本作为待验证的样本图片;C5. Extracting, from the first training set and the second training set, a first preset ratio (for example, 80%) of the ticket picture samples as the sample picture to be trained, and remaining the first training set and the second training set. a sample of the bill image as a sample image to be verified;
C6、利用提取的各个待训练的样本图片进行模型训练,以生成所述区域识别模型,并利用各个待验证的样本图片对生成的所述区域识别模型进行验证;C6: performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;
C7、若验证通过率大于等于预设阈值(例如,98%),则训练完成,或者,若验证通过率小于预设阈值,则增加票据图片样本的数量,并重复执行步骤C2、C3、C4、C5、C6。C7. If the verification pass rate is greater than or equal to a preset threshold (for example, 98%), the training is completed, or if the verification pass rate is less than the preset threshold, increase the number of ticket picture samples, and repeat steps C2, C3, and C4. , C5, C6.
步骤S20,根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对各个所述待识别字段的目标行字符区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字
段的目标行字符区域包含的字符信息。Step S20: Determine, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and call a corresponding character recognition model for each of the target line character regions of the to-be-identified field. Character recognition to identify each of the to-be-identified words
The character information contained in the target line character area of the segment.
本实施例中,在利用区域识别模型识别出各个待识别字段的目标行字符区域后,可根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对识别出的各个所述待识别字段的目标行字符区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字段的目标行字符区域包含的字符信息,完成整个票据图片的字符信息识别。In this embodiment, after the target line character area of each to-be-identified field is identified by using the area recognition model, the character recognition corresponding to each of the to-be-identified fields may be determined according to a predetermined mapping relationship between the to-be-identified field and the character recognition model. a model, in response to the identified target line character regions of each of the to-be-identified fields, calling a corresponding character recognition model for character recognition to respectively identify character information included in a target line character region of each of the to-be-identified fields, completing the entire The character information of the ticket picture is identified.
在一种可选的实施方式中,所述字符识别模型为时间递归神经网络模型(Long-Short Term Memory,LSTM),针对一个待识别字段对应的字符识别模型的训练过程如下:In an optional implementation manner, the character recognition model is a Long-Short Term Memory (LSTM), and the training process for a character recognition model corresponding to a field to be identified is as follows:
D1、针对该待识别字段,获取预设数量(例如,10万)的票据图片样本,票据图片样本中仅包含一行该待识别字段的字符信息,字体为黑色,背景为白色,并将各个票据图片样本的名称命名为其所包含的该待识别字段的字符信息;D1. For the to-be-identified field, obtain a preset number (for example, 100,000) of ticket picture samples, where the ticket picture sample contains only one line of character information of the to-be-identified field, the font is black, the background is white, and each ticket is The name of the picture sample is named as the character information of the field to be identified contained therein;
D2、将所述票据图片样本按照X:Y(例如,8:2)的比例分成第一数据集和第二数据集,第一数据集中的票据图片样本数量大于第二数据集中的票据图片样本数量,第一数据集作为训练集,第二数据集作为测试集,其中,X大于0,Y大于0;D2. The bill picture samples are divided into a first data set and a second data set according to a ratio of X:Y (for example, 8:2), and the number of bill picture samples in the first data set is larger than the bill picture sample in the second data set. Quantity, the first data set as a training set, and the second data set as a test set, where X is greater than 0 and Y is greater than 0;
D3、将第一数据集中的票据图片样本送入时间递归神经网络模型进行模型训练,每隔一段时间或预设次数的迭代(例如每进行1000次迭代),对训练得到的模型使用第二数据集进行测试,以评估当前训练的模型效果。测试时,使用训练得到的模型对第二数据集中的票据图片样本进行字符信息识别,并和测试的票据图片样本的名称做对比,以计算识别的结果和标注结果的误差,误差计算采用编辑距离作为计算标准。若训练得到的模型在测试时对票据图片样本的字符信息识别误差出现发散,则调整训练参数并重新训练,使训练时模型对票据图片样本的字符信息识别的误差能够收敛。当误差收敛后,结束模型训练,生成的模型作为最终的该待识别字段对应的字符识别模型。D3. The sample of the bill image in the first data set is sent to the time recurrent neural network model for model training, and the second data is used for the trained model every certain period of time or a preset number of iterations (for example, every 1000 iterations). The set is tested to evaluate the effect of the currently trained model. During the test, the trained model is used to identify the character information of the ticket image sample in the second data set, and compares with the name of the tested ticket picture sample to calculate the error of the recognition result and the labeling result, and the error calculation uses the editing distance. As a calculation standard. If the training model obtains divergence of the character information recognition error of the ticket picture sample during the test, the training parameters are adjusted and retrained, so that the error of the character information recognition of the ticket picture sample can be converged during the training. After the error converges, the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
与现有技术相比,本实施例通过票据图片中各个待识别字段对应的区域识别模型对各个待识别字段在所述票据图片中的行字符区域进行区域识别,识别出包含字符信息且固定宽度为预设值的小框,并将所包含的字符信息处于同一行的小框按顺序拼接形成包含字符信息的目标行字符区域,调用与待识别字段对应的字符识别模型对该目标行字符区域进行字符识别。由于识别出的包含字符信息的行字符区域为统一固定预设值的宽度,这样,可以将字符信息具体到更小的子区域,并对包含字符信息的各个子区域有一个很好的逼近,在利用字符识别模型进行字符识别时的目标行字符区域中除字符信息之外的其它干扰因素会少很多,从而降低票据信息识别的错误率。Compared with the prior art, in this embodiment, the area identification model corresponding to each to-be-identified field in the ticket picture performs area identification on each line character area in the ticket picture, and identifies the character information and the fixed width. a small frame of preset values, and the small boxes containing the character information in the same row are sequentially stitched to form a target line character region containing character information, and the character recognition model corresponding to the field to be recognized is called to the target line character region. Perform character recognition. Since the identified line character area containing the character information is the width of the unified fixed preset value, the character information can be specific to the smaller sub-area, and the sub-area containing the character information has a good approximation. In the target line character area when character recognition is performed by the character recognition model, there are much less interference factors than the character information, thereby reducing the error rate of the ticket information recognition.
在一可选的实施例中,在上述实施例的基础上,所述票据图片识别模型为深度卷积神经网络模型(例如,该深度卷积神经网络模型可以为在CaffeNet的环境下选取的基于深度卷积神经网络SSD(Single Shot MultiBox Detector)算法模型),该深度卷积神经网络模型由1个输入层、13个卷积层、5个池化层、2个全连接层、1个分类层构成。所述深度卷积神经网络模型的详细结构如下表1所示:In an optional embodiment, based on the foregoing embodiment, the ticket picture recognition model is a deep convolutional neural network model (for example, the deep convolutional neural network model may be selected based on a CaffeNet environment) Deep Spool (Single Shot MultiBox Detector) algorithm model, the deep convolutional neural network model consists of 1 input layer, 13 convolutional layers, 5 pooling layers, 2 fully connected layers, 1 classification Layer composition. The detailed structure of the deep convolutional neural network model is shown in Table 1 below:
表1Table 1
其中:Layer Name表示每一层的名称,Input表示输入层,Conv表示模型的卷积层,Conv1表示模型的第1个卷积层,MaxPool表示模型的最大值池化层,MaxPool1表示模型的第1个最大值池化层,Fc表示模型中的全连接层,Fc1表示模型中第1个全连接层,Softmax表示Softmax分类器;Batch Size表示当前层的输入图像数目;Kernel Size表示当前层卷积核的尺度(例如,Kernel Size可以等于3,表示卷积核的尺度为3x3);Stride Size表示卷积核的移动步长,即做完一次卷积之后移动到下一个卷积位置的距离;Pad Size表示对当前网络层之中的图像填充的大小。需要说明的是,本实施例中池化层的池化方式包括但不限于Mean pooling(均值采样)、Max pooling(最大值采样)、Overlapping(重叠采样)、L2pooling(均方采样)、Local Contrast Normalization(归一化采样)、Stochasticpooling(随即采样)、Def-pooling(形变约束采样)等等。Where: Layer Name indicates the name of each layer, Input indicates the input layer, Conv indicates the convolution layer of the model, Conv1 indicates the first convolution layer of the model, MaxPool indicates the maximum pooling layer of the model, and MaxPool1 indicates the model. 1 maximum pooling layer, Fc represents the fully connected layer in the model, Fc1 represents the first fully connected layer in the model, Softmax represents the Softmax classifier; Batch Size represents the number of input images of the current layer; Kernel Size represents the current layer volume The scale of the kernel (for example, the Kernel Size can be equal to 3, indicating that the scale of the convolution kernel is 3x3); the Stride Size indicates the moving step size of the convolution kernel, that is, the distance moved to the next convolution position after one convolution is completed. ; Pad Size indicates the size of the image fill in the current network layer. It should be noted that the pooling mode of the pooling layer in this embodiment includes, but is not limited to, Mean pooling, Max pooling, Overlapping, L2pooling, Local Contrast. Normalization, Stochasticpooling, Def-pooling, and more.
所述票据图片识别模型的训练过程如下:The training process of the ticket picture recognition model is as follows:
B1、为每一个预设票据类别(例如,预设票据类别可包括门诊票据和住院票据2种)准备预设数量(例如,1000张)的标注有对应的票据类别的票据图片样本;本实施例中,在训练之前,针对票据图片样本还做如下处理:
B1, for each preset ticket category (for example, the preset ticket category may include two types of outpatient bills and hospital bills), preparing a preset number (for example, 1000 sheets) of bill picture samples marked with corresponding bill categories; In the example, before the training, the bill image sample is processed as follows:
根据其高宽比信息以及印章的位置判断票据图片的转置情况,并做翻转调整:当高宽比大于1时,说明票据图片高宽颠倒,若印章位置在票据图片左侧,则对票据图像做顺时针旋转九十度处理,若印章位置在票据图片右侧,则对票据图像做逆时针旋转九十度处理;当高宽比小于1时,说明票据图片高宽未颠倒,若印章位置在票据图片下侧,则对票据图像做顺时针旋转一百八十度处理。According to the aspect ratio information and the position of the seal, the transposition of the bill picture is determined, and the flip adjustment is made: when the aspect ratio is greater than 1, the height and width of the bill picture are reversed, and if the stamp position is on the left side of the bill picture, the bill is The image is rotated clockwise by ninety degrees. If the stamp position is on the right side of the bill image, the bill image is rotated counterclockwise by ninety degrees. When the aspect ratio is less than 1, the bill image height and width are not reversed. The position is on the lower side of the ticket picture, and the ticket image is rotated clockwise by one hundred and eighty degrees.
找出标注存在严重问题的数据,比如关键位置信息缺失或超出整张图片范围,以及印章标注位置位于票据中央等明显标注错误的数据,对这些数据进行清理,确保数据标注准确无误。Find out the data with serious problems, such as the missing or beyond the entire image range, and the markedly located data in the center of the ticket, such as the wrongly marked data, to clean up the data to ensure that the data is accurate.
对经过翻转后的标注数据做修正,每个对象的标注数据指的是框出这个对象的矩形框的位置信息,用这个矩形框的左上角坐标(xmin,ymin)和右下角坐标(xmax,ymax)四个数来表示,如果xmax<xmin,则颠倒二者位置,对y坐标做同样的处理,以确保max>min。Correcting the inverted annotation data, the annotation data of each object refers to the position information of the rectangular frame of the object, and the coordinates of the upper left corner of the rectangle (xmin, ymin) and the coordinates of the lower right corner (xmax, Ymax) indicates four numbers. If xmax < xmin, reverse the position and do the same for the y coordinate to ensure max>min.
这样,能保证进行模型训练的票据图片样本均为高宽未颠倒且标注准确无误的票据图片,以利于后续的模型训练更加准确有效。In this way, it is ensured that the sample picture of the ticket for the model training is a picture of the ticket whose height and width are not reversed and marked accurately, so as to facilitate the subsequent model training to be more accurate and effective.
B2、将每一个预设票据类别对应的票据图片样本分为第一比例(例如,80%)的训练子集和第二比例(例如,20%)的验证子集,将各个训练子集中的票据图片样本进行混合以得到训练集,并将各个验证子集中的票据图片样本进行混合以得到验证集;B2, dividing the ticket picture sample corresponding to each preset ticket category into a training subset of a first ratio (for example, 80%) and a verification subset of a second ratio (for example, 20%), and respectively The ticket picture samples are mixed to obtain a training set, and the ticket picture samples in the respective verification subsets are mixed to obtain a verification set;
B3、利用所述训练集训练所述票据图片识别模型;B3. Training the ticket picture recognition model by using the training set;
B4、利用所述验证集验证训练的所述票据图片识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加每一个预设票据类别对应的票据图片样本的数量,并重新执行步骤B2、B3、B4。B4. Using the verification set to verify the accuracy of the ticket picture recognition model of the training, if the accuracy rate is greater than or equal to the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, then each one is added. The number of ticket picture samples corresponding to the preset ticket category is re-executed, and steps B2, B3, and B4 are re-executed.
此外,本申请还提供一种计算机可读存储介质,所述计算机可读存储介质存储有票据信息识别系统,所述票据信息识别系统可被至少一个处理器执行,以使所述至少一个处理器执行如上述实施例中的票据信息识别方法的步骤,该票据信息识别方法的步骤S10、S20、S30等具体实施过程如上文所述,在此不再赘述。Moreover, the present application also provides a computer readable storage medium storing a ticket information identification system, the ticket information identification system being executable by at least one processor to cause the at least one processor The specific implementation process of the steps S10, S20, and S30 of the ticket information identification method is as described above, and is not described here.
需要说明的是,在本文中,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者装置不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者装置所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、方法、物品或者装置中还存在另外的相同要素。It is to be understood that the term "comprises", "comprising", or any other variants thereof, is intended to encompass a non-exclusive inclusion, such that a process, method, article, or device comprising a series of elements includes those elements. It also includes other elements that are not explicitly listed, or elements that are inherent to such a process, method, article, or device. An element that is defined by the phrase "comprising a ..." does not exclude the presence of additional equivalent elements in the process, method, item, or device that comprises the element.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件来实现,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,空调器,或者网络设备等)执行本申请各个实施例所述的方法。Through the description of the above embodiments, those skilled in the art can clearly understand that the foregoing embodiment method can be implemented by means of software plus a necessary general hardware platform, and can also be implemented by hardware, but in many cases, the former is A better implementation. Based on such understanding, the technical solution of the present application, which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM, disk, The optical disc includes a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to perform the methods described in various embodiments of the present application.
以上参照附图说明了本申请的优选实施例,并非因此局限本申请的权利范围。上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。另外,虽然在
流程图中示出了逻辑顺序,但是在某些情况下,可以以不同于此处的顺序执行所示出或描述的步骤。The preferred embodiments of the present application have been described above with reference to the drawings, and are not intended to limit the scope of the application. The serial numbers of the embodiments of the present application are merely for the description, and do not represent the advantages and disadvantages of the embodiments. Also, although at
The logical order is shown in the flowchart, but in some cases the steps shown or described may be performed in a different order than the ones described herein.
本领域技术人员不脱离本申请的范围和实质,可以有多种变型方案实现本申请,比如作为一个实施例的特征可用于另一实施例而得到又一实施例。凡在运用本申请的技术构思之内所作的任何修改、等同替换和改进,均应在本申请的权利范围之内。
A person skilled in the art can implement the present application in various variants without departing from the scope and spirit of the present application. For example, the features of one embodiment can be used in another embodiment to obtain another embodiment. Any modifications, equivalent substitutions and improvements made within the technical concept of the application should be within the scope of the application.
Claims (20)
- 一种电子装置,其特征在于,所述电子装置包括存储器、处理器,所述存储器上存储有可在所述处理器上运行的票据信息识别系统,所述票据信息识别系统被所述处理器执行时实现如下步骤:An electronic device, comprising: a memory, a processor, on the memory, a ticket information recognition system operable on the processor, wherein the ticket information recognition system is The following steps are implemented during execution:在收到待处理的票据图片后,根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型,针对各个待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值的目标框,并将所包含的字符信息处于同一行的目标框按照识别的先后顺序拼接在一起形成包含字符信息的目标行字符区域;After receiving the picture of the ticket to be processed, determining a region recognition model corresponding to each field to be identified in the ticket image according to a predetermined mapping relationship between the field to be identified and the region identification model, and calling corresponding to each field to be identified The area recognition model performs area recognition on the line character area of the ticket picture, and identifies a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and the included character information is in the same line. The frames are spliced together in the order of recognition to form a target line character region containing character information;根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对各个所述待识别字段的目标行字符区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字段的目标行字符区域包含的字符信息。Determining, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and calling a corresponding character recognition model for character recognition for each target character region of the to-be-identified field And respectively identifying character information included in a target line character region of each of the to-be-identified fields.
- 如权利要求1所述的电子装置,其特征在于,所述根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型包括:The electronic device according to claim 1, wherein the determining the region recognition model corresponding to each to-be-identified field in the ticket image according to the mapping relationship between the predetermined to-be-identified field and the region recognition model comprises:A1、利用预先训练的票据图片识别模型对收到的票据图片的票据类别进行识别,并输出票据类别的识别结果;A1, using a pre-trained ticket picture recognition model to identify the ticket type of the received ticket picture, and outputting the recognition result of the ticket category;A2、利用预先确定的矫正规则对收到的票据图片进行倾斜矫正;A2, using a predetermined correction rule to perform tilt correction on the received bill image;A3、根据预先确定的票据类别与待识别字段的映射关系,确定识别的票据类别对应的待识别字段;A3. Determine, according to a mapping relationship between the predetermined ticket category and the to-be-identified field, a field to be identified corresponding to the identified ticket category;A4、根据预先确定的待识别字段与区域识别模型的映射关系,确定各个所述待识别字段对应的区域识别模型。A4. Determine, according to a predetermined mapping relationship between the to-be-identified field and the area recognition model, an area recognition model corresponding to each of the to-be-identified fields.
- 如权利要求1所述的电子装置,其特征在于,所述区域识别模型为卷积神经网络模型,针对一个待识别字段对应的区域识别模型的训练过程如下:The electronic device according to claim 1, wherein the region recognition model is a convolutional neural network model, and the training process for the region recognition model corresponding to a field to be identified is as follows:C1、针对该待识别字段,获取预设数量的票据图片样本;C1. Obtain a preset number of bill picture samples for the to-be-identified field;C2、在各个票据图片样本上每隔第一预设数量的像素,设置第二预设数量的不同高宽比且固定宽度为预设值的小框;C2, setting a second preset number of different height-to-width ratios and a fixed width to a preset value on each of the plurality of ticket image samples;C3、在各个票据图片样本上对包含该待识别字段的字符信息的小框进行标记;C3. Mark a small frame containing the character information of the to-be-identified field on each ticket picture sample;C4、将包含该待识别字段的字符信息的票据图片样本归入第一训练集,并将不包含该待识别字段的字符信息的票据图片样本归入第二训练集;C4. The ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;C5、分别从第一训练集和第二训练集中提取出第一预设比例的票据图片样本作为待训练的样本图片,并将第一训练集和第二训练集中剩余的票据图片样本作为待验证的样本图片;C5. Extracting, from the first training set and the second training set, the first preset ratio of the ticket picture samples as the sample picture to be trained, and using the remaining ticket picture samples in the first training set and the second training set as the to-be-verified Sample pictureC6、利用提取的各个待训练的样本图片进行模型训练,以生成所述区域识别模型,并利用各个待验证的样本图片对生成的所述区域识别模型进行验证;C6: performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;C7、若验证通过率大于或等于预设阈值,则训练完成,或者,若验证通过率小于预设阈值,则增加票据图片样本的数量,并重复执行步骤C2、C3、C4、C5、C6。 C7. If the verification pass rate is greater than or equal to the preset threshold, the training is completed, or if the verification pass rate is less than the preset threshold, the number of ticket picture samples is increased, and steps C2, C3, C4, C5, and C6 are repeatedly executed.
- 如权利要求2所述的电子装置,其特征在于,所述区域识别模型为卷积神经网络模型,针对一个待识别字段对应的区域识别模型的训练过程如下:The electronic device according to claim 2, wherein the region recognition model is a convolutional neural network model, and the training process for the region recognition model corresponding to a field to be identified is as follows:C1、针对该待识别字段,获取预设数量的票据图片样本;C1. Obtain a preset number of bill picture samples for the to-be-identified field;C2、在各个票据图片样本上每隔第一预设数量的像素,设置第二预设数量的不同高宽比且固定宽度为预设值的小框;C2, setting a second preset number of different height-to-width ratios and a fixed width to a preset value on each of the plurality of ticket image samples;C3、在各个票据图片样本上对包含该待识别字段的字符信息的小框进行标记;C3. Mark a small frame containing the character information of the to-be-identified field on each ticket picture sample;C4、将包含该待识别字段的字符信息的票据图片样本归入第一训练集,并将不包含该待识别字段的字符信息的票据图片样本归入第二训练集;C4. The ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;C5、分别从第一训练集和第二训练集中提取出第一预设比例的票据图片样本作为待训练的样本图片,并将第一训练集和第二训练集中剩余的票据图片样本作为待验证的样本图片;C5. Extracting, from the first training set and the second training set, the first preset ratio of the ticket picture samples as the sample picture to be trained, and using the remaining ticket picture samples in the first training set and the second training set as the to-be-verified Sample pictureC6、利用提取的各个待训练的样本图片进行模型训练,以生成所述区域识别模型,并利用各个待验证的样本图片对生成的所述区域识别模型进行验证;C6: performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;C7、若验证通过率大于或等于预设阈值,则训练完成,或者,若验证通过率小于预设阈值,则增加票据图片样本的数量,并重复执行步骤C2、C3、C4、C5、C6。C7. If the verification pass rate is greater than or equal to the preset threshold, the training is completed, or if the verification pass rate is less than the preset threshold, the number of ticket picture samples is increased, and steps C2, C3, C4, C5, and C6 are repeatedly executed.
- 如权利要求1所述的电子装置,其特征在于,所述字符识别模型为时间递归神经网络模型LSTM,针对一个待识别字段对应的字符识别模型的训练过程如下:The electronic device according to claim 1, wherein the character recognition model is a time recurrent neural network model LSTM, and the training process for a character recognition model corresponding to a field to be identified is as follows:D1、针对该待识别字段,获取预设数量的票据图片样本,票据图片样本中仅包含一行该待识别字段的字符信息,并将各个票据图片样本的名称命名为其包含的该待识别字段的字符信息;D1. For the to-be-identified field, obtain a preset number of ticket picture samples, where the ticket picture sample only contains one line of character information of the to-be-identified field, and name each ticket picture sample as the included identification field. Character information;D2、将所述票据图片样本按照X:Y的比例分成第一数据集和第二数据集,第一数据集中的票据图片样本数量大于第二数据集中的票据图片样本数量,将第一数据集作为训练集,第二数据集作为测试集,其中,X大于0,Y大于0;D2, dividing the bill picture sample into a first data set and a second data set according to a ratio of X:Y, wherein the number of bill picture samples in the first data set is greater than the number of bill picture samples in the second data set, and the first data set is As a training set, the second data set is used as a test set, where X is greater than 0 and Y is greater than 0;D3、将第一数据集中的票据图片样本送入预设的时间递归神经网络模型进行模型训练,每隔预设时间或预设次数的迭代,使用训练得到的模型对第二数据集中的票据图片样本进行字符信息识别,并将字符信息识别结果与测试的票据图片样本的名称进行比对,以计算字符信息识别结果的误差;若训练得到的模型对票据图片样本的字符信息识别的误差出现发散,则调整预设训练参数并重新训练模型,直至误差出现收敛,则结束模型训练,生成的模型作为最终的该待识别字段对应的字符识别模型。D3, sending the sample of the bill image in the first data set to a preset time recurrent neural network model for model training, using the trained model to view the bill image in the second data set every preset time or a preset number of iterations The sample performs character information recognition, and compares the character information recognition result with the name of the tested ticket picture sample to calculate the error of the character information recognition result; if the trained model scatters the error of the character information recognition of the ticket picture sample Then, the preset training parameters are adjusted and the model is retrained until the error converges, and the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
- 如权利要求2所述的电子装置,其特征在于,所述字符识别模型为时间递归神经网络模型LSTM,针对一个待识别字段对应的字符识别模型的训练过程如下:The electronic device according to claim 2, wherein the character recognition model is a time recurrent neural network model LSTM, and the training process for a character recognition model corresponding to a field to be identified is as follows:D1、针对该待识别字段,获取预设数量的票据图片样本,票据图片样本中仅包含一行该待识别字段的字符信息,并将各个票据图片样本的名称命名为其包含的该待识别字段的字符信息;D1. For the to-be-identified field, obtain a preset number of ticket picture samples, where the ticket picture sample only contains one line of character information of the to-be-identified field, and name each ticket picture sample as the included identification field. Character information;D2、将所述票据图片样本按照X:Y的比例分成第一数据集和第二数据集,第一数据集中的票据图片样本数量大于第二数据集中的票据图片样本数量,将第一数据集作为训练集,第二数据集作为测试集,其中,X大于0,Y大于0; D2, dividing the bill picture sample into a first data set and a second data set according to a ratio of X:Y, wherein the number of bill picture samples in the first data set is greater than the number of bill picture samples in the second data set, and the first data set is As a training set, the second data set is used as a test set, where X is greater than 0 and Y is greater than 0;D3、将第一数据集中的票据图片样本送入预设的时间递归神经网络模型进行模型训练,每隔预设时间或预设次数的迭代,使用训练得到的模型对第二数据集中的票据图片样本进行字符信息识别,并将字符信息识别结果与测试的票据图片样本的名称进行比对,以计算字符信息识别结果的误差;若训练得到的模型对票据图片样本的字符信息识别的误差出现发散,则调整预设训练参数并重新训练模型,直至误差出现收敛,则结束模型训练,生成的模型作为最终的该待识别字段对应的字符识别模型。D3, sending the sample of the bill image in the first data set to a preset time recurrent neural network model for model training, using the trained model to view the bill image in the second data set every preset time or a preset number of iterations The sample performs character information recognition, and compares the character information recognition result with the name of the tested ticket picture sample to calculate the error of the character information recognition result; if the trained model scatters the error of the character information recognition of the ticket picture sample Then, the preset training parameters are adjusted and the model is retrained until the error converges, and the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
- 如权利要求2所述的电子装置,其特征在于,所述票据图片识别模型为深度卷积神经网络模型,该深度卷积神经网络模型由1个输入层、13个卷积层、5个池化层、2个全连接层、1个分类层构成;所述票据图片识别模型的训练过程如下:The electronic device according to claim 2, wherein the bill picture recognition model is a deep convolutional neural network model, and the deep convolutional neural network model comprises one input layer, 13 convolution layers, and 5 pools. The layer, the two fully connected layers, and one sorting layer are formed; the training process of the bill picture recognition model is as follows:S1、为每一个预设票据类别准备预设数量的标注有对应的票据类别的票据图片样本;S1, preparing a preset number of bill picture samples marked with corresponding bill categories for each preset bill category;S2、将每一个预设票据类别对应的票据图片样本分为第一比例的训练子集和第二比例的验证子集,将各个训练子集中的票据图片样本进行混合以得到训练集,并将各个验证子集中的票据图片样本进行混合以得到验证集;S2, dividing the ticket picture sample corresponding to each preset ticket category into a training subset of the first ratio and a verification subset of the second ratio, mixing the ticket picture samples in each training subset to obtain a training set, and The sample of the ticket pictures in each verification subset is mixed to obtain a verification set;S3、利用所述训练集训练所述票据图片识别模型;S3. Train the ticket picture recognition model by using the training set;S4、利用所述验证集验证训练的所述票据图片识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加每一个预设票据类别对应的票据图片样本的数量,并重新执行步骤S2、S3、S4。S4. Using the verification set to verify the accuracy of the ticket picture recognition model of the training, if the accuracy rate is greater than or equal to the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, then each one is added. The number of ticket picture samples corresponding to the preset ticket category is determined, and steps S2, S3, and S4 are re-executed.
- 一种票据信息识别方法,其特征在于,所述票据信息识别方法包括:A method for identifying a bill information, wherein the bill information identifying method comprises:步骤一、在收到待处理的票据图片后,根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型,针对各个待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值的目标框,并将所包含的字符信息处于同一行的目标框按照识别的先后顺序拼接在一起形成包含字符信息的目标行字符区域;Step 1: After receiving the picture of the bill to be processed, determining the area recognition model corresponding to each field to be identified in the ticket picture according to the mapping relationship between the predetermined field to be identified and the area identification model, for each field to be identified, Invoking a corresponding area recognition model to perform area recognition on the line character area of the ticket picture, identifying a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and placing the included character information in the same The target frames of the rows are stitched together in the order of recognition to form a target line character region containing character information;步骤二、根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对各个所述待识别字段的目标行字符区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字段的目标行字符区域包含的字符信息。Step 2: Determine, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and call a corresponding character recognition model for each of the target line character regions of the to-be-identified field Character recognition is performed to respectively identify character information included in a target line character region of each of the to-be-identified fields.
- 如权利要求8所述的票据信息识别方法,其特征在于,所述根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型包括:The ticket information identifying method according to claim 8, wherein the determining the region recognition model corresponding to each to-be-identified field in the ticket image according to the mapping relationship between the predetermined to-be-identified field and the region recognition model comprises:A1、利用预先训练的票据图片识别模型对收到的票据图片的票据类别进行识别,并输出票据类别的识别结果;A1, using a pre-trained ticket picture recognition model to identify the ticket type of the received ticket picture, and outputting the recognition result of the ticket category;A2、利用预先确定的矫正规则对收到的票据图片进行倾斜矫正;A2, using a predetermined correction rule to perform tilt correction on the received bill image;A3、根据预先确定的票据类别与待识别字段的映射关系,确定识别的票据类别对应的待识别字段;A3. Determine, according to a mapping relationship between the predetermined ticket category and the to-be-identified field, a field to be identified corresponding to the identified ticket category;A4、根据预先确定的待识别字段与区域识别模型的映射关系,确定各个所述 待识别字段对应的区域识别模型。A4. Determine each of the foregoing according to a predetermined mapping relationship between the to-be-identified field and the area recognition model. The area identification model corresponding to the field to be identified.
- 如权利要求8所述的票据信息识别方法,其特征在于,所述区域识别模型为卷积神经网络模型,针对一个待识别字段对应的区域识别模型的训练过程如下:The ticket information identifying method according to claim 8, wherein the area recognition model is a convolutional neural network model, and the training process for the area recognition model corresponding to a field to be identified is as follows:C1、针对该待识别字段,获取预设数量的票据图片样本;C1. Obtain a preset number of bill picture samples for the to-be-identified field;C2、在各个票据图片样本上每隔第一预设数量的像素,设置第二预设数量的不同高宽比且固定宽度为预设值的小框;C2, setting a second preset number of different height-to-width ratios and a fixed width to a preset value on each of the plurality of ticket image samples;C3、在各个票据图片样本上对包含该待识别字段的字符信息的小框进行标记;C3. Mark a small frame containing the character information of the to-be-identified field on each ticket picture sample;C4、将包含该待识别字段的字符信息的票据图片样本归入第一训练集,并将不包含该待识别字段的字符信息的票据图片样本归入第二训练集;C4. The ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;C5、分别从第一训练集和第二训练集中提取出第一预设比例的票据图片样本作为待训练的样本图片,并将第一训练集和第二训练集中剩余的票据图片样本作为待验证的样本图片;C5. Extracting, from the first training set and the second training set, the first preset ratio of the ticket picture samples as the sample picture to be trained, and using the remaining ticket picture samples in the first training set and the second training set as the to-be-verified Sample pictureC6、利用提取的各个待训练的样本图片进行模型训练,以生成所述区域识别模型,并利用各个待验证的样本图片对生成的所述区域识别模型进行验证;C6: performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;C7、若验证通过率大于或等于预设阈值,则训练完成,或者,若验证通过率小于预设阈值,则增加票据图片样本的数量,并重复执行步骤C2、C3、C4、C5、C6。C7. If the verification pass rate is greater than or equal to the preset threshold, the training is completed, or if the verification pass rate is less than the preset threshold, the number of ticket picture samples is increased, and steps C2, C3, C4, C5, and C6 are repeatedly executed.
- 如权利要求9所述的票据信息识别方法,其特征在于,所述区域识别模型为卷积神经网络模型,针对一个待识别字段对应的区域识别模型的训练过程如下:The ticket information identifying method according to claim 9, wherein the area recognition model is a convolutional neural network model, and the training process for the area recognition model corresponding to a field to be identified is as follows:C1、针对该待识别字段,获取预设数量的票据图片样本;C1. Obtain a preset number of bill picture samples for the to-be-identified field;C2、在各个票据图片样本上每隔第一预设数量的像素,设置第二预设数量的不同高宽比且固定宽度为预设值的小框;C2, setting a second preset number of different height-to-width ratios and a fixed width to a preset value on each of the plurality of ticket image samples;C3、在各个票据图片样本上对包含该待识别字段的字符信息的小框进行标记;C3. Mark a small frame containing the character information of the to-be-identified field on each ticket picture sample;C4、将包含该待识别字段的字符信息的票据图片样本归入第一训练集,并将不包含该待识别字段的字符信息的票据图片样本归入第二训练集;C4. The ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;C5、分别从第一训练集和第二训练集中提取出第一预设比例的票据图片样本作为待训练的样本图片,并将第一训练集和第二训练集中剩余的票据图片样本作为待验证的样本图片;C5. Extracting, from the first training set and the second training set, the first preset ratio of the ticket picture samples as the sample picture to be trained, and using the remaining ticket picture samples in the first training set and the second training set as the to-be-verified Sample pictureC6、利用提取的各个待训练的样本图片进行模型训练,以生成所述区域识别模型,并利用各个待验证的样本图片对生成的所述区域识别模型进行验证;C6: performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;C7、若验证通过率大于或等于预设阈值,则训练完成,或者,若验证通过率小于预设阈值,则增加票据图片样本的数量,并重复执行步骤C2、C3、C4、C5、C6。C7. If the verification pass rate is greater than or equal to the preset threshold, the training is completed, or if the verification pass rate is less than the preset threshold, the number of ticket picture samples is increased, and steps C2, C3, C4, C5, and C6 are repeatedly executed.
- 如权利要求8所述的票据信息识别方法,其特征在于,所述字符识别模型为时间递归神经网络模型LSTM,针对一个待识别字段对应的字符识别模型的训练过程如下:The ticket information identifying method according to claim 8, wherein the character recognition model is a time recurrent neural network model LSTM, and the training process for a character recognition model corresponding to a field to be identified is as follows:D1、针对该待识别字段,获取预设数量的票据图片样本,票据图片样本中仅 包含一行该待识别字段的字符信息,并将各个票据图片样本的名称命名为其包含的该待识别字段的字符信息;D1. For the to-be-identified field, obtain a preset number of ticket picture samples, and only the ticket picture sample Character information including a row of the to-be-identified field, and naming the name of each ticket picture sample as the character information of the to-be-identified field;D2、将所述票据图片样本按照X:Y的比例分成第一数据集和第二数据集,第一数据集中的票据图片样本数量大于第二数据集中的票据图片样本数量,将第一数据集作为训练集,第二数据集作为测试集,其中,X大于0,Y大于0;D2, dividing the bill picture sample into a first data set and a second data set according to a ratio of X:Y, wherein the number of bill picture samples in the first data set is greater than the number of bill picture samples in the second data set, and the first data set is As a training set, the second data set is used as a test set, where X is greater than 0 and Y is greater than 0;D3、将第一数据集中的票据图片样本送入预设的时间递归神经网络模型进行模型训练,每隔预设时间或预设次数的迭代,使用训练得到的模型对第二数据集中的票据图片样本进行字符信息识别,并将字符信息识别结果与测试的票据图片样本的名称进行比对,以计算字符信息识别结果的误差;若训练得到的模型对票据图片样本的字符信息识别的误差出现发散,则调整预设训练参数并重新训练模型,直至误差出现收敛,则结束模型训练,生成的模型作为最终的该待识别字段对应的字符识别模型。D3, sending the sample of the bill image in the first data set to a preset time recurrent neural network model for model training, using the trained model to view the bill image in the second data set every preset time or a preset number of iterations The sample performs character information recognition, and compares the character information recognition result with the name of the tested ticket picture sample to calculate the error of the character information recognition result; if the trained model scatters the error of the character information recognition of the ticket picture sample Then, the preset training parameters are adjusted and the model is retrained until the error converges, and the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
- 如权利要求9所述的票据信息识别方法,其特征在于,所述字符识别模型为时间递归神经网络模型LSTM,针对一个待识别字段对应的字符识别模型的训练过程如下:The ticket information identifying method according to claim 9, wherein the character recognition model is a time recurrent neural network model LSTM, and the training process for a character recognition model corresponding to a field to be identified is as follows:D1、针对该待识别字段,获取预设数量的票据图片样本,票据图片样本中仅包含一行该待识别字段的字符信息,并将各个票据图片样本的名称命名为其包含的该待识别字段的字符信息;D1. For the to-be-identified field, obtain a preset number of ticket picture samples, where the ticket picture sample only contains one line of character information of the to-be-identified field, and name each ticket picture sample as the included identification field. Character information;D2、将所述票据图片样本按照X:Y的比例分成第一数据集和第二数据集,第一数据集中的票据图片样本数量大于第二数据集中的票据图片样本数量,将第一数据集作为训练集,第二数据集作为测试集,其中,X大于0,Y大于0;D2, dividing the bill picture sample into a first data set and a second data set according to a ratio of X:Y, wherein the number of bill picture samples in the first data set is greater than the number of bill picture samples in the second data set, and the first data set is As a training set, the second data set is used as a test set, where X is greater than 0 and Y is greater than 0;D3、将第一数据集中的票据图片样本送入预设的时间递归神经网络模型进行模型训练,每隔预设时间或预设次数的迭代,使用训练得到的模型对第二数据集中的票据图片样本进行字符信息识别,并将字符信息识别结果与测试的票据图片样本的名称进行比对,以计算字符信息识别结果的误差;若训练得到的模型对票据图片样本的字符信息识别的误差出现发散,则调整预设训练参数并重新训练模型,直至误差出现收敛,则结束模型训练,生成的模型作为最终的该待识别字段对应的字符识别模型。D3, sending the sample of the bill image in the first data set to a preset time recurrent neural network model for model training, using the trained model to view the bill image in the second data set every preset time or a preset number of iterations The sample performs character information recognition, and compares the character information recognition result with the name of the tested ticket picture sample to calculate the error of the character information recognition result; if the trained model scatters the error of the character information recognition of the ticket picture sample Then, the preset training parameters are adjusted and the model is retrained until the error converges, and the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
- 如权利要求9所述的票据信息识别方法,其特征在于,所述票据图片识别模型为深度卷积神经网络模型,该深度卷积神经网络模型由1个输入层、13个卷积层、5个池化层、2个全连接层、1个分类层构成;所述票据图片识别模型的训练过程如下:The ticket information identifying method according to claim 9, wherein the bill image recognition model is a deep convolutional neural network model, and the deep convolutional neural network model comprises an input layer, 13 convolution layers, and 5 The pooling layer, the two fully connected layers, and one sorting layer are formed; the training process of the bill image recognition model is as follows:S1、为每一个预设票据类别准备预设数量的标注有对应的票据类别的票据图片样本;S1, preparing a preset number of bill picture samples marked with corresponding bill categories for each preset bill category;S2、将每一个预设票据类别对应的票据图片样本分为第一比例的训练子集和第二比例的验证子集,将各个训练子集中的票据图片样本进行混合以得到训练集,并将各个验证子集中的票据图片样本进行混合以得到验证集;S2, dividing the ticket picture sample corresponding to each preset ticket category into a training subset of the first ratio and a verification subset of the second ratio, mixing the ticket picture samples in each training subset to obtain a training set, and The sample of the ticket pictures in each verification subset is mixed to obtain a verification set;S3、利用所述训练集训练所述票据图片识别模型;S3. Train the ticket picture recognition model by using the training set;S4、利用所述验证集验证训练的所述票据图片识别模型的准确率,若准确率大于或者等于预设准确率,则训练结束,或者,若准确率小于预设准确率,则增加每一个预设票据类别对应的票据图片样本的数量,并重新执行步骤S2、S3、S4。 S4. Using the verification set to verify the accuracy of the ticket picture recognition model of the training, if the accuracy rate is greater than or equal to the preset accuracy rate, the training ends, or if the accuracy rate is less than the preset accuracy rate, then each one is added. The number of ticket picture samples corresponding to the preset ticket category is determined, and steps S2, S3, and S4 are re-executed.
- 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有票据信息识别系统,所述票据信息识别系统可被至少一个处理器执行,以使所述至少一个处理器执行如下步骤:A computer readable storage medium, characterized in that the computer readable storage medium stores a ticket information identification system, the ticket information identification system being executable by at least one processor to cause the at least one processor to execute The following steps:在收到待处理的票据图片后,根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型,针对各个待识别字段,调用对应的区域识别模型对所述票据图片的行字符区域进行区域识别,从所述票据图片上识别出包含字符信息且固定宽度为预设值的目标框,并将所包含的字符信息处于同一行的目标框按照识别的先后顺序拼接在一起形成包含字符信息的目标行字符区域;After receiving the picture of the ticket to be processed, determining a region recognition model corresponding to each field to be identified in the ticket image according to a predetermined mapping relationship between the field to be identified and the region identification model, and calling corresponding to each field to be identified The area recognition model performs area recognition on the line character area of the ticket picture, and identifies a target frame containing the character information and having a fixed width as a preset value from the ticket picture, and the included character information is in the same line. The frames are spliced together in the order of recognition to form a target line character region containing character information;根据预先确定的待识别字段与字符识别模型的映射关系,确定各个所述待识别字段对应的字符识别模型,针对各个所述待识别字段的目标行字符区域,调用对应的字符识别模型进行字符识别,以分别识别出各个所述待识别字段的目标行字符区域包含的字符信息。Determining, according to a predetermined mapping relationship between the to-be-identified field and the character recognition model, a character recognition model corresponding to each of the to-be-identified fields, and calling a corresponding character recognition model for character recognition for each target character region of the to-be-identified field And respectively identifying character information included in a target line character region of each of the to-be-identified fields.
- 如权利要求15所述的计算机可读存储介质,其特征在于,所述根据预先确定的待识别字段与区域识别模型的映射关系,确定所述票据图片中各个待识别字段对应的区域识别模型包括:The computer readable storage medium according to claim 15, wherein the determining, according to a predetermined mapping relationship between the to-be-identified field and the region recognition model, the region recognition model corresponding to each of the to-be-identified fields in the ticket image comprises: :A1、利用预先训练的票据图片识别模型对收到的票据图片的票据类别进行识别,并输出票据类别的识别结果;A1, using a pre-trained ticket picture recognition model to identify the ticket type of the received ticket picture, and outputting the recognition result of the ticket category;A2、利用预先确定的矫正规则对收到的票据图片进行倾斜矫正;A2, using a predetermined correction rule to perform tilt correction on the received bill image;A3、根据预先确定的票据类别与待识别字段的映射关系,确定识别的票据类别对应的待识别字段;A3. Determine, according to a mapping relationship between the predetermined ticket category and the to-be-identified field, a field to be identified corresponding to the identified ticket category;A4、根据预先确定的待识别字段与区域识别模型的映射关系,确定各个所述待识别字段对应的区域识别模型。A4. Determine, according to a predetermined mapping relationship between the to-be-identified field and the area recognition model, an area recognition model corresponding to each of the to-be-identified fields.
- 如权利要求15所述的计算机可读存储介质,其特征在于,所述区域识别模型为卷积神经网络模型,针对一个待识别字段对应的区域识别模型的训练过程如下:The computer readable storage medium according to claim 15, wherein the region recognition model is a convolutional neural network model, and the training process for the region recognition model corresponding to a field to be identified is as follows:C1、针对该待识别字段,获取预设数量的票据图片样本;C1. Obtain a preset number of bill picture samples for the to-be-identified field;C2、在各个票据图片样本上每隔第一预设数量的像素,设置第二预设数量的不同高宽比且固定宽度为预设值的小框;C2, setting a second preset number of different height-to-width ratios and a fixed width to a preset value on each of the plurality of ticket image samples;C3、在各个票据图片样本上对包含该待识别字段的字符信息的小框进行标记;C3. Mark a small frame containing the character information of the to-be-identified field on each ticket picture sample;C4、将包含该待识别字段的字符信息的票据图片样本归入第一训练集,并将不包含该待识别字段的字符信息的票据图片样本归入第二训练集;C4. The ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;C5、分别从第一训练集和第二训练集中提取出第一预设比例的票据图片样本作为待训练的样本图片,并将第一训练集和第二训练集中剩余的票据图片样本作为待验证的样本图片;C5. Extracting, from the first training set and the second training set, the first preset ratio of the ticket picture samples as the sample picture to be trained, and using the remaining ticket picture samples in the first training set and the second training set as the to-be-verified Sample pictureC6、利用提取的各个待训练的样本图片进行模型训练,以生成所述区域识别模型,并利用各个待验证的样本图片对生成的所述区域识别模型进行验证;C6: performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;C7、若验证通过率大于或等于预设阈值,则训练完成,或者,若验证通过率小于预设阈值,则增加票据图片样本的数量,并重复执行步骤C2、C3、C4、C5、C6。 C7. If the verification pass rate is greater than or equal to the preset threshold, the training is completed, or if the verification pass rate is less than the preset threshold, the number of ticket picture samples is increased, and steps C2, C3, C4, C5, and C6 are repeatedly executed.
- 如权利要求16所述的计算机可读存储介质,其特征在于,所述区域识别模型为卷积神经网络模型,针对一个待识别字段对应的区域识别模型的训练过程如下:The computer readable storage medium according to claim 16, wherein the region recognition model is a convolutional neural network model, and the training process for the region recognition model corresponding to a field to be identified is as follows:C1、针对该待识别字段,获取预设数量的票据图片样本;C1. Obtain a preset number of bill picture samples for the to-be-identified field;C2、在各个票据图片样本上每隔第一预设数量的像素,设置第二预设数量的不同高宽比且固定宽度为预设值的小框;C2, setting a second preset number of different height-to-width ratios and a fixed width to a preset value on each of the plurality of ticket image samples;C3、在各个票据图片样本上对包含该待识别字段的字符信息的小框进行标记;C3. Mark a small frame containing the character information of the to-be-identified field on each ticket picture sample;C4、将包含该待识别字段的字符信息的票据图片样本归入第一训练集,并将不包含该待识别字段的字符信息的票据图片样本归入第二训练集;C4. The ticket picture sample containing the character information of the to-be-identified field is classified into the first training set, and the ticket picture sample that does not include the character information of the to-be-identified field is classified into the second training set;C5、分别从第一训练集和第二训练集中提取出第一预设比例的票据图片样本作为待训练的样本图片,并将第一训练集和第二训练集中剩余的票据图片样本作为待验证的样本图片;C5. Extracting, from the first training set and the second training set, the first preset ratio of the ticket picture samples as the sample picture to be trained, and using the remaining ticket picture samples in the first training set and the second training set as the to-be-verified Sample pictureC6、利用提取的各个待训练的样本图片进行模型训练,以生成所述区域识别模型,并利用各个待验证的样本图片对生成的所述区域识别模型进行验证;C6: performing model training by using the extracted sample images to be trained to generate the region recognition model, and verifying the generated region recognition model by using each sample image to be verified;C7、若验证通过率大于或等于预设阈值,则训练完成,或者,若验证通过率小于预设阈值,则增加票据图片样本的数量,并重复执行步骤C2、C3、C4、C5、C6。C7. If the verification pass rate is greater than or equal to the preset threshold, the training is completed, or if the verification pass rate is less than the preset threshold, the number of ticket picture samples is increased, and steps C2, C3, C4, C5, and C6 are repeatedly executed.
- 如权利要求15所述的计算机可读存储介质,其特征在于,所述字符识别模型为时间递归神经网络模型LSTM,针对一个待识别字段对应的字符识别模型的训练过程如下:The computer readable storage medium according to claim 15, wherein the character recognition model is a time recurrent neural network model LSTM, and the training process for a character recognition model corresponding to a field to be identified is as follows:D1、针对该待识别字段,获取预设数量的票据图片样本,票据图片样本中仅包含一行该待识别字段的字符信息,并将各个票据图片样本的名称命名为其包含的该待识别字段的字符信息;D1. For the to-be-identified field, obtain a preset number of ticket picture samples, where the ticket picture sample only contains one line of character information of the to-be-identified field, and name each ticket picture sample as the included identification field. Character information;D2、将所述票据图片样本按照X:Y的比例分成第一数据集和第二数据集,第一数据集中的票据图片样本数量大于第二数据集中的票据图片样本数量,将第一数据集作为训练集,第二数据集作为测试集,其中,X大于0,Y大于0;D2, dividing the bill picture sample into a first data set and a second data set according to a ratio of X:Y, wherein the number of bill picture samples in the first data set is greater than the number of bill picture samples in the second data set, and the first data set is As a training set, the second data set is used as a test set, where X is greater than 0 and Y is greater than 0;D3、将第一数据集中的票据图片样本送入预设的时间递归神经网络模型进行模型训练,每隔预设时间或预设次数的迭代,使用训练得到的模型对第二数据集中的票据图片样本进行字符信息识别,并将字符信息识别结果与测试的票据图片样本的名称进行比对,以计算字符信息识别结果的误差;若训练得到的模型对票据图片样本的字符信息识别的误差出现发散,则调整预设训练参数并重新训练模型,直至误差出现收敛,则结束模型训练,生成的模型作为最终的该待识别字段对应的字符识别模型。D3, sending the sample of the bill image in the first data set to a preset time recurrent neural network model for model training, using the trained model to view the bill image in the second data set every preset time or a preset number of iterations The sample performs character information recognition, and compares the character information recognition result with the name of the tested ticket picture sample to calculate the error of the character information recognition result; if the trained model scatters the error of the character information recognition of the ticket picture sample Then, the preset training parameters are adjusted and the model is retrained until the error converges, and the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
- 如权利要求16所述的计算机可读存储介质,其特征在于,所述字符识别模型为时间递归神经网络模型LSTM,针对一个待识别字段对应的字符识别模型的训练过程如下:The computer readable storage medium according to claim 16, wherein the character recognition model is a time recurrent neural network model LSTM, and the training process for a character recognition model corresponding to a field to be identified is as follows:D1、针对该待识别字段,获取预设数量的票据图片样本,票据图片样本中仅包含一行该待识别字段的字符信息,并将各个票据图片样本的名称命名为其包含的该待识别字段的字符信息;D1. For the to-be-identified field, obtain a preset number of ticket picture samples, where the ticket picture sample only contains one line of character information of the to-be-identified field, and name each ticket picture sample as the included identification field. Character information;D2、将所述票据图片样本按照X:Y的比例分成第一数据集和第二数据集, 第一数据集中的票据图片样本数量大于第二数据集中的票据图片样本数量,将第一数据集作为训练集,第二数据集作为测试集,其中,X大于0,Y大于0;D2, dividing the bill picture sample into a first data set and a second data set according to a ratio of X:Y, The number of ticket picture samples in the first data set is greater than the number of ticket picture samples in the second data set, the first data set is used as a training set, and the second data set is used as a test set, where X is greater than 0 and Y is greater than 0;D3、将第一数据集中的票据图片样本送入预设的时间递归神经网络模型进行模型训练,每隔预设时间或预设次数的迭代,使用训练得到的模型对第二数据集中的票据图片样本进行字符信息识别,并将字符信息识别结果与测试的票据图片样本的名称进行比对,以计算字符信息识别结果的误差;若训练得到的模型对票据图片样本的字符信息识别的误差出现发散,则调整预设训练参数并重新训练模型,直至误差出现收敛,则结束模型训练,生成的模型作为最终的该待识别字段对应的字符识别模型。 D3, sending the sample of the bill image in the first data set to a preset time recurrent neural network model for model training, using the trained model to view the bill image in the second data set every preset time or a preset number of iterations The sample performs character information recognition, and compares the character information recognition result with the name of the tested ticket picture sample to calculate the error of the character information recognition result; if the trained model scatters the error of the character information recognition of the ticket picture sample Then, the preset training parameters are adjusted and the model is retrained until the error converges, and the model training is ended, and the generated model is used as the final character recognition model corresponding to the to-be-identified field.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710930679.8A CN107798299B (en) | 2017-10-09 | 2017-10-09 | Bill information identification method, electronic device and readable storage medium |
CN201710930679.8 | 2017-10-09 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019071660A1 true WO2019071660A1 (en) | 2019-04-18 |
Family
ID=61533966
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2017/108735 WO2019071660A1 (en) | 2017-10-09 | 2017-10-31 | Bill information identification method, electronic device, and readable storage medium |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN107798299B (en) |
WO (1) | WO2019071660A1 (en) |
Cited By (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110147791A (en) * | 2019-05-20 | 2019-08-20 | 上海联影医疗科技有限公司 | Character recognition method, device, equipment and storage medium |
CN110298347A (en) * | 2019-05-30 | 2019-10-01 | 长安大学 | A kind of recognition methods of the automobile exhaust analyzer screen based on GrayWorld and PCA-CNN |
CN110503105A (en) * | 2019-09-02 | 2019-11-26 | 苏州美能华智能科技有限公司 | Character identifying method, training data acquisition methods, device and medium |
CN110766050A (en) * | 2019-09-19 | 2020-02-07 | 北京捷通华声科技股份有限公司 | Model generation method, text recognition method, device, equipment and storage medium |
CN110941717A (en) * | 2019-11-22 | 2020-03-31 | 深圳马可孛罗科技有限公司 | Passenger ticket rule analysis method and device, electronic equipment and computer readable medium |
CN110991456A (en) * | 2019-12-05 | 2020-04-10 | 北京百度网讯科技有限公司 | Bill identification method and device |
CN111192031A (en) * | 2019-12-26 | 2020-05-22 | 腾讯科技(深圳)有限公司 | Electronic bill generation method and device, electronic equipment and readable storage medium |
CN111223481A (en) * | 2020-01-09 | 2020-06-02 | 腾讯科技(深圳)有限公司 | Information extraction method and device, computer readable storage medium and electronic equipment |
CN111259889A (en) * | 2020-01-17 | 2020-06-09 | 平安医疗健康管理股份有限公司 | Image text recognition method and device, computer equipment and computer storage medium |
CN111325207A (en) * | 2020-03-05 | 2020-06-23 | 中国银行股份有限公司 | Bill identification method and device based on preprocessing |
CN111414908A (en) * | 2020-03-16 | 2020-07-14 | 湖南快乐阳光互动娱乐传媒有限公司 | Method and device for recognizing caption characters in video |
CN111666932A (en) * | 2020-05-27 | 2020-09-15 | 平安科技(深圳)有限公司 | Document auditing method and device, computer equipment and storage medium |
CN111695558A (en) * | 2020-04-28 | 2020-09-22 | 深圳市跨越新科技有限公司 | Logistics waybill picture rectification method and system based on YoloV3 model |
CN111738326A (en) * | 2020-06-16 | 2020-10-02 | 中国工商银行股份有限公司 | Sentence granularity marking training sample generation method and device |
CN111814833A (en) * | 2020-06-11 | 2020-10-23 | 浙江大华技术股份有限公司 | Training method of bill processing model, image processing method and image processing equipment |
CN112270224A (en) * | 2020-10-14 | 2021-01-26 | 招商银行股份有限公司 | Insurance responsibility analysis method and device and computer readable storage medium |
CN112633275A (en) * | 2020-12-22 | 2021-04-09 | 航天信息股份有限公司 | Multi-bill mixed-shooting image correction method and system based on deep learning |
CN112686262A (en) * | 2020-12-28 | 2021-04-20 | 广州博士信息技术研究院有限公司 | Method for extracting structured data and rapidly archiving handbooks based on image recognition technology |
CN112699871A (en) * | 2020-12-23 | 2021-04-23 | 平安银行股份有限公司 | Method, system, device and computer readable storage medium for field content identification |
CN113205041A (en) * | 2021-04-29 | 2021-08-03 | 百度在线网络技术(北京)有限公司 | Structured information extraction method, device, equipment and storage medium |
CN113283421A (en) * | 2021-06-24 | 2021-08-20 | 中国平安人寿保险股份有限公司 | Information identification method, device, equipment and storage medium |
CN113408516A (en) * | 2021-06-25 | 2021-09-17 | 京东数科海益信息科技有限公司 | Bill recognition device and method |
CN113553883A (en) * | 2020-04-24 | 2021-10-26 | 上海高德威智能交通系统有限公司 | Bill image identification method and device and electronic equipment |
CN113762292A (en) * | 2020-06-03 | 2021-12-07 | 杭州海康威视数字技术股份有限公司 | Training data acquisition method and device and model training method and device |
CN114328831A (en) * | 2021-12-24 | 2022-04-12 | 江苏银承网络科技股份有限公司 | Bill information identification and error correction method and device |
CN111626279B (en) * | 2019-10-15 | 2023-06-02 | 西安网算数据科技有限公司 | Negative sample labeling training method and highly-automatic bill identification method |
CN118134576A (en) * | 2024-05-08 | 2024-06-04 | 山东工程职业技术大学 | Digital electronic invoice management method and system based on artificial intelligence |
Families Citing this family (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108446621A (en) * | 2018-03-14 | 2018-08-24 | 平安科技(深圳)有限公司 | Bank slip recognition method, server and computer readable storage medium |
CN108595544A (en) * | 2018-04-09 | 2018-09-28 | 深源恒际科技有限公司 | A kind of document picture classification method |
CN108564035B (en) * | 2018-04-13 | 2020-09-25 | 杭州睿琪软件有限公司 | Method and system for identifying information recorded on document |
CN108629560A (en) * | 2018-04-18 | 2018-10-09 | 平安科技(深圳)有限公司 | Task distributing method, electronic equipment and storage medium |
CN108664897A (en) * | 2018-04-18 | 2018-10-16 | 平安科技(深圳)有限公司 | Bank slip recognition method, apparatus and storage medium |
CN108717543B (en) * | 2018-05-14 | 2022-01-14 | 北京市商汤科技开发有限公司 | Invoice identification method and device and computer storage medium |
CN110674831B (en) * | 2018-06-14 | 2023-01-06 | 佛山市顺德区美的电热电器制造有限公司 | Data processing method and device and computer readable storage medium |
CN110619252B (en) * | 2018-06-19 | 2022-11-04 | 百度在线网络技术(北京)有限公司 | Method, device and equipment for identifying form data in picture and storage medium |
CN108846379A (en) * | 2018-07-03 | 2018-11-20 | 南京览笛信息科技有限公司 | Face list recognition methods, system, terminal device and storage medium |
CN108960245B (en) * | 2018-07-13 | 2022-04-19 | 广东工业大学 | Tire mold character detection and recognition method, device, equipment and storage medium |
CN109214382A (en) * | 2018-07-16 | 2019-01-15 | 顺丰科技有限公司 | A kind of billing information recognizer, equipment and storage medium based on CRNN |
CN109255300B (en) * | 2018-08-14 | 2023-12-01 | 中国平安财产保险股份有限公司 | Bill information extraction method, bill information extraction device, computer equipment and storage medium |
CN109214385B (en) * | 2018-08-15 | 2021-06-08 | 腾讯科技(深圳)有限公司 | Data acquisition method, data acquisition device and storage medium |
CN109271980A (en) * | 2018-08-28 | 2019-01-25 | 上海萃舟智能科技有限公司 | A kind of vehicle nameplate full information recognition methods, system, terminal and medium |
CN109492143A (en) * | 2018-09-21 | 2019-03-19 | 平安科技(深圳)有限公司 | Image processing method, device, computer equipment and storage medium |
CN109784339A (en) * | 2018-12-13 | 2019-05-21 | 平安普惠企业管理有限公司 | Picture recognition test method, device, computer equipment and storage medium |
CN109815949A (en) * | 2018-12-20 | 2019-05-28 | 航天信息股份有限公司 | Invoice publicity method and system neural network based |
CN109858275A (en) * | 2018-12-20 | 2019-06-07 | 航天信息股份有限公司 | Invoice publicity method and system neural network based |
CN109598272B (en) * | 2019-01-11 | 2021-08-06 | 北京字节跳动网络技术有限公司 | Character line image recognition method, device, equipment and medium |
CN109858420A (en) * | 2019-01-24 | 2019-06-07 | 国信电子票据平台信息服务有限公司 | A kind of bill processing system and processing method |
CN109902737A (en) * | 2019-02-25 | 2019-06-18 | 厦门商集网络科技有限责任公司 | A kind of bill classification method and terminal |
CN110119741B (en) * | 2019-04-08 | 2022-09-27 | 浙江大学宁波理工学院 | Card image information identification method with background |
CN110956739A (en) * | 2019-05-09 | 2020-04-03 | 杭州睿琪软件有限公司 | Bill identification method and device |
CN110288755B (en) * | 2019-05-21 | 2023-05-23 | 平安银行股份有限公司 | Invoice checking method based on text recognition, server and storage medium |
CN110334596B (en) * | 2019-05-30 | 2024-02-02 | 平安科技(深圳)有限公司 | Invoice picture summarizing method, electronic device and readable storage medium |
CN110490193B (en) * | 2019-07-24 | 2022-11-08 | 西安网算数据科技有限公司 | Single character area detection method and bill content identification method |
CN110503054B (en) * | 2019-08-27 | 2022-09-23 | 广东工业大学 | Text image processing method and device |
CN110598686B (en) * | 2019-09-17 | 2023-08-04 | 携程计算机技术(上海)有限公司 | Invoice identification method, system, electronic equipment and medium |
CN110866495B (en) * | 2019-11-14 | 2022-06-28 | 杭州睿琪软件有限公司 | Bill image recognition method, bill image recognition device, bill image recognition equipment, training method and storage medium |
CN111104481B (en) * | 2019-12-17 | 2023-10-10 | 东软集团股份有限公司 | Method, device and equipment for identifying matching field |
CN111242790B (en) * | 2020-01-02 | 2020-11-17 | 平安科技(深圳)有限公司 | Risk identification method, electronic device and storage medium |
CN111461099A (en) * | 2020-03-27 | 2020-07-28 | 重庆农村商业银行股份有限公司 | Bill identification method, system, equipment and readable storage medium |
CN111695559B (en) * | 2020-04-28 | 2023-07-18 | 深圳市跨越新科技有限公司 | YoloV3 model-based waybill picture information coding method and system |
CN111563502B (en) * | 2020-05-09 | 2023-12-15 | 腾讯科技(深圳)有限公司 | Image text recognition method and device, electronic equipment and computer storage medium |
CN111695439B (en) * | 2020-05-20 | 2024-05-10 | 平安科技(深圳)有限公司 | Image structured data extraction method, electronic device and storage medium |
CN111931664B (en) * | 2020-08-12 | 2024-01-12 | 腾讯科技(深圳)有限公司 | Mixed-pasting bill image processing method and device, computer equipment and storage medium |
CN112115932B (en) * | 2020-08-19 | 2023-11-14 | 泰康保险集团股份有限公司 | Text extraction method and device, electronic equipment and storage medium |
CN112308036A (en) * | 2020-11-25 | 2021-02-02 | 杭州睿胜软件有限公司 | Bill identification method and device and readable storage medium |
CN112434689A (en) * | 2020-12-01 | 2021-03-02 | 天冕信息技术(深圳)有限公司 | Method, device and equipment for identifying information in picture and storage medium |
CN114627456A (en) * | 2020-12-10 | 2022-06-14 | 航天信息股份有限公司 | Bill text information detection method, device and system |
CN113205049A (en) * | 2021-05-07 | 2021-08-03 | 开放智能机器(上海)有限公司 | Document identification method and identification system |
CN113762152A (en) * | 2021-09-07 | 2021-12-07 | 上海盈策信息技术有限公司 | Bill verification method, system, equipment and medium |
CN116702024B (en) * | 2023-05-16 | 2024-05-28 | 见知数据科技(上海)有限公司 | Method, device, computer equipment and storage medium for identifying type of stream data |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104112128A (en) * | 2014-06-19 | 2014-10-22 | 中国工商银行股份有限公司 | Digital image processing system applied to bill image character recognition and method |
CN105260733A (en) * | 2015-09-11 | 2016-01-20 | 北京百度网讯科技有限公司 | Method and device for processing image information |
CN105654127A (en) * | 2015-12-30 | 2016-06-08 | 成都数联铭品科技有限公司 | End-to-end-based picture character sequence continuous recognition method |
CN107220648A (en) * | 2017-04-11 | 2017-09-29 | 平安科技(深圳)有限公司 | The character identifying method and server of Claims Resolution document |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120201472A1 (en) * | 2011-02-08 | 2012-08-09 | Autonomy Corporation Ltd | System for the tagging and augmentation of geographically-specific locations using a visual data stream |
US9398210B2 (en) * | 2011-02-24 | 2016-07-19 | Digimarc Corporation | Methods and systems for dealing with perspective distortion in connection with smartphone cameras |
US8582873B2 (en) * | 2011-06-16 | 2013-11-12 | Tandent Vision Science, Inc. | Use of an object database in an image process |
-
2017
- 2017-10-09 CN CN201710930679.8A patent/CN107798299B/en active Active
- 2017-10-31 WO PCT/CN2017/108735 patent/WO2019071660A1/en active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104112128A (en) * | 2014-06-19 | 2014-10-22 | 中国工商银行股份有限公司 | Digital image processing system applied to bill image character recognition and method |
CN105260733A (en) * | 2015-09-11 | 2016-01-20 | 北京百度网讯科技有限公司 | Method and device for processing image information |
CN105654127A (en) * | 2015-12-30 | 2016-06-08 | 成都数联铭品科技有限公司 | End-to-end-based picture character sequence continuous recognition method |
CN107220648A (en) * | 2017-04-11 | 2017-09-29 | 平安科技(深圳)有限公司 | The character identifying method and server of Claims Resolution document |
Cited By (41)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110147791A (en) * | 2019-05-20 | 2019-08-20 | 上海联影医疗科技有限公司 | Character recognition method, device, equipment and storage medium |
CN110298347B (en) * | 2019-05-30 | 2022-11-01 | 长安大学 | Method for identifying automobile exhaust analyzer screen based on GrayWorld and PCA-CNN |
CN110298347A (en) * | 2019-05-30 | 2019-10-01 | 长安大学 | A kind of recognition methods of the automobile exhaust analyzer screen based on GrayWorld and PCA-CNN |
CN110503105A (en) * | 2019-09-02 | 2019-11-26 | 苏州美能华智能科技有限公司 | Character identifying method, training data acquisition methods, device and medium |
CN110766050A (en) * | 2019-09-19 | 2020-02-07 | 北京捷通华声科技股份有限公司 | Model generation method, text recognition method, device, equipment and storage medium |
CN110766050B (en) * | 2019-09-19 | 2023-05-23 | 北京捷通华声科技股份有限公司 | Model generation method, text recognition method, device, equipment and storage medium |
CN111626279B (en) * | 2019-10-15 | 2023-06-02 | 西安网算数据科技有限公司 | Negative sample labeling training method and highly-automatic bill identification method |
CN110941717A (en) * | 2019-11-22 | 2020-03-31 | 深圳马可孛罗科技有限公司 | Passenger ticket rule analysis method and device, electronic equipment and computer readable medium |
CN110941717B (en) * | 2019-11-22 | 2023-08-11 | 深圳马可孛罗科技有限公司 | Passenger ticket rule analysis method and device, electronic equipment and computer readable medium |
CN110991456A (en) * | 2019-12-05 | 2020-04-10 | 北京百度网讯科技有限公司 | Bill identification method and device |
CN110991456B (en) * | 2019-12-05 | 2023-07-07 | 北京百度网讯科技有限公司 | Bill identification method and device |
CN111192031A (en) * | 2019-12-26 | 2020-05-22 | 腾讯科技(深圳)有限公司 | Electronic bill generation method and device, electronic equipment and readable storage medium |
CN111192031B (en) * | 2019-12-26 | 2023-06-23 | 腾讯科技(深圳)有限公司 | Electronic bill generation method and device, electronic equipment and readable storage medium |
CN111223481A (en) * | 2020-01-09 | 2020-06-02 | 腾讯科技(深圳)有限公司 | Information extraction method and device, computer readable storage medium and electronic equipment |
CN111223481B (en) * | 2020-01-09 | 2023-10-13 | 腾讯科技(深圳)有限公司 | Information extraction method, information extraction device, computer readable storage medium and electronic equipment |
CN111259889A (en) * | 2020-01-17 | 2020-06-09 | 平安医疗健康管理股份有限公司 | Image text recognition method and device, computer equipment and computer storage medium |
CN111325207A (en) * | 2020-03-05 | 2020-06-23 | 中国银行股份有限公司 | Bill identification method and device based on preprocessing |
CN111414908A (en) * | 2020-03-16 | 2020-07-14 | 湖南快乐阳光互动娱乐传媒有限公司 | Method and device for recognizing caption characters in video |
CN111414908B (en) * | 2020-03-16 | 2023-08-29 | 湖南快乐阳光互动娱乐传媒有限公司 | Method and device for recognizing caption characters in video |
CN113553883A (en) * | 2020-04-24 | 2021-10-26 | 上海高德威智能交通系统有限公司 | Bill image identification method and device and electronic equipment |
CN111695558B (en) * | 2020-04-28 | 2023-08-04 | 深圳市跨越新科技有限公司 | Logistics shipping list picture correction method and system based on YoloV3 model |
CN111695558A (en) * | 2020-04-28 | 2020-09-22 | 深圳市跨越新科技有限公司 | Logistics waybill picture rectification method and system based on YoloV3 model |
CN111666932B (en) * | 2020-05-27 | 2023-07-14 | 平安科技(深圳)有限公司 | Document auditing method, device, computer equipment and storage medium |
CN111666932A (en) * | 2020-05-27 | 2020-09-15 | 平安科技(深圳)有限公司 | Document auditing method and device, computer equipment and storage medium |
CN113762292A (en) * | 2020-06-03 | 2021-12-07 | 杭州海康威视数字技术股份有限公司 | Training data acquisition method and device and model training method and device |
CN113762292B (en) * | 2020-06-03 | 2024-02-02 | 杭州海康威视数字技术股份有限公司 | Training data acquisition method and device and model training method and device |
CN111814833B (en) * | 2020-06-11 | 2024-06-07 | 浙江大华技术股份有限公司 | Training method of bill processing model, image processing method and image processing equipment |
CN111814833A (en) * | 2020-06-11 | 2020-10-23 | 浙江大华技术股份有限公司 | Training method of bill processing model, image processing method and image processing equipment |
CN111738326A (en) * | 2020-06-16 | 2020-10-02 | 中国工商银行股份有限公司 | Sentence granularity marking training sample generation method and device |
CN112270224A (en) * | 2020-10-14 | 2021-01-26 | 招商银行股份有限公司 | Insurance responsibility analysis method and device and computer readable storage medium |
CN112633275B (en) * | 2020-12-22 | 2023-07-18 | 航天信息股份有限公司 | Multi-bill mixed shooting image correction method and system based on deep learning |
CN112633275A (en) * | 2020-12-22 | 2021-04-09 | 航天信息股份有限公司 | Multi-bill mixed-shooting image correction method and system based on deep learning |
CN112699871A (en) * | 2020-12-23 | 2021-04-23 | 平安银行股份有限公司 | Method, system, device and computer readable storage medium for field content identification |
CN112699871B (en) * | 2020-12-23 | 2023-11-14 | 平安银行股份有限公司 | Method, system, device and computer readable storage medium for identifying field content |
CN112686262A (en) * | 2020-12-28 | 2021-04-20 | 广州博士信息技术研究院有限公司 | Method for extracting structured data and rapidly archiving handbooks based on image recognition technology |
CN113205041A (en) * | 2021-04-29 | 2021-08-03 | 百度在线网络技术(北京)有限公司 | Structured information extraction method, device, equipment and storage medium |
CN113283421A (en) * | 2021-06-24 | 2021-08-20 | 中国平安人寿保险股份有限公司 | Information identification method, device, equipment and storage medium |
CN113283421B (en) * | 2021-06-24 | 2024-03-01 | 中国平安人寿保险股份有限公司 | Information identification method, device, equipment and storage medium |
CN113408516A (en) * | 2021-06-25 | 2021-09-17 | 京东数科海益信息科技有限公司 | Bill recognition device and method |
CN114328831A (en) * | 2021-12-24 | 2022-04-12 | 江苏银承网络科技股份有限公司 | Bill information identification and error correction method and device |
CN118134576A (en) * | 2024-05-08 | 2024-06-04 | 山东工程职业技术大学 | Digital electronic invoice management method and system based on artificial intelligence |
Also Published As
Publication number | Publication date |
---|---|
CN107798299A (en) | 2018-03-13 |
CN107798299B (en) | 2020-02-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019071660A1 (en) | Bill information identification method, electronic device, and readable storage medium | |
CN107766809B (en) | Electronic device, bill information identification method, and computer-readable storage medium | |
WO2019174130A1 (en) | Bill recognition method, server, and computer readable storage medium | |
WO2019104879A1 (en) | Information recognition method for form-type image, electronic device and readable storage medium | |
CN109829453B (en) | Method and device for recognizing characters in card and computing equipment | |
WO2019205376A1 (en) | Vehicle damage determination method, server, and storage medium | |
WO2019037259A1 (en) | Electronic device, method and system for categorizing invoices, and computer-readable storage medium | |
WO2018205467A1 (en) | Automobile damage part recognition method, system and electronic device and storage medium | |
US20200410074A1 (en) | Identity authentication method and apparatus, electronic device, and storage medium | |
CN112699775B (en) | Certificate identification method, device, equipment and storage medium based on deep learning | |
CN111814785B (en) | Invoice recognition method, training method of relevant model, relevant equipment and device | |
US11710210B1 (en) | Machine-learning for enhanced machine reading of non-ideal capture conditions | |
US20150379341A1 (en) | Robust method to find layout similarity between two documents | |
CN110288755A (en) | The invoice method of inspection, server and storage medium based on text identification | |
WO2021139494A1 (en) | Animal body online claim settlement method and apparatus based on monocular camera, and storage medium | |
CN111553251B (en) | Certificate four-corner defect detection method, device, equipment and storage medium | |
US20220092353A1 (en) | Method and device for training image recognition model, equipment and medium | |
WO2019056503A1 (en) | Store monitoring evaluation method, device and storage medium | |
CN111160395A (en) | Image recognition method and device, electronic equipment and storage medium | |
CN113673500A (en) | Certificate image recognition method and device, electronic equipment and storage medium | |
CN112396047B (en) | Training sample generation method and device, computer equipment and storage medium | |
US20210264583A1 (en) | Detecting identification tampering using ultra-violet imaging | |
CN110647931A (en) | Object detection method, electronic device, system, and medium | |
CN111241974B (en) | Bill information acquisition method, device, computer equipment and storage medium | |
US10896339B2 (en) | Detecting magnetic ink character recognition codes |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17928326 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 29/09/2020) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 17928326 Country of ref document: EP Kind code of ref document: A1 |