[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN111401252B - Book spine matching method and equipment of book checking system based on vision - Google Patents

Book spine matching method and equipment of book checking system based on vision Download PDF

Info

Publication number
CN111401252B
CN111401252B CN202010187361.7A CN202010187361A CN111401252B CN 111401252 B CN111401252 B CN 111401252B CN 202010187361 A CN202010187361 A CN 202010187361A CN 111401252 B CN111401252 B CN 111401252B
Authority
CN
China
Prior art keywords
spine
matching
module
feature
checking system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010187361.7A
Other languages
Chinese (zh)
Other versions
CN111401252A (en
Inventor
蔡君
张立安
廖丽平
谭志坚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Xingxi Intelligent Technology Co ltd
Guangdong Polytechnic Normal University
Original Assignee
Guangdong Xingxi Intelligent Technology Co ltd
Guangdong Polytechnic Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Xingxi Intelligent Technology Co ltd, Guangdong Polytechnic Normal University filed Critical Guangdong Xingxi Intelligent Technology Co ltd
Priority to CN202010187361.7A priority Critical patent/CN111401252B/en
Publication of CN111401252A publication Critical patent/CN111401252A/en
Application granted granted Critical
Publication of CN111401252B publication Critical patent/CN111401252B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • G06V10/464Salient features, e.g. scale invariant feature transforms [SIFT] using a plurality of salient features, e.g. bag-of-words [BoW] representations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/757Matching configurations of points or features
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a spine matching method based on a visual book checking system, wherein an algorithm for improving the spine matching precision in the book checking system comprises a mobile robot and a computer; the mobile robot acquires images in real time through the camera and transmits the images to the computer in real time for image processing; an xinxi-point model; a fast nearest neighbor algorithm FLANN of the high-dimensional data; false detection rejection model XingxiJudge; and (3) carrying out feature extraction on the spine image, and conveying the features to a FLANN feature matcher to be matched with a spine feature library. However, the method has error matching, after the results of a plurality of suspicious targets are matched, the suspicious targets are formed into 'picture pairs' one by one and sent into a false detection rejection model XingxiJudge, whether the detection is wrong or not is judged, if false detection exists, the suspicious targets are rejected until a unique target is left. Finally, the algorithm obviously improves the matching quality and has better real-time performance and feature matching accuracy. The time of the matching process is effectively reduced, and the accuracy of spine matching is improved.

Description

Book spine matching method and equipment of book checking system based on vision
Technical Field
The invention relates to the technical field of automatic book checking, in particular to a spine matching method and equipment of a visual book checking system.
Background
In the book checking work of each large and medium-sized library, the book checking work is the work which needs to be performed every a period of time. At present, paper books are still an integral part of book resources. For millions of books in large and medium libraries, the routine inventory work is very labor and material consuming. Therefore, the book checking system based on computer vision is applied to book checking work of various large and medium libraries, and the checking work efficiency is improved. The system shoots the spine image by the mobile robot through various technologies such as image processing, deep learning and the like, performs characteristic extraction and characteristic matching on the spine image to finish spine matching, and is one of important components of a book checking system.
Spine matching is carried out, and spine feature extraction is firstly required to be carried out on spine images. Currently, the mainstream spine feature extraction algorithm includes SIFT, SURF, ORB. The SIFT extracted features are robust, but not as real-time as SURF, ORB. The ORB feature extraction algorithm runs far better than SIFT and SURF and can be used for real-time feature detection, but ORB is not scale-change robust.
After the spine feature extraction is performed on the image, spine feature matching is required. Because the library inventory system needs to process a large amount of data, the matching algorithm selects the FLANN for matching, the FLANN can select a proper algorithm to process a large amount of data according to the data, and the FLANN is much faster than other nearest neighbor searches. In the spine matching process, the larger the matched data volume is, the more false checks are generated, and the common spine matching technology is completed, but the problems are still not solved.
Disclosure of Invention
The present invention aims to solve at least one of the technical problems existing in the prior art. Therefore, the invention discloses a spine matching method of a book checking system based on vision, which comprises the following steps:
step 1, a mobile robot acquires spine image information in real time through a camera;
step 2, extracting features of the acquired spine image, and extracting feature points and descriptors through a preset algorithm;
step 3, training grids to extract corner points by taking virtual three-dimensional objects as data sets;
step 4, automatically marking characteristic points, wherein real spine picture data are adopted, and the network trained in the step 3 is used for extracting angular points;
step 5, auditing the real spine picture data to obtain a new picture and forming a picture pair with a known position relationship, inputting the picture pair into a network, and extracting feature points and descriptors;
the features extracted through the preset algorithm are sent to a Flann feature matcher for rough matching, and a feature subset is obtained;
and 6, traversing the feature subsets by the features, calculating the distance to obtain a result set, sending elements in the result set into a false detection rejection model XingxiJudge one by one, judging whether false detection exists, if false detection exists, performing second matching to realize rejection false matching operation, optimizing parameters in the model through a false detection rejection network model, and matching the spine.
Further, the preset algorithm is to design two networks, namely a BaseDetector network, for detecting corner points of the basic geometric image; the other is the XingxiPoint model network for extracting feature points and descriptors.
Furthermore, the step 5 further includes that the result of feature matching may obtain a corresponding relationship list of two feature sets, and before the Flann invokes the matching function, training a matcher to achieve the purpose of improving the matching speed, and matching features of the query spine set with a trainer one by one, where feature points of the query spine set may match a plurality of suspicious targets.
Still further, the step 6 further includes: and in the later stage, verifying the correctness of the matching by a false detection rejection model XingxiJudge, and rejecting until the optimal matching is obtained if false detection exists.
Further, the XingxiPoint model is composed of a BACKBONE module, a PPN module, a ROIPOOL module and a KP module; the KP module is responsible for generating feature points and descriptors; the BACKBONE module is respectively connected with the PPN module and the ROIPOOL module, the PPN module is connected with the ROIPOOL module, and the ROIPOOL module is connected with the KP module.
Further, performing noise reduction treatment on the spine data set A; automatically marking characteristic points of the data set A by affine transformation and ORB algorithm to obtain a spine data set G with strong supervision marking; initializing an XingxiPoint model; training the XingxiPoint by using a data set G, and selecting a small batch gradient descent algorithm by an optimizer; using the XingxiPoint to query the spine Q to obtain a series of feature points and descriptors; in a library spine library, matching a plurality of suspicious targets of Q by adopting a BF matching algorithm.
The invention further discloses an electronic device, comprising: a processor; and a memory for storing executable instructions of the processor; wherein the processor is configured to perform the spine matching method of the vision-based book checking system described above via execution of the executable instructions.
The invention further discloses a computer readable storage medium, on which a computer program is stored, which when being executed by a processor, implements the spine matching method of the vision-based book checking system.
In summary, compared with the spine matching method in the existing visual book checking system, the spine matching method has the following advantages:
firstly, more stable spine feature points can be generated, and the matching precision is improved.
For feature point extraction, an ORB feature point extraction algorithm is generally adopted, but the feature points are not stable due to the fact that the actual photographed spine is not ideal, and the subsequent matching precision is directly affected; according to the XingxiPoint model based on deep learning, affine transformation in the actual shooting spine can be fitted under the condition that manual labeling is not needed, the model is insensitive to illumination and angles of actual shooting, and the subsequent spine matching precision is effectively improved.
And secondly, self-adaptive similarity measurement can be carried out on the matching result, and a final target can be accurately taken out.
In spine matching, similarity between a spine and suspicious results is generally queried by cosine distance measurement, and then the similarity is ranked according to the similarity; once the suspicious results are compared with the images, or the calculation accuracy is insufficient, the final target is wrong, which is unacceptable; the false detection rejection model XingxiJudge provided by the invention essentially changes fixed similarity measurement into non-fixed self-adaptive similarity measurement, thereby improving the accuracy of a final target.
Drawings
The invention will be further understood from the following description taken in conjunction with the accompanying drawings. The components in the figures are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the embodiments. In the figures, like reference numerals designate corresponding parts throughout the different views.
FIG. 1 is a schematic diagram of the structure of the XingxiPoint model of the present invention;
FIG. 2 is a schematic illustration of the present invention the XingxiJudge model structure schematic diagram.
Detailed Description
The present invention will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present invention more apparent. It should be understood that the detailed description is presented by way of example only and is not intended to limit the scope of the invention.
Example 1
As shown in fig. 1-2, this embodiment proposes a spine matching method of a visual book checking system, including the steps of:
s1, performing noise reduction treatment on a spine data set A;
preprocessing a spine image, wherein a Gaussian filter is adopted for noise reduction processing in the embodiment; the Gaussian filter is a linear smoothing filter and has good inhibition effect on noise subjected to normal distribution. The gaussian filter is a weighting matrix whose weight expression is as follows:
Figure BDA0002414662360000041
wherein W is ij Is the weight, i and j are the pixel index, K is the normalization constant, and σ is the standard deviation of the Gaussian distribution.
S2, automatically marking characteristic points of the data set A by using affine transformation and ORB algorithm to obtain a spine data set G with strong supervision marking;
affine transformation is performed on each image in the spine image dataset, and 2000 transformations are adopted in the embodiment to obtain 2000 transformed images. And respectively extracting characteristic points on the images by using ORB, so that 2000 characteristic point locating images can be obtained, the 2000 characteristic point locating images are subjected to inverse affine transformation and then accumulated together to obtain a final characteristic point locating image, namely the characteristic point marking of the original spine image, without manual marking, and the cost is saved. The data set G can be obtained by repeating the steps.
S3, initializing an XingxiPoint model;
the XingxiPoint model is essentially a variant of the target detection model FasterRCNN model;
comprises a BACKBONE module, a PPN module, a ROIPOOL module and a KP module. Wherein, the BACKBONE module is responsible for generating a characteristic diagram of the spine; any deep convolutional network can be used as a back bone module, and the common VGG network is adopted in this embodiment.
The PPN module is a full convolution network and is responsible for suggesting suspicious feature points to the ROIPOOL module and corresponds to the regional suggestion module RPN in the FasterRCNN model; wherein Anchor in the RPN module is characterized by two points. In the embodiment, the Anchor in the RPN is changed from two-point characterization to single-point characterization and is used as a PPN module. Note that a single point is herein referred to as a suspicious feature point.
The ROIPOOL module is responsible for receiving the suggested feature points of the PPN module, extracting corresponding feature blocks from the BACKBONE module and sending the feature blocks to the KP module. The structure of the ROIPOOL module in this embodiment is consistent with the ROIPOOL in the target detection model FaterRCNN.
The KP module is actually a regressor and is responsible for regressing real feature point coordinates according to the labeling information, and a feature block corresponding to the feature point is given as a descriptor.
The connection mode of each module is as follows: the BACKBONE module is respectively connected with the PPN module and the ROIPOOL module, the PPN module is connected with the ROIPOOL module, and the ROIPOOL module is connected with the KP module.
S4, training the XingxiPoint by using a data set G, and selecting a small batch gradient descent algorithm by an optimizer;
small batch gradient descent is a compromise to batch gradient descent and random gradient descent. The idea is to calculate the average gradient of the batch_size samples in each iteration, and then optimize the parameters. In this embodiment, we take batch_size=10, train 1000 spine images, and then the optimization process pseudo-code is:
Figure BDA0002414662360000051
where θ is the XingxiPoint model parameter, α is the learning rate, and the term following α is the average gradient of the batch_size samples.
S5, using the XingxiPoint for inquiring the spine Q to obtain a series of feature points and descriptors;
s6, matching a plurality of suspicious targets of the Q in a library spine library by adopting a BF matching algorithm;
and obtaining a corresponding relation list of the two feature sets through spine matching. The first set of features is named a dataset and the second set is a query set. BF matching is a similarity search between high-dimensional vectors through a distance function, and a k-d tree is an algorithm for fast and accurately finding a high-dimensional spatial index structure and approximate query of neighbors of a query point. Given a query point and a query distance threshold, all data from the dataset that is less than the query distance threshold from the query point is found. K data closest to the query point are found from the dataset, setting k=2, which is the nearest neighbor query. After the feature points and descriptors are obtained, each point to be queried in the set is found, and 2-neighbor query is carried out on the feature point set of the target object obtained by matching, namely, the nearest neighbor and the next nearest neighbor are obtained.
Defining nearest neighbors:
given a multidimensional space
Figure BDA0002414662360000053
Handle->
Figure BDA0002414662360000054
Becomes one sample point. The nearest neighbor of a given sample set E and sample points d, d is that any sample point d 'E is satisfied such that all neighbors (E, d, d'). The distance measure is as follows
Figure BDA0002414662360000052
Wherein d is i Is the i-th component of vector d.
The BF matching algorithm trains out a matcher. The matcher designs an index tree of the feature set, and matches each feature point of the query set with the matcher to obtain a plurality of suspicious targets.
S7, matching a plurality of suspicious targets to form a picture pair one by one, sending the picture pair into a false detection rejection model XingxiJudge, judging whether false detection exists or not, and rejecting if false detection exists. Finally leaving the unique target.
And matching a plurality of suspicious targets to form a picture pair one by one, and sending the picture pair into a false detection rejection model XingxiJudge. The false detection rejection model is used for screening 'picture pairs' to obtain the best matched spine image. And (3) finding out the first two key points of a plurality of suspicious targets closest to the spine image of the data set by extracting key points of the spine, judging as false detection if the ratio of the second closest distance divided by the closest distance is smaller than a certain threshold value, and adopting a rejecting operation.
The false detection rejection model XingxiJudge trains a classifier. For the two classification problems, the false detection is marked as a negative sample as 0, the positive detection is marked as a positive sample as 1, i.e. for class y, there is
y∈{0,1}
Probability h for positive samples θ (x) So that
0≤h θ (x)≤1
Where θ is the parameter to be optimized such that in case of sample x for unknown class 0 Classification time h θ (x 0 ) Probability that the sample is a positive sample. The classification criteria are as follows:
Figure BDA0002414662360000061
due to the high dimension of the feature space and the millions of spine images to be matched, similar distances entail a large number of mismatches, and the corresponding ratio is also high. So by raising this threshold, the number of matching points will be reduced, and the accuracy of spine matching can be effectively improved.
S8, after a unique target is left by the false detection rejection model XingxiJudge, finally outputting the spine to be queried.
Example two
The invention discloses a spine matching method of a book checking system based on vision, which comprises the following steps:
step 1, a mobile robot acquires spine image information in real time through a camera;
step 2, extracting features of the acquired spine image, and extracting feature points and descriptors through a preset algorithm;
step 3, training grids to extract corner points by taking virtual three-dimensional objects as data sets;
step 4, automatically marking characteristic points, wherein real spine picture data are adopted, and the network trained in the step 3 is used for extracting angular points;
step 5, auditing the real spine picture data to obtain a new picture and forming a picture pair with a known position relationship, inputting the picture pair into a network, and extracting feature points and descriptors;
the features extracted through the preset algorithm are sent to a Flann feature matcher for rough matching, and a feature subset is obtained;
and 6, traversing the feature subsets by the features, calculating the distance to obtain a result set, sending elements in the result set into a false detection rejection model XingxiJudge one by one, judging whether false detection exists, if false detection exists, performing second matching to realize rejection false matching operation, optimizing parameters in the model through a false detection rejection network model, and matching the spine.
Further, the preset algorithm is to design two networks, namely a BaseDetector network, for detecting corner points of the basic geometric image; the other is the XingxiPoint model network for extracting feature points and descriptors.
Furthermore, the step 5 further includes that the result of feature matching may obtain a corresponding relationship list of two feature sets, and before the Flann invokes the matching function, training a matcher to achieve the purpose of improving the matching speed, and matching features of the query spine set with a trainer one by one, where feature points of the query spine set may match a plurality of suspicious targets.
Still further, the step 6 further includes: and in the later stage, verifying the correctness of the matching by a false detection rejection model XingxiJudge, and rejecting until the optimal matching is obtained if false detection exists.
Further, the XingxiPoint model is composed of a BACKBONE module, a PPN module, a ROIPOOL module and a KP module; the KP module is responsible for generating feature points and descriptors; the BACKBONE module is respectively connected with the PPN module and the ROIPOOL module, the PPN module is connected with the ROIPOOL module, and the ROIPOOL module is connected with the KP module.
Further, performing noise reduction treatment on the spine data set A; automatically marking characteristic points of the data set A by affine transformation and ORB algorithm to obtain a spine data set G with strong supervision marking; initializing an XingxiPoint model; training the XingxiPoint by using a data set G, and selecting a small batch gradient descent algorithm by an optimizer; using the XingxiPoint to query the spine Q to obtain a series of feature points and descriptors; in a library spine library, matching a plurality of suspicious targets of Q by adopting a BF matching algorithm.
The invention further discloses an electronic device, comprising: a processor; and a memory for storing executable instructions of the processor; wherein the processor is configured to perform the spine matching method of the vision-based book checking system described above via execution of the executable instructions.
The invention further discloses a computer readable storage medium, on which a computer program is stored, which when being executed by a processor, implements the spine matching method of the vision-based book checking system.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article or apparatus that comprises the element.
It will be appreciated by those skilled in the art that embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
While the invention has been described above with reference to various embodiments, it should be understood that many changes and modifications can be made without departing from the scope of the invention. It is therefore intended that the foregoing detailed description be regarded as illustrative rather than limiting, and that it be understood that it is the following claims, including all equivalents, that are intended to define the spirit and scope of this invention. The above examples should be understood as illustrative only and not limiting the scope of the invention. Various changes and modifications to the present invention may be made by one skilled in the art after reading the teachings herein, and such equivalent changes and modifications are intended to fall within the scope of the invention as defined in the appended claims.

Claims (7)

1. A spine matching method of a book checking system based on vision is characterized by comprising the following steps:
step 1, a mobile robot acquires spine image information in real time through a camera;
step 2, extracting features of the acquired spine image, extracting feature points and descriptors through a preset algorithm, wherein the preset algorithm is to design two networks, namely a BaseDetector network, and the BaseDetector network is used for detecting corner points of the basic geometric image; the other is the XingxiPoint model network for extracting feature points and descriptors;
step 3, training grids to extract corner points by taking virtual three-dimensional objects as data sets;
step 4, automatically marking characteristic points, wherein real spine picture data are adopted, and the grids trained in the step 3 are used for extracting angular points;
step 5, auditing the real spine picture data to obtain a new picture and forming a picture pair with a known position relationship, inputting the picture pair into a network, and extracting feature points and descriptors; the features extracted through the preset algorithm are sent to a Flann feature matcher for rough matching, and a feature subset is obtained;
and 6, traversing the feature subsets by the features, calculating the distance to obtain a result set, sending elements in the result set into a false detection rejection model XingxiJudge one by one, judging whether false detection exists, if false detection exists, performing second matching to realize rejection false matching operation, optimizing parameters in the model through a false detection rejection network model, and matching the spine.
2. The spine matching method of the vision-based book checking system of claim 1, wherein the step 5 further comprises the step that the corresponding relation list of two feature sets is obtained by the feature matching result, before the matching function is called, the Flann trains a matcher to achieve the purpose of improving the matching speed, then inquires that features of the spine set are matched with a trainer one by one, and inquires that feature points of the spine set match a plurality of suspicious targets.
3. The spine matching method of a vision-based book checking system of claim 1, wherein said step 6 further comprises: and in the later stage, verifying the correctness of the matching by a false detection rejection model XingxiJudge, and rejecting until the optimal matching is obtained if false detection exists.
4. The spine matching method of a vision-based book checking system of claim 1, wherein the XingxiPoint model is composed of a back box module, a PPN module, a roiool module and a KP module; the KP module is responsible for generating feature points and descriptors; the BACKBONE module is respectively connected with the PPN module and the ROIPOOL module, the PPN module is connected with the ROIPOOL module, and the ROIPOOL module is connected with the KP module.
5. The spine matching method of a vision-based book checking system as claimed in claim 1, wherein the spine data set a is subjected to noise reduction; automatically marking characteristic points of the data set A by affine transformation and ORB algorithm to obtain a spine data set G with strong supervision marking; initializing an XingxiPoint model; training the XingxiPoint by using a data set G, and selecting a small batch gradient descent algorithm by an optimizer; using the XingxiPoint to query the spine Q to obtain a series of feature points and descriptors; in a library spine library, matching a plurality of suspicious targets of Q by adopting a BF matching algorithm.
6. An electronic device, comprising:
a processor; the method comprises the steps of,
a memory for storing executable instructions of the processor;
wherein the processor is configured to perform the spine matching method of the vision-based book checking system of any one of claims 1-5 via execution of the executable instructions.
7. A computer readable storage medium having stored thereon a computer program, which when executed by a processor, implements the spine matching method of the vision-based book checking system of any one of claims 1-5.
CN202010187361.7A 2020-03-17 2020-03-17 Book spine matching method and equipment of book checking system based on vision Active CN111401252B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010187361.7A CN111401252B (en) 2020-03-17 2020-03-17 Book spine matching method and equipment of book checking system based on vision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010187361.7A CN111401252B (en) 2020-03-17 2020-03-17 Book spine matching method and equipment of book checking system based on vision

Publications (2)

Publication Number Publication Date
CN111401252A CN111401252A (en) 2020-07-10
CN111401252B true CN111401252B (en) 2023-07-07

Family

ID=71428838

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010187361.7A Active CN111401252B (en) 2020-03-17 2020-03-17 Book spine matching method and equipment of book checking system based on vision

Country Status (1)

Country Link
CN (1) CN111401252B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113487598A (en) * 2021-07-26 2021-10-08 中国科学院国家空间科学中心 Bookbinding error detection system based on computer vision

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104966081B (en) * 2015-06-04 2018-05-01 广州美读信息技术有限公司 Spine image-recognizing method
CN106886707B (en) * 2017-03-21 2020-06-09 广东网金控股股份有限公司 Image authentication method and device
CN110427966A (en) * 2019-06-17 2019-11-08 青岛星科瑞升信息科技有限公司 One kind rejecting error hiding feature point methods based on characteristic point local feature

Also Published As

Publication number Publication date
CN111401252A (en) 2020-07-10

Similar Documents

Publication Publication Date Title
US9141871B2 (en) Systems, methods, and software implementing affine-invariant feature detection implementing iterative searching of an affine space
CN109002562B (en) Instrument recognition model training method and device and instrument recognition method and device
CN109740633B (en) Image similarity calculation method and device and storage medium
JP2018205937A (en) Image retrieval device and program
Lepsøy et al. Statistical modelling of outliers for fast visual search
Chatfield et al. Efficient retrieval of deformable shape classes using local self-similarities
CN109726746A (en) A kind of method and device of template matching
US20210117648A1 (en) 3-dimensional model identification
CN111753119A (en) Image searching method and device, electronic equipment and storage medium
AU2013271337A1 (en) Biometric verification
WO2015146113A1 (en) Identification dictionary learning system, identification dictionary learning method, and recording medium
CN112733969B (en) Object class identification method and device and server
CN113298146A (en) Image matching method, device, equipment and medium based on feature detection
CN111325276A (en) Image classification method and device, electronic equipment and computer-readable storage medium
CN110442749B (en) Video frame processing method and device
CN111401252B (en) Book spine matching method and equipment of book checking system based on vision
CN112446428B (en) Image data processing method and device
JP6017277B2 (en) Program, apparatus and method for calculating similarity between contents represented by set of feature vectors
CN110263196B (en) Image retrieval method, image retrieval device, electronic equipment and storage medium
JP5616310B2 (en) Image matching apparatus and image matching program
CN111738194A (en) Evaluation method and device for similarity of face images
CN108563692B (en) Target duplicate checking method and device
CN113688708A (en) Face recognition method, system and storage medium based on probability characteristics
CN112149566A (en) Image processing method and device, electronic equipment and storage medium
CN117372787B (en) Image multi-category identification method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant