[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US20220358791A1 - Identity information processing method and device based on fundus image - Google Patents

Identity information processing method and device based on fundus image Download PDF

Info

Publication number
US20220358791A1
US20220358791A1 US17/619,875 US202017619875A US2022358791A1 US 20220358791 A1 US20220358791 A1 US 20220358791A1 US 202017619875 A US202017619875 A US 202017619875A US 2022358791 A1 US2022358791 A1 US 2022358791A1
Authority
US
United States
Prior art keywords
fundus image
dimensional feature
feature vector
distance
eye
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US17/619,875
Other versions
US11893831B2 (en
Inventor
Andre Wang HE
JianHao Xiong
Meng Fu
Xiaopei ZHU
Xin Zhao
Chao He
Dalei Zhang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Eaglevision Medical Technology Co Ltd
Original Assignee
Shanghai Eaglevision Medical Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Eaglevision Medical Technology Co Ltd filed Critical Shanghai Eaglevision Medical Technology Co Ltd
Assigned to SHANGHAI EAGLEVISION MEDICAL TECHNOLOGY CO., LTD. reassignment SHANGHAI EAGLEVISION MEDICAL TECHNOLOGY CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: FU, MENG, HE, Andre Wang, HE, CHAO, XIONG, Jianhao, ZHANG, DALEI, ZHAO, XIN, ZHU, Xiaopei
Publication of US20220358791A1 publication Critical patent/US20220358791A1/en
Application granted granted Critical
Publication of US11893831B2 publication Critical patent/US11893831B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/193Preprocessing; Feature extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • G06F21/32User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/09Supervised learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • G06V10/449Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters
    • G06V10/451Biologically inspired filters, e.g. difference of Gaussians [DoG] or Gabor filters with interaction between the filter responses, e.g. cortical complex cells
    • G06V10/454Integrating the filters into a hierarchical structure, e.g. convolutional neural networks [CNN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/14Vascular patterns
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/18Eye characteristics, e.g. of the iris
    • G06V40/197Matching; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/50Maintenance of biometric data or enrolment thereof
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/20ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/40ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing

Definitions

  • the present disclosure relates to the field of image information processing, and in particular to an identity information processing method and device based on a fundus image.
  • a fundus image (also referred to as a retinal image) captured using a fundus camera can show human tissues such as the macula, optic disc, and blood vessels.
  • the fundus image is highly unique because the vascular orientation, bifurcations, and the shape of the optic disc are different from person to person, but the fundus image of a particular person does not change significantly with age.
  • the first step is to determine whether the identity information of the user is already stored in the database. If fundus images are stored in the database, all the fundus images already stored in the database need to be separately compared with the current fundus image. This approach is very inefficient.
  • an implementation of the present disclosure provides an identity information comparison method based on a fundus image, including:
  • the neural network is trained with triple sample data.
  • the triple sample data includes a first fundus image sample, a second fundus image sample, and a third fundus image sample.
  • the second fundus image sample and the first fundus image sample are fundus images of the same person, and the third fundus image sample and the first fundus image sample are fundus images of different people.
  • the neural network separately extracts multi-dimensional feature vectors of the first fundus image sample, the second fundus image sample, and the third fundus image sample; calculates a first distance between the second fundus image sample and the first fundus image sample, and a second distance between the third fundus image sample and the first fundus image sample according to the three extracted multi-dimensional feature vectors; obtains a loss value according to the first distance and the second distance; and adjusts parameters of the neural network according to the loss value.
  • the adjusting parameters of the neural network according to the loss value includes:
  • the determining, according to a comparison result, whether there is a matching between the currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database includes:
  • the fundus image includes a left-eye fundus image and a right-eye fundus image.
  • the database is configured to store user data.
  • Each group of user data includes a first pre-stored multi-dimensional feature vector corresponding to a left eye and a second pre-stored multi-dimensional feature vector corresponding to a right eye.
  • the multi-dimensional feature vector includes a first multi-dimensional feature vector corresponding to the left-eye fundus image and a second multi-dimensional feature vector corresponding to the right-eye fundus image.
  • the fundus image includes a left-eye fundus image and a right-eye fundus image
  • the database is configured to store user data
  • each group of user data includes one pre-stored multi-dimensional feature vector.
  • the obtaining a multi-dimensional feature vector representing the identity of a user includes:
  • the fundus image is a left-eye fundus image or a right-eye fundus image
  • the database is configured to store user data
  • each group of user data includes a first pre-stored multi-dimensional feature vector corresponding to a left eye or a second pre-stored multi-dimensional feature vector corresponding to a right eye.
  • the present disclosure further provides an identity verification method based on a fundus image, including:
  • the present disclosure further provides an identity information storage method based on a fundus image, including:
  • the present disclosure further provides a method for training a fundus image recognition model, including:
  • the training data includes a first fundus image sample, a second fundus image sample, and a third fundus image sample
  • the second fundus image sample and the first fundus image sample are fundus images of the same person
  • the third fundus image sample and the first fundus image sample are fundus images of different people
  • the recognizing the first fundus image sample, the second fundus image sample, and the third fundus image sample by using a fundus image recognition model to obtain a loss value includes:
  • the adjusting parameters of the fundus image recognition model according to the loss value includes:
  • an implementation of the present disclosure further provides a device for an identity information comparison based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity information comparison method based on a fundus image.
  • an implementation of the present disclosure further provides a device for identity verification based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity verification method based on a fundus image.
  • an implementation of the present disclosure further provides a device for an identity information storage based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity information storage method based on a fundus image.
  • an implementation of the present disclosure further provides a device for training a fundus image recognition model, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing method for training a fundus image recognition model.
  • a neural network is first used to convert a fundus image of a user into a multi-dimensional feature vector representing the identity of the user, and then abstract feature information related to personal features of the user is extracted by using characteristics of the neural network.
  • implementations of the disclosure may determine by comparing multi-dimensional feature vectors whether there is data matching with the current user in a database.
  • the database does not need to store a fundus image. It is also unnecessary to recognize pre-stored fundus images repeatedly every time a new fundus image is obtained, so that the efficiency of operation of an identity information comparison can be improved.
  • a neural network provided according to the implementation of the present disclosure is trained with triple training data and a corresponding loss function, so that a distance between feature vectors extracted by the neural network from different fundus images of the same eye can be decreased, and a distance between feature vectors extracted from fundus images of different eyes can be increased during the training process.
  • distances between feature vectors repeatedly extracted by the neural network from the same fundus image are small enough, and distances to feature vectors of other fundus images are large enough, that is, the information has particular uniqueness, so that feature vectors extracted by the neural network from fundus images can be used as identity information of a user.
  • FIG. 1 is a flowchart of an identity information comparison method according to an embodiment of the present disclosure
  • FIG. 2 is a flowchart of an identity information storage method according to an embodiment of the present disclosure
  • FIG. 3 is a schematic diagram of extracting identity information by using a neural network according to an embodiment of the present disclosure
  • FIG. 4 is another schematic diagram of extracting identity information by using a neural network according to an embodiment of the present disclosure
  • FIG. 5 is a flowchart of a neural network training method according to an embodiment of the present disclosure.
  • FIG. 6 shows an image patch in a fundus image
  • FIG. 7 shows a segmentation result of the image patch shown in FIG. 6 .
  • FIG. 8 shows a fundus blood vessel image
  • Embodiments of the present disclosure provide an identity information storage method based on a fundus image.
  • the method may be performed by an electronic device such as a computer or a server. As shown in FIG. 1 , the method includes the following steps S 1 to S 3 .
  • a fundus image is recognized by a neural network to obtain a multi-dimensional feature vector representing the identity of a user.
  • the neural network extracts feature information during recognition of an image. For different tasks, the neural network focuses on different content, and different feature information is extracted. For example, during the execution of a classification task, the neural network extracts corresponding feature information (usually a multi-dimensional feature vector) according to a category of a fundus image, and then performs classification according to the feature information.
  • the neural network used is configured to extract the multi-dimensional feature vector representing the identity of a user, rather than to perform a classification task or an image segmentation task.
  • multi-dimensional feature vectors extracted by the neural network from fundus images of people should be different.
  • multi-dimensional feature vectors extracted at different times should be the same (or approximately the same or similar).
  • the neural network in this application may be a deep convolutional neural network (CNN), and extracted multi-dimensional feature vectors can be normalized by setting an appropriate loss function and using a back propagation (BP) algorithm. Then, a trained CNN model can be used to extract feature vectors, which are usually high-dimensionality vectors, from a fundus image.
  • CNN deep convolutional neural network
  • BP back propagation
  • the neural network To enable the neural network to extract expected content, the neural network should be trained beforehand. There are many training methods, and different training data are used for different training methods. Details are described below in detail.
  • the obtained multi-dimensional feature vector is compared with each multi-dimensional feature vector pre-stored in a database.
  • the database may be categorized into a database configured based on a graphics processing unit (GPU) and a database configured based on a central processing unit (CPU).
  • a multi-dimensional feature vector pre-stored in the database may alternatively be the multi-dimensional feature vector extracted by the neural network in step S 1 from another fundus image.
  • a comparison result is used to represent a similarity between the two feature vectors.
  • a similarity between multi-dimensional vectors may be determined based on Euclidean distance, cosine similarity, normalized Euclidean distance, or the like.
  • the Euclidean distance measures an absolute distance between points in a multi-dimensional space, which is preferred when data is very dense and continuous. Since the calculation is based on absolute values of features of different dimensions, Euclidean measurement needs to ensure that indicators of different dimensions are at the same scale level.
  • Mahalanobis distance may be used.
  • the Mahalanobis distance is a distance based on sample distribution. For example, for two populations of normal distribution with an average value of a and b, respectively but having different variances, the probability of which distribution a sample point A belongs to in a distribution space is higher determines that A belongs to this distribution.
  • multi-dimensional feature vectors extracted by the neural network from fundus images of the same person captured at different moments and in different environments are usually not identical. Therefore, using a distance calculation method to measure a similarity between two feature vectors has a certain fault tolerance, and the distance is preferably a Euclidean distance.
  • step S 3 whether there is a matching between the currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database is determined according to a comparison result.
  • Different comparison methods may result in different comparison results. For example, whether each pre-stored multi-dimensional feature vector matches the currently obtained multi-dimensional feature vector may be determined based on the distance or the angle.
  • the distance between the two feature vectors is less than a set threshold, it can be determined that the two feature vectors have a sufficiently high similarity, indicating that the two match each other.
  • a neural network is first used to convert a fundus image of a user into a multi-dimensional feature vector used for representing the identity of the user, and then abstract feature information related to personal features of the user is extracted using characteristics of the neural network.
  • abstract feature information related to personal features of the user is extracted using characteristics of the neural network.
  • whether there is data matching the current user in a database can be determined by comparing multi-dimensional feature vectors.
  • the database does not need to store fundus images. It is also unnecessary to recognize pre-stored fundus images repeatedly every time a new fundus image is obtained, thereby improving the efficiency of an identity information comparison operation.
  • This scheme may be used to process a fundus image of one eye of a user. That is, the fundus image recognized in step S 1 is a left-eye fundus image or a right-eye fundus image.
  • feature vectors stored in the database are from fundus images of one eye.
  • the fundus images in step S 1 include a left-eye fundus image and a right-eye fundus image.
  • the two fundus images are separately recognized by the neural network, to obtain a first multi-dimensional feature vector corresponding to the left-eye fundus image and a second multi-dimensional feature vector corresponding to the right-eye fundus image.
  • each group of data stored in the database includes two pre-stored multi-dimensional feature vectors, which may be correspondingly compared in step S 2 .
  • fundus images of two eyes of a user are uniformly processed. Recognized fundus images include a left-eye fundus image and a right-eye fundus image.
  • the fundus images of two eyes are separately recognized by using the neural network in step S 1 , and a first multi-dimensional feature vector corresponding to the left-eye fundus image and a second multi-dimensional feature vector corresponding to the right-eye fundus image are output.
  • the two feature vectors are then combined, which may be performed in multiple ways. For example, two 1024-dimensional feature vectors may be connected into one 2048-dimensional feature vector.
  • the pre-stored data in the database is a combined multi-dimensional feature vector, and is a result of combining two feature vectors in advance.
  • the currently combined dimensional feature vector is compared with a pre-stored multi-dimensional feature vector in the database, to further determine, according to a comparison result, whether there is matching data. In this embodiment, it only needs to perform one comparison to determine whether there are matching data of the two eyes.
  • the method uses the foregoing comparison embodiment to establish and manage an identity information database. As shown in FIG. 2 , the method includes the following steps S 1 A to S 6 A.
  • a fundus image of a user is acquired.
  • the fundus image may be a fundus image of either or both eyes.
  • step S 2 A the fundus image is recognized using a neural network to obtain a multi-dimensional feature vector representing the identity of the user.
  • a neural network to obtain a multi-dimensional feature vector representing the identity of the user.
  • step S 3 A the obtained multi-dimensional feature vector is compared with each multi-dimensional feature vector pre-stored in a database. Reference may be made to the foregoing step S 2 , which will not be described herein.
  • step S 4 A whether there is a matching between the currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database is determined according to a comparison result.
  • step S 5 A is performed, or otherwise step S 6 A is performed.
  • the database may initially be empty, and has no pre-stored multi-dimensional feature vector, thus there is no matching content.
  • a plurality of pieces of data may be imported into the initial database in advance.
  • the data uses individuals as basic modules.
  • each piece of user data includes the name of the user and a multi-dimensional feature vector of the user (which is obtained based on a fundus image of the user), and may additionally include personal information such as the gender, age, diopter of the user, and the like.
  • the currently obtained multi-dimensional feature vector is stored in the database.
  • the system may prompt whether it is necessary to add the currently obtained multi-dimensional feature vector to the original database as a new piece of data. If YES is chosen, a request to memory is sent, and after the request succeeds, the new piece of data is added to the original database. Further, the system may further prompt whether to continue to input related supplementary information, for example, personal information such as the gender, age, the diopter, and the like of the user. After the information is entered, the information and the corresponding multi-dimensional feature vectors are stored as related data items.
  • the pre-stored multi-dimensional feature vector matching the currently obtained multi-dimensional feature vector in the database is replaced with the currently obtained multi-dimensional feature vector.
  • the system may prompt whether it is necessary to overwrite existing data with the currently obtained multi-dimensional feature vector. If YES is chosen, overwriting is performed. This step is an optional operation.
  • the system may only prompt that current user data already exists, and a replacement operation is not required.
  • a neural network is first used to convert a fundus image of a user into a multi-dimensional feature vector used for representing the identity of the user, and then abstract feature information related to personal features of the user is extracted using characteristics of the neural network.
  • whether there is data matching the current user in a database can be determined by comparing multi-dimensional feature vectors, so as to store the multi-dimensional feature vector.
  • the database does not need to store a fundus image. It is also not necessary to recognize an existing fundus image by using the neural network during the storage of every piece of identity information of the user, thereby improving the efficiency of storing identity information of users.
  • each piece of information of the user in the database established according to the scheme may include the name of the user, a multi-dimensional feature vector (a first multi-dimensional feature vector of a left eye and/or a second multi-dimensional feature vector of a right eye or a combination of the two), the gender, age, diopter of the user, and the like.
  • the multi-dimensional feature vector represents the identity information of the user.
  • step S 4 A that is, whether the identity information of the current user is stored in the database.
  • a corresponding identity determination operation can be performed. For example, at step S 5 A, it is determined that the current user is not a known user, and therefore the current user is not allowed to perform a subsequent operation; and at step S 6 A, it is determined that the current user is a known user, and therefore the current user is allowed to perform a subsequent operation.
  • the subsequent operation may be unlocking or logging in an electronic device, or the like, to implement identity verification, authentication, or the like.
  • An embodiment of the present disclosure further provides a device of an identity information comparison based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity information comparison method based on a fundus image.
  • An embodiment of the present disclosure further provides a device of an identity information storage based on a fundus image, including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity information storage method based on a fundus image.
  • An embodiment of the present disclosure further provides a device of an identity verification based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity verification method based on a fundus image.
  • aspects to extract features may include: providing an optimized intra-class distance, to make the intra-class distance smaller; and providing an optimized inter-class distance, to make the inter-class distance more distinguishable.
  • the intra-class distance refers to a distance between sample points of each pattern in the same class; and the inter-class distance refers to a distance between different classes.
  • FIG. 3 shows a training framework of a softmax-based classification model.
  • the framework of the classification model may be applied to training and extraction of identity features of fundus images, and on this basis, various auxiliary loss functions are provided to improve the characteristics of the identity features of the fundus images.
  • a fundus image Imgi is fed into a deep CNN, to obtain a multi-dimensional feature vector x 1 , and then passes through a fully connected layer and is processed by softmax, to obtain a score vector.
  • the softmax loss function is shown as follows:
  • N is a batch size
  • n is a quantity of classes
  • x i denotes a feature vector of an i th fundus image of a batch
  • y 1 denotes an actual identity class (label) of the fundus image
  • W j denotes a vector of a j th column of a weight W of the fully connected layer
  • b j is a bias value
  • a score vector s i obtained for the fundus image Imgi is 0.1, 0.2, 0.7). It is determined that Imgi is a fundus image of a user with a label of 3.
  • the obtained score vector s i is 0.1, 0.5, 0.4).
  • a model trained using this loss function is used to extract a feature vector that can be used for correct classification.
  • x i is a feature vector of an i th image in one batch.
  • y i is an identity label of a user to which the fundus image x i belongs.
  • Cyi ⁇ Rd denotes the center of a feature vector of an identity classification label of a y i th user, and d denotes a dimensionality of a feature vector.
  • a scalar ⁇ [0, 1] may be used to control a learning rate of the center. This scheme is distinguished in that the difference between internal classes can be reduced to the greatest extent while ensuring that features of different classes can be separated without complex recombination of training sets.
  • a softmax loss (L s ) and a contrastive loss (L c ) is used to train the deep CNN to perform discriminative feature learning, not only the accuracy of classification is taken into consideration, but also an inter-class distance m (margin) is added to expand a decision boundary, so that positive samples have a higher similarity, and negative samples have a lower similarity.
  • a softmax layer with n channels is introduced after the structure of the deep CNN to obtain a recognition module for classifying an identity class of an input fundus image, where n is the number of user identity classes of the user, and is defined as an identification loss function:
  • verification loss function which is in fact a contrastive loss
  • fundus images in a training set should be first paired to obtain image pairs, and it is labeled whether each image pair includes identity features of the same user.
  • a fundus image pair is input into the deep CNN to obtain image features. Identification loss is separately calculated for the fundus image pair. That is, user identity classification is performed on two input fundus images, and verification is performed on the fundus image pair at the same time, to verify whether two input fundus images of the fundus image pair are the same.
  • m (margin) defines an inter-class margin, which acts on the loss value only when a distance in the image pair with different identities is within the distance m. This effectively restricts an inter-class distance, and therefore the inter-class distance is more distinguishable.
  • embodiments of the present disclosure further provide a method for training a triple data-based model.
  • the method includes the following steps S 1 B to S 3 B.
  • training data is acquired, where the training data includes a first fundus image sample, a second fundus image sample, and a third fundus image sample.
  • the training data used in this embodiment is triple sample data.
  • the second fundus image sample and the first fundus image sample are fundus images of the same eye of the same person.
  • the third fundus image sample and the first fundus image sample are fundus images of different people.
  • a data set is first prepared.
  • the data set may include fundus images of n eyes. Each eye corresponds to m fundus images. That is, there are a total of n*m fundus images in the data set.
  • one fundus image is first randomly selected from the n*m fundus images, which is referred to as Anchor. Then another fundus image of the same eye as Anchor is selected, which is referred to as Positive (denoted as x ⁇ p ); and then one fundus image of an eye different from Anchor is selected, which is referred to as Negative (denoted as x ⁇ n ), so that a group of training data (Anchor, Positive, Negative) is obtained.
  • a fundus image may be preprocessed first to allow the trained neural network to be more accurate during recognition of the fundus image. Specifically, each fundus image may first be cropped. Because a captured fundus image originally has a relatively large black background, the fundus images may first be trimmed to remove a large number of black pixels in the background, until all fundus images are cropped to the smallest rectangles that can contain the entire circular fundus.
  • all fundus images may be cropped to a uniform format, for example, a size of 224*224 pixels.
  • Fundus images used for model training and recognition may adopt a uniform format of 224*224 pixels and RGB color channels.
  • the first fundus image sample, the second fundus image sample, and the third fundus image sample are recognized by the neural network to obtain a loss value.
  • a first distance between the second fundus image sample and the first fundus image sample, and a second distance between the third fundus image sample and the first fundus image sample are calculated according to a preset loss function, and a loss value is obtained according to the first distance and the second distance.
  • the neural network separately extracts features of the foregoing three fundus images, to obtain three multi-dimensional feature vectors, which are respectively denoted as: f (x i a ), f (x i p ), and f (x i n ). Then a first distance between f (x i a ) and f (x i p ), and a second distance between f (x i a ) and f (x i n ) are calculated.
  • the first distance and the second distance in this embodiment are Euclidean distances.
  • the loss value is calculated by using the first distance and the second distance, which may be performed by the following relational expression of the loss function.
  • a denotes a preset value which is a minimum distance between the first distance and the second distance. + denotes that when a value in [ ] is greater than 0, this value is taken as the loss value, and when the value in [ ] is less than 0, the loss is 0.
  • step S 3 B parameters of the neural network are adjusted according to the loss value.
  • the loss value is used as a benchmark to perform BP to update the parameters of the neural network.
  • the distance between Anchor and Positive needs to be smaller, and the distance between Anchor and Negative needs to be larger, to eventually obtain a smallest interval a between the first distance and the second distance.
  • data enhancement may be performed on the training data before training.
  • a process of the data enhancement may use rotation, translation, magnification, and principal component analysis (PCA) color enhancement.
  • Multiple fundus images using random enhancement parameters may be generated by performing data enhancement on each fundus image.
  • fundus images after data enhancement may adopt a uniform format of a 224*224 pixels and RGB color channels.
  • a fundus image may first be cropped, and then subjected to data enhancement, or the fundus image may first be subjected to data enhancement, and then cropped, there is no limitation thereto.
  • a neural network training method training is performed with triple training data and a corresponding loss function, so that a distance between feature vectors extracted by a neural network from different fundus images of the same eye can be gradually decreased, and a distance between feature vectors extracted from fundus images of different eyes can be increased.
  • distances between feature vectors repeatedly extracted by the neural network from the same fundus image are small enough, and distances to feature vectors of other fundus images are large enough, that is, the information has particular uniqueness, so that feature vectors extracted by the neural network from fundus images can be used as identity information of a user.
  • the fundus image may first be segmented, to obtain a fundus feature image as training data.
  • fundus features may be extracted by using a computer vision algorithm or a machine learning algorithm.
  • fundus features in the fundus image are extracted by using a segmental neural network, to obtain a probability plot or a binary image containing a fundus feature confidence level.
  • the fundus image may be divided into a plurality of image patches, the size of which is set according to the size of the fundus image.
  • the size of the image patch obtained after division should be significantly smaller than the size of the entire fundus image. For example, if the fundus image has a size of 1000*1000 (pixels), the size of the image patch obtained after division is 100*100 (pixels).
  • a blood vessel image in each image patch is segmented using a preset segmentation model to obtain segmented image patches.
  • the segmentation model may be a neural network such as a fully convolutional network (FCN), SegNet or DeepLab. Before use, the segmentation model should be trained by the sample data to provide the segmentation model with a particular semantic segmentation capability. The training may be performed using sample image patches manually labeled with a blood vessel region.
  • the segmentation model extracts features of a blood vessel image in the image patch, and segments the blood vessel image into patches according to the extracted features with the blood vessel image highlighted.
  • highlighting For example, various pixel values clearly different from the background may be used to indicate positions of blood vessels or the like.
  • the image patch shown in FIG. 6 is input into the segmentation model to obtain a segmented image patch shown in FIG. 7 .
  • the segmentation model used in this embodiment outputs a binary image.
  • the binary image uses two pixel values to respectively express the background and the blood vessel image, to intuitively highlight positions of blood vessels. Segmented image patches are spliced to form a fundus blood vessel image, to obtain an image shown in FIG. 8 , and then the image shown in FIG. 8 is used as training data.
  • a similar method may be used to extract other features, for example: features such as the optic disc, macula, and retina.
  • features such as the optic disc, macula, and retina.
  • the fundus feature image may be advanced indirect features (or referred to as abstract features) in the fundus feature image, for example, positions and directions of blood vessel bifurcations, positions and directions of blood vessel intersections, and blood vessel vectograph.
  • advanced indirect features or referred to as abstract features
  • the foregoing indirect features may be extracted from the fundus image as training data.
  • the present disclosure may use a form of hardware only embodiments, software only embodiments, or embodiments with a combination of software and hardware. Therefore, the present disclosure may use a form of hardware only embodiments, software only embodiments, or embodiments with a combination of software and hardware. Moreover, the present disclosure may use a form of a computer program product that is implemented on one or more computer-usable storage media (including but not limited to a disk memory, a compact disc read-only memory (CD-ROM), an optical memory, and the like) that include computer usable program code.
  • computer-usable storage media including but not limited to a disk memory, a compact disc read-only memory (CD-ROM), an optical memory, and the like
  • These computer program instructions may be provided for a general-purpose computer, a dedicated computer, an embedded processor, or a processor of any other programmable data processing device to generate a machine, so that the instructions executed by a computer or a processor of any other programmable data processing device generate an apparatus for implementing a specific function in one or more processes in the flowcharts and/or in one or more blocks in the block diagrams.
  • These computer program instructions may be stored in a computer readable memory that can instruct the computer or any other programmable data processing device to work in a specific manner, so that the instructions stored in the computer readable memory generate an artifact that includes an instruction apparatus.
  • the instruction apparatus implements a specific function in one or more processes in the flowcharts and/or in one or more blocks in the block diagrams.
  • These computer program instructions may be loaded onto a computer or another programmable data processing device, so that a series of operations and steps are performed on the computer or the another programmable device, thereby generating computer-implemented processing. Therefore, the instructions executed on the computer or the another programmable device provide steps for implementing a specific function in one or more processes in the flowcharts and/or in one or more blocks in the block diagrams.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Evolutionary Computation (AREA)
  • Ophthalmology & Optometry (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Biophysics (AREA)
  • Computer Security & Cryptography (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Hardware Design (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Vascular Medicine (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Epidemiology (AREA)
  • Primary Health Care (AREA)
  • Public Health (AREA)
  • Eye Examination Apparatus (AREA)
  • Image Analysis (AREA)
  • Collating Specific Patterns (AREA)

Abstract

The present disclosure provides a method and device to process identity information based on a fundus image. An identity information comparison method includes: recognizing a fundus image by using a neural network to obtain a multi-dimensional feature vector representing the identity of a user; comparing the obtained multi-dimensional feature vector with each multi-dimensional feature vector stored in a database; and determining, according to a comparison result, whether there is a matching between the currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database.

Description

    TECHNICAL FIELD
  • The present disclosure relates to the field of image information processing, and in particular to an identity information processing method and device based on a fundus image.
  • BACKGROUND
  • A fundus image (also referred to as a retinal image) captured using a fundus camera can show human tissues such as the macula, optic disc, and blood vessels. The fundus image is highly unique because the vascular orientation, bifurcations, and the shape of the optic disc are different from person to person, but the fundus image of a particular person does not change significantly with age.
  • Current technologies use fundus images to identify identity information. Accordingly a database of identity information based on fundus images needs to be established. An important question is what information to store during the establishment of a database. Most of the literatures, similar to a fingerprint or facial recognition technologies, identify and extract information about shape features of some key points using computer vision methods. Such information is stored in the database for subsequent comparison. However, the analysis of a fundus image is very different from that of a fingerprint image and a face image, and it is very difficult to determine key points and to find corresponding key points between two fundus images, resulting in poor usability. Therefore, in some existing technologies, a fundus image of a user is directly stored. That is, the fundus image is directly used as identity information of the user.
  • In a database with a large amount of identity information of users, when the identity information of a current user is obtained, the first step is to determine whether the identity information of the user is already stored in the database. If fundus images are stored in the database, all the fundus images already stored in the database need to be separately compared with the current fundus image. This approach is very inefficient.
  • SUMMARY
  • In view of this, an implementation of the present disclosure provides an identity information comparison method based on a fundus image, including:
  • recognizing a fundus image by using a neural network to obtain a multi-dimensional feature vector representing the identity of a user;
  • comparing the obtained multi-dimensional feature vector with each multi-dimensional feature vector pre-stored in a database; and
  • determining, according to a comparison result, whether there is a matching between the currently obtained multi-dimensional feature vector and a multi-dimensional feature vector matching pre-stored in the database.
  • Optionally, the neural network is trained with triple sample data. The triple sample data includes a first fundus image sample, a second fundus image sample, and a third fundus image sample. The second fundus image sample and the first fundus image sample are fundus images of the same person, and the third fundus image sample and the first fundus image sample are fundus images of different people.
  • Optionally, in a process of training the neural network, the neural network separately extracts multi-dimensional feature vectors of the first fundus image sample, the second fundus image sample, and the third fundus image sample; calculates a first distance between the second fundus image sample and the first fundus image sample, and a second distance between the third fundus image sample and the first fundus image sample according to the three extracted multi-dimensional feature vectors; obtains a loss value according to the first distance and the second distance; and adjusts parameters of the neural network according to the loss value.
  • Optionally, the adjusting parameters of the neural network according to the loss value includes:
  • feeding back the loss value to the neural network, to enable the neural network to adjust the parameters according to the loss value to decrease the first distance and increase the second distance until the first distance is smaller than the second distance by a preset value.
  • Optionally, the determining, according to a comparison result, whether there is a matching between the currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database includes:
  • calculating a distance between each pre-stored multi-dimensional feature vector and the currently obtained multi-dimensional feature vector; and
  • determining, according to the distance, whether each pre-stored multi-dimensional feature vector matches the currently obtained multi-dimensional feature vector.
  • Optionally, the fundus image includes a left-eye fundus image and a right-eye fundus image. The database is configured to store user data. Each group of user data includes a first pre-stored multi-dimensional feature vector corresponding to a left eye and a second pre-stored multi-dimensional feature vector corresponding to a right eye. The multi-dimensional feature vector includes a first multi-dimensional feature vector corresponding to the left-eye fundus image and a second multi-dimensional feature vector corresponding to the right-eye fundus image.
  • Optionally, the fundus image includes a left-eye fundus image and a right-eye fundus image, the database is configured to store user data, and each group of user data includes one pre-stored multi-dimensional feature vector.
  • The obtaining a multi-dimensional feature vector representing the identity of a user includes:
  • acquiring the first multi-dimensional feature vector corresponding to the left-eye fundus image and the second multi-dimensional feature vector corresponding to the right-eye fundus image that are output by the neural network; and
  • combining the first multi-dimensional feature vector and the second multi-dimensional feature vector to obtain the multi-dimensional feature vector representing the identity of the user.
  • Optionally, the fundus image is a left-eye fundus image or a right-eye fundus image, the database is configured to store user data, and each group of user data includes a first pre-stored multi-dimensional feature vector corresponding to a left eye or a second pre-stored multi-dimensional feature vector corresponding to a right eye.
  • The present disclosure further provides an identity verification method based on a fundus image, including:
  • acquiring a fundus image of a user; and
  • determining, by using the foregoing identity information comparison method based on a fundus image, whether there is a matching between a multi-dimensional feature vector of the fundus image and a multi-dimensional feature vector pre-stored in a database, to complete verification of the identity of the user.
  • The present disclosure further provides an identity information storage method based on a fundus image, including:
  • acquiring a fundus image of a user; and
  • determining, by using the foregoing identity information comparison method based on a fundus image, whether there is a matching between a multi-dimensional feature vector of the fundus image and a multi-dimensional feature vector pre-stored in a database; and
  • when there is no matching between a currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database, storing the currently obtained multi-dimensional feature vector in the database as identity information of the user.
  • The present disclosure further provides a method for training a fundus image recognition model, including:
  • acquiring training data, where the training data includes a first fundus image sample, a second fundus image sample, and a third fundus image sample, the second fundus image sample and the first fundus image sample are fundus images of the same person, and the third fundus image sample and the first fundus image sample are fundus images of different people;
  • recognizing the first fundus image sample, the second fundus image sample, and the third fundus image sample by using a fundus image recognition model to obtain a loss value; and
  • adjusting parameters of the fundus image recognition model according to the loss value.
  • Optionally, the recognizing the first fundus image sample, the second fundus image sample, and the third fundus image sample by using a fundus image recognition model to obtain a loss value includes:
  • calculating a first distance between the second fundus image sample and the first fundus image sample;
  • calculating a second distance between the third fundus image sample and the first fundus image sample; and
  • obtaining the loss value according to the first distance and the second distance.
  • Optionally, the adjusting parameters of the fundus image recognition model according to the loss value includes:
  • feeding back the loss value to the fundus image recognition model; and
  • adjusting the parameters according to the loss value to decrease the first distance and increase the second distance, until the first distance is smaller than the second distance by a preset value.
  • Correspondingly, an implementation of the present disclosure further provides a device for an identity information comparison based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity information comparison method based on a fundus image.
  • Correspondingly, an implementation of the present disclosure further provides a device for identity verification based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity verification method based on a fundus image.
  • Correspondingly, an implementation of the present disclosure further provides a device for an identity information storage based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity information storage method based on a fundus image.
  • Correspondingly, an implementation of the present disclosure further provides a device for training a fundus image recognition model, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing method for training a fundus image recognition model.
  • According to the method and device of the identity information comparison based on a fundus image provided in the present disclosure, a neural network is first used to convert a fundus image of a user into a multi-dimensional feature vector representing the identity of the user, and then abstract feature information related to personal features of the user is extracted by using characteristics of the neural network. During comparison, implementations of the disclosure may determine by comparing multi-dimensional feature vectors whether there is data matching with the current user in a database. In this implementation, the database does not need to store a fundus image. It is also unnecessary to recognize pre-stored fundus images repeatedly every time a new fundus image is obtained, so that the efficiency of operation of an identity information comparison can be improved.
  • A neural network provided according to the implementation of the present disclosure is trained with triple training data and a corresponding loss function, so that a distance between feature vectors extracted by the neural network from different fundus images of the same eye can be decreased, and a distance between feature vectors extracted from fundus images of different eyes can be increased during the training process. After training, distances between feature vectors repeatedly extracted by the neural network from the same fundus image are small enough, and distances to feature vectors of other fundus images are large enough, that is, the information has particular uniqueness, so that feature vectors extracted by the neural network from fundus images can be used as identity information of a user.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • To describe the technical schemes in specific embodiments of the present disclosure or the existing technology more clearly, the following briefly introduces the accompanying drawings required for describing the specific embodiments or the existing technology. Apparently, the accompanying drawings in the following description show some embodiments of the present disclosure, and a person having ordinary skills in the art may still derive other drawings from these accompanying drawings without creative efforts.
  • FIG. 1 is a flowchart of an identity information comparison method according to an embodiment of the present disclosure;
  • FIG. 2 is a flowchart of an identity information storage method according to an embodiment of the present disclosure;
  • FIG. 3 is a schematic diagram of extracting identity information by using a neural network according to an embodiment of the present disclosure;
  • FIG. 4 is another schematic diagram of extracting identity information by using a neural network according to an embodiment of the present disclosure;
  • FIG. 5 is a flowchart of a neural network training method according to an embodiment of the present disclosure;
  • FIG. 6 shows an image patch in a fundus image;
  • FIG. 7 shows a segmentation result of the image patch shown in FIG. 6; and
  • FIG. 8 shows a fundus blood vessel image.
  • DETAILED DESCRIPTION
  • The following clearly and completely describes the technical schemes of the present disclosure with reference to the accompanying drawings. Apparently, the described embodiments are merely some rather than all of the embodiments of the present disclosure. All other embodiments obtained by a person having ordinary skills in the art based on the embodiments of the present disclosure without creative efforts shall fall within the protection scope of the present disclosure.
  • In addition, the technical features involved in different embodiments of the present disclosure described below can be combined with each other as long as they do not constitute a conflict between them.
  • Embodiments of the present disclosure provide an identity information storage method based on a fundus image. The method may be performed by an electronic device such as a computer or a server. As shown in FIG. 1, the method includes the following steps S1 to S3.
  • At step S1, a fundus image is recognized by a neural network to obtain a multi-dimensional feature vector representing the identity of a user. The neural network extracts feature information during recognition of an image. For different tasks, the neural network focuses on different content, and different feature information is extracted. For example, during the execution of a classification task, the neural network extracts corresponding feature information (usually a multi-dimensional feature vector) according to a category of a fundus image, and then performs classification according to the feature information.
  • In this embodiment, the neural network used is configured to extract the multi-dimensional feature vector representing the identity of a user, rather than to perform a classification task or an image segmentation task. For different people (users), multi-dimensional feature vectors extracted by the neural network from fundus images of people should be different. For the same eye of the same person, during repeated recognition, multi-dimensional feature vectors extracted at different times should be the same (or approximately the same or similar). The neural network in this application may be a deep convolutional neural network (CNN), and extracted multi-dimensional feature vectors can be normalized by setting an appropriate loss function and using a back propagation (BP) algorithm. Then, a trained CNN model can be used to extract feature vectors, which are usually high-dimensionality vectors, from a fundus image.
  • To enable the neural network to extract expected content, the neural network should be trained beforehand. There are many training methods, and different training data are used for different training methods. Details are described below in detail.
  • At step S2, the obtained multi-dimensional feature vector is compared with each multi-dimensional feature vector pre-stored in a database. According to a configuration position of the database, the database may be categorized into a database configured based on a graphics processing unit (GPU) and a database configured based on a central processing unit (CPU). A multi-dimensional feature vector pre-stored in the database may alternatively be the multi-dimensional feature vector extracted by the neural network in step S1 from another fundus image.
  • There are many ways to compare two feature vectors. A comparison result is used to represent a similarity between the two feature vectors. A similarity between multi-dimensional vectors may be determined based on Euclidean distance, cosine similarity, normalized Euclidean distance, or the like. The Euclidean distance measures an absolute distance between points in a multi-dimensional space, which is preferred when data is very dense and continuous. Since the calculation is based on absolute values of features of different dimensions, Euclidean measurement needs to ensure that indicators of different dimensions are at the same scale level.
  • In some specific scenarios, Mahalanobis distance may be used. The Mahalanobis distance is a distance based on sample distribution. For example, for two populations of normal distribution with an average value of a and b, respectively but having different variances, the probability of which distribution a sample point A belongs to in a distribution space is higher determines that A belongs to this distribution.
  • Because different constraint distances are used to establish a feature space, different comparison methods should be adopted when comparing the results.
  • During practical application, multi-dimensional feature vectors extracted by the neural network from fundus images of the same person captured at different moments and in different environments are usually not identical. Therefore, using a distance calculation method to measure a similarity between two feature vectors has a certain fault tolerance, and the distance is preferably a Euclidean distance.
  • In another optional embodiment, it is also feasible to calculate an angle between each pre-stored multi-dimensional feature vector and the currently obtained multi-dimensional feature vector to measure a similarity therebetween.
  • At step S3, whether there is a matching between the currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database is determined according to a comparison result. Different comparison methods may result in different comparison results. For example, whether each pre-stored multi-dimensional feature vector matches the currently obtained multi-dimensional feature vector may be determined based on the distance or the angle.
  • Taking distance as an example, when the distance between the two feature vectors is less than a set threshold, it can be determined that the two feature vectors have a sufficiently high similarity, indicating that the two match each other.
  • According to an identity information comparison method based on a fundus image provided in the embodiments of the present disclosure, a neural network is first used to convert a fundus image of a user into a multi-dimensional feature vector used for representing the identity of the user, and then abstract feature information related to personal features of the user is extracted using characteristics of the neural network. During comparison, whether there is data matching the current user in a database can be determined by comparing multi-dimensional feature vectors. In this scheme, the database does not need to store fundus images. It is also unnecessary to recognize pre-stored fundus images repeatedly every time a new fundus image is obtained, thereby improving the efficiency of an identity information comparison operation.
  • This scheme may be used to process a fundus image of one eye of a user. That is, the fundus image recognized in step S1 is a left-eye fundus image or a right-eye fundus image. Correspondingly, feature vectors stored in the database are from fundus images of one eye.
  • This scheme may be extended to process fundus images of two eyes of a user. That is, the fundus images in step S1 include a left-eye fundus image and a right-eye fundus image. The two fundus images are separately recognized by the neural network, to obtain a first multi-dimensional feature vector corresponding to the left-eye fundus image and a second multi-dimensional feature vector corresponding to the right-eye fundus image. Correspondingly, each group of data stored in the database includes two pre-stored multi-dimensional feature vectors, which may be correspondingly compared in step S2.
  • When comparison is separately performed for two eyes, it is possible that a feature vector of one eye matches the data in the database whereas a feature vector of the other eye does not match the data in the database. Such a result is acceptable in some application scenarios, for example, during establishment or update of the database. However, such a result is not allowed in certain application scenarios, for example, during identity verification.
  • In a preferred embodiment, according to this scheme, fundus images of two eyes of a user are uniformly processed. Recognized fundus images include a left-eye fundus image and a right-eye fundus image. The fundus images of two eyes are separately recognized by using the neural network in step S1, and a first multi-dimensional feature vector corresponding to the left-eye fundus image and a second multi-dimensional feature vector corresponding to the right-eye fundus image are output. The two feature vectors are then combined, which may be performed in multiple ways. For example, two 1024-dimensional feature vectors may be connected into one 2048-dimensional feature vector.
  • Correspondingly, the pre-stored data in the database is a combined multi-dimensional feature vector, and is a result of combining two feature vectors in advance. During comparison in step S2, the currently combined dimensional feature vector is compared with a pre-stored multi-dimensional feature vector in the database, to further determine, according to a comparison result, whether there is matching data. In this embodiment, it only needs to perform one comparison to determine whether there are matching data of the two eyes.
  • An identity information storage method based on a fundus image is described below. The method uses the foregoing comparison embodiment to establish and manage an identity information database. As shown in FIG. 2, the method includes the following steps S1A to S6A.
  • At step S1A, a fundus image of a user is acquired. According to the foregoing introduction, the fundus image may be a fundus image of either or both eyes.
  • At step S2A, the fundus image is recognized using a neural network to obtain a multi-dimensional feature vector representing the identity of the user. Reference may be made to the foregoing step S1 for details, which will not be described herein.
  • At step S3A, the obtained multi-dimensional feature vector is compared with each multi-dimensional feature vector pre-stored in a database. Reference may be made to the foregoing step S2, which will not be described herein.
  • At step S4A, whether there is a matching between the currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database is determined according to a comparison result. When there is no pre-stored multi-dimensional feature vector matching the currently obtained multi-dimensional feature vector in the database, step S5A is performed, or otherwise step S6A is performed.
  • It needs to be noted that the database may initially be empty, and has no pre-stored multi-dimensional feature vector, thus there is no matching content. A plurality of pieces of data may be imported into the initial database in advance. The data uses individuals as basic modules. For example, each piece of user data includes the name of the user and a multi-dimensional feature vector of the user (which is obtained based on a fundus image of the user), and may additionally include personal information such as the gender, age, diopter of the user, and the like.
  • If it is determined after one-by-one comparison that there is no pre-stored multi-dimensional feature vector matching the multi-dimensional feature vector of the current user in the database, it indicates that identity information of the current user is not stored in the database; or otherwise it indicates that the identity information of the current user has been stored in the database.
  • At step S5A, the currently obtained multi-dimensional feature vector is stored in the database. Specifically, when a multi-dimensional feature vector is not found in an original database, the system may prompt whether it is necessary to add the currently obtained multi-dimensional feature vector to the original database as a new piece of data. If YES is chosen, a request to memory is sent, and after the request succeeds, the new piece of data is added to the original database. Further, the system may further prompt whether to continue to input related supplementary information, for example, personal information such as the gender, age, the diopter, and the like of the user. After the information is entered, the information and the corresponding multi-dimensional feature vectors are stored as related data items.
  • At step S6A, the pre-stored multi-dimensional feature vector matching the currently obtained multi-dimensional feature vector in the database is replaced with the currently obtained multi-dimensional feature vector. Specifically, when a multi-dimensional feature vector matching the current multi-dimensional feature vector is found in the database, the system may prompt whether it is necessary to overwrite existing data with the currently obtained multi-dimensional feature vector. If YES is chosen, overwriting is performed. This step is an optional operation. As an alternative, when a multi-dimensional feature vector matching the current multi-dimensional feature vector is found in the database, the system may only prompt that current user data already exists, and a replacement operation is not required.
  • According to the identity information storage method based on a fundus image provided in the embodiments of the present disclosure, a neural network is first used to convert a fundus image of a user into a multi-dimensional feature vector used for representing the identity of the user, and then abstract feature information related to personal features of the user is extracted using characteristics of the neural network. During storage, whether there is data matching the current user in a database can be determined by comparing multi-dimensional feature vectors, so as to store the multi-dimensional feature vector. In this scheme, the database does not need to store a fundus image. It is also not necessary to recognize an existing fundus image by using the neural network during the storage of every piece of identity information of the user, thereby improving the efficiency of storing identity information of users.
  • As exemplary description, each piece of information of the user in the database established according to the scheme may include the name of the user, a multi-dimensional feature vector (a first multi-dimensional feature vector of a left eye and/or a second multi-dimensional feature vector of a right eye or a combination of the two), the gender, age, diopter of the user, and the like. The multi-dimensional feature vector represents the identity information of the user.
  • A process of establishing and managing a database is described in the foregoing embodiments. Some changes may be further made on this basis to obtain an identity verification scheme. Specifically, according to the conclusion of step S4A, that is, whether the identity information of the current user is stored in the database, a corresponding identity determination operation can be performed. For example, at step S5A, it is determined that the current user is not a known user, and therefore the current user is not allowed to perform a subsequent operation; and at step S6A, it is determined that the current user is a known user, and therefore the current user is allowed to perform a subsequent operation. The subsequent operation may be unlocking or logging in an electronic device, or the like, to implement identity verification, authentication, or the like.
  • An embodiment of the present disclosure further provides a device of an identity information comparison based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity information comparison method based on a fundus image.
  • An embodiment of the present disclosure further provides a device of an identity information storage based on a fundus image, including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity information storage method based on a fundus image.
  • An embodiment of the present disclosure further provides a device of an identity verification based on a fundus image, the device including: at least one processor, and a memory communicatively connected to the at least one processor, where the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to carry out the foregoing identity verification method based on a fundus image.
  • How to obtain a model for generating a feature vector is described below with reference to FIG. 3 to FIG. 5 (the neural network in the foregoing embodiments). Aspects to extract features may include: providing an optimized intra-class distance, to make the intra-class distance smaller; and providing an optimized inter-class distance, to make the inter-class distance more distinguishable. The intra-class distance refers to a distance between sample points of each pattern in the same class; and the inter-class distance refers to a distance between different classes. To achieve this objective and extract better feature vectors, there are the following optional embodiments.
  • In a first optional embodiment, FIG. 3 shows a training framework of a softmax-based classification model. The framework of the classification model may be applied to training and extraction of identity features of fundus images, and on this basis, various auxiliary loss functions are provided to improve the characteristics of the identity features of the fundus images. A fundus image Imgi is fed into a deep CNN, to obtain a multi-dimensional feature vector x1, and then passes through a fully connected layer and is processed by softmax, to obtain a score vector. The softmax loss function is shown as follows:
  • L 1 = - 1 N i = 1 N log e w y i T x i + b y i i = 1 n e w j T x i + b j ,
  • where N is a batch size, n is a quantity of classes, xi denotes a feature vector of an ith fundus image of a batch, y1 denotes an actual identity class (label) of the fundus image, Wj denotes a vector of a jth column of a weight W of the fully connected layer,
  • e w v i T x i + b y i i = 1 n e w j T x i + b j ( 0 , TagBox[",", "NumberComma", Rule[SyntaxForm, "0"]] 1 )
  • denotes a probability value that xi is predicted as the actual identity class of the fundus image, and bj is a bias value.
  • For example, a classification model (1, 2, 3) with n=3 is defined, where classes 1, 2, and 3 denote identity labels of three users. A score vector si obtained for the fundus image Imgi is 0.1, 0.2, 0.7). It is determined that Imgi is a fundus image of a user with a label of 3. A contribution of Imgi to a loss value is L1=−log(0.7)=0.3567. However, if the obtained score vector si is 0.1, 0.5, 0.4). A contribution of Imgi to the loss value is L1=−log(0.4)=0.9163. As shown, when a probability value that the model predicts the fundus image to be a correct identity class is large, the contribution to the loss value is small, or otherwise, the contribution to the loss value is large. A model trained using this loss function is used to extract a feature vector that can be used for correct classification.
  • It is highly crucial for an identity recognition task to minimize a difference between internal classes while ensuring that features of different classes are distinguishable. To achieve this, the following loss function may be set:
  • L = L s + λ L c = - i = 1 m log e w y i T x i + b y i i = 1 n e w j T x i + b j + λ 2 i = 1 m x i - c yi 2 2 .
  • Joint supervision of a softmax loss (Ls) and a center loss (Lc) is used to train the deep CNN to perform discriminative feature learning. The center loss can effectively characterize an intra-class change. A scalar λ is used for striking a balance between Ls and Lc. When λ has different values, feature vectors have different distributions.
  • During training, a mini-batch of fundus images are input for each iteration, and m is the number of images input in each iteration. xi is a feature vector of an ith image in one batch. yi is an identity label of a user to which the fundus image xi belongs. Cyi∈Rd denotes the center of a feature vector of an identity classification label of a yi th user, and d denotes a dimensionality of a feature vector. During training, with every iteration, the center Cyi of a feature vector of each user label is updated. The center Cyi is calculated by averaging features of corresponding classification labels. Next, to prevent a small number of incorrectly labeled samples from causing a major disturbance, a scalar α∈[0, 1] may be used to control a learning rate of the center. This scheme is distinguished in that the difference between internal classes can be reduced to the greatest extent while ensuring that features of different classes can be separated without complex recombination of training sets.
  • In a second optional embodiment, based on that joint supervision of a softmax loss (Ls) and a contrastive loss (Lc) is used to train the deep CNN to perform discriminative feature learning, not only the accuracy of classification is taken into consideration, but also an inter-class distance m (margin) is added to expand a decision boundary, so that positive samples have a higher similarity, and negative samples have a lower similarity.
  • As shown in FIG. 4, a softmax layer with n channels is introduced after the structure of the deep CNN to obtain a recognition module for classifying an identity class of an input fundus image, where n is the number of user identity classes of the user, and is defined as an identification loss function:

  • Ident(f,t,θ id)=−Σi=1 n p i log {circumflex over (p)} i=−log {circumflex over (p)} t.
  • The foregoing formula is in fact a cross-entropy loss function, where f is a feature vector, t is a user identity class corresponding to the feature vector f, θid is a parameter of the softmax layer, and pi is a target probability distribution. For the user identity class t, pt=1, and for the remaining, pt=0. pi is a predicted probability distribution. During training, since the network integrates the following verification loss function, which is in fact a contrastive loss, before training is performed using the following loss function, fundus images in a training set should be first paired to obtain image pairs, and it is labeled whether each image pair includes identity features of the same user. For feature vectors (fi, li) and (f1, lj) of an input image pair, if li=lj, it is labeled that the label is yij=0; or otherwise, it is labeled that the label is yij=1, as shown below:
  • Verif ( f i , f j , y ij , θ ve ) = { 1 2 f i - f j 2 2 if y ij = 1 1 2 max ( 0 , m - f i - f j 2 ) 2 if y ij = - 1 .
  • During training, a fundus image pair is input into the deep CNN to obtain image features. Identification loss is separately calculated for the fundus image pair. That is, user identity classification is performed on two input fundus images, and verification is performed on the fundus image pair at the same time, to verify whether two input fundus images of the fundus image pair are the same. When the input fundus image pair have the same identity, that is, yij=0, a feature vector of the fundus image pair is (fi, fj), Verif(fi, fj, yi,j, θve)=½∥fi−fj2 2. When the input fundus image pair do not have the same identity, that is, yij=1, a feature vector of the fundus image pair is (fi, fj), Verif(fi, fj, yij, θve)=½(max(0, m−∥fi−fj2))2, . . . , where L2 normalization is used for a similarity distance of the feature vector (fi, f1). L1/L2 normalization or a cosine similarity is adopted, and m (margin) defines an inter-class margin, which acts on the loss value only when a distance in the image pair with different identities is within the distance m. This effectively restricts an inter-class distance, and therefore the inter-class distance is more distinguishable.
  • In a third optional embodiment, as shown in FIG. 5, embodiments of the present disclosure further provide a method for training a triple data-based model. The method includes the following steps S1B to S3B.
  • At step S1B, training data is acquired, where the training data includes a first fundus image sample, a second fundus image sample, and a third fundus image sample. The training data used in this embodiment is triple sample data. The second fundus image sample and the first fundus image sample are fundus images of the same eye of the same person. The third fundus image sample and the first fundus image sample are fundus images of different people.
  • In a specific embodiment, a data set is first prepared. As an example, the data set may include fundus images of n eyes. Each eye corresponds to m fundus images. That is, there are a total of n*m fundus images in the data set. To obtain training data, one fundus image is first randomly selected from the n*m fundus images, which is referred to as Anchor. Then another fundus image of the same eye as Anchor is selected, which is referred to as Positive (denoted as x−p); and then one fundus image of an eye different from Anchor is selected, which is referred to as Negative (denoted as x−n), so that a group of training data (Anchor, Positive, Negative) is obtained.
  • Before a neural network is trained by using the training data, a fundus image may be preprocessed first to allow the trained neural network to be more accurate during recognition of the fundus image. Specifically, each fundus image may first be cropped. Because a captured fundus image originally has a relatively large black background, the fundus images may first be trimmed to remove a large number of black pixels in the background, until all fundus images are cropped to the smallest rectangles that can contain the entire circular fundus.
  • In a specific embodiment, all fundus images may be cropped to a uniform format, for example, a size of 224*224 pixels. Fundus images used for model training and recognition may adopt a uniform format of 224*224 pixels and RGB color channels.
  • At step S2B, the first fundus image sample, the second fundus image sample, and the third fundus image sample are recognized by the neural network to obtain a loss value. A first distance between the second fundus image sample and the first fundus image sample, and a second distance between the third fundus image sample and the first fundus image sample are calculated according to a preset loss function, and a loss value is obtained according to the first distance and the second distance.
  • Specifically, the neural network separately extracts features of the foregoing three fundus images, to obtain three multi-dimensional feature vectors, which are respectively denoted as: f (xi a), f (xi p), and f (xi n). Then a first distance between f (xi a) and f (xi p), and a second distance between f (xi a) and f (xi n) are calculated. The first distance and the second distance in this embodiment are Euclidean distances.
  • The loss value is calculated by using the first distance and the second distance, which may be performed by the following relational expression of the loss function.

  • Σi N[∥f(x i a)−f(x i p)∥2 2 −∥f(x i a)−f(x i n2 2 +a]+,
  • where a denotes a preset value which is a minimum distance between the first distance and the second distance. + denotes that when a value in [ ] is greater than 0, this value is taken as the loss value, and when the value in [ ] is less than 0, the loss is 0.
  • At step S3B, parameters of the neural network are adjusted according to the loss value. The loss value is used as a benchmark to perform BP to update the parameters of the neural network.
  • During actual training, a large amount of the foregoing triple sample data needs to be used, to continuously adjust the parameters of the neural network, until the loss function converges. In a process of loss transmission of the neural network, the distance between Anchor and Positive needs to be smaller, and the distance between Anchor and Negative needs to be larger, to eventually obtain a smallest interval a between the first distance and the second distance.
  • To improve the robustness of the neural network, data enhancement may be performed on the training data before training. A process of the data enhancement may use rotation, translation, magnification, and principal component analysis (PCA) color enhancement. Multiple fundus images using random enhancement parameters may be generated by performing data enhancement on each fundus image. For example, fundus images after data enhancement may adopt a uniform format of a 224*224 pixels and RGB color channels. During actual operations, a fundus image may first be cropped, and then subjected to data enhancement, or the fundus image may first be subjected to data enhancement, and then cropped, there is no limitation thereto.
  • According to a neural network training method provided in the embodiments of the present disclosure, training is performed with triple training data and a corresponding loss function, so that a distance between feature vectors extracted by a neural network from different fundus images of the same eye can be gradually decreased, and a distance between feature vectors extracted from fundus images of different eyes can be increased. After training, distances between feature vectors repeatedly extracted by the neural network from the same fundus image are small enough, and distances to feature vectors of other fundus images are large enough, that is, the information has particular uniqueness, so that feature vectors extracted by the neural network from fundus images can be used as identity information of a user.
  • In an optional embodiment, to further eliminate interference image information unrelated to fundus recognition and improve the recognition capability of the neural network, before training, the fundus image may first be segmented, to obtain a fundus feature image as training data.
  • After the fundus image is obtained, fundus features may be extracted by using a computer vision algorithm or a machine learning algorithm. For example, fundus features in the fundus image are extracted by using a segmental neural network, to obtain a probability plot or a binary image containing a fundus feature confidence level. As shown in FIG. 6, the fundus image may be divided into a plurality of image patches, the size of which is set according to the size of the fundus image. In most cases, the size of the image patch obtained after division should be significantly smaller than the size of the entire fundus image. For example, if the fundus image has a size of 1000*1000 (pixels), the size of the image patch obtained after division is 100*100 (pixels).
  • A blood vessel image in each image patch is segmented using a preset segmentation model to obtain segmented image patches. The segmentation model may be a neural network such as a fully convolutional network (FCN), SegNet or DeepLab. Before use, the segmentation model should be trained by the sample data to provide the segmentation model with a particular semantic segmentation capability. The training may be performed using sample image patches manually labeled with a blood vessel region.
  • The segmentation model extracts features of a blood vessel image in the image patch, and segments the blood vessel image into patches according to the extracted features with the blood vessel image highlighted. There are a plurality of ways for highlighting. For example, various pixel values clearly different from the background may be used to indicate positions of blood vessels or the like.
  • The image patch shown in FIG. 6 is input into the segmentation model to obtain a segmented image patch shown in FIG. 7. The segmentation model used in this embodiment outputs a binary image. The binary image uses two pixel values to respectively express the background and the blood vessel image, to intuitively highlight positions of blood vessels. Segmented image patches are spliced to form a fundus blood vessel image, to obtain an image shown in FIG. 8, and then the image shown in FIG. 8 is used as training data.
  • In an optional embodiment, a similar method may be used to extract other features, for example: features such as the optic disc, macula, and retina. Through the extraction of fundus features, interference image information unrelated to fundus identity recognition can be greatly eliminated, thereby significantly improving the recognition performance of the model.
  • There may be advanced indirect features (or referred to as abstract features) in the fundus feature image, for example, positions and directions of blood vessel bifurcations, positions and directions of blood vessel intersections, and blood vessel vectograph. After an original fundus image is acquired, the foregoing indirect features may be extracted from the fundus image as training data.
  • Therefore, the present disclosure may use a form of hardware only embodiments, software only embodiments, or embodiments with a combination of software and hardware. Therefore, the present disclosure may use a form of hardware only embodiments, software only embodiments, or embodiments with a combination of software and hardware. Moreover, the present disclosure may use a form of a computer program product that is implemented on one or more computer-usable storage media (including but not limited to a disk memory, a compact disc read-only memory (CD-ROM), an optical memory, and the like) that include computer usable program code.
  • The present disclosure is described with reference to the flowcharts and/or block diagrams of the method, the device (system), and the computer program product according to the embodiments of the present disclosure. It should be understood that computer program instructions may be used to implement each process and/or each block in the flowcharts and/or the block diagrams and a combination of a process and/or a block in the flowcharts and/or the block diagrams. These computer program instructions may be provided for a general-purpose computer, a dedicated computer, an embedded processor, or a processor of any other programmable data processing device to generate a machine, so that the instructions executed by a computer or a processor of any other programmable data processing device generate an apparatus for implementing a specific function in one or more processes in the flowcharts and/or in one or more blocks in the block diagrams.
  • These computer program instructions may be stored in a computer readable memory that can instruct the computer or any other programmable data processing device to work in a specific manner, so that the instructions stored in the computer readable memory generate an artifact that includes an instruction apparatus. The instruction apparatus implements a specific function in one or more processes in the flowcharts and/or in one or more blocks in the block diagrams.
  • These computer program instructions may be loaded onto a computer or another programmable data processing device, so that a series of operations and steps are performed on the computer or the another programmable device, thereby generating computer-implemented processing. Therefore, the instructions executed on the computer or the another programmable device provide steps for implementing a specific function in one or more processes in the flowcharts and/or in one or more blocks in the block diagrams.
  • Obviously, the foregoing embodiments are merely examples for clear description, rather than a limitation to implementations. For a person having ordinary skills in the art, other changes or variations in different forms may also be made based on the foregoing description. All implementations cannot and do not need to be exhaustively listed herein. Obvious changes or variations that are derived there from still fall within the protection scope of the disclosure of the present disclosure.

Claims (21)

1. A method of identity information comparison based on a fundus image, the method comprising:
recognizing a fundus image by using a neural network to obtain a multi-dimensional feature vector representing the identity of a user;
comparing the obtained multi-dimensional feature vector with each pre-stored multi-dimensional feature vector in a database; and
determining, according to the comparison result, whether there is a matching between a currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database.
2. The method of claim 1, wherein the neural network is trained using triple sample data, wherein the triple sample data comprises a first fundus image sample, a second fundus image sample, and a third fundus image sample, the second fundus image sample and the first fundus image sample are fundus images of a same person, and the third fundus image sample and the first fundus image sample are fundus images of different people.
3. The method of claim 2, wherein in a process of training the neural network, the neural network separately extracts multi-dimensional feature vectors of the first fundus image sample, the second fundus image sample and the third fundus image sample, calculates a first distance between the second fundus image sample and the first fundus image sample and a second distance between the third fundus image sample and the first fundus image sample according to the three extracted multi-dimensional feature vectors, obtains a loss value according to the first distance and the second distance, and adjusts parameters of the neural network according to the loss value.
4. The method of claim 3, wherein the adjusting parameters of the neural network according to the loss value comprises:
feeding back the loss value to the neural network, to enable the neural network to adjust the parameters according to the loss value to decrease the first distance and increase the second distance until the first distance is smaller than the second distance by a preset value.
5. The method of claim 1, wherein the determining, according to the comparison result, whether there is a matching between a currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database comprises:
calculating a distance between each pre-stored multi-dimensional feature vector and the currently obtained multi-dimensional feature vector respectively; and
determining, according to the distance, whether each pre-stored multi-dimensional feature vector matches the currently obtained multi-dimensional feature vector.
6. The method of claim 1, wherein the fundus image comprises a left-eye fundus image and a right-eye fundus image, the database is configured to store user data, each group of user data comprises a first pre-stored multi-dimensional feature vector corresponding to a left eye and a second pre-stored multi-dimensional feature vector corresponding to a right eye, and the multi-dimensional feature vector comprises a first multi-dimensional feature vector corresponding to the left-eye fundus image and a second multi-dimensional feature vector corresponding to the right-eye fundus image.
7. The method of claim 1, wherein the fundus image comprises a left-eye fundus image and a right-eye fundus image, the database is configured to store user data, and each group of user data comprises one pre-stored multi-dimensional feature vector; and
the obtaining a multi-dimensional feature vector representing the identity of a user comprises:
acquiring the first multi-dimensional feature vector corresponding to the left-eye fundus image and the second multi-dimensional feature vector corresponding to the right-eye fundus image that are output by the neural network; and
combining the first multi-dimensional feature vector and the second multi-dimensional feature vector to obtain the multi-dimensional feature vector representing the identity of the user.
8. The method of claim 1, wherein the fundus image is a left-eye fundus image or a right-eye fundus image, the database is configured to store user data, and each group of user data comprises a first pre-stored multi-dimensional feature vector corresponding to a left eye or a second pre-stored multi-dimensional feature vector corresponding to a right eye.
9. A method of identity verification based on a fundus image, the method comprising:
acquiring a fundus image of a user; and
determining, by using the method of claim 1, whether there is a matching between a multi-dimensional feature vector of the fundus image and a multi-dimensional feature vector matching pre-stored in a database, to complete verification of the identity of the user.
10. The method of claim 9, further comprising:
when there is no pre-stored multi-dimensional feature vector matching a currently obtained multi-dimensional feature vector in the database, storing the currently obtained multi-dimensional feature vector in the database as identity information of the user.
11. (canceled)
12. (canceled)
13. (canceled)
14. A device of identity information comparison based on a fundus image, the device comprising: at least one processor, and a memory communicatively connected to the at least one processor, wherein the memory stores an instruction executable by one processor, which when executed by the at least one processor, causes the at least one processor to perform operations comprising:
recognizing a fundus image by using a neural network to obtain a multi-dimensional feature vector representing the identity of a user;
comparing the obtained multi-dimensional feature vector with each pre-stored multi-dimensional feature vector in a database; and
determining, according to the comparison result, whether there is a matching between a currently obtained multi-dimensional feature vector and a multi-dimensional feature vector pre-stored in the database.
15. The device of claim 14, wherein the neural network is trained with triple sample data, the triple sample data comprises a first fundus image sample, a second fundus image sample, and a third fundus image sample, the second fundus image sample and the first fundus image sample are fundus images of a same person, and the third fundus image sample and the first fundus image sample are fundus images of different people.
16. The device of claim 15, wherein in a process of training the neural network, the neural network separately extracts multi-dimensional feature vectors of the first fundus image sample, the second fundus image sample and the third fundus image sample, calculates a first distance between the second fundus image sample and the first fundus image sample and a second distance between the third fundus image sample and the first fundus image sample according to the three extracted multi-dimensional feature vectors, obtains a loss value according to the first distance and the second distance, and adjusts parameters of the neural network according to the loss value.
17. The device of claim 16, wherein the adjusting parameters of the neural network according to the loss value comprises:
feeding back the loss value to the neural network, to enable the neural network to adjust the parameters according to the loss value to decrease the first distance and increase the second distance until the first distance is smaller than the second distance by a preset value.
18. The device of claim 14, wherein the determining, according to the comparison result, whether there is a prestored multi-dimensional feature vector matching a currently obtained multi-dimensional feature vector in the database comprises:
calculating a distance between each prestored multi-dimensional feature vector and the currently obtained multi-dimensional feature vector respectively; and
determining, according to the distance, whether each prestored multi-dimensional feature vector matches the currently obtained multi-dimensional feature vector.
19. The device of claim 14, wherein the fundus image comprises a left-eye fundus image and a right-eye fundus image, the database is configured to store user data, each group of user data comprises a first prestored multi-dimensional feature vector corresponding to a left eye and a second prestored multi-dimensional feature vector corresponding to a right eye, and the multi-dimensional feature vector comprises a first multi-dimensional feature vector corresponding to the left-eye fundus image and a second multi-dimensional feature vector corresponding to the right-eye fundus image.
20. The device of claim 14, wherein the fundus image comprises a left-eye fundus image and a right-eye fundus image, the database is configured to store user data, and each group of user data comprises one prestored multi-dimensional feature vector; and
wherein the obtaining a multi-dimensional feature vector representing the identity of a user comprises:
acquiring the first multi-dimensional feature vector corresponding to the left-eye fundus image and the second multi-dimensional feature vector corresponding to the right-eye fundus image that are output by the neural network; and
combining the first multi-dimensional feature vector and the second multi-dimensional feature vector to obtain the multi-dimensional feature vector representing the identity of the user.
21. The device of claim 14, wherein the fundus image is a left-eye fundus image or a right-eye fundus image, the database is configured to store user data, and each group of user data comprises a first prestored multi-dimensional feature vector corresponding to a left eye or a second prestored multi-dimensional feature vector corresponding to a right eye.
US17/619,875 2019-06-28 2020-04-08 Identity information processing method and device based on fundus image Active 2040-07-17 US11893831B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN201910578335.4 2019-06-28
CN201910578335.4A CN110263755B (en) 2019-06-28 2019-06-28 Eye ground image recognition model training method, eye ground image recognition method and eye ground image recognition device
PCT/CN2020/083625 WO2020258981A1 (en) 2019-06-28 2020-04-08 Identity information processing method and device based on fundus image

Publications (2)

Publication Number Publication Date
US20220358791A1 true US20220358791A1 (en) 2022-11-10
US11893831B2 US11893831B2 (en) 2024-02-06

Family

ID=67923118

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/619,875 Active 2040-07-17 US11893831B2 (en) 2019-06-28 2020-04-08 Identity information processing method and device based on fundus image

Country Status (4)

Country Link
US (1) US11893831B2 (en)
EP (1) EP3992844A4 (en)
CN (1) CN110263755B (en)
WO (1) WO2020258981A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20230316722A1 (en) * 2021-06-29 2023-10-05 Inspur Suzhou Intelligent Technology Co., Ltd. Image recognition method and apparatus, and device and readable storage medium
US20230334085A1 (en) * 2019-12-23 2023-10-19 Microsoft Technology Licensing, Llc Machine learning multiple features of depicted item

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110263755B (en) 2019-06-28 2021-04-27 上海鹰瞳医疗科技有限公司 Eye ground image recognition model training method, eye ground image recognition method and eye ground image recognition device
CN110717907A (en) * 2019-10-06 2020-01-21 浙江大学 Intelligent hand tumor detection method based on deep learning
CN111178403B (en) * 2019-12-16 2023-10-17 北京迈格威科技有限公司 Method, device, electronic equipment and storage medium for training attribute identification model
CN111178221A (en) * 2019-12-24 2020-05-19 珠海格力电器股份有限公司 Identity recognition method and device
CN111553436B (en) * 2020-04-30 2024-03-22 上海鹰瞳医疗科技有限公司 Training data generation method, model training method and equipment
CN113011253B (en) * 2021-02-05 2023-04-21 中国地质大学(武汉) Facial expression recognition method, device, equipment and storage medium based on ResNeXt network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150071504A1 (en) * 2008-12-12 2015-03-12 At&T Intellectual Property I, L.P. System and method for matching faces
US20150265144A1 (en) * 2012-11-08 2015-09-24 The Johns Hopkins University System and method for detecting and classifying severity of retinal disease
US20180140180A1 (en) * 2016-11-22 2018-05-24 Delphinium Clinic Ltd. Method and system for classifying optic nerve head
US20190130566A1 (en) * 2015-04-06 2019-05-02 IDx, LLC Systems and methods for feature detection in retinal images
US20200405148A1 (en) * 2019-06-27 2020-12-31 Bao Tran Medical analysis system

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102254189A (en) * 2011-08-16 2011-11-23 武汉大学 Method for identifying face expression based on cloud model
CN102496007B (en) 2011-12-02 2014-06-11 陈中山 Human body identity identification instrument
CN103605962A (en) * 2013-11-19 2014-02-26 四川长虹电器股份有限公司 Between-class and within-class distance based human face verification method
CN106022317A (en) * 2016-06-27 2016-10-12 北京小米移动软件有限公司 Face identification method and apparatus
US10922393B2 (en) * 2016-07-14 2021-02-16 Magic Leap, Inc. Deep neural network for iris identification
CN106228142B (en) * 2016-07-29 2019-02-15 西安电子科技大学 Face verification method based on convolutional neural networks and Bayesian decision
EP3336774B1 (en) * 2016-12-13 2020-11-25 Axis AB Method, computer program product and device for training a neural network
CN107247940A (en) * 2017-06-18 2017-10-13 天津大学 A kind of joint embedding grammar of face recognition and cluster
CN109390053B (en) * 2017-08-02 2021-01-08 上海市第六人民医院 Fundus image processing method, fundus image processing apparatus, computer device, and storage medium
CN109101913A (en) * 2018-08-01 2018-12-28 北京飞搜科技有限公司 Pedestrian recognition methods and device again
CN109614613B (en) * 2018-11-30 2020-07-31 北京市商汤科技开发有限公司 Image description statement positioning method and device, electronic equipment and storage medium
CN109902665A (en) * 2019-03-28 2019-06-18 北京达佳互联信息技术有限公司 Similar face retrieval method, apparatus and storage medium
CN110427804B (en) * 2019-06-18 2022-12-09 中山大学 Iris identity verification method based on secondary transfer learning
CN110263755B (en) * 2019-06-28 2021-04-27 上海鹰瞳医疗科技有限公司 Eye ground image recognition model training method, eye ground image recognition method and eye ground image recognition device
CN110414423A (en) * 2019-07-25 2019-11-05 上海鹰瞳医疗科技有限公司 Personal identification method and equipment

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150071504A1 (en) * 2008-12-12 2015-03-12 At&T Intellectual Property I, L.P. System and method for matching faces
US20150265144A1 (en) * 2012-11-08 2015-09-24 The Johns Hopkins University System and method for detecting and classifying severity of retinal disease
US20190130566A1 (en) * 2015-04-06 2019-05-02 IDx, LLC Systems and methods for feature detection in retinal images
US20180140180A1 (en) * 2016-11-22 2018-05-24 Delphinium Clinic Ltd. Method and system for classifying optic nerve head
US20200405148A1 (en) * 2019-06-27 2020-12-31 Bao Tran Medical analysis system

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Oinonen, et al., "IDENTITY VERIFICATION BASED ON VESSEL MATCHING FROM FUNDUS IMAGES," Proceedings of 2010 IEEE 17th International Conference on Image Processing (Year: 2010) *
Waheed et al., "Person identification using vascular and non-vascular retinal features," Computers and Electrical Engineering 53 (2016) 359–371 (Year: 2016) *
Xiaodong Li, "VISUAL NAVIGATION IN UNMANNED AIR VEHICLES WITH SIMULTANEOUS LOCATION AND MAPPING (SLAM)," PhD Thesis, Cranfield University, January 2014 (Year: 2014) *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20230334085A1 (en) * 2019-12-23 2023-10-19 Microsoft Technology Licensing, Llc Machine learning multiple features of depicted item
US12093305B2 (en) * 2019-12-23 2024-09-17 Microsoft Technology Licensing, Llc Machine learning multiple features of depicted item
US20230316722A1 (en) * 2021-06-29 2023-10-05 Inspur Suzhou Intelligent Technology Co., Ltd. Image recognition method and apparatus, and device and readable storage medium
US12026933B2 (en) * 2021-06-29 2024-07-02 Inspur Suzhou Intelligent Technology Co., Ltd. Image recognition method and apparatus, and device and readable storage medium

Also Published As

Publication number Publication date
CN110263755B (en) 2021-04-27
US11893831B2 (en) 2024-02-06
EP3992844A4 (en) 2023-05-24
WO2020258981A1 (en) 2020-12-30
CN110263755A (en) 2019-09-20
EP3992844A1 (en) 2022-05-04

Similar Documents

Publication Publication Date Title
US11893831B2 (en) Identity information processing method and device based on fundus image
US9684850B2 (en) Biological information processor
CN112232117A (en) Face recognition method, face recognition device and storage medium
EP4246431A1 (en) Face image quality assessment method and apparatus, computer device and storage medium
EP3680794B1 (en) Device and method for user authentication on basis of iris recognition
CN105117692A (en) Real-time face identification method and system based on deep learning
KR20120114564A (en) Apparatus and method for selecting representative fingerprint template
US9514354B2 (en) Facial analysis by synthesis and biometric matching
CN104036254A (en) Face recognition method
Proença et al. A reminiscence of “mastermind”: Iris/periocular biometrics by “in-set” CNN iterative analysis
KR20210033940A (en) How to Train Neural Networks for Human Facial Recognition
CN116363712B (en) Palmprint palm vein recognition method based on modal informativity evaluation strategy
Mousavi et al. A distinctive landmark-based face recognition system for identical twins by extracting novel weighted features
TWI325568B (en) A method for face varification
CN112926557B (en) Method for training multi-mode face recognition model and multi-mode face recognition method
KR20210018586A (en) Method and apparatus for identifying video content using biometric features of characters
CN114519898A (en) Biological characteristic multi-mode fusion recognition method and device, storage medium and equipment
Quintiliano et al. Face recognition based on eigeneyes
US11562173B2 (en) Method, device, and computer program product for model updating
CN111428670B (en) Face detection method, face detection device, storage medium and equipment
KR101501410B1 (en) Method for detecting a position of eye
CN113627476B (en) Face clustering method and system based on feature standardization
Shaker et al. Identification Based on Iris Detection Technique.
Zhou et al. Eye localization based on face alignment
Aly et al. Adaptive feature selection and data pruning for 3D facial expression recognition using the Kinect

Legal Events

Date Code Title Description
AS Assignment

Owner name: SHANGHAI EAGLEVISION MEDICAL TECHNOLOGY CO., LTD., CHINA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HE, ANDRE WANG;XIONG, JIANHAO;FU, MENG;AND OTHERS;REEL/FRAME:058410/0008

Effective date: 20211110

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE