WO2021196955A1 - 图像识别方法及相关装置、设备 - Google Patents
图像识别方法及相关装置、设备 Download PDFInfo
- Publication number
- WO2021196955A1 WO2021196955A1 PCT/CN2021/078748 CN2021078748W WO2021196955A1 WO 2021196955 A1 WO2021196955 A1 WO 2021196955A1 CN 2021078748 W CN2021078748 W CN 2021078748W WO 2021196955 A1 WO2021196955 A1 WO 2021196955A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- recognized
- medical image
- feature representation
- medical
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 54
- 210000000056 organ Anatomy 0.000 claims abstract description 148
- 238000000605 extraction Methods 0.000 claims abstract description 73
- 238000011176 pooling Methods 0.000 claims description 37
- 238000012545 processing Methods 0.000 claims description 31
- 238000001514 detection method Methods 0.000 claims description 26
- 210000004185 liver Anatomy 0.000 claims description 25
- 238000012549 training Methods 0.000 claims description 17
- 230000004927 fusion Effects 0.000 claims description 12
- 210000003734 kidney Anatomy 0.000 claims description 11
- 238000007499 fusion processing Methods 0.000 claims description 10
- 238000007781 pre-processing Methods 0.000 claims description 10
- 210000000952 spleen Anatomy 0.000 claims description 9
- 238000003384 imaging method Methods 0.000 description 20
- 238000010586 diagram Methods 0.000 description 8
- 238000009792 diffusion process Methods 0.000 description 5
- 238000002597 diffusion-weighted imaging Methods 0.000 description 5
- 210000003240 portal vein Anatomy 0.000 description 5
- 230000009286 beneficial effect Effects 0.000 description 4
- 238000003908 quality control method Methods 0.000 description 4
- 238000002591 computed tomography Methods 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000002595 magnetic resonance imaging Methods 0.000 description 3
- 238000001356 surgical procedure Methods 0.000 description 3
- 210000001015 abdomen Anatomy 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000003745 diagnosis Methods 0.000 description 2
- 230000008034 disappearance Effects 0.000 description 2
- 201000010099 disease Diseases 0.000 description 2
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 2
- 230000003902 lesion Effects 0.000 description 2
- 238000011477 surgical intervention Methods 0.000 description 2
- 210000001519 tissue Anatomy 0.000 description 2
- 238000010200 validation analysis Methods 0.000 description 2
- 230000003044 adaptive effect Effects 0.000 description 1
- 210000003484 anatomy Anatomy 0.000 description 1
- 238000002583 angiography Methods 0.000 description 1
- 210000001367 artery Anatomy 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000017531 blood circulation Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 210000004072 lung Anatomy 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000008816 organ damage Effects 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 238000012552 review Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/03—Recognition of patterns in medical or anatomical images
- G06V2201/031—Recognition of patterns in medical or anatomical images of internal organs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/07—Target detection
Definitions
- This application relates to the field of artificial intelligence technology, in particular to an image recognition method and related devices and equipment.
- scan image categories often include timing-related pre-contrast scan, early arterial phase, late arterial phase, portal phase, delay phase, etc.
- scan image categories can also include scan parameters related T1-weighted inverse imaging, T1-weighted in-phase imaging, T2-weighted imaging, diffusion-weighted imaging, surface diffusion coefficient imaging, etc.
- the radiologist is usually required to identify the scanned image category of the scanned medical image to ensure that the required medical image is obtained; or, during hospitalization or outpatient treatment, the doctor is usually required to review the scanned medical image Recognize, determine the scanned image category of each medical image, and then read the image.
- the above-mentioned method of manually identifying the scanned image category of the medical image has low efficiency, and is subject to subjective influence and is difficult to ensure accuracy. Therefore, how to improve the efficiency and accuracy of image recognition has become an urgent problem to be solved.
- This application provides an image recognition method and related devices and equipment.
- the first aspect of the present application provides an image recognition method, including: acquiring at least one scanned medical image to be recognized, and respectively determining the target area corresponding to the target organ in each medical image to be recognized; Perform feature extraction on regional image data to obtain the individual feature representation of each medical image to be recognized; fuse the individual feature representations of at least one medical image to be recognized to obtain a global feature representation; use the individual feature representation of each medical image to be recognized And the global feature representation to determine the scanned image category to which each medical image to be recognized belongs.
- the feature extraction of the image data of each target area is performed to obtain each
- the individual feature representation of the medical image to be recognized can eliminate interference from other organs, which is conducive to improving the accuracy of recognition, and the individual feature representations of at least one medical image to be recognized are merged to obtain a global feature representation, and then each medical image to be recognized
- the individual feature representation and the global feature representation of the image can not only represent the characteristics of the medical image to be recognized, but also the difference of other medical images to be recognized, so that the individual feature representation and global feature representation of each medical image to be recognized can be used.
- fusing the individual feature representations of at least one medical image to be identified to obtain a global feature representation includes: performing global pooling processing on the individual feature representations of at least one medical image to be identified to obtain a global feature representation.
- the global feature representation can be obtained quickly and conveniently, which can help improve the efficiency of subsequent image recognition.
- subjecting at least one individual feature representation of the medical image to be identified to global pooling processing to obtain the global feature representation includes: subjecting at least one individual feature representation of the medical image to be identified to global maximum pooling processing to obtain the first global feature representation; And, performing global average pooling processing on the individual feature representations of at least one medical image to be recognized to obtain a second global feature representation; and performing stitching processing on the first global feature representation and the second global feature representation to obtain a global feature representation.
- the first global feature representation is obtained, and performing global average pooling processing on the individual feature representation of at least one medical image to be recognized to obtain the second Global feature representation, so that the first global feature representation and the second global feature representation are stitched together to obtain the global feature representation, so it can help to accurately represent the difference between each medical image to be recognized and other medical images to be recognized. , Which can help improve the accuracy of subsequent image recognition.
- using the individual feature representation and global feature representation of each medical image to be recognized to determine the scanned image category to which each medical image to be recognized belongs includes: using the individual feature representation and global feature representation of each medical image to be recognized to obtain each A final feature representation of the medical image to be recognized, using the final feature representation of each medical image to be recognized to determine the scanned image category to which each medical image to be recognized belongs.
- the final feature representation obtained by using the individual feature representation and the global feature representation of each medical image to be recognized can not only represent the characteristics of the medical image to be recognized, but also the difference of other medical images to be recognized, so as to use each
- the final feature representation of each medical image to be recognized can improve the accuracy of image recognition when determining the scanned image category to which each medical image to be recognized belongs.
- using the individual feature representation and the global feature representation of each medical image to be recognized to obtain the final feature representation of each medical image to be recognized includes: stitching the individual feature representation and the global feature representation of each medical image to be recognized respectively , Get the final feature representation corresponding to the medical image to be recognized.
- the final feature representation corresponding to the medical image to be recognized can be quickly obtained, which can help improve the efficiency of subsequent image recognition.
- performing feature extraction on the image data of each target area to obtain the individual feature representation of each medical image to be recognized includes: using the feature extraction sub-network of the recognition network to perform feature extraction on the image data of each target area to obtain each The individual feature representations of the medical images to be recognized; the individual feature representations of at least one medical image to be recognized are fused to obtain the global feature representation, and the individual feature representation and the global feature representation of each medical image to be recognized are used to obtain each
- the final feature representation of the medical image includes: using the fusion module of the recognition network to fuse the individual feature representation of at least one medical image to be recognized to obtain a global feature representation, and use the individual feature representation and global feature representation of each medical image to be recognized, Obtain the final feature representation of each medical image to be recognized; use the final feature expression of each medical feature to be recognized to determine the scanned image category to which each medical image to be recognized belongs, including: using the classification sub-network of the recognition network to Recognizing the final feature of the medical image means performing predictive classification to obtain the scanned image category to
- the individual feature representation of each medical image to be recognized is obtained, and the fusion module of the recognition network is used to extract the features of at least one medical image to be recognized.
- the individual feature representations are fused to obtain a global feature representation.
- the individual feature representation and global feature representation of each medical image to be recognized are used to obtain the final feature representation of each medical image to be recognized, so that the classification sub-network of the recognition network is used for each
- the final feature of the medical image to be recognized indicates that the predicted classification is performed to obtain the scanned image category to which each medical image to be recognized belongs. Therefore, the scanned image category to which the medical image to be recognized belongs can be finally obtained through the recognition network, which can further improve the efficiency of image recognition .
- the recognition network is obtained by training with sample medical images, and the number of sample medical images used for each training of the recognition network is not fixed.
- the number of sample medical images used in each training recognition network is not fixed, and the number of sample medical images can be randomized, which can help to accurately scan image categories under different scanning protocols in different institutions.
- Image recognition can improve the accuracy of image recognition.
- the feature extraction sub-network includes at least one set of sequentially connected dense convolution blocks and pooling layers; and/or, the recognition network includes a preset number of feature extraction sub-networks; the feature extraction sub-network of the recognition network is used for each target Performing feature extraction on the image data of the region to obtain the individual feature representation of each medical image to be recognized includes: inputting the image data of each target region into a corresponding feature extraction sub-network for feature extraction, and obtaining the individual of each medical image to be recognized Feature representation.
- the feature extraction sub-network includes at least one set of dense convolutional blocks and pooling layers that are connected in sequence. Therefore, the dense convolutional block connection strategy is adopted, that is, the features under each convolutional layer are tightly spliced with the next layer and transmitted Each subsequent layer can effectively alleviate the problem of gradient disappearance, strengthen feature propagation and feature reuse, and can greatly reduce the number of parameters; and the recognition network is set to include a preset number of feature extraction sub-networks, and The image data of each target area is input into a corresponding feature extraction sub-network for feature extraction, and the individual feature representation of each medical image to be recognized is obtained. The feature extraction operation of the image data of at least one target area can be processed in parallel. Conducive to improving the efficiency of image recognition.
- respectively determining the target area corresponding to the target organ in each medical image to be recognized includes: using an organ detection network to detect at least one medical image to be recognized to obtain first position information of the target organ and information about the target organ.
- the second position information of the adjacent organ; the first position information and the second position information are used to determine the target area corresponding to the target organ.
- the organ detection network is used to detect at least one medical image to be recognized to obtain the first position information of the target organ and the second position information of the adjacent organs of the target organ, so that not only the target to be recognized can be considered Organs can also consider the surrounding organs, so that the first position information and the second position information can be used to determine the target area corresponding to the target organ, which can ensure that the shape of the organ changes under surgical treatment, etc.
- the target area corresponding to the target organ is obtained by positioning, so the robustness of image recognition can be improved.
- the medical image to be recognized is a three-dimensional image
- the organ detection network is used to detect at least one medical image to be recognized to obtain the first position information of the target organ and the second position information of the adjacent organs of the target organ. It also includes: dividing each medical image to be identified along the coronal plane to obtain multiple three-dimensional sub-images; projecting each sub-image in a direction perpendicular to the coronal plane to obtain the corresponding two-dimensional sub-image; using organ detection
- the network detects at least one medical image to be identified, and obtains the first position information of the target organ and the second position information of the adjacent organs of the target organ. Dimension sub-images are detected to obtain first position information and second position information.
- each medical image to be recognized is divided along the coronal plane to obtain multiple three-dimensional sub-images, and each sub-image is projected in a direction perpendicular to the coronal plane to obtain the corresponding Therefore, the organ detection network is used to detect the two-dimensional sub-image corresponding to at least one medical image to be recognized, and the first position information and the second position information are obtained, which can further improve the target area location corresponding to the target organ Accuracy.
- the target organ is the liver
- the adjacent organs include at least one of the kidney and the spleen
- the first position information includes at least one set of diagonal vertex positions of the corresponding area of the target organ and the size of the corresponding area.
- the second position information includes at least one vertex position of the corresponding area adjacent to the organ.
- setting the target organ as the liver, and setting the adjacent organ to include at least one of the kidney and the spleen can help locate the target area corresponding to the liver; setting the first position information to include the area corresponding to the target organ At least one set of diagonal vertex positions and the size of the corresponding area, and the second position information is set to include at least one vertex position adjacent to the corresponding area of the organ, which can facilitate accurate positioning of the target area corresponding to the target organ.
- the method further includes at least one of the following: scanning at least one medical image to be recognized according to its scan Sort the image categories; if the scanned image categories of the medical images to be recognized are repeated, the first warning information is output to remind the scanner; if there is no preset scanned image category in the scanned image categories of at least one medical image to be recognized, then Output the second warning message to remind the scanner.
- the scanned image category to which each medical image to be recognized belongs execute sorting of at least one medical image to be recognized according to its scanned image category, which can improve the convenience of doctors in reading the image;
- the first warning information is output to remind the scanner
- the preset scanned image category does not exist in the scan image category of at least one medical image to be recognized
- the second warning information is output to remind the scanner.
- the method further includes: preprocessing the image data of each target region, wherein the preprocessing includes the following At least one: adjusting the image size of the target area to a preset size, and normalizing the image intensity of the target area to a preset range.
- the image data of each target area is preprocessed, and the preprocessing includes at least one of the following: adjusting the image size of the target area to a preset size, and normalizing the image intensity of the target area to The preset range can help improve the accuracy of subsequent image recognition.
- the second aspect of the present application provides an image recognition device, including: a region acquisition module, a feature extraction module, a fusion processing module, and a category determination module.
- the region acquisition module is configured to acquire at least one scanned medical image to be identified, and respectively determine The target region corresponding to the target organ in each medical image to be recognized;
- the feature extraction module is configured to extract features from the image data of each target region to obtain the individual feature representation of each medical image to be recognized; configuration of the fusion processing module In order to fuse the individual feature representations of at least one medical image to be identified to obtain a global feature representation;
- the category determination module is configured to use the individual feature representation and the global feature representation of each medical image to be identified to determine which medical image belongs to Scanned image category.
- a third aspect of the present application provides an electronic device including a memory and a processor coupled to each other, and the processor is configured to execute program instructions stored in the memory to implement the image recognition method in the first aspect.
- the fourth aspect of the present application provides a computer-readable storage medium having program instructions stored thereon, and the program instructions implement the image recognition method in the first aspect when the program instructions are executed by a processor.
- the feature extraction is performed on the image data of each target area to obtain each
- the individual feature representations of a medical image to be recognized can eliminate interference from other organs, which is beneficial to improve the accuracy of recognition, and the individual feature representations of at least one medical image to be recognized are merged to obtain a global feature representation, and then each to be recognized
- the individual feature representation and global feature representation of medical images can not only represent the characteristics of the medical image to be recognized, but also the differences of other medical images to be recognized, so that the individual feature representation and global feature representation of each medical image to be recognized can be used.
- FIG. 1 is a schematic flowchart of an embodiment of an image recognition method according to the present application.
- FIG. 2 is a schematic diagram of the state of the process of determining the scanned image category to which the medical image to be recognized belongs;
- FIG. 3 is a schematic flowchart of an embodiment of step S11 in FIG. 1;
- FIG. 4 is a schematic diagram of the framework of an embodiment of the image recognition device of the present application.
- FIG. 5 is a schematic diagram of the framework of an embodiment of the electronic device of the present application.
- Fig. 6 is a schematic diagram of a framework of an embodiment of a computer-readable storage medium according to the present application.
- system and "network” in this article are often used interchangeably in this article.
- the term “and/or” in this article is only an association relationship describing the associated objects, which means that there can be three relationships, for example, A and/or B, which can mean: A alone exists, A and B exist at the same time, exist alone B these three situations.
- the character "/” in this text generally indicates that the associated objects before and after are in an "or” relationship.
- "many” in this document means two or more than two.
- the execution subject of the image recognition method may be an image recognition device.
- the image recognition method may be executed by a terminal device or a server or other processing equipment.
- the terminal device may be a user equipment (User Equipment, UE). ), mobile devices, user terminals, terminals, cellular phones, cordless phones, personal digital assistants (PDAs), handheld devices, computing devices, in-vehicle devices, wearable devices, etc.
- the image recognition method can be implemented by a processor calling computer-readable instructions stored in the memory.
- FIG. 1 is a schematic flowchart of an embodiment of an image recognition method according to the present application. Specifically, it can include the following steps:
- Step S11 Obtain at least one scanned medical image to be recognized, and respectively determine the target area corresponding to the target organ in each medical image to be recognized.
- the medical images to be recognized may include CT images and MR images, which are not limited here.
- the medical image to be recognized may be obtained by scanning the abdomen, chest and other areas, and may be specifically set according to actual application conditions, which is not limited here. For example, when the liver, spleen, and kidney are the target organs that need diagnosis and treatment, the abdomen can be scanned to obtain medical images to be identified; or, when the heart and lungs are the target organs that need diagnosis and treatment, the chest can be scanned. Obtain the medical image to be recognized, and other situations can be deduced by analogy, so we will not give examples one by one here.
- the scanning mode may be plain scanning, enhanced scanning, etc., which are not limited here.
- the medical image to be recognized may be a three-dimensional image, and the target area corresponding to the target organ in the medical image to be recognized may be a three-dimensional area, which is not limited here.
- the target organ can be set according to the actual application.
- the target organ can be the liver; or when the doctor needs to judge whether the kidney has lesions and the extent of the disease, the target organ
- the device can be the kidney, and other conditions can be set according to the actual application, so we will not give examples one by one here.
- an organ detection network for detecting target organs can be pre-trained, so that the organ detection network can be directly used to detect each medical image to be identified to obtain the corresponding target area.
- Step S12 Perform feature extraction on the image data of each target area respectively to obtain the individual feature representation of each medical image to be recognized.
- the image data of each target area may also be preprocessed.
- the preprocessing may include The image size of the area is adjusted to a preset size (for example, 32*256*256).
- the preprocessing may also include normalizing the image intensity of the target area to a preset range (for example, the range of 0 to 1).
- the preset ratio For example, the gray value corresponding to 99.9%
- the normalized clamp value is used as the normalized clamp value, so that the contrast of the image data of the target area can be enhanced, which is beneficial to improve the accuracy of subsequent image recognition.
- a recognition network in order to improve the convenience of feature extraction, can also be pre-trained.
- the recognition network can include a feature extraction sub-network for feature extraction, so that the feature extraction sub-network can be used to analyze the image of each target area.
- the data is feature-extracted, and the individual feature representation of each medical image to be recognized is obtained.
- the feature extraction sub-network includes at least a set of sequentially connected dense convolution blocks (Dense Block) and a pooling layer.
- the features of each layer of the dense convolution block are closely connected to the next layer. Splicing and transferring each layer afterwards makes the transfer of features and gradients more effective.
- the feature extraction sub-network may include three sets of sequentially connected dense convolution blocks and pooling layers, where, except for the pooling layer contained in the last set of adaptive pooling, the pooling layers contained in other groups It is the maximum pooling; in addition, the feature extraction sub-network may also include a group, two groups, four groups, and other groups of dense convolution blocks (Dense Block) and a pooling layer connected in sequence, which are not limited here.
- the recognition network may specifically include a preset number of feature extraction sub-networks, so that the image data of each target area can be input into a corresponding feature extraction sub-network for feature extraction, and each target area can be extracted. Recognize the individual feature representations of medical images, and then the feature extraction operations of the image data of each target area can be processed in parallel, so the efficiency of feature extraction can be improved, and the efficiency of subsequent image recognition can be improved.
- the preset number can be greater than Or equal to the category of the scanned image.
- the preset number can be set to an integer greater than or equal to 5 , For example, 5, 6, 7, etc., which are not limited here; or, when the scanned image category includes T1-weighted inverse imaging, T1-weighted in-phase imaging, T2-weighted imaging, diffusion-weighted imaging, surface diffusion related to tracing parameters In coefficient imaging, the preset number can be set to an integer greater than or equal to 5, for example, 5, 6, 7, etc., which are not limited here; or, when the scanned image category includes both T1 weighted inversion related to the tracing parameter Imaging, T1-weighted in-phase imaging, T2-weighted imaging, diffusion-weighted imaging, surface diffusion coefficient imaging, as well as timing-related pre-contrast scan, early arterial, late arterial, portal phase, and delay periods.
- the preset number can be set It is an integer greater than or equal to 10, for example, 10, 11, 12, and so on.
- the early arterial phase can indicate that the portal vein has not been enhanced
- the late arterial phase can indicate that the portal vein has been enhanced
- the portal phase can indicate that the portal vein has been fully enhanced and the liver vessels have been enhanced by forward blood flow, and the liver soft cell tissue has been under the markers.
- the delay period can indicate that the portal vein and arteries are in an enhanced state and weaker than the portal stage, and the liver soft cell tissue is in an enhanced state and weaker than the portal stage.
- Other scan image categories will not be illustrated here.
- Figure 2 is a schematic diagram of the state of the process of determining the scanned image category to which the medical image to be recognized belongs. As shown in Figure 2, rectangular boxes filled with different gray levels represent medical image 1 to medical image to be recognized The individual feature representation 1, individual feature representation 2, individual feature representation 3, ..., individual feature representation n extracted from the image data of the target region corresponding to the target organ in n.
- Step S13 Fusion of individual feature representations of at least one medical image to be identified to obtain a global feature representation.
- the recognition network may also include a fusion module, so that the fusion module can be used to fuse the individual feature representations of at least one medical image to be recognized to obtain a global feature representation.
- the individual feature representation of at least one medical image to be recognized may be subjected to global pooling processing to obtain a global feature representation.
- the individual feature representation of at least one medical image to be recognized may be subjected to global maximum pooling (Global Max Pooling, GMP) processing to obtain a first global feature representation
- GMP global maximum pooling
- the individual feature representation of at least one medical image to be recognized can be global
- GAP global average pooling
- individual feature representation 1, individual feature representation 2, individual feature representation 3, ..., individual feature representation n are respectively subjected to global maximum pooling and global average pooling to obtain the first global feature representation ( The oblique line in FIG. 2 fills the rectangular frame) and the second global feature representation (the grid line fills the rectangular frame in FIG. 2), and the first global feature representation and the second global feature representation are spliced to obtain the global feature representation.
- Step S14 Use the individual feature representation and the global feature representation of each medical image to be recognized to determine the scanned image category to which each medical image to be recognized belongs.
- the individual feature representation and global feature representation of each medical image to be recognized can be used to obtain the final feature representation of each medical image to be recognized, and then the final feature representation of each medical image to be recognized can be used to determine each medical image to be recognized.
- the category of the scanned image to which the medical image belongs so that the final feature representation can not only represent the characteristics of the medical image to be recognized, but also the difference of other medical images to be recognized, and then use the final feature representation of each medical image to be recognized to determine each
- the accuracy of image recognition can be improved.
- the fusion module in the recognition network can be used to use the individual feature representation and global feature representation of each medical image to be recognized to obtain each to be recognized The final feature representation of the medical image.
- the individual feature representation and the global feature representation of each medical image to be recognized can also be spliced to obtain the final feature representation corresponding to the medical image to be recognized.
- Figure 2 Please refer to Figure 2 in combination. As shown in Figure 2, the first global feature represented by a rectangular box filled with diagonal lines is represented and the second global feature represented by a rectangular box filled with grid lines is represented respectively and represented by a rectangular box filled with different gray levels. The individual feature representations of is spliced, and the final feature representation corresponding to each medical image to be recognized can be obtained.
- the recognition network may also include a classification sub-network, so that the classification sub-network can be used to predict and classify the final feature representation of each medical image to be recognized, and obtain the scanned image category to which each medical image to be recognized belongs.
- the classification sub-network can include a fully connected layer and a softmax layer, so that the fully connected layer can be used to connect the final feature representation of each medical image to be recognized, and the softmax layer can be used for probability normalization Therefore, the probability value that each medical image to be recognized belongs to each scanned image category is obtained, so the scanned image category corresponding to the maximum probability value can be used as the scanned image category to which the medical image to be recognized belongs.
- the recognition network including the feature extraction sub-network, the fusion module and the classification sub-network may be obtained by training using sample medical images.
- the feature extraction sub-network can be used to perform feature extraction on the image data of the target area annotated in each sample medical image to obtain the individual feature representation of each sample medical image
- the fusion module can be used to extract the individual characteristics of at least one sample medical image.
- Feature representations are fused to obtain a global feature representation.
- the individual feature representation and global feature representation of each sample medical image are used to obtain the final feature representation of each sample medical image, and then the classification sub-network is used to determine the final feature of each sample medical image.
- the number of sample medical images used for each training of the recognition network may not be fixed.
- the sample medical images used for each training of the recognition network may belong to the same object, and the number of scanned image categories to which the sample medical images used for each training of the recognition network belongs may not be fixed.
- the sample medical images used in a certain training recognition network belong to T1-weighted inverse imaging, T1-weighted in-phase imaging, and T2-weighted imaging
- the sample medical images used in another training recognition network belong to diffusion-weighted imaging and surface diffusion coefficient imaging.
- the specific settings can be set according to the actual application situation. I will not give examples one by one here, so that the number of sample medical images can be randomized, which can help to accurately scan image categories when different institutions and different scanning protocols are missing. Image recognition can improve the robustness of the recognition network.
- the above-mentioned trained recognition network can be set in an image post-processing workstation, a filming workstation, a computer-aided image reading system, etc., so as to realize automatic recognition of medical images to be recognized and improve recognition efficiency.
- all medical images to be recognized that belong to the same object in a scan process can be recognized in one recognition process, so that the performance of the recognition network can be fully verified;
- all medical images to be recognized that belong to the same object in one scan can be recognized in one recognition process, so that the difference between each medical image to be recognized and all other medical images to be recognized can be considered , which in turn can help improve the accuracy of recognition.
- At least one medical image to be recognized is obtained by scanning the same object. Therefore, in order to facilitate the doctor to read the image, after obtaining the scanned image category to which each medical image to be recognized belongs, at least one medical image to be recognized may be The images are sorted according to their scan image category, for example, T1-weighted inverse imaging, T1-weighted in-phase imaging, plain scan before angiography, early arterial, late arterial, portal phase, delay phase, T2-weighted imaging, diffusion-weighted imaging, The preset order of surface diffusion coefficient imaging is sorted. In addition, the preset order can also be set according to the doctor’s habits, which is not limited here, so as to improve the convenience of doctors in reading the film.
- the sorted at least one medical image to be recognized can also be displayed in a window corresponding to the number of medical images to be recognized. For example, if the number of medical images to be recognized is 5, it can be displayed in 5 display windows. Medical images to be recognized. Therefore, it is possible to reduce the time for doctors to look through multiple medical images to be identified for comparison back and forth, and to improve the efficiency of image reading.
- At least one medical image to be recognized is obtained by scanning the same object. Therefore, in order to perform quality control during the scanning process, after obtaining the scanned image category to which each medical image to be recognized belongs, it can also be determined Identify whether there is a repetition in the scanned image category of the medical image, and when there is a repetition, output first warning information to remind the scanner. For example, if there are two medical images to be recognized whose scan image categories are both "delay period", it can be considered that the scan quality is out of compliance during the scanning process. Therefore, in order to remind the scanner, the first warning message can be output. Therefore, it is possible to output the warning reason (for example, there are medical images to be recognized with repeated scan image categories, etc.).
- the preset scanned image category is "portal phase"
- the second warning message can be output, specifically, the reason for the warning can be output (for example, there is no portal vein image in the medical image to be identified, etc.). Therefore, the image quality control can be realized during the scanning process, so that when it is contrary to the actual situation, the error can be corrected in time, and the second registration of the patient can be avoided.
- the feature extraction is performed on the image data of each target area to obtain each
- the individual feature representations of a medical image to be recognized can eliminate interference from other organs, which is beneficial to improve the accuracy of recognition, and the individual feature representations of at least one medical image to be recognized are merged to obtain a global feature representation, and then each to be recognized
- the individual feature representation and global feature representation of medical images can not only represent the characteristics of the medical image to be recognized, but also the differences of other medical images to be recognized, so that the individual feature representation and global feature representation of each medical image to be recognized can be used.
- FIG. 3 is a schematic flowchart of an embodiment of step S11 in FIG. 1.
- FIG. 3 is a schematic flowchart of an embodiment of determining the target region corresponding to the target organ in each medical image to be recognized, which may specifically include the following steps:
- Step S111 Use the organ detection network to detect at least one medical image to be identified, to obtain first position information of the target organ and second position information of the adjacent organs of the target organ.
- the backbone network of the organ detection network can adopt an efficient net. In other implementation scenarios, the backbone network of the organ detection network can also adopt other networks, which is not limited here.
- the target organ may be set according to actual conditions.
- the target organ may be the liver, and the adjacent organs of the target organ may include at least one of the kidney and the spleen.
- the first position information of the target organ may include at least one set of diagonal vertex positions (for example, position coordinates) of the corresponding area of the target organ and the size (for example, length, width, etc.) of the corresponding area.
- the second position information may at least include at least one vertex position (for example, position coordinates) of the corresponding region of the adjacent organ.
- the medical image to be recognized can be a three-dimensional image.
- each medical image to be recognized can be divided along the coronal plane to obtain multiple three-dimensional sub-images, and Project each sub-image in the direction perpendicular to the coronal plane to obtain the corresponding two-dimensional sub-image, so that subsequent identification and detection can be performed based on multiple two-dimensional sub-images obtained by the projection.
- the organ detection network can be used to At least one two-dimensional sub-image corresponding to the medical image to be recognized is detected to obtain the first position information and the second position information, so that it can be accurately when the size of the target organ is abnormal or the shape of the target organ changes after surgical intervention
- the first location information of the target organ and the second location information of the adjacent organ of the target organ are determined. For example, when the target organ is the liver, when the liver size is abnormal or the liver morphology changes (such as partial loss) after surgical intervention, the positions of the liver apex and liver apex cannot be stably represented.
- Organs detection can be performed on two two-dimensional sub-images, and the detection results on multiple two-dimensional sub-images can be combined to obtain the first position information of the liver and the second position information of the kidney, spleen, etc., which can effectively avoid the apex and top of the liver. Possible detection error due to unstable position.
- Step S112 Use the first position information and the second position information to determine the target area corresponding to the target organ.
- the geographic correlation between the target organ and the adjacent organs in the anatomical structure can be considered, so the first location information and the second location information are used, It can accurately determine the target area corresponding to the target organ.
- the first position information may include the upper left and lower left vertices of the corresponding area of the liver, the height and width of the corresponding area
- the second position information may include the right side of the corresponding area of adjacent organs such as the spleen and kidney.
- the lower vertex therefore, the target area corresponding to the liver can be obtained by cropping the medical image to be recognized according to the first position information and the second position information.
- Other scenes can be deduced by analogy, so I won't give examples one by one here.
- the organ detection network is used to detect at least one medical image to be recognized, and the first position information of the target organ and the second position information of the adjacent organs of the target organ are obtained.
- the target organ that needs to be identified can also consider the surrounding adjacent organs, so that the first location information and the second location information can be used to determine the target area corresponding to the target organ, which can ensure that the morphology of the organ changes after surgical treatment, etc.
- the target area corresponding to the target organ can also be located, so the robustness of image recognition can be improved.
- FIG. 4 is a schematic diagram of a framework of an embodiment of an image recognition device 40 of the present application.
- the image recognition device 40 includes a region acquisition module 41, a feature extraction module 42, a fusion processing module 43, and a category determination module 44.
- the region acquisition module 41 is configured to acquire at least one scanned medical image to be recognized, and to determine each medical image to be recognized.
- the target region in the image corresponding to the target organ is configured to extract features from the image data of each target region to obtain the individual feature representation of each medical image to be recognized;
- the fusion processing module 43 is configured to at least The individual feature representations of a medical image to be recognized are fused to obtain a global feature representation;
- the category determination module 44 is configured to use the individual feature representation and the global feature representation of each medical image to be recognized to determine the scanned image to which each medical image to be recognized belongs category.
- the feature extraction is performed on the image data of each target area to obtain each
- the individual feature representations of a medical image to be recognized can eliminate interference from other organs, which is beneficial to improve the accuracy of recognition, and the individual feature representations of at least one medical image to be recognized are merged to obtain a global feature representation, and then each to be recognized
- the individual feature representation and global feature representation of medical images can not only represent the characteristics of the medical image to be recognized, but also the differences of other medical images to be recognized, so that the individual feature representation and global feature representation of each medical image to be recognized can be used.
- the fusion processing module 43 is configured to perform global pooling processing on the individual feature representation of at least one medical image to be recognized to obtain a global feature representation.
- the global feature representation can be obtained quickly and conveniently, which can help improve the efficiency of subsequent image recognition.
- the fusion processing module 43 includes a first pooling sub-module configured to perform global maximum pooling processing on the individual feature representation of at least one medical image to be recognized to obtain the first global feature representation.
- the fusion processing module 43 includes The second pooling sub-module is configured to perform global average pooling processing on the individual feature representations of at least one medical image to be recognized to obtain a second global feature representation.
- the fusion processing module 43 includes a splicing processing sub-module configured to combine the first global The feature representation and the second global feature representation are spliced to obtain a global feature representation.
- the first global feature representation is obtained by subjecting at least one individual feature representation of the medical image to be identified to global maximum pooling processing, and the individual feature representation of at least one medical image to be identified is subject to global average pooling processing ,
- the second global feature representation so that the first global feature representation and the second global feature representation are stitched together to obtain the global feature representation, so it can help to accurately represent each medical image to be recognized and other medical images to be recognized. The difference between them can help improve the accuracy of subsequent image recognition.
- the category determination module 44 includes a feature processing sub-module and a category determination sub-module.
- the feature processing sub-module is configured to use the individual feature representation and the global feature representation of each medical image to be recognized to obtain each medical image to be recognized.
- the final feature representation of the category determination sub-module is configured to use the final feature representation of each medical image to be recognized to determine the scanned image category to which each medical image to be recognized belongs.
- the final feature representation obtained by using the individual feature representation and the global feature representation of each medical image to be recognized can not only represent the characteristics of the medical image to be recognized, but also the differences of other medical images to be recognized. Therefore, when the final feature representation of each medical image to be recognized is used to determine the scanned image category to which each medical image to be recognized belongs, the accuracy of image recognition can be improved.
- the feature processing sub-module is configured to respectively perform stitching processing on the individual feature representation and the global feature representation of each medical image to be recognized to obtain the final feature representation corresponding to the medical image to be recognized.
- the final feature representation corresponding to the medical image to be recognized can be quickly obtained, which can help improve the efficiency of subsequent image recognition.
- the feature extraction module 42 is configured to use the feature extraction sub-network of the recognition network to perform feature extraction on the image data of each target area to obtain the individual feature representation of each medical image to be recognized
- the fusion processing module 43 is configured to The fusion module of the recognition network is used to fuse the individual feature representations of at least one medical image to be recognized to obtain a global feature representation.
- the feature processing sub-module is configured to use the fusion module of the recognition network to use the individual feature representation and the global feature of each medical image to be recognized Feature representation, the final feature representation of each medical image to be recognized is obtained, and the category determination sub-module is configured to use the classification sub-network of the recognition network to predict and classify the final feature representation of each medical image to be recognized to obtain each medical image to be recognized The category of the scanned image to which it belongs.
- the feature extraction sub-network of the recognition network is used to perform feature extraction on the image data of each target area to obtain the individual feature representation of each medical image to be recognized
- the fusion module of the recognition network is used to combine at least one
- the individual feature representations of the recognized medical images are fused to obtain a global feature representation
- the individual feature representations and global feature representations of each medical image to be recognized are used to obtain the final feature representation of each medical image to be recognized, thereby using the classifier of the recognition network
- the network predicts and classifies the final feature representation of each medical image to be recognized, and obtains the scanned image category to which each medical image to be recognized belongs. Therefore, the scanned image category to which the medical image to be recognized belongs can be finally obtained through the recognition network, which can further improve The efficiency of image recognition.
- the recognition network is obtained by training with sample medical images, and the number of sample medical images used for each training of the recognition network is not fixed.
- the number of sample medical images used for each training recognition network is not fixed, and the number of sample medical images can be randomized, which can help when the types of scanned images are missing under different institutions and different scanning protocols.
- the image recognition can also be performed accurately, and the accuracy of the image recognition can be improved.
- the feature extraction sub-network includes at least one set of dense convolutional blocks and pooling layers connected in sequence; and/or, the recognition network includes a preset number of feature extraction sub-networks, and the feature extraction module 42 is configured to The image data of a target area are respectively input into a corresponding feature extraction sub-network for feature extraction, and the individual feature representation of each medical image to be recognized is obtained.
- the feature extraction sub-network includes at least one set of dense convolutional blocks and pooling layers that are connected in sequence. Therefore, the dense convolutional block connection strategy is adopted, that is, the features under each convolutional layer are close to the next layer. After splicing and transferring each layer, it can effectively alleviate the problem of gradient disappearance, strengthen feature propagation and feature reuse, and can greatly reduce the number of parameters; and the recognition network is set to include a preset number of feature extractors Network, and input the image data of each target area into a corresponding feature extraction sub-network for feature extraction, and obtain the individual feature representation of each medical image to be recognized.
- the feature extraction operation of the image data of at least one target area can be processed in parallel , It can help improve the efficiency of image recognition.
- the area acquisition module 41 includes an organ detection sub-module configured to detect at least one medical image to be identified using an organ detection network to obtain first position information of the target organ and adjacent organs of the target organ.
- the area acquisition module 41 includes an area determination sub-module configured to use the first location information and the second location information to determine the target area corresponding to the target organ.
- the organ detection network is used to detect at least one medical image to be recognized, and the first position information of the target organ and the second position information of the adjacent organs of the target organ are obtained.
- the target organ that needs to be identified can also consider the surrounding adjacent organs, so that the first location information and the second location information can be used to determine the target area corresponding to the target organ, which can ensure that the morphology of the organ changes after surgical treatment, etc.
- the target area corresponding to the target organ can also be located, so the robustness of image recognition can be improved.
- the medical image to be recognized is a three-dimensional image
- the region acquisition module 41 further includes an image division sub-module configured to divide each medical image to be recognized along the coronal plane to obtain multiple three-dimensional sub-images.
- the region acquisition module 41 also includes an image projection sub-module, configured to project each sub-image in a direction perpendicular to the coronal plane to obtain a corresponding two-dimensional sub-image, and the organ detection sub-module is configured to use the organ detection network to identify at least one The two-dimensional sub-image corresponding to the medical image is detected to obtain the first position information and the second position information.
- each medical image to be recognized is divided along the coronal plane to obtain multiple three-dimensional sub-images, and each sub-image is performed in a direction perpendicular to the coronal plane. Projection to obtain the corresponding two-dimensional sub-image, so that the organ detection network is used to detect the two-dimensional sub-image corresponding to at least one medical image to be recognized, and the first position information and the second position information are obtained, which can further improve the correspondence of the target organ The accuracy of the target area positioning.
- the target organ is the liver
- the adjacent organs include at least one of the kidney and the spleen
- the first position information includes at least one set of diagonal vertex positions and corresponding areas of the corresponding area of the target organ
- the second position information includes at least one vertex position of the corresponding area adjacent to the organ.
- setting the target organ as the liver, and setting the adjacent organ to include at least one of the kidney and the spleen can help locate the target area corresponding to the liver; and set the first position information to include the target organ. At least one set of diagonal vertex positions of the organ corresponding area and the size of the corresponding area, and the second position information is set to include at least one vertex position adjacent to the organ corresponding area, which can facilitate accurate positioning of the target area corresponding to the target organ.
- the image recognition device 40 further includes an image sorting module configured to sort the at least one medical image to be recognized according to its scanned image category; the image recognition device 40 further includes a first output module configured to When the scanned image category of the image is repeated, the first warning information is output to remind the scanner; the image recognition device 40 further includes a second output module configured to have no preset scan in the scanned image category of the at least one medical image to be recognized In the image category, the second warning message is output to remind the scanner.
- the scan image category to which each medical image to be recognized belongs is determined, it is executed to sort at least one medical image to be recognized according to its scan image category, which can improve the convenience of doctor reading;
- the first warning information is output to remind the scanner, and when the preset scanned image category does not exist in the scanned image category of the at least one medical image to be recognized, the second warning information is output to remind Scanners can achieve image quality control during the scanning process, so that when it is contrary to reality, they can correct errors in time to avoid the second registration of patients.
- the image recognition device 40 further includes a preprocessing module configured to preprocess the image data of each target area, wherein the preprocessing includes at least one of the following: adjusting the image size of the target area to a preset Size, normalize the image intensity of the target area to the preset range.
- the image data of each target area is preprocessed, and the preprocessing includes at least one of the following: adjusting the image size of the target area to a preset size, and adjusting the image intensity of the target area It is normalized to the preset range, so it can help improve the accuracy of subsequent image recognition.
- FIG. 5 is a schematic diagram of a framework of an embodiment of an electronic device 50 of the present application.
- the electronic device 50 includes a memory 51 and a processor 52 coupled to each other, and the processor 52 is configured to execute program instructions stored in the memory 51 to implement the steps of any of the above-mentioned image recognition method embodiments.
- the electronic device 50 may include but is not limited to a microcomputer and a server.
- the electronic device 50 may also include mobile devices such as a notebook computer and a tablet computer, which are not limited herein.
- the processor 52 is configured to control itself and the memory 51 to implement the steps of any of the above-mentioned image recognition method embodiments.
- the processor 52 may also be referred to as a CPU (Central Processing Unit, central processing unit).
- the processor 52 may be an integrated circuit chip with signal processing capabilities.
- the processor 52 may also be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application specific integrated circuit (ASIC), a field-programmable gate array (Field-Programmable Gate Array, FPGA), or other Programmable logic devices, discrete gates or transistor logic devices, discrete hardware components.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
- the processor 52 may be jointly implemented by an integrated circuit chip.
- the above solution can improve the efficiency and accuracy of image recognition.
- FIG. 6 is a schematic diagram of a framework of an embodiment of the computer-readable storage medium 60 of the present application.
- the computer-readable storage medium 60 stores program instructions 601 that can be executed by the processor, and the program instructions 601 are used to implement the steps of any of the foregoing image recognition method embodiments.
- the above solution can improve the efficiency and accuracy of image recognition.
- the disclosed method and device can be implemented in other ways.
- the device implementation described above is only illustrative, for example, the division of modules or units is only a logical function division, and there may be other divisions in actual implementation, for example, units or components can be combined or integrated. To another system, or some features can be ignored, or not implemented.
- the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
- the units described as separate components may or may not be physically separate, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of this embodiment.
- the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
- the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
- the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
- the technical solution of the present application essentially or the part that contributes to the existing technology or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium , Including several instructions to make a computer device (which can be a personal computer, a server, or a network device, etc.) or a processor execute all or part of the steps of the methods in the various embodiments of the present application.
- the aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disks or optical disks and other media that can store program codes. .
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Health & Medical Sciences (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Human Computer Interaction (AREA)
- Evolutionary Biology (AREA)
- Molecular Biology (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Image Analysis (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Hardware Redundancy (AREA)
Abstract
Description
Claims (16)
- 一种图像识别方法,包括:获取至少一个扫描得到的待识别医学图像,并分别确定每个所述待识别医学图像中与目标脏器对应的目标区域;分别对每一所述目标区域的图像数据进行特征提取,得到每个待识别医学图像的个体特征表示;将所述至少一个待识别医学图像的个体特征表示进行融合,得到全局特征表示;利用每一所述待识别医学图像的个体特征表示和所述全局特征表示,确定每一所述待识别医学图像所属的扫描图像类别。
- 根据权利要求1所述的图像识别方法,其中,所述将所述至少一个待识别医学图像的个体特征表示进行融合,得到全局特征表示包括:将所述至少一个待识别医学图像的个体特征表示进行全局池化处理,得到所述全局特征表示。
- 根据权利要求2所述的图像识别方法,其中,所述将所述至少一个待识别医学图像的个体特征表示进行全局池化处理,得到所述全局特征表示包括:将所述至少一个待识别医学图像的个体特征表示进行全局最大池化处理,得到第一全局特征表示;以及,将所述至少一个待识别医学图像的个体特征表示进行全局平均池化处理,得到第二全局特征表示;将所述第一全局特征表示和所述第二全局特征表示进行拼接处理,得到所述全局特征表示。
- 根据权利要求1所述的图像识别方法,其中,所述利用每一所述待识别医学图像的个体特征表示和所述全局特征表示,确定所述待识别医学图像所属的扫描图像类别包括:利用每一所述待识别医学图像的个体特征表示和所述全局特征表示,得到每一所述待识别医学图像的最终特征表示;利用每个所述待识别医学特征的最终特征表示,确定每一所述待识别医学图像所属的扫描图像类别。
- 根据权利要求4所述的图像识别方法,其中,所述利用每一所述待识别医学图像的个体特征表示和所述全局特征表示,得到每一所述待识别医学图像的最终特征表示包括:分别将每一所述待识别医学图像的个体特征表示和所述全局特征表示进行拼接处理,得到所述待识别医学图像对应的最终特征表示。
- 根据权利要求4所述的图像识别方法,其中,所述分别对每一所述目标区域的图像数据进行特征提取,得到每个待识别医学图像的个体特征表示包括:利用识别网络的特征提取子网络对每一所述目标区域的图像数据进行特征提取,得到每个待识别医学图像的个体特征表示;所述将所述至少一个待识别医学图像的个体特征表示进行融合,得到全局特征表示,利用每一所述待识别医学图像的个体特征表示和所述全局特征表示,得到每一所述待识别医学图像的最终特征表示包括:利用所述识别网络的融合模块将所述至少一个待识别医学图像的个体特征表示进行融合,得到全局特征表示,并利用每一所述待识别医学图像的个体特征表示和所述全局特征表示,得到每一所述待识别医学图像的最终特征表示;所述利用每个所述待识别医学特征的最终特征表示,确定每一所述待识别医学图像所属的扫描图像类别,包括:利用所述识别网络的分类子网络对每一所述待识别医学图像的最终特征表示进行预测分类,得到每一所述待识别医学图像所属的扫描图像类别。
- 根据权利要求6所述的图像识别方法,其中,所述识别网络是利用样本医学图像训练得到的,每次训练所述识别网络所使用的所述样本医学图像数量不固定。
- 根据权利要求6或7所述的图像识别方法,其中,所述特征提取子网络包括至少一组顺序连接的稠密卷积块和池化层;和/或,所述识别网络包括预设数量个特征提取子网络;所述利用识别网络的特征提取子网络对每一所述目标区域的图像数据进行特征提取,得到每个待识别医学图像的个体特征表示包括:将每一所述目标区域的图像数据分别输入对应一个所述特征提取子网络进行特征提取,得到每个待识别医学图像的个体特征表示。
- 根据权利要求1至8所述的图像识别方法,其中,所述分别确定每个所述待识别医学图像中与目标脏器对应的目标区域包括:利用脏器检测网络分别对所述至少一个待识别医学图像进行检测,得到所述目标脏器第一位置信息和所述目标脏器的毗邻脏器的第二位置信息;利用所述第一位置信息和所述第二位置信息,确定所述目标脏器对应的目标区域。
- 根据权利要求9所述的图像识别方法,其中,所述待识别医学图像为三维图像,所述利用脏器检测网络分别对所述至少一个待识别医学图像进行检测,得到所述目标脏器第一位置信息和所述目标脏器的毗邻脏器的第二位置信息之前,所述方法还包括:将每一所述待识别医学图像沿冠状面进行划分,得到多个三维子图像;将每一所述子图像沿垂直于冠状面的方向进行投影,得到对应的二维子图像;所述利用脏器检测网络分别对所述至少一个待识别医学图像进行检测,得到所述目标脏器第一位置信息和所述目标脏器的毗邻脏器的第二位置信息包括:利用所述脏器检测网络对所述至少一个待识别医学图像对应的所述二维子图像进行检测,得到所述第一位置信息和所述第二位置信息。
- 根据权利要求9或10所述的图像识别方法,其中,所述目标脏器为肝脏,所述毗邻脏器包括肾脏、脾脏中的至少一者;和/或,所述第一位置信息包括所述目标脏器对应区域的至少一组对角顶点位置和对应区域的尺寸,所述第二位置信息至少包括所述毗邻脏器对应区域的至少一个顶点位置。
- 根据权利要求1至11任一项所述的图像识别方法,其中,所述利用每一所述待识别医学图像的个体特征表示和所述全局特征表示,确定每一所述待识别医学图像所属的扫描图像类别之后,所述方法还包括以下至少一者:将所述至少一个待识别医学图像按照其扫描图像类别进行排序;若所述待识别医学图像的扫描图像类别存在重复,则输出第一预警信息,以提示扫描人员;若所述至少一个待识别医学图像的扫描图像类别中不存在预设扫描图像类别,则输出第二预警信息,以提示扫描人员。
- 根据权利要求1至12任一项所述的图像识别方法,其中,所述分别对每一所述目标区域的图像数据进行特征提取,得到每个待识别医学图像的个体特征表示之前,所述方法还包括:对每一所述目标区域的图像数据进行预处理,其中,所述预处理包括以下至少一种:将所述目标区域的图像尺寸调整至预设尺寸,将所述目标区域的图像强度归一化至预设 范围。
- 一种图像识别装置,包括:区域获取模块,配置为获取至少一个扫描得到的待识别医学图像,并分别确定每个所述待识别医学图像中与目标脏器对应的目标区域;特征提取模块,配置为分别对每一所述目标区域的图像数据进行特征提取,得到每个待识别医学图像的个体特征表示;融合处理模块,配置为将所述至少一个待识别医学图像的个体特征表示进行融合,得到全局特征表示;类别确定模块,配置为利用每一所述待识别医学图像的个体特征表示和所述全局特征表示,确定每一所述待识别医学图像所属的扫描图像类别。
- 一种电子设备,包括相互耦接的存储器和处理器,所述处理器配置为执行所述存储器中存储的程序指令,以实现权利要求1至13任一项所述的图像识别方法。
- 一种计算机可读存储介质,其上存储有程序指令,所述程序指令被处理器执行时实现权利要求1至13任一项所述的图像识别方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020227004540A KR20220031695A (ko) | 2020-03-31 | 2021-03-02 | 이미지 인식 방법 및 관련 장치, 기기 |
JP2021577453A JP2022539162A (ja) | 2020-03-31 | 2021-03-02 | 画像識別方法およびその関連装置、機器 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010246970.5 | 2020-03-31 | ||
CN202010246970.5A CN111507381B (zh) | 2020-03-31 | 2020-03-31 | 图像识别方法及相关装置、设备 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021196955A1 true WO2021196955A1 (zh) | 2021-10-07 |
Family
ID=71874107
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/078748 WO2021196955A1 (zh) | 2020-03-31 | 2021-03-02 | 图像识别方法及相关装置、设备 |
Country Status (5)
Country | Link |
---|---|
JP (1) | JP2022539162A (zh) |
KR (1) | KR20220031695A (zh) |
CN (1) | CN111507381B (zh) |
TW (1) | TWI776426B (zh) |
WO (1) | WO2021196955A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115082772A (zh) * | 2022-07-19 | 2022-09-20 | 小米汽车科技有限公司 | 地点识别方法、装置、车辆、存储介质及芯片 |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111507381B (zh) * | 2020-03-31 | 2024-04-02 | 上海商汤智能科技有限公司 | 图像识别方法及相关装置、设备 |
CN112464701B (zh) * | 2020-08-26 | 2023-06-30 | 北京交通大学 | 基于轻量化特征融合ssd的人员是否佩戴口罩检测方法 |
CN112329844A (zh) * | 2020-11-03 | 2021-02-05 | 上海商汤智能科技有限公司 | 图像目标分类方法及相关装置、设备、存储介质 |
CN112633086B (zh) * | 2020-12-09 | 2024-01-26 | 西安电子科技大学 | 基于多任务EfficientDet的近红外行人监测方法、系统、介质、设备 |
CN112819811A (zh) * | 2021-02-24 | 2021-05-18 | 上海商汤智能科技有限公司 | 图像分析方法及相关装置、电子设备、存储介质 |
CN113065017A (zh) * | 2021-03-30 | 2021-07-02 | 上海商汤智能科技有限公司 | 医学图像显示方法及相关装置、设备、存储介质 |
CN113362314B (zh) * | 2021-06-18 | 2022-10-18 | 北京百度网讯科技有限公司 | 医学图像识别方法、识别模型训练方法及装置 |
CN113361637A (zh) * | 2021-06-30 | 2021-09-07 | 杭州东方通信软件技术有限公司 | 基站机房的安全隐患识别方法及装置 |
CN113516758B (zh) * | 2021-07-07 | 2024-10-29 | 上海商汤善萃医疗科技有限公司 | 图像显示方法及相关装置和电子设备、存储介质 |
CN114049937B (zh) * | 2021-11-22 | 2024-07-23 | 上海商汤善萃医疗科技有限公司 | 图像评测方法及相关装置、电子设备和存储介质 |
CN114154512A (zh) * | 2021-12-09 | 2022-03-08 | 京东科技信息技术有限公司 | 小样本学习处理方法、装置、设备及存储介质 |
CN114332032A (zh) * | 2021-12-30 | 2022-04-12 | 上海商汤智能科技有限公司 | 图像识别方法、装置、电子设备及存储介质 |
CN114529710A (zh) * | 2021-12-31 | 2022-05-24 | 谢传淼 | 目标区域检测方法、装置及电子设备 |
CN114663715B (zh) * | 2022-05-26 | 2022-08-26 | 浙江太美医疗科技股份有限公司 | 医学图像质控、分类模型训练方法、装置及计算机设备 |
CN115294110B (zh) * | 2022-09-30 | 2023-01-06 | 杭州太美星程医药科技有限公司 | 扫描期的识别方法、装置、电子设备和存储介质 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9305349B2 (en) * | 2013-06-28 | 2016-04-05 | Samsung Electronics Co., Ltd. | Apparatus and method for detecting lesion |
CN110321920A (zh) * | 2019-05-08 | 2019-10-11 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、计算机可读存储介质和计算机设备 |
CN110675940A (zh) * | 2019-08-01 | 2020-01-10 | 平安科技(深圳)有限公司 | 病理图像标注方法、装置、计算机设备及存储介质 |
CN110752028A (zh) * | 2019-10-21 | 2020-02-04 | 腾讯科技(深圳)有限公司 | 一种图像处理方法、装置、设备以及存储介质 |
CN110866897A (zh) * | 2019-10-30 | 2020-03-06 | 上海联影智能医疗科技有限公司 | 一种图像检测方法及计算机可读存储介质 |
CN111507381A (zh) * | 2020-03-31 | 2020-08-07 | 上海商汤智能科技有限公司 | 图像识别方法及相关装置、设备 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103974661B (zh) * | 2011-12-21 | 2016-08-24 | 株式会社日立制作所 | 医用图像诊断装置以及使用了医用图像诊断装置的相位决定方法 |
US11238975B2 (en) * | 2014-04-02 | 2022-02-01 | University Of Louisville Research Foundation, Inc. | Computer aided diagnosis system for classifying kidneys |
US10275877B2 (en) * | 2015-06-12 | 2019-04-30 | International Business Machines Corporation | Methods and systems for automatically determining diagnosis discrepancies for clinical images |
CN106204587B (zh) * | 2016-05-27 | 2019-01-08 | 浙江德尚韵兴图像科技有限公司 | 基于深度卷积神经网络和区域竞争模型的多器官分割方法 |
WO2019000455A1 (zh) * | 2017-06-30 | 2019-01-03 | 上海联影医疗科技有限公司 | 图像分割的方法及系统 |
CA3079438A1 (en) * | 2017-11-30 | 2019-06-06 | Nantomics, Llc | Detecting intratumor heterogeneity of molecular subtypes in pathology slide images using deep-learning |
US11721439B2 (en) * | 2018-03-08 | 2023-08-08 | Koninklijke Philips N.V. | Resolving and steering decision foci in machine learning-based vascular imaging |
CN109492529A (zh) * | 2018-10-08 | 2019-03-19 | 中国矿业大学 | 一种多尺度特征提取及全局特征融合的人脸表情识别方法 |
CN110335259B (zh) * | 2019-06-25 | 2021-08-31 | 腾讯科技(深圳)有限公司 | 一种医学影像识别方法、装置和存储介质 |
CN110570483B (zh) * | 2019-08-08 | 2023-12-22 | 上海联影智能医疗科技有限公司 | 扫描方法、装置、计算机设备和存储介质 |
CN110503642B (zh) * | 2019-08-23 | 2021-10-29 | 强联智创(北京)科技有限公司 | 一种基于dsa图像的定位方法及系统 |
CN110852987B (zh) * | 2019-09-24 | 2022-04-22 | 西安交通大学 | 基于深形态学的血管斑块检测方法、设备及存储介质 |
-
2020
- 2020-03-31 CN CN202010246970.5A patent/CN111507381B/zh active Active
-
2021
- 2021-03-02 WO PCT/CN2021/078748 patent/WO2021196955A1/zh active Application Filing
- 2021-03-02 JP JP2021577453A patent/JP2022539162A/ja active Pending
- 2021-03-02 KR KR1020227004540A patent/KR20220031695A/ko active Search and Examination
- 2021-03-16 TW TW110109423A patent/TWI776426B/zh active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9305349B2 (en) * | 2013-06-28 | 2016-04-05 | Samsung Electronics Co., Ltd. | Apparatus and method for detecting lesion |
CN110321920A (zh) * | 2019-05-08 | 2019-10-11 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、计算机可读存储介质和计算机设备 |
CN110675940A (zh) * | 2019-08-01 | 2020-01-10 | 平安科技(深圳)有限公司 | 病理图像标注方法、装置、计算机设备及存储介质 |
CN110752028A (zh) * | 2019-10-21 | 2020-02-04 | 腾讯科技(深圳)有限公司 | 一种图像处理方法、装置、设备以及存储介质 |
CN110866897A (zh) * | 2019-10-30 | 2020-03-06 | 上海联影智能医疗科技有限公司 | 一种图像检测方法及计算机可读存储介质 |
CN111507381A (zh) * | 2020-03-31 | 2020-08-07 | 上海商汤智能科技有限公司 | 图像识别方法及相关装置、设备 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115082772A (zh) * | 2022-07-19 | 2022-09-20 | 小米汽车科技有限公司 | 地点识别方法、装置、车辆、存储介质及芯片 |
Also Published As
Publication number | Publication date |
---|---|
CN111507381A (zh) | 2020-08-07 |
KR20220031695A (ko) | 2022-03-11 |
TW202139208A (zh) | 2021-10-16 |
CN111507381B (zh) | 2024-04-02 |
TWI776426B (zh) | 2022-09-01 |
JP2022539162A (ja) | 2022-09-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021196955A1 (zh) | 图像识别方法及相关装置、设备 | |
US20220222932A1 (en) | Training method and apparatus for image region segmentation model, and image region segmentation method and apparatus | |
US8369593B2 (en) | Systems and methods for robust learning based annotation of medical radiographs | |
CN111696083B (zh) | 一种图像处理方法、装置、电子设备及存储介质 | |
Zhang et al. | Intelligent scanning: Automated standard plane selection and biometric measurement of early gestational sac in routine ultrasound examination | |
WO2022062590A1 (zh) | 图像识别方法及装置、设备、存储介质和程序 | |
JP2022537866A (ja) | 画像分類方法、画像分類装置、画像処理方法、医療用電子機器、画像分類機器、及びコンピュータプログラム | |
EP3612981A1 (en) | Target detection in latent space | |
TW202125415A (zh) | 三維目標檢測及模型的訓練方法、設備、儲存媒體 | |
WO2019223123A1 (zh) | 病变部位识别方法及装置、计算机装置及可读存储介质 | |
Yao et al. | Advances on pancreas segmentation: a review | |
WO2022095258A1 (zh) | 图像目标分类方法、装置、设备、存储介质及程序 | |
CN111724371A (zh) | 一种数据处理方法、装置及电子设备 | |
CN111598144B (zh) | 图像识别模型的训练方法和装置 | |
Chupin et al. | Fully automatic segmentation of the hippocampus and the amygdala from MRI using hybrid prior knowledge | |
EP1771821A1 (en) | Projection views and orientation of chest radiographs | |
WO2019223121A1 (zh) | 病变部位识别方法及装置、计算机装置及可读存储介质 | |
CN116862930B (zh) | 适用于多模态的脑血管分割方法、装置、设备和存储介质 | |
Vittitoe et al. | Markov random field modeling in posteroanterior chest radiograph segmentation | |
Kaur et al. | Systematic survey on generative adversarial networks for brain tumor segmentation and classification | |
CN115035133A (zh) | 一种模型训练方法、图像分割方法及相关装置 | |
Huang et al. | Cross-dataset multiple organ segmentation from CT imagery using FBP-derived domain adaptation | |
CN113450306A (zh) | 提供骨折检测工具的方法 | |
KR20220095401A (ko) | 인공지능을 이용한 안면부 골절 판독 장치 및 방법 | |
CN113239978B (zh) | 医学图像预处理模型与分析模型的相关方法和装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21780594 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021577453 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20227004540 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21780594 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 13.04.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21780594 Country of ref document: EP Kind code of ref document: A1 |