CN111583204B - Organ positioning method of two-dimensional sequence magnetic resonance image based on network model - Google Patents
Organ positioning method of two-dimensional sequence magnetic resonance image based on network model Download PDFInfo
- Publication number
- CN111583204B CN111583204B CN202010344910.7A CN202010344910A CN111583204B CN 111583204 B CN111583204 B CN 111583204B CN 202010344910 A CN202010344910 A CN 202010344910A CN 111583204 B CN111583204 B CN 111583204B
- Authority
- CN
- China
- Prior art keywords
- organ
- image
- positioning
- sequence
- dimensional
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 210000000056 organ Anatomy 0.000 title claims abstract description 123
- 238000000034 method Methods 0.000 title claims abstract description 57
- 238000001514 detection method Methods 0.000 claims abstract description 55
- 238000012545 processing Methods 0.000 claims abstract description 34
- 238000012795 verification Methods 0.000 claims abstract description 25
- 238000007781 pre-processing Methods 0.000 claims abstract description 11
- 230000008569 process Effects 0.000 claims abstract description 8
- 238000012549 training Methods 0.000 claims description 33
- 230000004807 localization Effects 0.000 claims description 14
- 238000012216 screening Methods 0.000 claims description 9
- 238000010606 normalization Methods 0.000 claims description 7
- 238000002372 labelling Methods 0.000 claims description 5
- 238000013519 translation Methods 0.000 claims description 5
- 230000005489 elastic deformation Effects 0.000 claims description 3
- 230000007246 mechanism Effects 0.000 claims description 3
- 238000009966 trimming Methods 0.000 claims description 2
- 238000002360 preparation method Methods 0.000 abstract description 2
- 210000002307 prostate Anatomy 0.000 description 23
- 238000013527 convolutional neural network Methods 0.000 description 16
- 230000011218 segmentation Effects 0.000 description 16
- 238000010586 diagram Methods 0.000 description 10
- 230000000694 effects Effects 0.000 description 9
- 238000010200 validation analysis Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 6
- 238000012360 testing method Methods 0.000 description 6
- 238000000605 extraction Methods 0.000 description 5
- 239000010410 layer Substances 0.000 description 4
- 238000011176 pooling Methods 0.000 description 4
- 210000001519 tissue Anatomy 0.000 description 4
- 238000003384 imaging method Methods 0.000 description 3
- 230000003902 lesion Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 210000004907 gland Anatomy 0.000 description 2
- 210000000664 rectum Anatomy 0.000 description 2
- QLMNCUHSDAGQGT-UHFFFAOYSA-N sintofen Chemical compound N1=C(C(O)=O)C(=O)C=2C(OCCOC)=CC=CC=2N1C1=CC=C(Cl)C=C1 QLMNCUHSDAGQGT-UHFFFAOYSA-N 0.000 description 2
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 1
- 206010060862 Prostate cancer Diseases 0.000 description 1
- 208000000236 Prostatic Neoplasms Diseases 0.000 description 1
- 238000001574 biopsy Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000004195 computer-aided diagnosis Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000013136 deep learning model Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 230000001094 effect on targets Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 239000011229 interlayer Substances 0.000 description 1
- 210000003734 kidney Anatomy 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000013178 mathematical model Methods 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 230000000877 morphologic effect Effects 0.000 description 1
- 210000000496 pancreas Anatomy 0.000 description 1
- 208000023958 prostate neoplasm Diseases 0.000 description 1
- 238000011158 quantitative evaluation Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 238000013526 transfer learning Methods 0.000 description 1
- 238000012285 ultrasound imaging Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 238000011179 visual inspection Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30081—Prostate
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Medical Informatics (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a network model-based organ positioning method for a two-dimensional sequence magnetic resonance image, which comprises the following steps: data set preparation, preprocessing and expansion processing; constructing an improved organ preliminary positioning network model based on the fast R-CNN; optimizing and adjusting the organ primary positioning network model by using the verification set to obtain an organ positioning network model; performing organ preliminary positioning on the verification set image to obtain a plurality of target detection frames and corresponding credibility scores aiming at each two-dimensional slice image, and reserving target candidate frames with credibility greater than a threshold value in each image; constructing a spatial curve fitting model based on sequence correlation processing; preprocessing a two-dimensional sequence magnetic resonance image of an organ to be positioned, inputting the preprocessed two-dimensional sequence magnetic resonance image into an organ positioning network model, and obtaining a target candidate frame with the credibility being more than a threshold value in each image according to an organ initial positioning result; the organ is finally located based on the sequence correlation process. The pair of organs is accurately positioned and has good generalization performance.
Description
Technical Field
The invention relates to a method for positioning organs in a magnetic resonance image, in particular to a method for positioning organs in a two-dimensional sequence magnetic resonance image based on a network model.
Background
Magnetic Resonance (MR) imaging has now become the primary imaging modality for prostate-assisted diagnosis due to its superior spatial resolution and tissue contrast. Compared to transrectal ultrasound imaging (TRUS), it facilitates targeted biopsy and treatment with important implications for prostate tumor lesion localization, volume assessment, and staging of prostate cancer. However, the examination of MR images of the prostate is currently a visual examination by a radiologist on the basis of each slice image, and is therefore a rather time-consuming, cumbersome and subjective task.
Organ localization is important for many medical image processing tasks such as image registration, organ segmentation and lesion detection. An effective assessment of the initial position of the organ can greatly improve the performance of subsequent processing. For example, for organ segmentation, the initial positioning of the organ can focus the segmentation task on the region of interest, thereby improving the segmentation speed, reducing the memory storage, and reducing the risk of false positive segmentation.
At present, some semi-automatic or fully automatic methods applied to computer-aided diagnosis have been proposed in the segmentation and detection of various organs/tissues of medical images. However, due to the heterogeneity of image brightness, imaging artifacts, and signal intensity around the rectal coil caused by the differences in the scanners and scanning schemes, and the differences in size and shape of the glands themselves, as well as the inherent differences in low contrast between the glands and surrounding tissue structures, lack of strong boundaries, etc., prostate segmentation and detection still face significant challenges.
In recent years, a two-stage region-based target detection algorithm, such as Faster R-CNN, has been widely used in the field of medical image processing due to its excellent detection accuracy and good detection efficiency. But since the network was originally designed for target detection in natural images and there are still some limitations for the detection of small targets, it is not possible to uniquely and accurately locate organs in magnetic resonance images.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention provides an organ positioning method based on the improved two-dimensional sequence magnetic resonance image of Faster R-CNN, which fully utilizes the difference between the natural image characteristics and the medical image characteristics to optimize the positioning accuracy and the detection success rate of organs.
Therefore, the invention adopts the following technical scheme:
a organ positioning method of two-dimensional sequence magnetic resonance images based on a network model comprises the following steps:
s1, preparing a data set:
collecting a plurality of two-dimensional sequence magnetic resonance images about an organ, labeling organ regions in the images by a rectangular target box, and then dividing the images into a training set and a verification set;
s2, preprocessing a data set: sequentially carrying out pixel intensity maximum and minimum normalization processing, center cutting processing and image size normalization processing on each two-dimensional sequence MR image;
s3, performing data expansion processing on the training set image obtained in the step S2 and the rectangular target frame mark of the corresponding organ;
s4, constructing an improved organ positioning network model based on the master R-CNN, and comprising the following steps:
1) Constructing a target detection network architecture based on the improved Faster R-CNN, replacing a VGG16 architecture in the classic Faster R-CNN with ResNet-50 with a space attention mechanism, and training ResNet-50 by using an ImageNet large-scale natural data set to obtain an initial training weight parameter of the network;
2) Using the training set image expanded in the step S3 and a corresponding rectangular target frame label as the input of the network in the step 1), and performing iterative adjustment on the whole network architecture parameter by using a multitask loss function formed by classification loss and regression loss to complete network training and generate an organ preliminary positioning network model;
s5, labeling the verification set image preprocessed in the step S2 and the rectangular target frame corresponding to the verification set image and inputting the labeled verification set image and the rectangular target frame into the organ primary positioning network model generated in the step S4, and optimizing and adjusting the organ positioning model according to an output result to obtain an organ positioning network model;
s6, performing preliminary organ positioning on the verification set image preprocessed in the step S2 by using the network model constructed in the step S5 to obtain a plurality of target detection frames and corresponding credibility scores for each two-dimensional slice image, and reserving target candidate frames with credibility larger than a certain threshold value in each image;
s7, constructing a spatial curve fitting model based on sequence correlation processing:
1) Aiming at each two-dimensional slice image with the initial detection result in the S6, the axial slice images belonging to the same sequence are sorted in a front-back order;
2) Extracting key points of a target frame in the slice image with a single reliable target candidate frame from each sorted sequence, performing spatial curve fitting in the sequence direction by using the key points, and determining the best curve fitting mode to be least square quartic polynomial fitting;
s8, preprocessing a two-dimensional sequence magnetic resonance image of a certain organ to be positioned according to the method in the step S2, and inputting the preprocessed two-dimensional sequence magnetic resonance image into the organ positioning network model obtained in the step S5 to obtain a primary organ positioning result;
s9, obtaining a plurality of target detection frames and corresponding credibility scores of each two-dimensional slice image according to the organ initial positioning result output in the step S8, and reserving target candidate frames with credibility greater than a certain threshold value in each image;
s10, performing final positioning on the organ based on sequence correlation processing:
1) Sorting each two-dimensional slice image with the preliminary detection result in the S9 in sequence;
2) Aiming at the sorted sequence image, extracting key points of a target frame in a slice image with a single reliable target candidate frame, and performing spatial curve fitting in the sequence direction on the key points according to a least square quartic polynomial fitting mode obtained in S7;
3) Screening and judging a target frame closest to a fitting position in a two-dimensional image with multiple prediction frames as a final positioning frame of the organ by using the fitted space curve and the minimum space Euclidean distance, and updating space curve fitting parameters after each screening;
4) And fitting the two-dimensional images of the missing detection of the first end and the last end of the sequence in each sequence by using the finally updated space curve to complete the final positioning of the organ.
When the pixel intensity is normalized in step S2, the pixel value x of the pixel i i The normalized pixel values are: norm i =(x i -Pmin)/(Pmax-Pmin)×255,norm i ∈[0,255]Wherein Pmax and Pmin refer to the maximum and minimum pixel values of the slice image where the pixel i is located respectivelyThe value is obtained.
In step S2, when the center trimming process is performed, if the original size is W × H, the trimmed size is α W × α H, α is a scale factor, and α =2/3 is preferable.
Preferably, in step S2, the image size is normalized to 600 × 600.
In step S3, data expansion processing is performed through horizontal flipping, small-amplitude horizontal translation or vertical translation, small-angle random rotation, and elastic deformation. Preferably, the small angle is randomly rotated by an angle of-5 to 5 °.
In step S6, the threshold value is 0.80 to 0.90.
The invention is inspired by the fact that the fast R-CNN model obtains excellent performance in natural image target detection, and is established on the basis of the model, further improves and perfects the model by fully utilizing the central position prior information of an organ in an image and the evolution rule between sequence images, thereby realizing the accurate positioning of the prostate organ in a magnetic resonance image and laying a good foundation for medical image tasks such as subsequent organ registration, segmentation and the like. The improved two-dimensional sequence magnetic resonance organ positioning method based on the Faster R-CNN fully utilizes the difference between the natural image characteristics and the medical image characteristics to optimize the organ positioning accuracy and the detection success rate.
Experiments prove that on the premise of not having too much complicated image preprocessing work, the method can realize more accurate organ positioning and has good generalization performance. The method has the recall rate of 96.91 percent for the area of the organ and the positioning success rate of 99.39 percent for the organ.
Drawings
FIG. 1 is a block flow diagram of the method of the present invention;
figure 2 is an example of axial magnetic resonance images of the prostate from different medical centers used in the present invention;
fig. 3 is a specific structure of a part of network blocks in the present invention: (a) convolving the attention block; (b) an identification block; (c) a spatial attention module;
FIG. 4 is a partial preliminary test result based on improved Faster R-CNN: (a) An example of a slice image with multiple target detection results; (b) slice image examples lacking target detection results;
FIG. 5 is a schematic diagram of spatial curve fitting and updating based on sequence correlation, wherein: (1) Projection of key points of the rejected target frame on X-Z and X-Y planes; (2) drawing and adjusting a space fitting curve; (3) projection of the space fitting curve on an X-Z surface; (4) Projection of space fitting curve on X-Y plane
FIG. 6 is a partial inspection result plot of the method of the present invention on a validation set: wherein the red mark is the real outline of the prostate organ, and the yellow rectangular box is the organ positioning result of the algorithm;
FIG. 7 is a graph of axial slice test results of the method of the invention in one practical case: (a) the 5 th slice; (b) the 11 th slice; (c) slicing the 15 th slice; (d) the 18 th slice.
FIG. 8 is a flow chart of the model training and testing stages of the present invention.
Detailed Description
Because the initial positioning of the organ is very important for medical image processing such as organ segmentation, image registration, lesion detection and the like, effective evaluation of the initial position of the organ can help to remove most of interference information in the image background, thereby improving subsequent processing. The invention introduces a spatial attention module which focuses on organ position characteristics and correlation processing which improves the detection effect among sequences based on a two-stage detection algorithm, shows excellent detection effect on a test data set with higher heterogeneity, and lays a good foundation for the subsequent processing of medical images.
The method of the present invention will be described in detail with reference to specific examples.
Embodiment a two-dimensional sequence magnetic resonance image organ positioning method based on network model
This example describes the positioning method of the prostate organ in detail by taking the example of a prostate magnetic resonance data set with a small size and high heterogeneity. As shown in fig. 8, the method comprises the steps of:
s1, preparing a data set:
the prostate magnetic resonance image used in this example is from the prostate segmentation challenge match PROMISE12 dataset published at the MICCAI conference of 2012. The data set contains 80T 2-weighted magnetic resonance image cases from four medical centers, 50 of which have expert manual segmentation masks, each center using different acquisition equipment and acquisition protocols when acquiring images. Sample images from four centers are shown in fig. 2. Wherein 1.5T and 3T represent the magnetic field strength and the presence or absence of ERC represents whether a rectal coil is used for image acquisition.
This example divides the 50 prostate MR images with expert manual segmentation masks into a training set and a verification set in a ratio of 4.
S2, preprocessing a data set image:
(1) Preprocessing images of a training set and a verification set:
and performing the following four steps of basic preprocessing operation on the obtained training set and verification set images: (a) Because the interlayer resolution of the original three-dimensional image in the axial direction is higher than the in-layer resolution, the three-dimensional image in the original data is expanded according to the axial sequence to form a two-dimensional sequence MR image. (b) Secondly, since the image intensities from different acquisition centers are greatly different, intensity maximum and minimum normalization processing is performed for each two-dimensional slice image. E.g. pixel value x of pixel i i The normalized pixel values are: norm i =(x i -Pmin)/(Pmax-Pmin)×255,norm i ∈[0,255]And Pmax and Pmin respectively refer to the maximum value and the minimum value of the pixel of the slice image where the pixel i is located. (c) In addition, in consideration of a priori information that an organ is usually located at the center of an image, in order to enhance the detection performance for a small object, center clipping processing is performed on the image (if the original size is W × H, the post-clipping size is α W × α H, α is a scale factor, and in this embodiment, α = 2/3). (d) Finally, the image sizes are uniformly normalized, and normalized to 600 × 600 in this embodiment.
(2) Generation of training set and validation set target box labels
To train and validate the deep learning network with tagged data, image-level annotations, i.e., rectangular target box annotations, are generated based on each training and validation image using the segmentation masks (i.e., pixel-level annotations) corresponding to the training and validation set images. According to the method, after the center cutting and size normalization operation of (c) - (d) in (1) are carried out on the segmentation mask images corresponding to the training and verification images, the circumscribed rectangle of the real outline of the organ is used as the position label of the organ, and the training set images and the verification set images and the corresponding label images with the position labels are used as the input of a subsequent training and verification network.
For data sets without segmentation masks, rectangular target box labeling may be performed on regions of the organ in these images first.
S3, data expansion processing:
because extensive data sets are often needed for deep learning model training to better learn the generalization characteristics of image data, and medical images are often difficult to obtain due to problems of patient privacy and the like, the invention carries out data expansion processing on training set images and corresponding segmentation mask images through four modes of horizontal turning, small-amplitude horizontal or vertical translation, small-angle random rotation (-5 ℃) and elastic deformation. The number of the expanded images is 16 times of the number of the original images.
S4, constructing an improved organ preliminary positioning network model based on the master R-CNN:
(1) Building a network structure:
the invention adopts a network structure similar to that of Faster R-CNN, except that a ResNet-50 architecture with a space attention mechanism is used for replacing VGG16 in the original Faster R-CNN to realize the feature extraction of an input image and the classification and identification of targets in the image, as shown in FIG. 1. The specific structure of the convolution attention module and the embedding manner of the spatial attention module therein are shown in fig. 3 (a); the specific structure of the identification block and the spatial attention module is shown in fig. 3 (b), (c). The spatial attention module is added to better utilize the central position information of the organ in the image and enhance the sensitivity of the network to the position feature learning. With F ∈ R C×H×W When the feature map is input as an intermediate feature map, the calculation formula of the attention feature map added into the spatial attention module is as follows:
Attention(F)=sigmoid(f 7×7 [F avg ,F max ])F avg (F)=Global_avgpool(F),
F max (F)=Global_maxpool(F)
wherein, F avg ,F max ∈R 1×H×W ,f 7×7 Representing a convolution operation with a 7 x 7 convolution kernel.
(2) Network training strategy and loss function settings
1) Training strategy
Due to the limitation of the medical data set, the network is pre-trained on the ImageNet large-scale natural image set by using the idea of transfer learning, the pre-trained weight parameters are used as the initial weights of the method, and the network is further trained by using the labeled prostate data set on the basis. Parameters of beta 1=0.9, beta 2=0.999 and epsilon =10 are selected during the training process -8 Adam optimizer of (1), and using 2X 10 -5 A fixed learning rate.
2) Loss function setting
The Faster R-CNN network comprises two tasks of regression and classification, wherein a regressor is used for carrying out regression prediction on a candidate box, and a classifier is used for classifying objects in the candidate box. Therefore, the network training process uses a multitask loss function, and the calculation formula is as follows:
wherein L is cls 、L reg The classification loss and the regression loss are respectively expressed as follows:
i) Classification loss by using a two-class cross-entropy loss function, where p i Is an achor [ i ]]The probability of the prediction being the target,is an achor [ i ]]True tag ofI.e. satisfy
ii) regression loss:wherein,t i ={t x ,t y ,t w ,t h is a vector representing the prediction candidate box anchor [ i }]Parameterized coordinates of (a);and the coordinate vector of the corresponding real target frame.
(3) Network training
1) Feature extraction
Inputting the images of the training set obtained in the step 3 into a feature extraction network, along with the continuous reduction of the resolution of the convolutional layer, the network can better learn the global features (namely generalization features) of the images, and the addition of a residual error structure further optimizes the feature extraction process. As shown in fig. 1, the image undergoes four down-sampling operations with a step size of 2 in stages 1 to 4, and thus the resulting convolution feature map size is about 1/16 of the original input image size (taking an input image size of 600 × 600 as an example, the output convolution feature map size is 38 × 38).
2) Prediction of target candidate box
(a) Based on the convolution characteristic diagram obtained in the step 1), the RPN network firstly performs a convolution operation of 3 x 3 on the convolution characteristic diagram, then forms a plurality of anchors with different sizes and different proportions by taking each pixel in the convolution characteristic diagram as a center, and maps the anchors onto the convolution characteristic diagram to obtain a plurality of target candidate box suggestions. In the present embodiment, anchlors having dimensions {64,128,256,512} and aspect ratios {0.5,1,2} are selected in consideration of the morphological size of the organ itself and the proportion thereof in the image, thereby forming 4 × 3=12 anchlors at each pixel point.
(b) Performing ROI pooling operation based on the convolution feature map obtained in 1) and the target candidate frame suggestions with different sizes and sizes obtained in (a) to obtain target candidate frame feature maps with the same size;
(c) Performing further convolution operation on the target candidate frame feature map obtained in the step (b) by using a stage 5 shown in fig. 1, and then respectively obtaining a class probability prediction and a position offset prediction of each target candidate frame by using a full-connection layer;
(d) A non-maximum suppression (NMS) with a selected threshold is performed to remove the set of target candidate boxes with higher overlap in the same class.
(e) And (d) carrying out loss function evaluation on the classification and position prediction of the target candidate frame obtained through screening in the step (d) by using the label image corresponding to the training image, so that the parameters of the deep convolutional neural network are iteratively updated by using a gradient descent back propagation method, and the network parameters obtained after iteration to the maximum set times are used as the optimal network parameters to complete training and generate an organ initial positioning network model.
S5, applying the constructed network model to the primary positioning of a verification set:
and (3) inputting the preprocessed verification set image (without target frame marks) in the S2 into the constructed network model, and outputting to obtain a plurality of target detection frames and corresponding credibility scores for each two-dimensional slice image in the verification set image. And reserving the target candidate frames with the credibility larger than the threshold (the threshold is selected to be 0.80 in the method) in each image.
S6, performing final positioning on the organ based on sequence correlation processing:
a phenomenon that a plurality of target detection frames appear in the individual slice image as shown in (a) of fig. 4 and a phenomenon that a target detection frame is absent in (b) occur in the preliminary detection result of step S4 because this way ignores the characteristics of the medical image itself. For organ positioning detection, the position of an organ between two-dimensional sequence MR images of the same individual does not change greatly, and the evolution of the surface contour of the organ has a certain rule, so that the evolution of the surface contour of the organ can be described by using a mathematical model, and the method introduces sequence correlation processing to improve the problems. The method comprises the following steps:
(1) Two-dimensional image serialization:
and (5) aiming at each two-dimensional slice image with the initial detection result in the S4, the axial slice images belonging to the same sequence are well sorted according to the front and back sequence, and preparation is made for further sequence correlation processing.
(2) And finally determining a target frame:
taking the detection of the prostate organ as an example, each image contains the prostate organ and the organ is unique. Therefore, by using the uniqueness of the target and the micro-variability of the organ positions in the sequence images, firstly, aiming at each sequence in the verification set obtained in the step (1), three-dimensional space curve fitting is carried out on nine key points (including the central point, four vertexes and the midpoints of four sides of a rectangular frame) of a target frame in the slice image which belongs to the same sequence and has a single reliable target candidate frame, and the fitting mode adopts least square fourth-order polynomial fitting.
The following processes are performed for the two problems in fig. 4 (a) and (b), respectively:
i) The phenomenon of two or more target frames for individual two-dimensional images in each sequence (as shown in fig. 4 (a)): and screening and judging a target frame closest to the fitting position in the two-dimensional image with the multiple prediction frames as the position of the final organ by using the fitted space curve and the minimum space Euclidean distance, and updating the fitting parameters of the space curve after each screening. FIG. 5 is a schematic diagram illustrating this process, wherein the mathematical formula for the object box filtering is based on:
wherein, b * The organ prediction frame finally selected from the slice image is shown,respectively representing the fitting predicted values of the y direction and the z direction at the jth key point,and coordinate values in the y direction and the z direction at the jth key point of a certain target frame predicted by the network are respectively represented.
II) basal and basal prostate organs at the first and last ends of the sequence in each sequence may appear without target boxes due to the small target (as shown in fig. 4 (b):
due to the limitation of the target detection network to the small target detection, the method utilizes the space curve finally updated by each sequence to fit the two-dimensional images missing detection in the sequence to obtain the target positioning of the organ. Thereby, a final localization of the organ of the validation set image is achieved.
Validation set localization effect assessment
And (3) evaluating the detection effect of the method from two aspects of qualitative and quantitative by combining the positioning result of the verification set image obtained in the step (S6) and the organ true annotation of the verification set image in the step (S2):
from a qualitative perspective, fig. 6 (a) - (d) are graphs of the detection results obtained on the validation set. The red line shows the real outline of the organ, and the yellow rectangle frame shows the detection result of the prostate organ by the method of the present invention. From the detection result diagram, the method can realize more accurate and unique organ positioning, and simultaneously well avoids organs such as bladder, rectum and the like which are close to the organs in shape and size.
From the quantitative evaluation, multiple tests show that the recall rate of the method to the real area of the organ reaches 96.91%, the area recall rate is 50% and is taken as the standard of the positioning success rate, and the positioning success rate of the method to the organ in the verification set reaches 99.39%.
From the detection efficiency, the method only takes 0.3s (Intel Corei7 GHz processor) for organ positioning detection in a two-dimensional slice image, and can well meet the actual requirement in a medical task.
Preferably, the preprocessed verification set is input into the network model formed in step S4, the training network is further optimized and adjusted to obtain an optimized and adjusted network model, and then step S5 is performed.
Although the above description has been given by taking the location of a prostate organ as an example, it can be understood that the method is also applicable to the detection and location of tissues and organs such as kidney and pancreas.
Example two
The method for positioning the prostatic organ to be positioned by utilizing the method comprises the following steps of:
(1) Image preprocessing:
for a three-dimensional magnetic resonance prostate organ image of a certain case, two-dimensional serialization processing is firstly carried out. Then, maximum and minimum intensity normalization processing is performed to normalize the pixel values of the two-dimensional slice image to be between [0,255 ]. Finally, the image is cut properly according to the position and the proportion of the organ in the image, and the size of the image is normalized to 600 multiplied by 600.
(2) Based on the fast R-CNN improved organ positioning network model, the organ is initially positioned:
1) Feature extraction: inputting the two-dimensional slice image obtained in the step (1) into a convolutional neural network constructed in the step S4 to obtain a convolutional characteristic diagram about the image;
2) Generating a target candidate box suggestion: inputting the convolution characteristic diagram obtained in the step 1) into an RPN (resilient packet network) to generate a target candidate box suggestion;
3) ROI pooling (i.e. region of interest pooling): taking the convolution feature map output by 1) and the target candidate box suggestion output by 2) as input, mapping the target candidate box suggestion to the corresponding position of the feature map, and performing pooling operation on the mapped feature map to obtain a target area feature map with uniform size;
4) And (3) classification prediction: classifying according to the target region characteristic graph obtained in the step 3), distinguishing whether an object contained in the corresponding target candidate frame is an organ or a background, and obtaining a confidence score corresponding to the category;
5) And (3) regression of a bounding box: performing boundary frame regression on the target region characteristic graph obtained in the step 3) to refine the boundary frame, and finally obtaining the accurate position of the detection frame;
6) Performing a process with a threshold value N based on the target frames belonging to the organ and the positions thereof obtained in 4) and 5) t (N t Fixed value) to remove the target frame set with higher overlapping degree, and keep the target candidate frame with reliability higher than a certain fixed threshold (the recommended selection threshold is between 0.80 and 0.90) in each two-dimensional slice image. Therefore, the positions of the target detection frames in each two-dimensional slice image and the corresponding credibility scores are obtained.
(3) Final localization of organs based on sequence correlation processing:
1) Sorting the slice images according to the front and back sequence of the original sequence;
2) The least squares quartic polynomial spatial curve fit is performed on nine key points of the target frame (including the center point, four vertices and four edges of the rectangular frame) in the sequence images with a single reliable target candidate frame. And secondly, screening and judging a target frame closest to the fitting position in the two-dimensional image with the multiple prediction frames as a final positioning frame of the organ by using the fitted space curve and the minimum space Euclidean distance, and updating the fitting parameters of the space curve after each screening.
3) Aiming at the phenomenon that no target frame is possibly generated due to small target in organ detection at the head end and the tail end of the sequence, the target positioning of the organ is obtained by fitting the two-dimensional image lacking detection through the finally updated space curve, so that the organ positioning detection of the whole two-dimensional sequence image is realized.
In this embodiment, the map of the prostate organ location effect in the axial slice image is shown in fig. 7, wherein (a) - (d) show the test effect maps of the 5 th, 11 th, 15 th and 18 th slices of the axial serial slice, respectively. Through transverse comparison and observation of the positioning effect of the prostate organs in the four images, the method has a very good positioning effect on the prostate organs in the whole sequence slice, particularly the positioning of the organs at the first end and the last end of the sequence, which is little related to or incomparable with other algorithms; in addition, in a longitudinal view, the method can well exclude areas with similar shapes and sizes with prostate organs, such as bladder, rectum and the like, while realizing more accurate prostate positioning, and shows that the method has better characteristic learning effect on target organs.
In summary, the organ localization identification method for two-dimensional sequence magnetic resonance images has the advantages that a general organ localization method is constructed by using a magnetic resonance data set with small scale and high heterogeneity, and subjectivity of traditional slice-based visual inspection is well avoided. In addition, the invention realizes the preliminary detection and positioning of organs by improving the target detection algorithm, namely Faster R-CNN, used in the existing natural image processing; and then, the organ positioning effect of the head end and the tail end of the sequence is greatly improved by utilizing the relevance between the medical sequence slice images (namely the evolution rule of the position and the shape size of the whole organ between layers).
Claims (8)
1. A organ positioning method of two-dimensional sequence magnetic resonance images based on a network model comprises the following steps:
s1, preparing a data set:
collecting a plurality of two-dimensional sequence magnetic resonance images about an organ, labeling organ regions in the images by a rectangular target box, and then dividing the images into a training set and a verification set;
s2, preprocessing a data set: sequentially carrying out pixel intensity maximum and minimum normalization processing, center cutting processing and image size normalization processing on each two-dimensional sequence MR image;
s3, performing data expansion processing on the training set image obtained in the step S2 and the rectangular target frame marks of the corresponding organs;
s4, constructing an organ preliminary positioning network model based on fast R-CNN improvement, comprising the following steps:
1) Constructing an improved target detection network architecture based on the Faster R-CNN, replacing a VGG16 architecture in the classic Faster R-CNN with ResNet-50 with a space attention mechanism, and training the ResNet-50 by using an ImageNet large-scale natural data set to obtain an initial training weight parameter of the network;
2) Using the training set image expanded in the step S3 and a corresponding rectangular target frame label as the input of the network in the step 1), and performing iterative adjustment on the whole network architecture parameter by using a multitask loss function formed by classification loss and regression loss to complete network training and generate an organ preliminary positioning network model;
s5, labeling and inputting the verification set image preprocessed in the step S2 and the rectangular target box corresponding to the verification set image into the organ primary positioning network model generated in the step S4, and optimizing and adjusting the organ positioning model according to an output result to obtain an organ positioning network model;
s6, performing preliminary organ positioning on the verification set image preprocessed in the step S2 by using the network model constructed in the step S5 to obtain a plurality of target detection frames and corresponding credibility scores aiming at each two-dimensional slice image, and reserving target candidate frames with credibility greater than a certain threshold value in each image;
s7, constructing a spatial curve fitting model based on sequence correlation processing:
1) Aiming at each two-dimensional slice image with the initial detection result in the S6, the axial slice images belonging to the same sequence are sorted in a front-back order;
2) Extracting key points of a target frame in the slice image with a single reliable target candidate frame from each sorted sequence, performing spatial curve fitting in the sequence direction by using the key points, and determining the best curve fitting mode to be least square quartic polynomial fitting;
s8, preprocessing a two-dimensional sequence magnetic resonance image of a certain organ to be positioned according to the method in the step S2, and inputting the preprocessed two-dimensional sequence magnetic resonance image into the organ positioning network model obtained in the step S5 to obtain a primary positioning result of the organ;
s9, obtaining a plurality of target detection frames and corresponding credibility scores of each two-dimensional slice image according to the organ initial positioning result output in the step S8, and reserving target candidate frames with credibility greater than a certain threshold value in each image;
s10, performing final positioning on the organ based on sequence correlation processing:
1) Sorting each two-dimensional slice image with the preliminary detection result in the S9 in sequence;
2) Aiming at the sorted sequence image, extracting key points of a target frame in a slice image with a single reliable target candidate frame, and performing spatial curve fitting in the sequence direction on the key points according to a least square quartic polynomial fitting mode obtained in S7;
3) Screening and judging a target frame closest to a fitting position in a two-dimensional image with multiple prediction frames as a final positioning frame of the organ by using the fitted space curve and the minimum space Euclidean distance, and updating space curve fitting parameters after each screening;
4) And fitting the two-dimensional images of the missing detection of the first end and the last end of the sequence in each sequence by using the finally updated space curve to complete the final positioning of the organ.
2. The organ localization method according to claim 1, wherein: in step S2, when the pixel intensity is normalized, the pixel value x of the pixel i i The normalized pixel values are:
norm i =(x i -Pmin)/(Pmax-Pmin)×255,norm i ∈[0,255],
wherein Pmax and Pmin respectively refer to the maximum value and the minimum value of the pixel of the slice image where the pixel i is located.
3. The organ localization method according to claim 1, wherein: in step S2, when the center trimming process is performed, if the original size is W × H, the trimmed size is α W × α H, and α is a scaling factor.
4. The organ localization method according to claim 3, wherein: the α =2/3.
5. The organ localization method according to claim 1, wherein: in step S2, the image size is normalized to 600 × 600.
6. The organ localization method according to claim 1, wherein: and in the step S3, data expansion processing is carried out through horizontal overturning, small-amplitude horizontal translation or vertical translation, small-angle random rotation and elastic deformation.
7. The organ localization method according to claim 6, wherein: the angle of the small-angle random rotation is-5 degrees.
8. The organ localization method according to claim 1, wherein: the threshold value in step S6 is 0.80-0.90.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010344910.7A CN111583204B (en) | 2020-04-27 | 2020-04-27 | Organ positioning method of two-dimensional sequence magnetic resonance image based on network model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010344910.7A CN111583204B (en) | 2020-04-27 | 2020-04-27 | Organ positioning method of two-dimensional sequence magnetic resonance image based on network model |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111583204A CN111583204A (en) | 2020-08-25 |
CN111583204B true CN111583204B (en) | 2022-10-14 |
Family
ID=72125433
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010344910.7A Active CN111583204B (en) | 2020-04-27 | 2020-04-27 | Organ positioning method of two-dimensional sequence magnetic resonance image based on network model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111583204B (en) |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112164028A (en) * | 2020-09-02 | 2021-01-01 | 陈燕铭 | Pituitary adenoma magnetic resonance image positioning diagnosis method and device based on artificial intelligence |
CN112053342A (en) * | 2020-09-02 | 2020-12-08 | 陈燕铭 | Method and device for extracting and identifying pituitary magnetic resonance image based on artificial intelligence |
CN112598634B (en) * | 2020-12-18 | 2022-11-25 | 燕山大学 | CT image organ positioning method based on 3D CNN and iterative search |
CN113160144B (en) * | 2021-03-25 | 2023-05-26 | 平安科技(深圳)有限公司 | Target object detection method, target object detection device, electronic equipment and storage medium |
CN113436139A (en) * | 2021-05-10 | 2021-09-24 | 上海大学 | Small intestine nuclear magnetic resonance image identification and physiological information extraction system and method based on deep learning |
CN113256574B (en) * | 2021-05-13 | 2022-10-25 | 中国科学院长春光学精密机械与物理研究所 | Three-dimensional target detection method |
CN113674248B (en) * | 2021-08-23 | 2022-08-12 | 广州市番禺区中心医院(广州市番禺区人民医院、广州市番禺区心血管疾病研究所) | Magnetic resonance amide proton transfer imaging magnetic susceptibility detection method and related equipment |
CN113538435B (en) * | 2021-09-17 | 2022-02-18 | 北京航空航天大学 | Pancreatic cancer pathological image classification method and system based on deep learning |
CN114037714B (en) * | 2021-11-02 | 2024-05-24 | 大连理工大学人工智能大连研究院 | 3D MR and TRUS image segmentation method for prostate system puncture |
CN114820584B (en) * | 2022-05-27 | 2023-02-21 | 北京安德医智科技有限公司 | Lung focus positioner |
CN116777935B (en) * | 2023-08-16 | 2023-11-10 | 天津市肿瘤医院(天津医科大学肿瘤医院) | Deep learning-based method and system for automatically segmenting prostate whole gland |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101301207A (en) * | 2008-05-28 | 2008-11-12 | 华中科技大学 | Vascular angiography three-dimensional rebuilding method under dynamic model direction |
CN109727240A (en) * | 2018-12-27 | 2019-05-07 | 深圳开立生物医疗科技股份有限公司 | A kind of three-dimensional ultrasound pattern blocks tissue stripping means and relevant apparatus |
CN110009628A (en) * | 2019-04-12 | 2019-07-12 | 南京大学 | A kind of automatic testing method for polymorphic target in continuous two dimensional image |
CN110211097A (en) * | 2019-05-14 | 2019-09-06 | 河海大学 | Crack image detection method based on fast R-CNN parameter migration |
CN110503112A (en) * | 2019-08-27 | 2019-11-26 | 电子科技大学 | A kind of small target deteection of Enhanced feature study and recognition methods |
CN110610210A (en) * | 2019-09-18 | 2019-12-24 | 电子科技大学 | Multi-target detection method |
CN111027547A (en) * | 2019-12-06 | 2020-04-17 | 南京大学 | Automatic detection method for multi-scale polymorphic target in two-dimensional image |
-
2020
- 2020-04-27 CN CN202010344910.7A patent/CN111583204B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101301207A (en) * | 2008-05-28 | 2008-11-12 | 华中科技大学 | Vascular angiography three-dimensional rebuilding method under dynamic model direction |
CN109727240A (en) * | 2018-12-27 | 2019-05-07 | 深圳开立生物医疗科技股份有限公司 | A kind of three-dimensional ultrasound pattern blocks tissue stripping means and relevant apparatus |
CN110009628A (en) * | 2019-04-12 | 2019-07-12 | 南京大学 | A kind of automatic testing method for polymorphic target in continuous two dimensional image |
CN110211097A (en) * | 2019-05-14 | 2019-09-06 | 河海大学 | Crack image detection method based on fast R-CNN parameter migration |
CN110503112A (en) * | 2019-08-27 | 2019-11-26 | 电子科技大学 | A kind of small target deteection of Enhanced feature study and recognition methods |
CN110610210A (en) * | 2019-09-18 | 2019-12-24 | 电子科技大学 | Multi-target detection method |
CN111027547A (en) * | 2019-12-06 | 2020-04-17 | 南京大学 | Automatic detection method for multi-scale polymorphic target in two-dimensional image |
Non-Patent Citations (2)
Title |
---|
CBAM: Convolutional Block Attention Module;Sanghyun Woo,Jongchan Park,Joon-Young Lee,In So Kweon;《ECCV 2018》;20180718;第4-7页 * |
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks;Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun;《CVPR2016》;20160106;全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN111583204A (en) | 2020-08-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111583204B (en) | Organ positioning method of two-dimensional sequence magnetic resonance image based on network model | |
WO2021088747A1 (en) | Deep-learning-based method for predicting morphological change of liver tumor after ablation | |
CN110338840B (en) | Three-dimensional imaging data display processing method and three-dimensional ultrasonic imaging method and system | |
WO2021203795A1 (en) | Pancreas ct automatic segmentation method based on saliency dense connection expansion convolutional network | |
CN110363802B (en) | Prostate image registration system and method based on automatic segmentation and pelvis alignment | |
CN109215033A (en) | The method and system of image segmentation | |
CN106485695A (en) | Medical image Graph Cut dividing method based on statistical shape model | |
CN108364294A (en) | Abdominal CT images multiple organ dividing method based on super-pixel | |
CN109509193B (en) | Liver CT atlas segmentation method and system based on high-precision registration | |
CN108038513A (en) | A kind of tagsort method of liver ultrasonic | |
CN108062749B (en) | Identification method and device for levator ani fissure hole and electronic equipment | |
CN107274399A (en) | A kind of Lung neoplasm dividing method based on Hession matrixes and 3D shape index | |
CN112365464A (en) | GAN-based medical image lesion area weak supervision positioning method | |
CN112598613A (en) | Determination method based on depth image segmentation and recognition for intelligent lung cancer diagnosis | |
CN103544695B (en) | A kind of efficiently based on the medical image cutting method of game framework | |
CN111784653A (en) | Multi-scale network MRI pancreas contour positioning method based on shape constraint | |
CN111383759A (en) | Automatic pneumonia diagnosis system | |
CN110570430B (en) | Orbital bone tissue segmentation method based on volume registration | |
CN112258536B (en) | Integrated positioning and segmentation method for calluses and cerebellum earthworm parts | |
CN114565762B (en) | Weak supervision liver tumor segmentation method based on ROI and split fusion strategy | |
CN115115662A (en) | Knowledge and data fusion drive-based multi-modal medical image target segmentation method | |
WO2023133929A1 (en) | Ultrasound-based human tissue symmetry detection and analysis method | |
CN118691630B (en) | Optimized segmentation method for lung medical image | |
CN118212459B (en) | Esophageal structure, tumor contour and stage prediction method, device and medium | |
CN116934757B (en) | Method, equipment and storage medium for lung nodule false positive pruning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |