WO2020093042A1 - Neural networks for biomedical image analysis - Google Patents
Neural networks for biomedical image analysis Download PDFInfo
- Publication number
- WO2020093042A1 WO2020093042A1 PCT/US2019/059653 US2019059653W WO2020093042A1 WO 2020093042 A1 WO2020093042 A1 WO 2020093042A1 US 2019059653 W US2019059653 W US 2019059653W WO 2020093042 A1 WO2020093042 A1 WO 2020093042A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- recurrent
- segmentation
- net
- convolutional
- unit
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10116—X-ray image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10132—Ultrasound image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30024—Cell structures in vitro; Tissue sections in vitro
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30041—Eye; Retina; Ophthalmic
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30088—Skin; Dermal
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30096—Tumor; Lesion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30101—Blood vessel; Artery; Vein; Vascular
Definitions
- DPIA Digital Pathology Image Analysis
- the present disclosure describes techniques for using artificial neural networks and other machine-learning models for image processing tasks, including digital pathology image analysis.
- aspects of the subj ect matter described in this specification can be implemented as a computer-based method; a non-transitory, computer-readable medium storing computer- readable instructions to perform the computer-implemented method; and a computer- implemented system comprising a computer memory interoperably coupled with a hardware processor configured to perform the computer-implemented method/the instructions stored on the non-transitory, computer-readable medium.
- a residual unit can help when training deep architectures.
- feature accumulation with recurrent residual convolutional layers can assure better feature representation for segmentation tasks.
- feature accumulation can facilitate designing better U-Net architectures with the same or fewer number of network parameters and with better performance for medical image segmentation.
- the image segmentation neural network subsystem can include a plurality of encoding units arranged in succession so that each encoding unit after a first encoding unit is configured to process an input set of feature maps from a preceding encoding unit to generate an output set of feature maps having a lower dimensionality than the input set of feature maps, wherein the first encoding unit is configured to process a neural network input representing a data map to generate a first output feature map, and each encoding unit comprises a recurrent convolutional block or a recurrent-residual convolutional unit; and a plurality of decoding units arranged in succession so that each decoding unit after a first decoding unit is configured to process a first input set of feature maps from a preceding decoding unit and a second input set of feature maps from a corresponding encoding unit to
- Each encoding unit of the plurality of encoding units can include a recurrent convolutional block.
- the recurrent convolutional block can include a plurality of forward recurrent convolutional layers.
- Each encoding unit of the plurality of encoding units can include a recurrent- residual convolutional unit.
- the recurrent-residual convolutional unit can include a plurality of recurrent convolution layers having residual connectivity.
- the data map can be or include an input image.
- the final feature map can be or include a segmentation map for the data map.
- the system can further include a segmentation engine on the one or more data processing apparatuses, the segmentation engine configured to segment the data map using the segmentation map.
- the final feature map can be or include a density heap map for the data map.
- the data map can be an input image that depicts a slide of cells, and the neural network subsystem can be configured for use in a nuclei segmentation task to identify nuclei in the slide of cells.
- the data map can be an input image that depicts a slide of cells, and the neural network subsystem can be configured for use in an epithelium segmentation task to identify epithelium in the slide of cells.
- the data map can be an input image that depicts a slide of cells, and the neural network subsystem can be configured for use in a tubule segmentation task to identify tubules in the slide of cells.
- Some implementations of the subject matter disclosed herein include methods for processing a data map with a neural network subsystem having a plurality of encoder units and a plurality of decoder units, each decoder unit corresponding to a different encoder unit.
- the method can include actions of processing successive representations of the data map with the plurality of encoder units to generate a set of feature maps for the data map, each feature map having a lower dimensionality than the data map, each encoder unit comprising a recurrent convolutional block or a recurrent-residual convolutional unit; and upsampling the set of feature maps with the plurality of decoder units to generate a final feature map for the data map that has a higher dimensionality than feature maps in the set of feature maps.
- Each encoding unit of the plurality of encoding units can include a recurrent convolutional block.
- the recurrent convolutional block can include a plurality of forward recurrent convolutional layers.
- Each encoding unit of the plurality of encoding units can include a recurrent- residual convolutional unit.
- the recurrent-residual convolutional unit can include a plurality of recurrent convolution layers having residual connectivity.
- the data map can be or include a medical image.
- Some implementations of the subject matter disclosed herein include a system having one or more data processing apparatuses and an image segmentation neural network subsystem implemented on the data processing apparatus(es).
- the image segmentation neural network subsystem can include one or more first convolutional layers, one or more inception recurrent residual convolutional neural network (IRRCNN) blocks, and one or more transition blocks.
- IRRCNN inception recurrent residual convolutional neural network
- Each IRRCNN block can include an inception unit and a residual unit, the inception unit including recurrent convolutional layers that are merged by concatenation, the residual unit configured to sum input features to the IRRCNN block with an output of the inception unit.
- the neural network subsystem can be configured to process a data map to perform a classification task based on the data map.
- the neural network subsystem can further include a softmax layer.
- Some implementations of the subject matter disclosed herein include methods that include actions of obtaining a neural network input, the neural network input representing a data map; processing the neural network input with a neural network system to generate a classification for one or more items shown in the data map, the neural network system including one or more first convolutional layers, one or more inception recurrent residual convolutional (IRRCNN) blocks, and one or more transition blocks; and providing the classification for storage, processing, or presentation.
- IRRCNN inception recurrent residual convolutional
- FIG. 1 is a pictorial representation of an example of a densely connected recurrent convolutional (DCRC) block.
- DCRC densely connected recurrent convolutional
- FIGS. 3A-3C are images showing examples of three different types of cancer cells, including chronic lymphocytic leukemia (CLL) cells, follicular lymphoma (FL) cells, and mantle cell lymphoma (MCL) cells respectively.
- CLL chronic lymphocytic leukemia
- FL follicular lymphoma
- MCL mantle cell lymphoma
- FIGS. 4A-4C are images showing examples of non-overlapping patches from original samples.
- FIG. 5 is a graph showing example values for training and validation accuracy for lymphoma classification for 40 epochs.
- FIG. 6A is a graph showing examples of area under receiver operating characteristics (ROC) curve values for an image-based method.
- FIG. 6B is a graph showing examples of area under ROC curve values for a patch-based method.
- FIG. 7A shows images of example samples of tissue without invasive ductal carcinoma (IDC).
- FIG. 7B shows images of example samples of tissue with IDC.
- FIGS. 8A and 8B are diagrams showing examples of images of randomly- selected samples for first class samples and second class samples, respectively.
- FIG. 9 is a graph showing examples of training and accuracy data for IDC classification.
- FIG. 10 is a graph showing examples of area under ROC curve values for invasive ductal classification.
- FIGS. 11A-11B are diagrams showing examples of images for randomly- selected samples from nuclei segmentation dataset from ISMI-2017.
- FIG. 12 is a graph showing example values for training and validaiton accuracy for nuclei segmenation.
- FIGS. 13A-13C are diagrams showing examples of images for quantitative results for nuclei segmentation.
- FIGS. 14A-14B are diagrams showing examples of images of database samples from Epithelium segmentation.
- FIGS. 15A-15B are diagrams showing examples of images of database samples for epithelium segmentation.
- FIGS. 16A-16C are diagrams showing examples of experimental outputs for
- FIG. 17 is a graph showing an example of a plot of an under ROC curve for
- FIGS. 18A-18B are diagrams showing examples of images of database samples for Tubule segmentation.
- FIGS. 19A-19D are diagrams showing examples of patches from an input sample used for training and testing.
- FIGS. 20A-20D are diagrams of examples of images for quantitiave results for tubule segmentation.
- FIGS. 21A-21D are diagrams showing examples of images for quantitiave results for tubule segmentation.
- FIG. 22 is a graph showing an example of an ROC curve for Tubule segmentation.
- FIGS. 23A and 23B are drawings of examples of input samples and label masks with single pixel annotation, respectively.
- FIG. 24 is a graph showing examples of training and validation accuracy for lymphocyte detection.
- FIGS. 25A-25D are diagrams showing examples of qualitative results for lymphocyte detection with UD-Net.
- FIG. 26 is a diagram showing examples of an image from the dataset.
- FIGS. 27-28 are drawings of examples of patches for non-mitosis and mitosis cells, respectively.
- FIG. 29 shows a graph of example training and validation accuracy values for mitosis detection.
- FIG. 31 is an overall layer flow diagram of a presently-disclosed IRRCNN.
- FIG. 32 is a diagram of an example of an architecture for the Inception
- FIG. 34 is a graph showing example training and validation accuracy values for
- IRRCNN IRCNN, BIN, and BIRN on CIFAR-100.
- FIG. 36 is a diagram showing examples of sample images from the
- FIG. 40 is a diagram showing examples of images from the CU3D-100 dataset.
- FIG. 46 is a diagram showing an example of an RU-Net architecture with convolutional encoding and decoding units using recurrent convolutional layers (RCL) which is based on a U-Net architecture.
- RCL recurrent convolutional layers
- FIGS. 50A and 50B are diagrams showing example patches and corresponding outputs, respectively.
- FIGS. 51A and 51B are graphs showing examples of values for training and validation accuracy of the presently-disclosed RU-Net and R2U-Net models compared to the ResU-Net and U-Net models for 150 epochs.
- FIGS. 52A-52C are diagrams showing examples of experimental outputs for three different datasets for retina blood vessel segmentation using R2UNet.
- FIG. 53 is a diagram showing examples of AUC values for retina blood vessel segmentation for the best performance achieved with R2U-Net on three different datasets.
- FIG. 55 is a diagram illustrating a qualitative assessment of the presently- disclosed R2U-Net for the skin cancer segmentation task.
- FIG. 56 is a diagram showing experimental results for lung segmentation.
- FIG. 59 is a diagram showing examples of testing errors of the R2U-Net
- FIG. 62 is a diagram showing Inception-Recurrent Convolutional Neural
- IRCNN Network
- FIG. 65 is a graph showing examples of values for the training and validation loss of the IRCNN for both experiments using the CIFAR-100 dataset and data augmentation (with and without initialization and optimization).
- FIG. 66 is a graph showing example values for the training and testing accuracy of the IRCNN with LSUV and EVE.
- FIGS. 67 and 68 are graphs showing the model loss and accuracy for both training and validation phases, respectively.
- FIG. 69 is a graph showing example values for the testing accuracy of IRCNN, EIN, and EIRN on CIFAR-100 dataset.
- FIG. 70 is a diagram showing examples of images.
- FIG. 71 is a graph showing example values for validation accuracy of IRCNN, EIRN, EIN, and RCNN.
- the present disclosure applies advanced deep convolutional neural network (DCNN) techniques including IRRCNN-, DCRN-, R2U-Net-, and R2U-Net based regression models for solving different DPIA problems that are evaluated on different publicly available benchmark datasets related to seven unique tasks of DPIA. These tasks include: invasive ductal carcinoma detection, lymphoma classification, nuclei segmentation, epithelium segmentation, tubule segmentation, lymphocyte detection, and mitosis detection. Details of these various networks (e.g., IRRCNN, R2U-Net, RU-Net are described further below in this specification).
- DCNN advanced deep convolutional neural network
- DCRN Densely Connected Recurrent Convolutional Network
- DCNs Densely Connected Networks
- This architecture ensures the reuse of the features inside the model, providing better performance on different computer vision tasks which is empirically investigated on different datasets.
- DCRN Recurrent Connected Convolutional
- FIG. 1 is a pictorial representation of an example of a densely connected recurrent convolutional (DCRC) block.
- x, H,([X 0 , X I , X 2 - XI- I ]) (1) where [x 0 , Xi, x 2 ••• Xi_i] includes the concatenated features from 0, . , 1— 1 layers and
- ( ⁇ ) is a single tensor.
- — 1) are the inputs to the standard convolution layers and the 1 th recurrent convolution layers respectively.
- the w , k) and w , k) values are the weights of the standard convolutional layer and the recurrent convolutional layers of I th layer and k lh feature map respectively, and b ( k) is the bias.
- the recurrent convolution operations are performed with respect to t.
- the ground truth is created with a single pixel annotation, where the individual dot represents a cell.
- Datasets that can be used in various implementations contain around at least five to five hundred nuclei with a center pixel of the cell in input samples.
- each dot is represented with a Gaussian density.
- the R2U-Net model was applied to estimate the Gaussian densities from the input samples instead of computing the class or pixel level probability which is considered for DL-based classification and segmentation model respectively.
- This model is named University of Dayton Network (UD-Net). For each input sample, a density surface D(x) is generated with superposition of the Gaussian values.
- the objective is to regress this density surface for the corresponding input cell image I (x) .
- the goal is achieved with R2U-Net model with the mean squared error loss between the output heat maps and the target Gaussian density surface which is the ultimate loss function for regression problem.
- the model R2U- Net computes the density heat maps D(x).
- Implementations can advance DL approaches such as IRRCNN, DCRCN, and
- the images can include, for example, Lymphoma classification, Invasive ductal carcinoma (IDC) detection, Epithelium segmentation, Tubule segmentation, Nuclei segmentation, Lymphocyte detection, and Mitosis detection.
- IDC Invasive ductal carcinoma
- Keras and TensorFlow frameworks were used on a single GPU machine with 56G of RAM and an NIVIDIA GEFORCE GTX-980 Ti.
- FIGS. 4A-4C are images showing examples of non-overlapping patches from original samples.
- the actual database sample and first five non-overlapping patches from the original images are shown in FIGS. 4A-4C.
- the statistics of the original dataset and the number of samples after extracting non-overlapping patches are shown in Table 2.
- the performance of IRRCNN model can be evaluated with two different approaches: an entire image-based approach, and a patch-based approach.
- image-based approach the original sample is resized to 256x256.
- 8 and 32 samples per batch can be considered for image-based and patch- based methods respectively.
- the Stochastics Gradient Descent (SGD) optimization method is used with initial learning rate 0.01. The model was trained for only 40 epochs, where after 20 epochs the learning rate is decreased with the factor of 10.
- FIG. 5 is a graph showing example values for training and validation accuracy for lymphoma classification for 40 epochs.
- testing accuracy is computed with a testing dataset, which is totally different samples from training samples. Testing accuracies of around 92.12% and 99.8% were achieved for entire image based and patch-based method, as shown in Table 3. From this evaluation, it can be concluded that, as the number of samples increases, the performance of DL approach increases significantly. The highest accuracy is achieved in patch-based method and has an around 3.22% better performance compared to existing deep learning-based approaches for Lymphoma classification.
- FIG. 6A is a graph showing examples of area under ROC curve values for an image-based method.
- FIG. 6B is a graph showing examples of area under ROC curve values for a patch-based method.
- IDC Invasive Ductal Carcinoma
- FIG. 7A shows images of example samples of tissue without IDC.
- FIG. 7B shows images of example samples of tissue with IDC.
- the IRRCNN model with four IRRCNN and transition blocks was used for implementations of the present disclosure.
- FIG. 9 is a graph showing examples of training and accuracy data for invasive ductal carcinoma (IDC) classification.
- FIG. 10 is a graph showing examples of area under ROC curve values for invasive ductal classification.
- the testing result show around 0.9573 as Area under the ROC curve.
- the total testing time for 31508 samples is 109.585 seconds. Therefore, testing time per sample is 0.0035 sec.
- nuclei segmentation is a very import problem in the field of digital pathology for several reasons.
- R2U-Net (1 - 32- 64- 128- 256- 128- 64- 32- 1) were used with 4M network parameters.
- FIGS. 11 A-l 1B are diagrams showing examples of images for randomly-selected samples from a nuclei segmentation dataset from ISMI-2017.
- An Adam optimizer with a learning rate of 2xe-4 can be applied with a cross entropy loss, a batch size of two, and a number of epochs of 1000.
- FIG. 12 is a graph showing example values for training and validaiton accuracy for nuclei segmenation. From FIG 12, it can be observed that the model shows very high accuracy for training, however, around 98% accuracy was achieved during validation phase.
- the method shows around a 97.70% testing accuracy on testing dataset which is 20% of total samples.
- the experimental results are shown in Table 7. From Table 7, it can be seen that around a 3.31% better performance can be achieved as compared to existing deep learning-based approaches for nuclei segmentation on same dataset.
- FIGS. 14A-14B are diagrams showing examples of images of database samples from Epithelium segmentation.
- FIG. 14A shows the input samples
- FIG. 14B shows the corresponding binary masks for input samples.
- FIGS. 15A-15B are diagrams showing examples of images of database samples for epithelium segmentation.
- FIG. 15 A shows an input sample and ground truth of the corresponding samples.
- FIG. 15B shows extracted non-overlapping patches for input mages and output mask are shown.
- FIGS. 16A-16C are diagrams showing examples of experimental outputs for Epithelium segmentation.
- FIG. 16A shows inputs samples
- FIG. 16B shows the ground truth
- FIG. 16C shows the model outputs.
- FIG. 17 is a graph showing an example of a plot of an under ROC curve for Epithelium segmentation. Analysis shows that in the area under ROC curve of FIG. 17, the results have achieved a 92.02% area under ROC curve for epithelium segmentation.
- the R2U-Net is applied for Epithelium segmentation from whole slide images (WSI).
- the experiment has been done with an Epithelium segmentation dataset and achieved 90.50% and 92.54% for Fl-score and accuracy, respectively.
- a conclusion can be made that qualitative results demonstrate very accurate segmentation compared to ground truth
- the R2U-Net model was used, which is an end-to-end model consisting of encoding and decoding units.
- the total number of network parameters can be, for example, 1.107 million.
- FIGS. 18A-18B are diagrams showing examples of patches from an input sample used for training and testing. 80% of the patches were used for training, and the remaining 20% were used for testing.
- Database samples for tubule segmentation include input sample (FIG. 19A) and ground truth (FIG. 19B) of the corresponding samples. Extracted non-overlapping patches for input mages and output mask are shown in FIGS. 19C and 19D, respectively.
- the Adam optimizer was applied with a learning rate of 2e-4 and a cross entropy loss.
- a batch size of 16 and a number of epochs of 500 were used during training for tubule segmentation.
- FIGS. 20A-20D are diagrams of examples of images for quantitiave results for tubule segmentation.
- FIG. 20A shows the inputs samples.
- FIG. 20B shows the label masks.
- FIG. 20C shows the model outputs.
- FIG. 20D shows the only tubule part from benign images.
- FIGS. 21A-21D are diagrams showing examples of images for quantitiave results for tubule segmentation.
- FIG. 21 A shows the inputs samples.
- FIG. 21B shows the label masks.
- FIG. 21C shows the model outputs.
- FIG. 21D shows the only tubule part from benign images.
- FIG. 22 is a graph showing an example of an ROC curve for Tubule segmentation. Analysis can show that it is possible to achieve 90.45% for the area under ROC curve, which is shown in FIG. 22.
- the R2U-Net is applied for tubule segmentation from whole slide images (WSI). The performance of R2U-Net is analyzed on publicly available dataset for tubule segmentation. 90.13% and 90.31% were achieved for the Fl-score and the accuracy, respectively. Qualitative results demonstrate very accurate segmentation compared to ground truth.
- Lymphocyte is a very important part of the human’s immune system and a subtype of white blood cell (WBC). This type of cell is used to determine different types of cancer such as breast cancer and ovarian cancer.
- WBC white blood cell
- the dataset can be taken from published papers.
- the total number of samples are 100 with 100 center pixel annotated masks.
- the size of the image is 100x 100. 90% of patches were used for training, and the remaining 10% were used for testing.
- An Adam optimizer was applied, with a learning rate of 2xe-4 and cross entropy loss.
- a batch size of 32 and a number of epochs of 1000 were used.
- FIG. 24 is a graph showing examples of training and validation accuracy for lymphocyte detection.
- FIGS. 25A-25D are diagrams showing examples of qualitative results for lymphocyte detection with UD-Net.
- FIG. 25 A represents inputs samples
- FIG. 25B shows ground truth
- FIG. 25C shows the models outputs
- FIG. 25D shows the final outputs where the blue dots are for ground truth and green dots for model outputs.
- the R2U-Net is applied for tubule segmentation from whole slide images (WSI) for Lymphocyte detection. 90.23% accuracy for lymphocyte detection was achieved. A conclusion can be made that qualitative results demonstrate very accurate segmentation compared to ground truth.
- the cell growth rate can be determined with counting of mitotic events from the pathological images, which is an important aspect to determine the aggressiveness of breast cancer diagnosis.
- the manually counting process is applied in pathological practices that are very extremely difficult and time consuming. Therefore, an automatic mitosis detection approach has application in pathological practice.
- FIG. 26 is a diagram showing examples of an image from the dataset.
- the images in FIG. 26 include images for an input, an actual mask, a dilated mask, and a mask with target mitosis. As the number of mitosis cell are very less, different augmentation approaches were applied with ⁇ 0,45,90,135,180,215,270 ⁇ . In one study, 32x32 patches were extracted for the input images, with a total number of patches being 728,073. From the patches, 100,000 patches were randomly selected, with 80,000 patches being used for training and the remaining 20,000 patches used for testing.
- FIGS. 27-28 are drawings of examples of patches for non-mitosis and mitosis cells, respectively.
- the testing phase achieved 99.54% testing accuracy for mitosis detection.
- the experimental results show 99.68% area under ROC curve. It takes 138.94 seconds for 20,000 samples.
- the experimental results show higher recognition accuracy against most of the popular DCNN models including the RCNN.
- the performance of the IRRCNN approach was also investigated against the Equivalent Inception Network (EIN) and the Equivalent Inception Residual Network (EIRN) counterpart on the CIFAR-100 dataset. Improvement in classification accuracy of around 4.53%, 4.49% and 3.56% were reported as compared with the RCNN, EIN, and EIRN on the CIFAR-100 dataset respectively.
- the experiment has been conducted on the TinyImageNet-200 and CU3D-100 datasets where the IRRCNN provides better testing accuracy compared to the Inception Recurrent CNN (IRCNN), the EIN, and the EIRN.
- FIG. 30 is a diagram of a visual information processing pipeline of the human brain.
- vl though v4 represent the visual cortex areas.
- the visual context areas of vl though v4 process information using recurrent techniques.
- FIG. 31 is an overall layer flow diagram of a presently-disclosed IRRCNN.
- the IRRCNN includes the IRRCNN-Block, the IRRCNN-Transition block, and the Softmax layer at the end.
- the present disclosure provides an improved DCNN architecture based on Inception, Residual networks and the RCNN architecture. Therefore, the model can be called the Inception Recurrent Residual Convolutional Neural Network (IRRCNN).
- IRRCNN Inception Recurrent Residual Convolutional Neural Network
- An objective of this model is to improve recognition performance using the same number or fewer computational parameters when compared to alternative equivalent deep learning approaches.
- the inception-residual units utilized are based on Inception- v4.
- the Inception-v4 network is a deep learning model that concatenates the outputs of the convolution operations with different sized convolution kernels in the inception block.
- Inception-v4 is a simplified structure of Inception-v3, containing more inception modules using lower rank filters.
- Incpetion-v4 includes a residual concept in the inception network called the Inception-v4 Residual Network, which improves overall accuracy of recognition tasks.
- the outputs of the inception units are added to the inputs of the respective units.
- the overall structure of the presently-disclosed IRRCNN model is shown in FIG. 31. From FIG. 31, it can be seen that the overall model consists of several convolution layers, IRRCNN blocks, transition blocks, and a Softmax at the output layer.
- FIG. 32 is a diagram of an example of an architecture for the Inception Recurrent Residual Convolutional Neural Network (IRRCNN) block.
- the block consists of the inception unit at the top which contains recurrent convolutional layers that are merged by concatenation, and the residual units. A summation of the input features with the outputs of the inception unit can be seen at the end of the block.
- IRRCNN Inception Recurrent Residual Convolutional Neural Network
- a part of this presently-disclosed architecture is the IRRCNN block that includes RCLs, inception units, and residual units (shown in detail in FIG. 32).
- the inputs are fed into the input layer, then passed through inception units where RCLs are applied, and finally the outputs of the inception units are added to the inputs of the IRRCNN-block.
- the recurrent convolution operations perform with respect to the different sized kernels in the inception unit. Due to the recurrent structure within the convolution layer, the outputs at the present time step are added with the outputs of previous time step. The outputs at the present time step are then used as inputs for the next time step. The same operations are performed with respect to the time steps that are considered.
- / is the standard Rectified Linear Unit (ReLU) activation function.
- the performance of this model was also explored with the Exponential Linear Unit (ELU) activation function in the following experiments.
- the outputs y of the inception units for the different size kernels and average pooling layer are defined as yi xi ( ), ysxsM- and y xl (x) respectively.
- the final outputs of Inception Recurrent Convolutional Neural Networks (IRCNN) unit are defined as T(x b w L ) which can be expressed as
- 0 represents the concatenation operation with respect to the channel or feature map axis.
- the outputs of the IRCNN-unit are then added with the inputs of the IRRCNN-block.
- the residual operation of the IRRCNN-block can be expressed by the following equation.
- x i+1 refers to the inputs for the immediate next transition block
- x L represents the input samples of the IRRCNN-block
- iv represents the kernel weights of the I th IRRCNN-block
- T(x L , Wi) represents the outputs from of I th layer of the IRCNN-unit.
- the number of feature maps and the dimensions of the feature maps for the residual units are the same as in the IRRCNN-block shown in FIG. 32. Batch normalization is applied to the outputs of the IRRCNN-block. Eventually, the outputs of this IRRCNN-block are fed to the inputs of the immediate next transition block.
- transition block different operations are performed including convolution, pooling, and dropout, depending upon the placement of the transition block in the network.
- Inception units were not included in the transition block on the small-scale implementation for CIFAR-10 and CIFAR-100. However, inception units were applied to the transition block during the experiment using the TinyImageNet-200 dataset and for the large- scale model which is the equivalent model of Inception-v3.
- the down-sampling operations are performed in the transition block where max-pooling operations are performed with a 3x3 patch and a 2x2 stride.
- the non-overlapping max-pooling operation has a negative impact on model regularization. Therefore, overlapped max-pooling for regularizing the network were used, which is very important when training a deep network architecture.
- Late use of a pooling layer helps to increase the non-linearity of the features in the network, as this results in higher dimensional feature maps being passed through the convolution layers in the network.
- Two special pooling layers were applied in the model with three IRRCNN-blocks and a transition- block for the experiments that use the CIFAR-10 or CIFAR-100 dataset.
- the presently-disclosed IRRCNN model has been evaluated using four different benchmark datasets: CIFAR-10, CIFAR-100, TinyImageNet-200, and CU3D-100.
- the dataset statistics are provided in Table 12. Different validation and testing samples were used for the TinyImageNet-200 dataset. The entire experiment was conducted on a Linux environment running on a single GPU machine with an NVIDIA GTX-980TL
- the values b ⁇ , b 2 e [0,1] are exponential decay rates for moment estimation in Adam.
- the b 3 e [0,1) is an exponential decay rate for computing relative changes.
- the IRRCNN-block uses the 12— norm for a weight regularization of 0.002.
- the ReLU activation function was used in the first experiment, and the ELU activation is used in the second experiment. In both experiments, training occurred for the networks for 350 epochs with a batch size of 128 for CIFAR-10 and 100.
- the CIFAR- 10 dataset is a benchmark dataset for object classification.
- the dataset consists of 32x32 color images split into 50,000 samples for training, and the remaining 10,000 samples are used for testing (classification into one of 10 classes).
- the experiment was conducted with and without data augmentation. When using data augmentation, only random horizontal flipping was applied. This approach achieved around 8.41% testing error without data augmentation and 7.37% testing error with augmented data (only horizontal flipping) using SDG techniques.
- Table 13 Testing error (%) of the IRRCNN on CIFAR-10 object classification dataset without and with data augmentation.
- the dataset contains 50,000 samples for training and 10,000 samples for validation and testing. Each sample is a 32x32x3 image, and the dataset has 100 classes.
- the presently-disclosed IRRCNN model was studied with and without data augmentation. During the experiment with augmented data, the SGD and LSUV initialization approaches and the EVE optimization function were used. In both cases, the presently-disclosed technique shows better recognition accuracy compared with different DCNN models including the IRCNN. Examples of values for the validation accuracy of the IRRCNN model for both experiments on CIFAR-100 with data augmentation are shown in FIG. 34.
- the presently-disclosed IRRCNN model shows better performance in the both experiments when compared to the IRCNN, EIN, and EIRN models.
- the experimental results when using CIFAR-100 are shown in Table 14.
- the IRRCNN model provides better testing accuracy compared to many recently developed methods. A recognition accuracy of 72.78% was achieved with LSUV+EVE, which is around a 4.49% improvement compared to one of the baseline RCNN methods with almost the same number of parameters (-3.5M).
- Table 14 shows a testing error (%) of the IRRCNN on the CIFAR-100 object classification dataset without and with data augmentation (DA). For unbiased comparison, the accuracy provided by recent studies is listed in a similar experimental setting.
- Cl 00 refers to without data augmentation
- Cl 00+ refers to with data augmentation.
- FitNet4 LSUV initialization approach applied to the DCNN architecture
- FitNet4 achieved 70.04% classification accuracy on augmented data with mirroring and random shifts for CIFAR- 100.
- only random horizontal flipping for data augmentation was applied and achieved around 1.76% better recognition accuracy against FitNet4.
- FIG. 34 is a graph showing example training and validation accuracy values for IRRCNN, IRCNN, BIN, and BIRN on CIFAR-100.
- the vertical and horizontal axis represents accuracy and epochs respectively.
- the presently-disclosed model shows the best recognition accuracy in all cases.
- FIG. 34 The model accuracy for both training and validation are shown in FIG. 34. From the figures, it is observed that this presently-disclosed model shows lower loss and highest recognition accuracy compared to EIN and EIRN, which proves the necessity of the presently- disclosed models.
- FIG. 35 is a graph showing examples of values for testing accuracy of the IRRCNN model against IRCNN, EIN, and EIRN on the augmented CIFAR- 100 dataset. It can be summarized that the presently-disclosed IRRCNN provides around 1.02%, 4.49%, and 3.56% improved testing accuracy compared to IRCNN, EIN, and EIRN respectively.
- FIG. 36 is a diagram showing examples of sample images from the TinyImageNet-200 dataset.
- the IRRCNN model with two general convolution layers with a 3x3 kernel were used at the beginning of the network followed by sub-sampling layer with 3x3 convolution using a stride of 2x2.
- four IRRCNN blocks are used followed by four transition blocks.
- a global average pooling layer is used followed by a Softmax layer.
- FIGS. 37 A and 37B are graphs showing examples of accuracy values during training and validation, respectively, for the TinyImageNet-200 dataset.
- the graphs can result from experimentation with the IRRCNN, IRCNN, equivalent RCNN, EIN, and EIRN using the TinyImageNet-200 dataset.
- the presently-disclosed IRRCNN model provides better recognition accuracy during training compared to equivalent models including IRCNN, EIN, and EIRN with almost the same number of network parameters (-15M).
- 15M network parameters
- DCNN takes a lot of time and power when training a reasonably large model.
- the Inception-Residual networks with RCLs significantly reduce training time with faster convergence and better recognition accuracy.
- FIG. 38 is a graph showing examples of values for validation accuracy for various models on the Tiny-ImageNet dataset.
- FIG. 39 is a graph showing examples of values for the top 1% and top 5% testing accuracy on TinyImageNet-200 dataset. From the bar graph, the impact of recurrent connectivity is clearly observed, and a top-l% testing accuracy of 52.23% was achieved, whereas the EIRN and EIN show 51.14% and 45.63% top-l% testing accuracy. The same behavior is observed for Top-5% accuracy as well.
- the IRRCNN provides better testing accuracy when compared against all other models in both cases which absolutely displays the robustness of the presently-disclosed deep learning architecture.
- FIG. 40 is a diagram showing examples of images from the CU3D-100 dataset.
- the images in this dataset are three-dimensional views of real-world objects normalized for different positions, orientations, and scales.
- the rendered images have a 40° depth rotation about the y-axis (plus a horizontal flip), a 20° tilt rotation about x-axis, and an 80° overhead lighting rotation. 75% percent of the images were used for training, and the remaining 25% images were used for testing, which were selected randomly from the whole dataset.
- IRRCNN Inception-v3, and WRN, respectively.
- the IRRCNN model shows a 0.68% and 1.30% higher testing accuracy compared to Inception-v3 and WRN.
- a recently published paper with sparse neural networks with recurrent layers reported about 94.6% testing accuracy on the CU3D dataset which is around 5.24% less testing accuracy compare to this presently-disclosed IRRCNN model.
- the pre-trained ImageNet weights are used as initial weights for IRRCNN and Inception-v3 models, and only a few layers from the top of the models were trained. The pre-trained weights were taken from GitHub.
- split ratio 0.9, 0.7, 0.5, 0.3, and 0.1
- the number of training and validation samples are taken according to the split ratio, where the number of training samples are increased and the number of validation samples are decreased in the trials respectively.
- a split ratio of 0.9 refers to only 10% of samples (1423) being used for training, and the remaining 90% of samples (12815) are used for validation for the first trial.
- a split ratio of 0.7 means that 30% of the samples are used for training and the remaining 70% samples are used for validation for second trial, and so on.
- FIGS. 43A and 43B are graphs showing examples of errors versus split ratio for five different trials on CU3D-100 dataset for training and validation, respectively. It can be also observed from FIGS. 42A-42B that the models have converged after 22 epochs. Therefore, in each trial, 25 epochs were considered, and the errors here match the average training and validation errors of the last five epochs.
- FIGS. 43A-43B show that the presently- disclosed IRRCNN model shows fewer training and validation errors for five different trials in both cases.
- FIG. 44 is a graph showing examples of values for testing accuracy for different trials on CU3D-100 dataset. In each trial, the models have been tested with the remaining 25% of the samples, and the testing errors are shown in FIG. 44. From FIG. 44, it is seen that R2U-Net shows the lowest error for almost all trails compared to Inception-v3 and WRN.
- Ill -Recurrent Residual U-Net for Medical Image Segmentation
- the present disclosure describes using a Recurrent U-Net as well as a Recurrent Residual U-Net model, which are named RU-Net and R2U-Net respectively.
- the presently- disclosed models utilize the power of U-Net, Residual Networks, and Recurrent Convolutional Neural Networks (RCNNs).
- RCNNs Recurrent Convolutional Neural Networks
- the presently-disclosed models are tested on three benchmark datasets such as blood vessel segmentation in retina images, skin cancer segmentation, and lung lesion segmentation.
- the experimental results show superior performance on segmentation tasks compared to equivalent models including a variant of a fully connected convolutional neural network (FCN) called SegNet, U-Net, and the residual U-Net (ResU-Net).
- FCN convolutional neural network
- FIGS. 45A-45C are diagrams showing medical image segmentation examples.
- FIG. 45A shows retina blood vessel segmentation.
- FIG. 45B shows skin cancer lesion segmentation.
- FIG. 45 C shows lung segmentation.
- the present disclosure can be applied in different modalities of medical imaging including segmentation, classification, detection, registration, and medical information processing.
- the medical imaging comes from different imaging techniques such as Computer Tomography (CT), ultrasound, X-ray, and Magnetic Resonance Imaging (MRI).
- CT Computer Tomography
- MRI Magnetic Resonance Imaging
- CAD Computer-Aided Diagnosis
- This specification discloses two modified and improved segmentation models, one using recurrent convolution networks, and another using recurrent residual convolutional networks.
- the presently-disclosed models can be evaluated on different modalities of medical imaging as shown in FIG. 45.
- the present disclosure provides at last two deep-l earning models, including RU-Net and R2U-Net.
- experiments are conducted on three different modalities of medical imaging including retina blood vessel segmentation, skin cancer segmentation, and lung segmentation.
- performance evaluation of the presently- disclosed models is conducted for the patch-based method for retina blood vessel segmentation tasks and the end-to-end image-based approach for skin lesion and lung segmentation tasks.
- the network consists of two main parts: the convolutional encoding and decoding units.
- the basic convolution operations are performed followed by ReLU activation in both parts of the network.
- 2x2 max-pooling operations are performed.
- the convolution transpose (representing up-convolution, or de-convolution) operations are performed to up-sample the feature maps.
- the U-Net model provides several advantages for segmentation tasks. First, this model allows for the use of global location and context at the same time. Second, it works with very few training samples and provides better performance for segmentation tasks. Third, an end-to-end pipeline processes the entire image in the forward pass and directly produces segmentation maps. This ensures that U-Net preserves the full context of the input images, which is a major advantage when compared to patch-based segmentation approaches.
- x[ ⁇ t) and x, r(l,/) (t— 1) are the inputs to the standard convolution layers and the I th RCL respectively.
- the w and w k values are the weights of the standard convolutional layer and the RCL of the k lh feature map respectively, and b k is the bias.
- the outputs of the RCL are fed to the standard ReLU activation function / and are expressed:
- T(xi, iv, ) represents the outputs from of I th layer of the RCNN unit.
- the output of T(x 0 Wi ) is used for downsampling and upsampling layers in the convolutional encoding and decoding units of the RU-Net model respectively.
- the final outputs of the RCNN unit are passed through the residual unit that is shown FIG. 47D. Let’s consider that the output of the RRCNN-block to be x, +1 , and it can be calculated as follows:
- x L represents the input samples of the RRCNN-block.
- the x i+1 sample is the input for the immediate succeeding subsampling or up-sampling layers in the encoding and decoding convolutional units of the R2U-Net model.
- the number of feature maps and the dimensions of the feature maps for the residual units are the same as in the RRCNN-block shown in FIG. 47D.
- the presently-disclosed deep learning models are the building blocks of the stacked convolutional units shown in FIGS. 47B and 47D.
- FIG. 47 A The basic convolutional unit of this model is shown in FIG. 47 A.
- FIG. 47C the U-Net model with forward convolutional layers with residual connectivity is used, which is often called a residual U-net (or a ResU-Net) and is shown in FIG. 47C.
- the third architecture is the U-Net model with forward recurrent convolutional layers as shown in FIG. 47B, which is named RU-Net.
- the last architecture is the U-Net model with recurrent convolution layers with residual connectivity as shown in FIG. 47D, which is named R2U-Net.
- the pictorial representation of the unfolded RCL layers with respect to time step is shown in FIGS. 48A-48B.
- concatenation to the feature maps from the encoding unit to the decoding unit can be applied for the RU-Net and R2U-Net models.
- the U-Net model only shows the benefit during the training process in the form of better convergence.
- the presently-disclosed models show benefits for both training and testing phases due to the feature accumulation inside the model.
- the feature accumulation with respect to different time steps ensures better and stronger feature representation.
- the cropping and copying unit can be removed from the basic U-Net model and use only concatenation operations. Therefore, with all the above-mentioned changes, the presently-disclosed models are much better compared to equivalent SegNet, U-Net and ResU-Net models, which ensures better performance with the same or fewer number of network parameters.
- the presently-disclosed architectures when compared to U-Net.
- the first is the efficiency in terms of the number of network parameters.
- the presently-disclosed RU-Net and R2U-Net architectures are designed to have the same number of network parameters when compared to U-Net and ResU-Net, and the RU- Net and R2U-Net models show better performance on segmentation tasks.
- the recurrent and residual operations do not increase the number of network parameters. However, they do have a significant impact on training and testing performance which is shown through empirical evaluation with a set of experiments. This approach is also generalizable, as it can easily be applied to deep learning models based on SegNet, 3D-UNet, and V-Net with improved performance for segmentation tasks.
- FIGS. 49A-49C are diagrams showing example images from training datasets.
- the image in FIG. 49A was taken from the DRIVE dataset.
- the image in FIG. 49B was taken from the STARE dataset.
- the image from FIG. 49C was taken from the CHASE-DB1 dataset.
- FIG. 49A shows the original images.
- FIG. 49B shows the fields of view (FOV).
- FIG. 49C shows the target outputs.
- each convolutional block consists of two or three RCLs, where 3x3 convolutional kernels are applied, proceeded by ReLU activation layers, followed by a batch normalization layer. For down-sampling, a 2x2 max- pooling layer followed by a 1 xl convolutional layer is used between the convolutional blocks.
- each block consists of a convolutional transpose layer followed by two convolutional layers and a concatenation layer.
- the concatenation operations are performed between the features in the encoding and decoding units in the network.
- the features are then mapped to a single output feature map where l x l convolutional kernels are used with a sigmoid activation function.
- the segmentation region is generated with threshold (7) which is empirically set to 0.5 in the experiment.
- the architecture shown in the fourth row is used for retina blood vessel segmentation on the DRIVE dataset, as well as skin cancer segmentation. Also, the SegNet model was implemented with a similar architecture and a similar number of feature maps for impartial comparison in the cases of skin cancer lesions and lung segmentation. The architecture can be written as
- each convolutional block contains three convolutional layers and a batch normalization layer which requires a total of 14.94M network parameters.
- the architecture cab be written as 1 ⁇ 32(3) ⁇ 64(3) ⁇ 128(3) ⁇ 256(3) ⁇ 128(3) ⁇ 64(3) ⁇ 32(3) ⁇ 1 for the SegNet model (three convolutional layers and a batch normalization layer are used in each block) which requires a total of 1.7M network parameters.
- DRIVE retina blood vessel segmentation
- STARE retina blood vessel segmentation
- CHASE DBl CHASE DBl
- the DRIVE dataset consists of 40 color retina images in total, of which 20 samples are used for training and the remaining 20 samples are used for testing.
- the size of each original image is 565x584 pixels.
- the images were cropped to only contain the data from columns 9 through 574, which then makes each image 565x565 pixels.
- 190,000 randomly selected patches from 20 of the images in the DRIVE dataset were considered, where 171,000 patches were used for training, and the remaining 19,000 patches were used for validation.
- FIGS. 50A and 50B are diagrams showing example patches and corresponding outputs, respectively.
- the size of each patch is 48x48 for all three datasets shown in FIGS. 50A-50B.
- the second dataset, STARE contains 20 color images, and each image has a size of 700x605 pixels. Due to the small number of samples in the STARE dataset, two approaches are often applied for training and testing when using this dataset. First, training is sometimes performed with randomly selected samples from all 20 images.
- Another approach is the“leave-one-ouf’ method, where in each trial one image is selected for testing, and training is conducted on the remaining 19 samples. Therefore, there is no overlap between training and testing samples.
- the“leave-one-ouf’ approach can be used for the STARE dataset.
- the CHASE DBl dataset contains 28 color retina images, and the size of each image is 999x960 pixels. The images in this dataset were collected from both left and right eyes of 14 school children. The dataset is divided into two sets where samples are selected randomly. A 20-sample set is used for training and the remaining 8 samples are used for testing.
- the Skin Cancer Segmentation dataset is taken from the Kaggle competition on skin lesion segmentation that occurred in 2016. This dataset contains 900 images along with associated ground truth samples for training. Another set of 379 images is provided for testing. The original size of each sample was 700x900, which was rescaled to 128x128 for this implementation.
- the training samples include the original images, as well as corresponding target binary images containing cancer or non-cancer lesions. The target pixels are set to a value of either 255 or 0, denoting pixels inside or outside the target lesion respectively.
- DC and JA are calculated using the following Equations 18 and 19, respectively.
- FIGS. 51A and 51B are graphs showing examples of values for training and validation accuracy of the presently-disclosed RU-Net and R2U-Net models compared to the ResU-Net and U-Net models for 150 epochs.
- FIGS. 51A-51B the training and validation accuracy values that are shown are when using the DRIVE dataset.
- the presently-disclosed R2U-Net and RU-Net models provide better performance during both the training and validation phase when compared to the U-Net and ResU-Net models.
- Quantitative results are achieved with the four different models using the DRIVE dataset, and the results are shown in Table 17.
- the overall accuracy and AUC are considered when comparing the performance of the presently-disclosed methods in most cases.
- the results achieved with the presently-disclosed models with 0.841M network parameters (Table 16,“RBVS+LS” row) are higher than those obtained when using the state- of-the-art approaches in most cases.
- FIGS. 52A-52C are diagrams showing examples of experimental outputs for three different datasets for retina blood vessel segmentation using R2UNet.
- FIG. 52A shows input images in gray scale.
- FIG. 52B shows the ground truth.
- FIG. 52C shows the experimental outputs.
- the images correspond to the DRIVE, STARE, and CHASE DBl datasets, respectively.
- Table 17 shows experimental results of presently-disclosed approaches for retina blood vessel segmentation and comparison against other traditional and deep learning- based approaches.
- FIG. 53 is a diagram showing examples of AUC values for retina blood vessel segmentation for the best performance achieved with R2U-Net on three different datasets.
- the ROC for the highest AUCs for the R2U-Net (with 1.07M network parameters) model on each of the three-retina blood vessel segmentation datasets is shown in FIG. 53.
- this dataset is preprocessed with mean subtraction, and was normalized according to the standard deviation.
- the ADAM optimization technique with a learning rate of 2x l0 4 and binary cross entropy loss was used.
- MSE was calculated during the training and validation phase. In this case, 10% of the samples are used for validation during training with a batch size of 32 and 150 epochs.
- the training accuracy of the presently-disclosed R2U-Net and RU-Net models was compared with that of the ResU-Net and U-Net models for an end-to-end image based segmentation approach. The training and the validation accuracy for all four models are shown in FIGS. 54A-54B.
- FIGS. 54A-54B The training and the validation accuracy for all four models are shown in FIGS. 54A-54B.
- 54A and 54B are diagrams showing example values for training and validation accuracy, respectively, of R2U-Net, RU-Net, ResU-Net, and U-Net for skin lesion segmentation.
- the presently-disclosed RU-Net and R2U-Net models show better performance when compared with the equivalent U-Net and ResU-Net models.
- the quantitative results of this experiment were compared against existing methods as shown in Table 18.
- Table 18 shows experimental results of the presently-disclosed approaches for skin cancer lesion segmentation and comparison against other traditional and deep learning-based approaches.
- the presently-disclosed approaches were compared against recently published results using performance metrics including sensitivity, specificity, accuracy, AUC, and DC.
- the presently-disclosed R2U-Net model provides a testing accuracy 0.9472 with a higher AUC, which is 0.9430.
- the JA and DC are calculated for all models, and the R2U-Net model provides 0.9278 for JA, and 0.9627 for the DC for skin lesion segmentation.
- the RU-Net and R2U-Net models show higher accuracy and AUC compared to the VGG-16 and GoolgeNet models. In most cases, the RU-Net and R2U-Net models show better performance against equivalent SegNet, U-Net, and ResU-Net models for skin lesion segmentation. Some qualitative outputs of the SegNet, U-Net, and R2U-Net models for skin cancer lesion segmentation are shown for visual comparison in FIG. 55.
- FIG. 55 is a diagram illustrating a qualitative assessment of the presently- disclosed R2U-Net for the skin cancer segmentation task.
- the first column shows the input sample
- the second column shows ground truth
- the third column shows the outputs from the SegNet model
- the fourth column shows the outputs from the U-Net model
- the fifth column shows the results of the presently-disclosed R2U-Net model.
- the target lesions are segmented accurately with a similar shape in ground truth. However, if one closely observes the outputs in the first, second, and fourth rows of images in FIG.
- the presently-disclosed R2U-Net model provides a very similar output shape to the ground truth when compared to the outputs of the SegNet and U-Net models. If one observes the third row of images in FIG. 55, it can be clearly seen that the input image contains three lesions. One is a target lesion, and the other brighter lesions are not targets.
- the R2U-Net model segments the desired part of the image more accurately when compared to the SegNet and U-Net models.
- the fifth row clearly demonstrates that the R2U-Net model provides a very similar shape to the ground truth, which is a much better representation than those obtained from the SegNet and U-Net models. Thus, it can be stated that the R2U-Net model is more capable and robust for skin cancer lesion segmentation.
- Lung segmentation is very important for analyzing lung related diseases, and it can be applied to lung cancer segmentation and lung pattern classification for identifying other problems.
- the ADAM optimizer is used with a learning rate of 2 10 4 .
- the DI loss function according to Equation 20 was used. In this case, 10% of the samples were used for validation, with a batch size of 16 for 150 epochs.
- the presently-disclosed R2U-Net model showed 0.26 and 0.55 percent better testing accuracy compared to the equivalent SegNet and U-Net models respectively.
- the R2U-Net model provided 0.18 percent better accuracy against the ResU-Net model with the same number of network parameters.
- FIG. 56 is a diagram showing experimental results for lung segmentation.
- the first column shows the inputs
- the second column shows the ground truth
- the third column shows the outputs of SegNet
- the fourth column for the outputs of U-Net
- the fifth column for the outputs of R2U-Net. It can be visualized that the R2U-Net shows better segmentation results with internal details that are very similar to those displayed in the ground data. If one observes the input, ground truth, and output of the different approaches in the first and second rows, the outputs of the presently-disclosed approaches show better segmentation with more accurate internal details.
- the R2U-Net model clearly defines the inside hole in the left lung, whereas the SegNet and U-Net models do not capture this detail.
- the last row of images in FIG. 56 shows that the SegNet and U-Net models provide outputs that incorrectly capture parts of the image that are outside of the lesion.
- the R2U- Net model provides a much more accurate segmentation result.
- FIG. 56 is provided as heat maps which show the sharpness of the segmentation borders. These outputs show that ground truth tends to have a sharper boundary when compared to the model outputs.
- a split ratio 0.9 means that only 10 percent of the samples are used for training and remaining 90% of the samples are used for validation.
- a split ratio of 0.7 means that only 30% of the samples are used for training and the remaining 70% of the samples are used for validation.
- FIGS. 58A-58B are graphs showing examples of values for the performance of three different models (SegNet, U-Net and R2U-Net) for different numbers of training and validation samples.
- FIG. 58A shows the training DI coefficient errors (l-DI).
- FIG. 58B shows validation DI coefficient errors for five different trials.
- FIGS. 58A-58B show the training and validation DI coefficient errors (l-DI) with respect to the number of training and validation samples. In each trial, 150 epochs were considered, and the errors that presented are the average training and validation errors of the last twenty epochs.
- FIGS. 58A-58B show that the presently-disclosed R2U-Net model shows the lowest training and validation error for all of the tested split ratios, except for result where the split ratio is equal to 0.5 for the validation case.
- the error for the R2U-Net model is only slightly greater than that of the U-Net model.
- FIG. 59 is a diagram showing examples of testing errors of the R2U-Net, SegNet, and U-Net models for different split rations for the lung segmentation application. In each trial, the models were tested with the remaining 20% of the samples. The R2U-Net model shows the lowest error for almost all trials relative to the error obtained from the SegNet and U-Net models.
- the presently-disclosed models show better performance against SegNet when using both 1.07M and 13.34M network parameters, which is around 0.7M and 2.66M fewer when compared to SegNet.
- the model provides better performance with same or fewer number of network parameters compared to the SegNet, U-Net and ResU-Net model.
- the model possesses significant advantages in terms of memory and processing time.
- the computational time to segment per sample in testing phase is shown in Table 20 for all three datasets.
- the processing times during the testing phase for the STARE, CHASE DB, and DRIVE datasets were 6.42, 8.66, and 2.84 seconds per sample respectively. It can take around 90 seconds on average to segment an entire image (which is equivalent to a few thousand image patches).
- the presently-disclosed R2U-Net approach takes around 6 seconds per sample, which is an acceptable rate in a clinical use scenario.
- when executing skin cancer segmentation and lung segmentation entire images could be segments in 0.32 and 1.145 seconds respectively.
- Table 20 Computational time for processing an entire image during the testing phase
- the present disclosure includes an extension of the U-Net architecture using Recurrent Convolutional Neural Networks and Recurrent Residual Convolutional Neural Networks.
- the presently-disclosed models are called“RU-Net” and “R2U-Net” respectively. These models were evaluated using three different applications in the field of medical imaging including retina blood vessel segmentation, skin cancer lesion segmentation, and lung segmentation.
- the experimental results demonstrate that the presently- disclosed RU-Net and R2U-Net models show better performance in most of the cases for segmentation tasks with the same number of network parameters when compared to existing methods including the SegNet, U-Net, and residual U-Net (or ResU-Net) models on all three datasets.
- FIG 60 is a diagram showing examples of Recurrent Multilayer Perceptron (RMLP), Convolutional Neural Network (CNN), and Recurrent Convolutional Neural Network (RCNN) models.
- RMLP Recurrent Multilayer Perceptron
- CNN Convolutional Neural Network
- RCNN Recurrent Convolutional Neural Network
- the present disclosure provides a deep learning architecture which combines two most recently used models: a revised version of Inception network and the RCNN.
- the recurrent convolutional layers are incorporated within inception block, and the convolution operations are performed considering different time steps.
- FIG. 61 is a diagram showing an overall operational flow diagram of the presently-disclosed Inception Recurrent Convolutional Neural Network (IRCNN).
- the IRCNN includes an IRCNN block, a transition block, and a softmax layer.
- FIG. 62 is a diagram showing Inception-Recurrent Convolutional Neural Network (IRCNN) block with different convolutional layers with respect to different size of kernels.
- IRCNN Inception-Recurrent Convolutional Neural Network
- the presently-disclosed inception block with recurrent convolution layers is shown in FIG. 62.
- a goal of the DCNN architecture of the Inception and Residual networks is to implement large scale deep networks. As the model becomes larger and deeper, the computational parameters of the architecture are increased dramatically. Thus, the model becomes more complex to train and computationally expensive. In this scenario, the recurrent property of the present disclosure ensures better training and testing accuracy with less or equal computational parameters.
- a deep learning model called IRCNN is used with combination of the recently developed Inception-v4 and RCNN.
- the presently-disclosed architecture is based on several recently developed deep learning architectures, including Inception Nets and RCNNs. It tries to reduce the number of computational parameters, while providing better recognition accuracy.
- the IRCNN architecture consists of general convolution layers, IRCNN blocks, transition blocks, and a softmax layer at the end.
- the presently disclosure architecture provides recurrence in the Inception module, as shown in the IRCNN block in FIG. 62.
- a feature of Inceptionv4 is that it concatenates the outputs of multiple differently sized convolutional kernels in the inception block.
- Inception-v4 is a simplified version of Inception- v3, using lower rank filters and pooling layers.
- Inception-v4 however combines Residual concepts with Inception networks to improve the overall accuracy over Inception-v3.
- the outputs of inception layers are added with the inputs to the Inception Residual module.
- the present disclosure utilizes the inception concepts from Inception-v4.
- the IRCNN block performs recurrent convolution operations with different sized kernels (see FIG. 61).
- the inputs to the next time step are the sum of the convolutional outputs of the present time step and previous time steps. The same operations are repeated based on the number of time steps that are considered. As the input and output dimensions do not change, this is an accumulation of feature maps with respect to the time step are considered. This helps to strengthen the extraction of the target features.
- one of the path of inception block contains an average pooling operation is applied before the recurrent convolution layer.
- each Recurrent Convolution Layer (RCL) in the IRCNN block is similar to operations by others in the field. To describe these operations, consider a vectorized patch centered at (i,j) of an input sample on the k th feature map in the RCL unit.
- the 0[ /fe (t) refers the output of I th layer at time step t.
- the output can be expressed as:
- LRN Local Response Normalization
- the outputs of the IRCNN block with respect to the different kernel sizes l x l and 3 x 3, and average pooling operations followed by 1 c 1 are defined as z i xi (x), z 3x3 (x), and z ⁇ xl (x) respectively.
- the final output z out of the IRCNN-block can be expressed as:
- Q represents the concatenation operation with respect to the channel-axis on the output samples of inception layers.
- t 3 is used, which indicates that one forward and three recurrent convolutional layers have been used in each IRCNN-block (individual path) which is clearly demonstrated in FIG. 62.
- the outputs of the IRCNN-block becomes the inputs that are fed into the transition block.
- transition block three operations (convolution, pooling, and Dropout) are performed depending upon the placement of the block in the network. According to FIG. 61, all of the operations in the very first transition block have been applied. In the second transition block, only convolution with dropout operations has been used.
- the third transition block consists of convolution, global-average pooling, and drop-out layers.
- the global-average pooling layer is used as an alternative to a fully connected layer. There are several advantages of a global-average pooling layer. Firstly, it is very close in operation to convolution, hence enforcing correspondence between feature maps and categories. The feature maps can be easily interpreted as class confidence. Secondly, it does not need computational parameters, thus helping to avoid over-fitting of the network.
- Late use of the pooling layer is advantageous because it increases the number of non-linear hidden layers in the network. Therefore, only two special pooling layers have been applied in first and third transition block of this architecture. Special pooling is carried out with the max-pooling layer in this network (not all transition blocks have pooling layer).
- the max-pooling layers perform operations with a 3x3 patch and a 2 x 2 stride over the input samples. Since the non-overlapping max-pooling operation has a negative impact on model regularization, overlapped max-pooling for regularizing the network are used. This can facilitate training a deep network architecture. Eventually, a global-average pooling layer is used as an alternative of fully connected layers. Finally, a softmax logistic regression layer is used at the end of the IRCNN architecture.
- a model was used with four IRCNN blocks followed by transition layers, a fully connected layer, and softmax layer for the experiment on TinyImageNet-200 dataset.
- the number of feature maps in each of the forward convolution layers and RCLs in IRCNN blocks have almost doubled compared to the model that is used for CIFAR-100, which significantly increases number of network parameters to approximately 9.3M.
- EIN, and EIRN models are implemented with the same structure of IRCNN model with inception and inception-residual modules respectively.
- Batch Normalization BN is used instead of LRN in IRCNN, RCNN, EIN, and EIRN models.
- Equation 27 has been skipped, and the concatenation operation is performed directly on the output of Equation 26.
- BN is applied at the end of IRCNN block on z out .
- the impact of RCLs on the DenseNet model has been empirically investigated.
- An BN layer is used in the dense block with RCLs. Only 4 dense blocks have been used, with 4 layers in each block and a growth rate of 6. The experimental result shows significant improvement of training, validation, testing accuracies with DenseNet with RCLs against original DenseNet model.
- the first experiment trained the presently-disclosed IRCNN architecture using the stochastic gradient descent (SGD) optimization function with default initialization for deep networks found in Keras.
- the Nesterov momentum is set to 0.9 and decay to 9.99 c e ⁇ 7 .
- LSUV Layer- sequential unit-variance
- An improved version of the optimization function based on Adam known as EVE was also used.
- the value of the learning rate (T) is 1 x e— 4
- decay (g) is 1 x e— 4
- the (/? ! , /? 2 ) e [0,1) values are exponential decay rates for moment estimation in Adam.
- the b 3 o [0,1) is an exponential decay rate for computing relative changes.
- the k and K values are lower and upper thresholds for relative change, and e is a fuzzy factor.
- the h ⁇ norm was used with a value of 0.002 for weight regularization on each convolutional layer in the IRCNN block.
- MNIST is one of the most popular datasets for handwritten digits from 0-9, the dataset contains 28x28 pixel grayscale images with 60,000 training examples and 10,000 testing examples.
- the presently-disclosed model was trained with two IRCNN blocks (IRCNN-block 1 and IRCNN-block 2) and the ReLU activation function was used. The model was trained with 60,000 samples and 10,000 samples were used for the validation set. Eventually the trained network was tested with 10,000 testing examples. A test error of 0.32% was attained with the IRCNN and the SGD, and about 0.29% error are achieved for the IRCNN when initializing with LSUV and the EVE optimization function. The summary of the classification accuracies is given in Table 21.
- Table 21 Testing errors (%) on MNIST, CIFAR- 10(00), CIFAR- 100(000), and SVHN
- CIFAR-10 is an object classification benchmark consisting of 32 c 32 color images representing 10 classes. It is split into 50,000 samples for training and 10,000 samples for testing. The experiment was conducted with and without data augmentation. The entire experiment was conducted on models similar to the one shown in FIG. 61. Using the presently- disclosed approach, about 8.41% error was achieved without data augmentation, and 7.37% error was achieved with data augmentation using the SGD technique. These results are better than those of most of the recognized DCNN models stated in Table 21. Better performance is observed from the IRCNN with LSUV as the initialization approach and EVE as the optimization technique. The results show around 8.17% and 7.11% error without and with data augmentation respectively.
- FIG. 63 is a graph showing example values for training and validation loss.
- the training and validation loss apply to the experiment on CIFAR-l 0 of model.
- FIG. 64 is a graph showing examples of values for training and validation accuracy of IRCNN with SGD and LSUV+EVE.
- FIG. 65 is a graph showing examples of values for the training and validation loss of the IRCNN for both experiments using the CIFAR-l 00 dataset and data augmentation (with and without initialization and optimization).
- the IRCNN was used with a LSUV initialization approach and the EVE optimization function.
- the default initialization approach of Keras and the SGD optimization method are used in the second experiment. It is clearly shown that the presently-disclosed model has lower error rates in the both experiments, showing the effectiveness of the presently-disclosed IRCNN learning model.
- FIG. 66 is a graph showing example values for the training and testing accuracy of the IRCNN with LSUV and EVE.
- SVHN Street View House Numbers
- This dataset contains color images representing house numbers from Google Street View. This experiment considered the second version, which consists with 32 c 32 color examples. There are 73,257 samples are in the training set and 26,032 samples in testing set. In addition, this dataset has 531,131 extra samples that are used for training purposes. As single input samples of this dataset contain multiple digits, the main goal is to classify the central digit. Due to the huge variation of color and brightness, this dataset is much for difficult to classify compared to the MNIST dataset. In this case, experimentation occurred with the same model as is used in CIFAR-10 and CIFAR-100.
- the presently-disclosed architecture also performs well when compared to other recently proposed optimized architectures.
- An DCNN architecture is called FitNet4 has conducted experiment with LSUV initialization approach, and it only achieved 70.04% classification accuracy with data augmentation using mirroring and random shifts for CIFAR- 100.
- random horizontal flipping was only applied for data augmentation in this implementation and achieved about 1.72% better recognition accuracy against FitNet4.
- the Inception network has have implemented with the same number of layers and parameters as in the transition and Inception-block.
- FIGS. 67 and 68 are graphs showing the model loss and accuracy for both training and validation phases, respectively. From both figures, it can be clearly observed that the presently-disclosed model shows lower loss and the highest recognition accuracy during validation phase compared with EIN and EIRN, proving the effectiveness of the presently- disclosed model. It also demonstrates the advantage of recurrent layers in Inception networks.
- FIG. 69 is a graph showing example values for the testing accuracy of IRCNN, EIN, and EIRN on CIFAR-100 dataset. It can be summarized that the presently-disclosed model of IRCNN shows around 3.47% and 2.54% better testing accuracy compared to EIN and EIRN respectively.
- FIG. 70 is a diagram showing examples of images.
- FIG. 71 is a graph showing example values for validation accuracy of IRCNN, EIRN, EIN, and RCNN. The impact of transfer learning is observed based on FIG. 71.
- FIG. 72 is a graph showing example values for validation accuracy of DenseNet and DenseNet with a Recurrent Convolutional Layer (RCL).
- Table 22 shows the testing accuracy for all the models including RCNN and DenseNet.
- the IRCNN provides better performance compared to EIN, EIRN, and RCNN with almost same number of parameters for object recognition task on the TinyImageNet-200 dataset.
- Experiments were also conducted with DenseNet and DenseNet with RCL on the TinyImageNet-200 dataset.
- the experimental results show that DenseNet with RCLs provides about 0.38% improvement on Top-l% accuracy compared to DenseNet with only 1M net- work parameters.
- the experimental results show DenseNet with RCLs provides higher testing accuracy in both Top-l% and Top-5% compared against DenseNet model.
- Implementations of the subject matter and the functional operations described in this specification can be implemented in digital electronic circuitry, in tangibly embodied computer software or firmware, in computer hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them.
- Software implementations of the described subject matter can be implemented as one or more computer programs, that is, one or more modules of computer program instructions encoded on a tangible, non-transitory, computer-readable computer-storage medium for execution by, or to control the operation of, data processing apparatus.
- the program instructions can be encoded in/on an artificially generated propagated signal, for example, a machine-generated electrical, optical, or electromagnetic signal that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus.
- the computer-storage medium can be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of computer-storage mediums.
- Configuring one or more computers means that the one or more computers have installed hardware, firmware, or software (or combinations of hardware, firmware, and software) so that when the software is executed by the one or more computers, particular computing operations are performed.
- the term“real-time,”“real time,”“realtime,”“real (fast) time (RET),”“near(ly) real-time (NRT),”“quasi real-time,” or similar terms means that an action and a response are temporally proximate such that an individual perceives the action and the response occurring substantially simultaneously.
- the time difference for a response to display (or for an initiation of a display) of data following the individual’s action to access the data can be less than 1 ms, less than 1 sec., or less than 5 secs.
- the terms“data processing apparatus,”“computer,” or“electronic computer device” refer to data processing hardware and encompass all kinds of apparatus, devices, and machines for processing data, including by way of example, a programmable processor, a computer, or multiple processors or computers.
- the apparatus can also be, or further include special purpose logic circuitry, for example, a central processing unit (CPU), an FPGA (field programmable gate array), or an ASIC (application-specific integrated circuit).
- the data processing apparatus or special purpose logic circuitry (or a combination of the data processing apparatus or special purpose logic circuitry) can be hardware- or software-based (or a combination of both hardware- and software-based).
- the apparatus can optionally include code that creates an execution environment for computer programs, for example, code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of execution environments.
- code that constitutes processor firmware for example, code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of execution environments.
- the present disclosure contemplates the use of data processing apparatuses with or without conventional operating systems, for example LINUX, UNIX, WINDOWS, MAC OS, ANDROID, IOS, or any other suitable conventional operating system.
- a computer program which can also be referred to or described as a program, software, a software application, a module, a software module, a script, or code can be written in any form of programming language, including compiled or interpreted languages, or declarative or procedural languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment.
- a computer program can, but need not, correspond to a file in a file system.
- a program can be stored in a portion of a file that holds other programs or data, for example, one or more scripts stored in a markup language document, in a single file dedicated to the program in question, or in multiple coordinated files, for example, files that store one or more modules, sub-programs, or portions of code.
- a computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
- the methods, processes, or logic flows described in this specification can be performed by one or more programmable computers executing one or more computer programs to perform functions by operating on input data and generating output.
- the methods, processes, or logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, for example, a CPU, an FPGA, or an ASIC.
- Computers suitable for the execution of a computer program can be based on general or special purpose microprocessors, both, or any other kind of CPU.
- a CPU will receive instructions and data from and write to a memory.
- the essential elements of a computer are a CPU, for performing or executing instructions, and one or more memory devices for storing instructions and data.
- a computer will also include, or be operatively coupled to, receive data from or transfer data to, or both, one or more mass storage devices for storing data, for example, magnetic, magneto-optical disks, or optical disks.
- mass storage devices for storing data, for example, magnetic, magneto-optical disks, or optical disks.
- a computer need not have such devices.
- a computer can be embedded in another device, for example, a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a global positioning system (GPS) receiver, or a portable storage device, for example, a universal serial bus (USB) flash drive, to name just a few.
- PDA personal digital assistant
- GPS global positioning system
- USB universal serial bus
- Computer-readable media suitable for storing computer program instructions and data includes all forms of permanent/non permanent or volatile/non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, for example, random access memory (RAM), read-only memory (ROM), phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), erasable programmable read-only memory (EPROM), electrically erasable programmable read-only memory (EEPROM), and flash memory devices; magnetic devices, for example, tape, cartridges, cassettes, intemal/removable disks; magneto-optical disks; and optical memory devices, for example, digital video disc (DVD), CD-ROM, DVD+/-R, DVD-RAM, DVD-ROM, HD-DVD, and BLURAY, and other optical memory technologies.
- RAM random access memory
- ROM read-only memory
- PRAM phase change memory
- SRAM static random access memory
- DRAM dynamic random access memory
- EPROM erasable programmable read
- the memory can store various objects or data, including caches, classes, frameworks, applications, modules, backup data, jobs, web pages, web page templates, data structures, database tables, repositories storing dynamic information, and any other appropriate information including any parameters, variables, algorithms, instructions, rules, constraints, or references thereto. Additionally, the memory can include any other appropriate data, such as logs, policies, security or access data, reporting files, as well as others.
- the processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
- implementations of the subject matter described in this specification can be implemented on a computer having a display device, for example, a CRT (cathode ray tube), LCD (liquid crystal display), LED (Light Emitting Diode), or plasma monitor, for displaying information to the user and a keyboard and a pointing device, for example, a mouse, trackball, or trackpad by which the user can provide input to the computer.
- a display device for example, a CRT (cathode ray tube), LCD (liquid crystal display), LED (Light Emitting Diode), or plasma monitor
- a keyboard and a pointing device for example, a mouse, trackball, or trackpad by which the user can provide input to the computer.
- Input can also be provided to the computer using a touchscreen, such as a tablet computer surface with pressure sensitivity, a multi-touch screen using capacitive or electric sensing, or other type of touchscreen.
- a computer can interact with a user by sending documents to and receiving documents from a device that is used by the user; for example, by sending web pages to a web browser on a user’s client device in response to requests received from the web browser.
- GUI graphical user interface
- a GUI can be used in the singular or the plural to describe one or more graphical user interfaces and each of the displays of a particular graphical user interface. Therefore, a GUI can represent any graphical user interface, including but not limited to, a web browser, a touch screen, or a command line interface (CLI) that processes information and efficiently presents the information results to the user.
- a GUI can include a plurality of user interface (UI) elements, some or all associated with a web browser, such as interactive fields, pull-down lists, and butons. These and other UI elements can be related to or represent the functions of the web browser.
- UI user interface
- Implementations of the subject mater described in this specification can be implemented in a computing system that includes a back-end component, for example, as a data server, or that includes a middleware component, for example, an application server, or that includes a front-end component, for example, a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back-end, middleware, or front-end components.
- the components of the system can be interconnected by any form or medium of wireline or wireless digital data communication (or a combination of data communication), for example, a communication network.
- Examples of communication networks include a local area network (LAN), a radio access network (RAN), a metropolitan area network (MAN), a wide area network (WAN), Worldwide Interoperability for Microwave Access (WIMAX), a wireless local area network (WLAN) using, for example, 802.11 a/b/g/n or 802.20 (or a combination of 802.1 lx and 802.20 or other protocols consistent with this disclosure), all or a portion of the Internet, or any other communication system or systems at one or more locations (or a combination of communication networks).
- the network can communicate with, for example, Internet Protocol (IP) packets, Frame Relay frames, Asynchronous Transfer Mode (ATM) cells, voice, video, data, or other suitable information (or a combination of communication types) between network addresses.
- IP Internet Protocol
- ATM Asynchronous Transfer Mode
- the computing system can include clients and servers.
- a client and server are generally remote from each other and typically interact through a communication network.
- the relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
- any claimed implementation is considered to be applicable to at least a computer-implemented method; a non-transitory, computer-readable medium storing computer-readable instructions to perform the computer-implemented method; and a computer system comprising a computer memory interoperably coupled with a hardware processor configured to perform the computer-implemented method or the instructions stored on the non- transitory, computer-readable medium.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Evolutionary Computation (AREA)
- Public Health (AREA)
- Medical Informatics (AREA)
- Epidemiology (AREA)
- Primary Health Care (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Databases & Information Systems (AREA)
- Pathology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201862755097P | 2018-11-02 | 2018-11-02 | |
US62/755,097 | 2018-11-02 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020093042A1 true WO2020093042A1 (en) | 2020-05-07 |
Family
ID=70463392
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2019/059653 WO2020093042A1 (en) | 2018-11-02 | 2019-11-04 | Neural networks for biomedical image analysis |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2020093042A1 (en) |
Cited By (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111739028A (en) * | 2020-05-26 | 2020-10-02 | 华南理工大学 | Nail region image acquisition method, system, computing device and storage medium |
CN111881718A (en) * | 2020-06-08 | 2020-11-03 | 东北农业大学 | Yolo v3 target detection network-based cow posture identification method |
CN111931802A (en) * | 2020-06-16 | 2020-11-13 | 南京信息工程大学 | Pedestrian re-identification method based on fusion of middle-layer features of Simese network structure |
CN111951292A (en) * | 2020-09-01 | 2020-11-17 | 吉林大学 | Object surface reflection attribute extraction method, device, equipment and storage medium |
CN111951221A (en) * | 2020-07-13 | 2020-11-17 | 清影医疗科技(深圳)有限公司 | Glomerular cell image identification method based on deep neural network |
CN112053285A (en) * | 2020-09-01 | 2020-12-08 | 平安科技(深圳)有限公司 | Image processing method, image processing device, computer equipment and storage medium |
CN112054979A (en) * | 2020-09-14 | 2020-12-08 | 四川大学 | Radio automatic modulation identification method based on fuzzy dense convolution network |
CN112053363A (en) * | 2020-08-19 | 2020-12-08 | 苏州超云生命智能产业研究院有限公司 | Retinal vessel segmentation method and device and model construction method |
CN112070722A (en) * | 2020-08-14 | 2020-12-11 | 厦门骁科码生物科技有限公司 | Fluorescence in situ hybridization cell nucleus segmentation method and system |
CN112233117A (en) * | 2020-12-14 | 2021-01-15 | 浙江卡易智慧医疗科技有限公司 | New coronary pneumonia CT detects discernment positioning system and computing equipment |
CN112287998A (en) * | 2020-10-27 | 2021-01-29 | 佛山市南海区广工大数控装备协同创新研究院 | Method for detecting target under low-light condition |
CN112330682A (en) * | 2020-11-09 | 2021-02-05 | 重庆邮电大学 | Industrial CT image segmentation method based on deep convolutional neural network |
CN112541532A (en) * | 2020-12-07 | 2021-03-23 | 长沙理工大学 | Target detection method based on dense connection structure |
CN112561868A (en) * | 2020-12-09 | 2021-03-26 | 深圳大学 | Cerebrovascular segmentation method based on multi-view cascade deep learning network |
CN112767355A (en) * | 2021-01-20 | 2021-05-07 | 北京小白世纪网络科技有限公司 | Method and device for constructing thyroid nodule Tirads grading automatic identification model |
CN112990041A (en) * | 2021-03-25 | 2021-06-18 | 武汉大学 | Remote sensing image building extraction method based on improved U-net |
CN113065551A (en) * | 2021-03-19 | 2021-07-02 | 杭州迪英加科技有限公司 | Method for performing image segmentation using a deep neural network model |
CN113223002A (en) * | 2021-05-07 | 2021-08-06 | 西安智诊智能科技有限公司 | Blood vessel image segmentation method |
CN113240691A (en) * | 2021-06-10 | 2021-08-10 | 南京邮电大学 | Medical image segmentation method based on U-shaped network |
CN113299374A (en) * | 2021-06-03 | 2021-08-24 | 广东财经大学 | Thyroid nodule ultrasonic image automatic segmentation system based on deep learning |
CN113298826A (en) * | 2021-06-09 | 2021-08-24 | 东北大学 | Image segmentation method based on LA-Net network |
CN113516053A (en) * | 2021-05-28 | 2021-10-19 | 西安空间无线电技术研究所 | Ship target refined detection method with rotation invariance |
CN113671031A (en) * | 2021-08-20 | 2021-11-19 | 北京房江湖科技有限公司 | Wall hollowing detection method and device |
CN113762349A (en) * | 2021-08-11 | 2021-12-07 | 同济大学 | Lightweight aliasing dense network classification method and system for marine organisms |
US20210390338A1 (en) * | 2020-06-15 | 2021-12-16 | Dalian University Of Technology | Deep network lung texture recogniton method combined with multi-scale attention |
CN113837193A (en) * | 2021-09-23 | 2021-12-24 | 中南大学 | Zinc flotation froth image segmentation algorithm based on improved U-Net network |
CN114022745A (en) * | 2021-11-05 | 2022-02-08 | 光大科技有限公司 | Neural network model training method and device |
CN115455838A (en) * | 2022-09-26 | 2022-12-09 | 大连海事大学 | Time-course data-oriented high spatial resolution flow field reconstruction method |
CN115512368A (en) * | 2022-08-22 | 2022-12-23 | 华中农业大学 | Cross-modal semantic image generation model and method |
CN117351003A (en) * | 2023-11-24 | 2024-01-05 | 华南理工大学 | Multi-model integrated multi-phase MRI tumor classification method based on video actions |
CN117935253A (en) * | 2024-01-23 | 2024-04-26 | 北京邮电大学 | Method for measuring quantitative index of kidney pathology and correlation between measured index and clinical phenotype |
CN118429782A (en) * | 2024-07-05 | 2024-08-02 | 吉林大学 | Apatite fission track artificial intelligent identification method based on deep learning |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180089834A1 (en) * | 2016-09-29 | 2018-03-29 | Magic Leap, Inc. | Neural network for eye image segmentation and image quality estimation |
US20180218497A1 (en) * | 2017-01-27 | 2018-08-02 | Arterys Inc. | Automated segmentation utilizing fully convolutional networks |
-
2019
- 2019-11-04 WO PCT/US2019/059653 patent/WO2020093042A1/en active Application Filing
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180089834A1 (en) * | 2016-09-29 | 2018-03-29 | Magic Leap, Inc. | Neural network for eye image segmentation and image quality estimation |
US20180218497A1 (en) * | 2017-01-27 | 2018-08-02 | Arterys Inc. | Automated segmentation utilizing fully convolutional networks |
Non-Patent Citations (2)
Title |
---|
ALOM ET AL.: "Improved Inception-Residual Convolutional Neural Network for Object Recognition", ARXIV.ORG; 1712.09888, 28 December 2017 (2017-12-28), pages 1 - 18, XP055705977, Retrieved from the Internet <URL:https://arxiv.org/ftp/arxiv/papers/1712/1712.09888.pdf> [retrieved on 20191227] * |
ALOM ET AL.: "Recurrent Residual Convolutional Neural Network based on U-Net (R2U-Net) for Medical Image Segmentation", ARXIV.ORG; 1802.06955, 29 May 2018 (2018-05-29), pages 1 - 12, XP055705984, Retrieved from the Internet <URL:https://arxiv.org/ftp/arxiv/papers/1802/1802.06955.pdf> [retrieved on 20191226] * |
Cited By (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111739028A (en) * | 2020-05-26 | 2020-10-02 | 华南理工大学 | Nail region image acquisition method, system, computing device and storage medium |
CN111881718A (en) * | 2020-06-08 | 2020-11-03 | 东北农业大学 | Yolo v3 target detection network-based cow posture identification method |
US20210390338A1 (en) * | 2020-06-15 | 2021-12-16 | Dalian University Of Technology | Deep network lung texture recogniton method combined with multi-scale attention |
US11551029B2 (en) * | 2020-06-15 | 2023-01-10 | Dalian University Of Technology | Deep network lung texture recogniton method combined with multi-scale attention |
CN111931802A (en) * | 2020-06-16 | 2020-11-13 | 南京信息工程大学 | Pedestrian re-identification method based on fusion of middle-layer features of Simese network structure |
CN111951221A (en) * | 2020-07-13 | 2020-11-17 | 清影医疗科技(深圳)有限公司 | Glomerular cell image identification method based on deep neural network |
CN111951221B (en) * | 2020-07-13 | 2023-10-31 | 清影医疗科技(深圳)有限公司 | Glomerular cell image recognition method based on deep neural network |
CN112070722A (en) * | 2020-08-14 | 2020-12-11 | 厦门骁科码生物科技有限公司 | Fluorescence in situ hybridization cell nucleus segmentation method and system |
CN112053363B (en) * | 2020-08-19 | 2023-12-15 | 苏州超云生命智能产业研究院有限公司 | Retina blood vessel segmentation method, retina blood vessel segmentation device and model construction method |
CN112053363A (en) * | 2020-08-19 | 2020-12-08 | 苏州超云生命智能产业研究院有限公司 | Retinal vessel segmentation method and device and model construction method |
CN112053285A (en) * | 2020-09-01 | 2020-12-08 | 平安科技(深圳)有限公司 | Image processing method, image processing device, computer equipment and storage medium |
CN112053285B (en) * | 2020-09-01 | 2024-05-31 | 平安科技(深圳)有限公司 | Image processing method, device, computer equipment and storage medium |
CN111951292A (en) * | 2020-09-01 | 2020-11-17 | 吉林大学 | Object surface reflection attribute extraction method, device, equipment and storage medium |
CN112054979A (en) * | 2020-09-14 | 2020-12-08 | 四川大学 | Radio automatic modulation identification method based on fuzzy dense convolution network |
CN112054979B (en) * | 2020-09-14 | 2022-02-25 | 四川大学 | Radio automatic modulation identification method based on fuzzy dense convolution network |
CN112287998A (en) * | 2020-10-27 | 2021-01-29 | 佛山市南海区广工大数控装备协同创新研究院 | Method for detecting target under low-light condition |
CN112330682A (en) * | 2020-11-09 | 2021-02-05 | 重庆邮电大学 | Industrial CT image segmentation method based on deep convolutional neural network |
CN112330682B (en) * | 2020-11-09 | 2022-08-26 | 重庆邮电大学 | Industrial CT image segmentation method based on deep convolutional neural network |
CN112541532A (en) * | 2020-12-07 | 2021-03-23 | 长沙理工大学 | Target detection method based on dense connection structure |
CN112541532B (en) * | 2020-12-07 | 2024-03-01 | 长沙理工大学 | Target detection method based on dense connection structure |
CN112561868B (en) * | 2020-12-09 | 2021-12-07 | 深圳大学 | Cerebrovascular segmentation method based on multi-view cascade deep learning network |
CN112561868A (en) * | 2020-12-09 | 2021-03-26 | 深圳大学 | Cerebrovascular segmentation method based on multi-view cascade deep learning network |
CN112233117A (en) * | 2020-12-14 | 2021-01-15 | 浙江卡易智慧医疗科技有限公司 | New coronary pneumonia CT detects discernment positioning system and computing equipment |
CN112767355A (en) * | 2021-01-20 | 2021-05-07 | 北京小白世纪网络科技有限公司 | Method and device for constructing thyroid nodule Tirads grading automatic identification model |
CN113065551B (en) * | 2021-03-19 | 2023-08-08 | 杭州迪英加科技有限公司 | Method for performing image segmentation using deep neural network model |
CN113065551A (en) * | 2021-03-19 | 2021-07-02 | 杭州迪英加科技有限公司 | Method for performing image segmentation using a deep neural network model |
CN112990041A (en) * | 2021-03-25 | 2021-06-18 | 武汉大学 | Remote sensing image building extraction method based on improved U-net |
CN113223002A (en) * | 2021-05-07 | 2021-08-06 | 西安智诊智能科技有限公司 | Blood vessel image segmentation method |
CN113516053B (en) * | 2021-05-28 | 2024-05-14 | 西安空间无线电技术研究所 | Ship target refinement detection method with rotation invariance |
CN113516053A (en) * | 2021-05-28 | 2021-10-19 | 西安空间无线电技术研究所 | Ship target refined detection method with rotation invariance |
CN113299374B (en) * | 2021-06-03 | 2023-08-29 | 广东财经大学 | Thyroid nodule ultrasonic image automatic segmentation system based on deep learning |
CN113299374A (en) * | 2021-06-03 | 2021-08-24 | 广东财经大学 | Thyroid nodule ultrasonic image automatic segmentation system based on deep learning |
CN113298826A (en) * | 2021-06-09 | 2021-08-24 | 东北大学 | Image segmentation method based on LA-Net network |
CN113298826B (en) * | 2021-06-09 | 2023-11-14 | 东北大学 | Image segmentation method based on LA-Net network |
CN113240691A (en) * | 2021-06-10 | 2021-08-10 | 南京邮电大学 | Medical image segmentation method based on U-shaped network |
WO2022257408A1 (en) * | 2021-06-10 | 2022-12-15 | 南京邮电大学 | Medical image segmentation method based on u-shaped network |
CN113762349B (en) * | 2021-08-11 | 2024-03-29 | 同济大学 | Marine organism-oriented lightweight aliasing dense network classification method and system |
CN113762349A (en) * | 2021-08-11 | 2021-12-07 | 同济大学 | Lightweight aliasing dense network classification method and system for marine organisms |
CN113671031A (en) * | 2021-08-20 | 2021-11-19 | 北京房江湖科技有限公司 | Wall hollowing detection method and device |
CN113837193A (en) * | 2021-09-23 | 2021-12-24 | 中南大学 | Zinc flotation froth image segmentation algorithm based on improved U-Net network |
CN113837193B (en) * | 2021-09-23 | 2023-09-01 | 中南大学 | Zinc flotation froth image segmentation method based on improved U-Net network |
CN114022745A (en) * | 2021-11-05 | 2022-02-08 | 光大科技有限公司 | Neural network model training method and device |
CN115512368A (en) * | 2022-08-22 | 2022-12-23 | 华中农业大学 | Cross-modal semantic image generation model and method |
CN115512368B (en) * | 2022-08-22 | 2024-05-10 | 华中农业大学 | Cross-modal semantic generation image model and method |
CN115455838A (en) * | 2022-09-26 | 2022-12-09 | 大连海事大学 | Time-course data-oriented high spatial resolution flow field reconstruction method |
CN115455838B (en) * | 2022-09-26 | 2023-09-01 | 大连海事大学 | High-spatial-resolution flow field reconstruction method for time-course data |
CN117351003B (en) * | 2023-11-24 | 2024-02-23 | 华南理工大学 | Multi-model integrated multi-phase MRI tumor classification method based on video actions |
CN117351003A (en) * | 2023-11-24 | 2024-01-05 | 华南理工大学 | Multi-model integrated multi-phase MRI tumor classification method based on video actions |
CN117935253A (en) * | 2024-01-23 | 2024-04-26 | 北京邮电大学 | Method for measuring quantitative index of kidney pathology and correlation between measured index and clinical phenotype |
CN118429782A (en) * | 2024-07-05 | 2024-08-02 | 吉林大学 | Apatite fission track artificial intelligent identification method based on deep learning |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020093042A1 (en) | Neural networks for biomedical image analysis | |
Zhang et al. | Modified U-Net for plant diseased leaf image segmentation | |
Rezaoana et al. | Detection and classification of skin cancer by using a parallel CNN model | |
Al-Bander et al. | Multiscale sequential convolutional neural networks for simultaneous detection of fovea and optic disc | |
Halder et al. | Adaptive morphology aided 2-pathway convolutional neural network for lung nodule classification | |
Korbar et al. | Looking under the hood: Deep neural network visualization to interpret whole-slide image analysis outcomes for colorectal polyps | |
Wang et al. | Frnet: an end-to-end feature refinement neural network for medical image segmentation | |
Prusty et al. | A novel transfer learning technique for detecting breast cancer mammograms using VGG16 bottleneck feature | |
Zhao et al. | BASCNet: Bilateral adaptive spatial and channel attention network for breast density classification in the mammogram | |
EP3896649A1 (en) | Medical image synthesis of abnormality patterns associated with covid-19 | |
Prasad et al. | Lung cancer detection and classification using deep neural network based on hybrid metaheuristic algorithm | |
Das et al. | Inception-based deep learning architecture for tuberculosis screening using chest X-rays | |
El-Assiouti et al. | RegionInpaint, Cutoff and RegionMix: Introducing Novel Augmentation Techniques for Enhancing the Generalization of Brain Tumor Identification. | |
de Araújo et al. | Automated detection of segmental glomerulosclerosis in kidney histopathology | |
Durai et al. | Automatic severity grade classification of diabetic retinopathy using deformable ladder Bi attention U-net and deep adaptive CNN | |
Ren | Investigation of convolutional neural network architectures for image-based feature learning and classification | |
Hossain et al. | Recognition of tuberculosis on medical X-ray images utilizing MobileNet transfer learning | |
Kim et al. | Prediction of locations in medical images using orthogonal neural networks | |
Franco-Barranco et al. | Current Progress and Challenges in Large-Scale 3D Mitochondria Instance Segmentation | |
Santos et al. | A Method Based on Deep Neural Network for Instance Segmentation of Retinal Lesions Caused by Diabetic Retinopathy | |
Bou | Deep Learning models for semantic segmentation of mammography screenings | |
Asaad et al. | An integrative framework for brain tumor segmentation and classification using neuraclassnet | |
Jones | Developing Novel Computer Aided Diagnosis Schemes for Improved Classification of Mammography Detected Masses | |
Hu et al. | Pulmonary nodules recognition based on parallel cross-convolution | |
Mudduluru | Developing and Applying Hybrid Deep Learning Models for Computer-Aided Diagnosis of Medical Image Data |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19877914 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19877914 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19877914 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 081121) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19877914 Country of ref document: EP Kind code of ref document: A1 |