[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US12136164B2 - Method and system for rule based display of sets of images using image content derived parameters - Google Patents

Method and system for rule based display of sets of images using image content derived parameters Download PDF

Info

Publication number
US12136164B2
US12136164B2 US18/229,101 US202318229101A US12136164B2 US 12136164 B2 US12136164 B2 US 12136164B2 US 202318229101 A US202318229101 A US 202318229101A US 12136164 B2 US12136164 B2 US 12136164B2
Authority
US
United States
Prior art keywords
dicom
study
primary
studies
parametera
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US18/229,101
Other versions
US20230386125A1 (en
Inventor
Malte Westerhoff
Detlev Stalling
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
PME IP Pty Ltd
Original Assignee
PME IP Pty Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/831,975 external-priority patent/US8976190B1/en
Priority claimed from US16/052,571 external-priority patent/US10540803B2/en
Application filed by PME IP Pty Ltd filed Critical PME IP Pty Ltd
Priority to US18/229,101 priority Critical patent/US12136164B2/en
Publication of US20230386125A1 publication Critical patent/US20230386125A1/en
Application granted granted Critical
Publication of US12136164B2 publication Critical patent/US12136164B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/08Volume rendering
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/05Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves 
    • A61B5/055Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves  involving electronic [EMR] or nuclear [NMR] magnetic resonance, e.g. magnetic resonance imaging
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/7475User input or interface means, e.g. keyboard, pointing device, joystick
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/02Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
    • A61B6/03Computed tomography [CT]
    • A61B6/032Transmission computed tomography [CT]
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/02Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
    • A61B6/03Computed tomography [CT]
    • A61B6/037Emission tomography
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/46Arrangements for interfacing with the operator or the patient
    • A61B6/461Displaying means of special interest
    • A61B6/465Displaying means of special interest adapted to display user selection data, e.g. graphical user interface, icons or menus
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/46Arrangements for interfacing with the operator or the patient
    • A61B6/461Displaying means of special interest
    • A61B6/466Displaying means of special interest adapted to display 3D data
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • A61B6/5205Devices using data or image processing specially adapted for radiation diagnosis involving processing of raw data to produce diagnostic data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • G06F3/147Digital output to display device ; Cooperation and interconnection of the display device with other functional units using display panels
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/001Texturing; Colouring; Generation of texture or colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/60Editing figures and text; Combining figures or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/36Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
    • G09G5/363Graphics controllers
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/20ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/40ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/70ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/08Detecting, measuring or recording devices for evaluating the respiratory organs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/04Indexing scheme for image data processing or generation, in general involving 3D image data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/24Indexing scheme for image data processing or generation, in general involving graphical user interfaces [GUIs]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/41Medical
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2219/00Indexing scheme for manipulating 3D models or images for computer graphics
    • G06T2219/20Indexing scheme for editing of 3D models
    • G06T2219/2012Colour editing, changing, or manipulating; Use of colour codes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2219/00Indexing scheme for manipulating 3D models or images for computer graphics
    • G06T2219/20Indexing scheme for editing of 3D models
    • G06T2219/2024Style variation
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2340/00Aspects of display data processing
    • G09G2340/12Overlay of images, i.e. displayed pixel being the result of switching between the corresponding input pixels
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2380/00Specific applications
    • G09G2380/08Biomedical applications

Definitions

  • the images printed on films would be ‘hung’ in front of a light box.
  • the ‘hanging’ would often follow a specific protocol.
  • a particular organization or doctor may choose for a two-view chest X-Ray with a two-view prior exam, that the films be hung from left to right as follows: Frontal view of current examination, lateral view of current examination, frontal view of prior examination, lateral view of prior examination.
  • the doctor may hang mammography exams with the corresponding views of current and prior next to each other, if that was more appropriate for the diagnostic workflow in that case.
  • the organization or doctor developed a traditional ‘Hanging Protocol’.
  • the film and the light box are often being replaced by computer systems, called PACS (Picture Archiving and Communication System). PACS systems can mimic the Hanging Protocols.
  • CT Computer Tomography
  • MRI Magnetic Resonance Imaging
  • PET Positron Emission Tomography
  • the invention pertains to digital data processing and, more particularly, by way of example, to the visualization of image data.
  • Three dimensional (3D) and four dimensional (4D) image data is routinely acquired with CT, MRI, PET, confocal microscopes, 3D ultrasound devices, and other imaging devices.
  • the medical imaging market is just one example of a market that uses these devices.
  • the visualization of image data market is growing rapidly, with new CT scanners collecting larger amounts of data more quickly than previous generation CT scanners.
  • the invention has application to areas including medical imaging, atmospheric studies, astrophysics and geophysics.
  • Embodiments of this invention therefore provide a computer-based analytic framework whereby image-based information from a variety of different sources can be integrated to provide increased ability to display relevant information, e.g., to display information for a physician to diagnose and evaluate a patient's condition.
  • a physician can incur increased likelihood of confusion of image-based information from different patients.
  • a diagnostician e.g., physician
  • Such inadvertent conflation can produce misdiagnosis or mistaken non-diagnosis.
  • the outcome can be serious, e.g., misdiagnoses of a patient can result in increased chance of morbidity and/or mortality.
  • a First Study is first selected for review by a physician or diagnostician. Selection of a Study will generally be based on some particular characteristic. Such characteristic can be anatomical, disease-based, or both. Once a First Study is selected, an Additional Candidate Study can be selected based on the anatomical location of the First Study. Therefore, if the First Study is a Chest X-Ray, an Additional Candidate Study can be a Chest CT scan, MRI, positron-emission tomography (PET) scan, or other image of the chest.
  • PET positron-emission tomography
  • an Additional Candidate Study could be a series of X-Ray images taken after ingestion of a contrast agent (such as barium). It can be appreciated that such anatomically selected Additional Candidate Studies can be applied to any organ, organ system, or tissue.
  • Additional Candidate Studies can be selected based on the type of disorder or disease being evaluated. For example, in a case in which a patient has had a diagnosis of cancer of one organ (e.g., lung), it can be desirable for Additional Candidate Studies to be targeted to identification of metastases in another organ.
  • a First Study is a Chest X-Ray
  • an Additional Candidate Study can be of the lymphatic system, head and neck, or various abdominal quadrants.
  • Such Additional Candidate Studies may be X-ray, CT scans, MRI scans, PET scans, vascular visualizations (e.g., with injected contrast media) or histological images taken during a biopsy. Because the degree of detail (i.e., “granularity”) obtained using different imaging techniques may vary widely it can be desirable to have a Rule Based process whereby the granularity of an Additional Candidate Study is increased over that of the First Study.
  • a Chest X-Ray is a two-dimensional image in which the entirety of the chest and lungs is represented as a flat image.
  • An Additional Candidate Study could be a CT scan, where “2-dimensional” images are acquired at a series of different “depths” (e.g., “slices”) through the organ. If the 2-dimensional images are of sufficient quality to produce a 3-dimensional image of the organ with desirable degree of granularity, then the Additional Candidate Study can be depicted and displayed along with the image of the First Study.
  • General Rule 2 for selecting an Additional Candidate Study therefore can be:
  • a method or system uses a rule derived basis to display image sets.
  • the selection of the images to be displayed, the layout of the images, i.e., the hanging, as well as the rendering parameters and styles can be determined using a rule derived basis.
  • the user is presented with images displayed based on their preferences without having to first manually adjust parameters. Accordingly, there is a time saving in not displaying images initially in a non-rule derived basis.
  • the parameters used in the rules can be derived from meta data stored in the data files, such as the DICOM parameters, but they can also be derived from the image content using one or more Convolutional Neural Networks (CNN).
  • CNN Convolutional Neural Networks
  • Each CNN is pre-trained to derive relevant aspects about the image.
  • the CNN is applied to the images of the Study, and the output of the CNN is used to define Image Content based parameters.
  • Image Content Based Parameters are (i) finer granular anatomic information, e.g. whether or not a particular organ is covered by a particular study, or (ii) whether or not a particular medical condition is present, such as a fracture or bleeding.
  • FIG. 1 depicts a flow chart showing the steps of applying various rules to the selected Study, according to an embodiment of the invention
  • FIG. 2 depicts the resulting display for an example study, according to an embodiment of the invention.
  • FIG. 3 shows an example of a user interface to specify rules including a dialog box to configure Study Selection rules, according to an embodiment of the invention.
  • FIG. 4 depicts a line drawing of an illustration of the human spine, with the vertebrae labeled according to standard terminology in human anatomy.
  • FIG. 5 A is a line drawing of an illustration of Study S 1 of a different part of the spine that may have been taken at different point in time to FIGS. 5 B- 5 D .
  • Study S 1 is a scan of the lumbar spine containing all lumbar vertebrae L1, L2, L3, L4, and L5.
  • FIG. 5 B is a line drawing of an illustration of Study S 2 of a different part of the spine that may have been taken at different point in time to FIGS. 5 A, 5 C, and 5 D .
  • Study S 2 is a scan of the cervical spine and does not contain any lumbar vertebrae.
  • FIG. 5 C is a line drawing of an illustration of Study S 3 of a different part of the spine that may have been taken at different point in time to FIGS. 5 A, 5 B, and 5 D .
  • Study S 3 is a scan of vertebrae extending from lumbar to thoracic spine and also contains all five lumbar vertebrae (L1, L2, L3, L4, and L5).
  • FIG. 5 D is a line drawing of an illustration of Study S 4 of a different part of the spine that may have been taken at different point in time to FIGS. 5 A- 5 C .
  • Study S 4 is a scan of the thoracic spine but it also contains lumbar vertebrae L1 and L2.
  • FIG. 6 A is a line drawing corresponding to FIG. 5 A and depicts the result of applying a particular neuronal network to Study S 1 taken from one patient at a specific time, depicting different sections of the spine.
  • the Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow, according to an embodiment of the present invention.
  • FIG. 6 B is a line drawing corresponding to FIG. 5 B and depicts the result of applying a particular neuronal network to Study S 2 taken from one patient at a specific time, depicting different sections of the spine.
  • the Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow, according to an embodiment of the present invention.
  • FIG. 6 C is a line drawing corresponding to FIG. 5 C and depicts the result of applying a particular neuronal network to Study S 3 taken from one patient at a specific time, depicting different sections of the spine.
  • the Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow, according to an embodiment of the present invention.
  • FIG. 6 D is a line drawing corresponding to FIG. 5 D and depicts the result of applying a particular neuronal network to Study S 4 taken from one patient at a specific time, depicting different sections of the spine.
  • the Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow, according to an embodiment of the present invention.
  • FIG. 7 depicts a subset of the DICOM tags and Image Content Based Parameters extracted from the Studies S 1 , S 2 , S 3 , S 4 shown in FIG. 8 , namely Modality, BodyPartExamined, and Vertebrae, according to an embodiment of the present invention.
  • FIG. 8 depicts an example for a Study Selection Rule according to an embodiment of the present invention.
  • the Rule uses the Image Content Based Parameter Vertebrae.
  • the table shows the result of the selection if Study S 1 was loaded by a user as primary study.
  • Studies S 3 and S 4 would be selected for comparison, because they have common anatomy with the primary study, and Study S 2 would not be selected. As is obvious in this example, this could not be achieved using a rule based on the DICOM tag BodyPartExamined alone.
  • FIG. 9 A is a line drawing of a current study which has been loaded by a user into an exemplary hanging protocol showing FIGS. 9 A- 9 D .
  • FIG. 9 B is a line drawing of a current study which has been loaded by a user into an exemplary hanging protocol showing FIGS. 9 A- 9 D .
  • FIG. 9 C is a line drawing of a prior study which has been loaded into an exemplary hanging protocol showing FIGS. 9 A- 9 D , where FIG. 9 C is one of two series of relevant prior studies identified by a Study Selection Rule as containing different but overlapping parts of the anatomy of the same patient, according to an embodiment of the present invention.
  • FIG. 9 D is a line drawing of a prior study which has been loaded into an exemplary hanging protocol showing FIGS. 9 A- 9 D , where FIG. 9 D is one of two series of relevant prior studies identified by a Study Selection Rule as containing different but overlapping parts of the anatomy of the same patient, according to an embodiment of the present invention.
  • transitional phrase “consisting of” excludes any element, step, or ingredient not specified in the claim, but does not exclude additional components or steps that are unrelated to the invention such as impurities ordinarily associated with a composition.
  • a Study will be used to refer to the set of images produced by an examination.
  • a Study consists of one or more images.
  • the images can be grouped into one or more image series.
  • Each image, each series, and the whole Study can have different parameters attached.
  • For medical images these can be defined by the Digital Imaging and Communication in Medicine (DICOM) standard.
  • DICOM Digital Imaging and Communication in Medicine
  • Some or all of the images in a Study can form one or more three dimensional “Volumes.”
  • 3D modalities such as CT or MRI
  • each individual image in the series corresponds to a volume, but that is not a requirement.
  • a cardiac CT may contain multiple 3D volumes covering the heart, each corresponding to a different point in the cardiac cycle, and all of the images belonging to all of these volumes being grouped into the same series.
  • Hanning Protocol will be used to refer to specific conventions how X-Ray films are arranged (hung) at a light box.
  • Display Protocol will be used to refer to the way images are displayed in a computer system, specifically the selection of the images to be displayed, the layout of the images, as well as the rendering parameters and styles.
  • View will be used to refer to data corresponding to a digital image view of a Set of Images rendered with a given set of rendering parameters and rendering modes.
  • Viewport will be used to refer to the logical part of the screen on the client computer in which a particular View is displayed, for example the user interface on the client computer can contain four rectangular Viewports 1160 of which three show a frontal, left, and bottom view respectively of a particular data, while the fourth viewer might show a 2D cross section through the same or a different data set.
  • Sets of Images or “Image Set” will be used to refer to one or more images, selected based on the rules.
  • Protocol Selection Rules will be used to refer to the rules used to select the layout of the images to be displayed.
  • Image Set Rules will be used to refer to the rules used to form Image Sets 1165 from the images of one or more Study by applying selection, sorting, and breaking rules.
  • Style Rules will be used to refer to the rules to determine which rendering type, rendering style, and rendering parameters are used for a particular Image Set 1165 in a particular viewer.
  • Volume Rendering will be used to refer to Volume Rendering techniques including shaded Volume Rendering techniques, maximum intensity projection (MIP), oblique slicing or multi-planar reformats (MPR), axial/sagittal and coronal slice display, and thick slices (also called slabs).
  • MIP maximum intensity projection
  • MPR multi-planar reformats
  • axial/sagittal and coronal slice display and thick slices (also called slabs).
  • thick slices also called slabs.
  • Volume Rendering is used to display 3D images from 3D image data sets, where a typical 3D image data set is a large number of 2D slice images acquired by a CT or MRI scanner and stored in a data structure.
  • anatomical characteristic will be selected from the group consisting of one or more of spine, chest, abdomen, breast, shoulder, trapezius, arm, elbow, wrist, finger, pelvis, hip, fibula, knee, tibula, ankle, foot, neck, head, temporomandibular junction, face, brain, dentition, sinus, adrenals, retina, pituitary, and prostate.
  • the anatomical characteristic can include the Body Part Examined.
  • An anatomical characteristic can be either natural or pathologic.
  • a natural anatomical characteristic of a patient would be the presence of seven cervical vertebrae.
  • a pathologic anatomical characteristic of a patient would be the presence of only six cervical vertebrae.
  • anatomical feature refers to a medical condition, e.g., whether a fracture or bleeding is present in a given image or volume.
  • An anatomical feature can be a fractured fibula, a herniated disc, urethral bleeding, e.g. bleeding with benign prostate hyperplasia, lacerated breast, Gun Shot Wound (GSW) to the chest, infection by Treponema per pneumonia giving rise to YAWS lesion in left distal leg.
  • GSW Gun Shot Wound
  • An anatomical feature is pathologic.
  • ParameterA is fibula then the anatomical feature can be a fractured fibula.
  • disease based characteristic can be selected from the type of disorder or disease being evaluated, e.g., a diagnosis of lung cancer.
  • the disease based characteristic can include the Body Part Examined.
  • a disease based characteristic is pathologic.
  • the phrase “carried out using Convolutional Neural Networks” means that CNN is used to select or identify based on an anatomical characteristic ParameterZ.
  • ParameterZ is SPINE
  • other secondary studies with the same anatomical characteristic can be selected and CNN can generate one or more ImageContentBased parameters from one or more of these secondary studies.
  • the presence of the ImageContentBased parameters can be used to generate a final list for display. This way the rule would not select a cervical spine scan for comparison when the current study is a lumbar spine, but it could select a prior thoracic spine scan for comparison, if that scan did have an overlap with the current scan of the lumbar spine.
  • the traditional ‘Hanging Protocol’ is either not intuitive, cannot display the information in a manner in which it can be reviewed or is not the most efficient way to display images.
  • Alternative ways of rendering the acquired images can be more efficient or more appropriate for displaying the information. Examples include Volume Rendering techniques or maximum intensity projections of stacks of cross-sectional images, rendering of oblique slices, rendering of thick slices or slabs, or rendering of fused images (e.g. in PET/CT).
  • Specialized diagnostic workstations that are often specific to a clinical application area are used to provide appropriate rendering of the acquired images. As organizations and doctors require better and faster visualization methods that allow users to interact with the image data in real-time, the requirements and demands for displaying the data will increase.
  • FIG. 2 depicts an example study where the rules have created two Sets of Images.
  • One Set of Images consists of a series of CT images forming a 3D volume, which is depicted in a volume rendered style in the Viewport 1160 in the upper left and in three orthogonal cross sections in the three other viewports in the left half of the screen.
  • the second Set of Images consist of one chest X-Ray, assigned to a single Viewport 1160 covering the right half of the screen and rendering the X-Ray in 2D style.
  • Appropriate data windows have been chosen by the rules to highlight the vasculature in the 3D rendering, as this is a study with contrast, as the rules can determine by the StudyDescription containing the word ‘contrast’.
  • FIG. 1 is a flow chart showing how the rules are used to create the two Sets of Images shown in FIG. 2 .
  • a primary Study 1105 which can be manually selected by a user.
  • the Study Selection Rules 1115 can identify additional candidate studies 1110 .
  • the second set of studies 1125 which includes the candidate studies 1110 and the primary Study 1105 are available to be loaded into Viewports 1160 .
  • step (ii) 1140 the Protocol Selection Rules 1135 select a Display Protocol 1145 from the Available Display Protocols 1130 based on DICOM Parameters and Abstract Tags present in the second studies 1125 .
  • step (iii) 1155 Image Set Rules 1150 are used to define a plurality of Image Sets 1165 from the second studies 1125 .
  • the one or more Viewports 1160 are defined in the Display Protocol 1145 .
  • step (iv) 1175 Viewport Assignment Rules 1170 assign one or more Image Sets 1165 to one or more Viewports 1160 .
  • Style Rules 1180 define a rendering style and rendering parameters.
  • steps (i) through (v) are performed by a server processor running a render server program with an interface shown in FIG. 3 in which the rules (Study Selection Rules 1115 , Protocol Selection Rules 1135 , Image Set Rules 1150 , Viewport Assignment Rules 1170 , and the one or more Style Rules 1180 ) are used to automatically select and display the Image Sets 1165 in the Viewports 1160 .
  • the rules (Study Selection Rules 1115 , Protocol Selection Rules 1135 , Image Set Rules 1150 , Viewport Assignment Rules 1170 , and the one or more Style Rules 1180 ) are used to automatically select and display the Image Sets 1165 in the Viewports 1160 .
  • a render server program is described in U.S. application Ser. No. 13/831,967, entitled “Multi-User Mult-GPU Render Server Apparatus and Methods”, inventors M. Westerhoff et al., which was filed Mar. 15, 2013, is herein expressly incorporated by reference in its entirety.
  • a rule based render server program is described in U.S. application Ser. No. 13/831,982, entitled “Method and System for Transferring Data to Improve Responsiveness when Sending Large Data Sets”, inventors D Stalling et al., which was filed Mar. 15, 2013, is herein incorporated by reference in its entirety.
  • the system can be connected to a network, e.g. in a hospital, with data being sent to the system from Imaging Modalities, such as CT Scanners or an X-Ray machine, from other computer systems, such as an image archive or PACS system, e.g. using the DICOM network protocol and file format or other suitable network protocols, such as HTTP, HTTPS, SMB and other suitable file formats, such as TIFF, PNG, JPEG. Data can also be inserted into the system by using a CD or DVD, or a USB Memory Stick or other portable media.
  • the system can also query other systems, such as an image archive, and retrieve data, using suitable network protocols and file formats, such as DICOM, or WADO.
  • the images and the volumes of the Study are being processed individually by one or more Convolutional Neural Network (CNN).
  • CNNs can be used for images and volumes respectively, and pre-selection rules can be used to determine which images or volumes to process with which CNN.
  • DICOM tag Modality can be used to process CT images with a different CNN than MRI images.
  • Student Selection Parameters will be used to refer to one or more parameters chosen from the group of DICOM Parameters, Abstract Tags, and Image Content Based Parameters.
  • CNN convolutional Neural Network
  • Alex Krizhevsky et al. ImageNet Classification with Deep Convolutional Neural Networks , In: ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 25 (Eds. F. Pereira, C. J. C. Burges, L. Bottou and K. Q. Weinberger), Curran Associates, Inc., 2012, pp. 1097-1105; and Christian Szegedy et al, Going Deeper with Convolutions, In: COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2015, each of which is incorporated herein by reference and for all purposes.
  • CVPR COMPUTER VISION AND PATTERN RECOGNITION
  • the parameters used in the rules can be derived from an image content using one or more CNN.
  • CNN can be used in the context of localization and object detection.
  • a CNN consists of an input layer, one or more hidden layers and an output layer.
  • optimizing the performance of a CNN can be accomplished by increasing the depth or the number of levels of the network and its width or the number of units at each level.
  • the width defines the region of space within which visual stimuli affect the firing of a single neuron or the receptive field. Given the availability of a large amount of labeled training data it is possible to train higher quality models.
  • the depth and width of the CNN can be maximized, while constraining the computational requirement.
  • an additional 1 ⁇ 1 convolutional layers can be added to the receptive field.
  • an additional n ⁇ n convolutional layers can be added to the receptive field.
  • filters can be used to reduce the dimension and thereby constrain computational demands.
  • the outputs of multiple nodes at one layer can be combined into a single node in the next layer to constrain the computational demands.
  • a resulting matrix of the CNN would include sparse clustering between regions of dense clustering.
  • reconfiguring a matrix containing sparse clustering between regions of dense clustering into two or more relatively dense submatrices can be used to constrain the computational demands.
  • max-pooling in which a matrix is partitioned into a set of non-overlapping submatrices and the maximum for each submatrix is output can be used to constrain the computational demands.
  • filtering is followed by rectified linear activation.
  • the optimal network topology can be constructed layer by layer by analyzing the correlation statistics of the activations of the previous layer and clustering neurons with highly correlated outputs and the institution of multi-scale processing.
  • each CNN is pre-trained to produce one or more output channels that represent relevant aspects of the input images or volumes. These output channels of the CNNs are referred to as “Image Content Based Parameters” in the following.
  • the CNN is applied to the images of the Study.
  • the output of the CNN is used to define Image Content based parameters.
  • an Image Content Based Parameter includes finer granular anatomic information. In an embodiment of the present invention, an Image Content Based Parameter includes whether a particular organ is covered by a particular study. In an embodiment of the present invention, an Image Content Based Parameter includes whether a particular medical condition is present. In an embodiment of the present invention, an Image Content Based Parameter includes a fracture. In an embodiment of the present invention, an Image Content Based Parameter includes a fracture of a specific bone. In an embodiment of the present invention, an Image Content Based Parameter includes a fracture of a tibia. In an embodiment of the present invention, an Image Content Based Parameter includes bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes arterial bleeding.
  • an Image Content Based Parameter includes arterial bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes external venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes internal venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes external venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes internal venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes varicose internal venal bleeding. As can be appreciated by a person of ordinary skill Image Content Based Parameter can cover a variety of medical conditions and their anatomic locations.
  • FIG. 4 depicts an illustration of the human spine, with the vertebrae (C1 130 , C2 135 , C3 140 , C4 145 , C5 150 , C6 152 , C7 154 , Th1 156 , Th2 158 , Th3 160 , Th4 162 , Th5 164 , Th6 166 , Th7 168 , Th8 170 , Th9 172 , Th10 174 , Th11 1176 , Th12 178 , L1 180 , L2 182 , L3 184 , L4 186 , L5 188 , Os sacrum 190 and coccyx 191 ) labeled according to standard terminology in human anatomy.
  • the vertebrae C1 130 , C2 135 , C3 140 , C4 145 , C5 150 , C6 152 , C7 154 , Th1 156 , Th2 158 , Th3 160 , Th4 162 , Th5
  • FIGS. 5 A- 5 D are schematic representations and are not present in any original images.
  • FIG. 5 A depicts Study S 1
  • FIG. 5 B depicts Study S 2
  • FIG. 5 C depicts Study S 3
  • FIG. 5 D depicts Study S 4 showing different parts of the spine that may have been taken at different time points.
  • FIG. 5 A depicts Study S 1
  • FIG. 5 B depicts Study S 2
  • FIG. 5 C depicts Study S 3
  • FIG. 5 D depicts Study S 4 showing different parts of the spine that may have been taken at different time points.
  • FIG. 5 A depicts an illustration of Study S 1 , a scan of the thoracic-sacrum region of the spine containing a portion of a thoracic vertebra 178 , all lumbar vertebrae L1 180 , L2 182 , L3 184 , L4 186 , and L5 188 , and a portion of the sacrum 190 .
  • FIG. 5 B depicts an illustration of Study S 2 , a scan of the cervical spine showing cervical vertebrae C1 (Atlas) 130 , C2 (Axis) 135 , C3 140 , C4 145 , C5 150 , C6 152 , C7 154 , Th1 156 , and a portion of Th2 158 .
  • FIG. 1 cervical vertebrae
  • FIG. 5 B does not contain any lumbar vertebrae.
  • FIG. 5 C depicts an illustration of Study S 3 , a scan of vertebrae extending from lumbar to sacral regions of the spine including a portion of C7 154 , Th1 156 , Th2 158 , Th3 160 , Th4 162 , Th5 164 , Th6 166 , Th7 168 , Th8 170 , Th9 172 , Th10 174 , Th11 176 , Th12 178 , L1 180 , L2 182 , L3 184 , L4 186 , L5 188 , and portion of Os sacrum 190 .
  • FIG. 5 C depicts an illustration of Study S 3 , a scan of vertebrae extending from lumbar to sacral regions of the spine including a portion of C7 154 , Th1 156 , Th2 158 , Th3 160 , Th4 162 , Th5 164 , Th6 166
  • FIG. 5 C contains all five lumbar vertebrae (L1 180 , L2 182 , L3 184 , L4 186 , and L5 188 ).
  • FIG. 5 D depicts an illustration of Study S 4 , a scan of the thoracic and lumbar regions of the spine including a portion of C7 154 , Th1 156 , Th2 158 , Th3 160 , Th4 162 , Th5 164 , Th6 166 , Th7 168 , Th8 170 , Th9 172 , Th10 174 , Th11 176 , Th12 178 , L1 180 , L2 182 and a portion of L3 184 .
  • applying a Convolutional Neuronal Network (CNN) to Study S 1 results in FIG. 6 A which corresponds to FIG. 5 A which recognizes L1 180 , L2 182 , L3 184 , L4 186 , and L5 188 in Study S 1 and outputs the Image Content Based Parameters ⁇ L1, L2, L3, L4, L5 ⁇ .
  • applying a Convolutional Neuronal Network (CNN) to Study S 2 results in FIG. 6 B which corresponds to FIG.
  • FIG. 6 D which corresponds to FIG. 5 D which recognizes Th1 156 , Th2 158 , Th3 160 , Th4 162 , Th5 164 , Th6 166 , Th7 168 , Th8 170 , Th9 172 , Th10 174 , Th11 176 , Th12 178 , L1 180 , L2 182 in Study S 4 and outputs the Image Content Based Parameters ⁇ T1, T2, T3, T4 T5, T6, T7, T8, T9, T10, T11, T12, L1, L2 ⁇ . That is, based on the image the CNN recognizes vertebrae and outputs the Image Content Based Parameters.
  • these Image Content Based Parameters can then be used to select which of Studies S 2 , S 3 and S 4 can help a medical practitioner who has measured Study S 1 make appropriate comparisons and diagnoses.
  • the CNN analysis of Study S 1 was able to recognize L1 180 , L2 182 , L3 184 , L4 186 , and L5 188 in Study S 1 and output Image Content Based Parameters ⁇ L1, L2, L3, L4, L5 ⁇ the medical professional can be interested in displaying other studies that display the lumber vertebra L1, L2, L3, L4, L5.
  • the CNN analysis of Study S 2 did not recognize lumber vertebra L1, L2, L3, L4, L5.
  • any labels and any grayscale coding ( 405 corresponds with cervical vertebrae, 410 corresponds with thoracic vertebrae, 415 correspond with lumbar vertebrae, 420 corresponds with sacrum and 425 corresponds with the coccyx) in FIG. 6 are not present in any original images.
  • FIG. 6 A shows that of the labeled vertebra 178 , 180 , 182 , 184 , 186 , 188 , and 190 all lumbar vertebrae 180 , 182 , 184 , 186 , and 188 appear as labeled by arrow.
  • FIG. 6 A shows that of the labeled vertebra 178 , 180 , 182 , 184 , 186 , 188 , and 190 all lumbar vertebrae 180 , 182 , 184 , 186 , and 188 appear as labeled by arrow.
  • FIG. 1 shows that of the labeled vertebra 178 , 180 , 182 , 184 ,
  • FIG. 6 B shows that all of the labeled vertebra 130 , 135 , 140 , 145 , 150 , 152 , 154 , 156 , and 158 appear as labeled by arrow.
  • FIG. 6 C shows that of the labeled vertebra 154 , 156 , 158 , 160 , 162 , 164 , 166 , 168 , 170 , 172 , 174 , 176 , 178 , 180 , 182 , 184 , 186 , 188 , and 190 appearing as labeled by arrow are 156 , 158 , 160 , 162 , 164 , 166 , 168 , 170 , 172 , 174 , 176 , 178 , 180 , 182 , 184 , 186 , and 188 .
  • FIG. 6 D shows that of the labeled vertebra 154 , 156 , 158 , 160 , 162 , 164 , 166 , 168 , 170 , 172 , 174 , 176 , 178 , 180 , 182 , and 184 appearing as labeled by arrow are 156 , 158 , 160 , 162 , 164 , 166 , 168 , 170 , 172 , 174 , 176 , 178 , 180 , and 182 .
  • the Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow.
  • FIG. 8 depicts an example for a Study Selection Rule according to an embodiment of the present invention.
  • the Study Selection Rule is as follows:
  • each CNN is pre-trained to produce one or more output channels that represent relevant aspects of the input images or volumes.
  • These output channels of the CNNs are referred to herein as “Image Content Based Parameters”.
  • Image Content Based Parameters can be anatomical parameters. For example, they can be more fine granular than the information stored in DICOM parameters.
  • the DICOM parameter BodyPartExamined may specify “SPINE”, whilst an Image Content Based Parameters can be defined for each vertebrae, such as L1, L2, L3 and so forth for the first, second, and third vertebrae in the lumbar spine.
  • Image Content Based Parameters can also be defined for medical conditions, such as whether a fracture or bleeding is present in a given image or volume.
  • the Image Content Based Parameters computed for a Study are stored in an appropriate form, e.g. in a database, a text file, or as private DICOM tags.
  • Image Content Based Parameters can also be computed at a later time than Study Insertion, e.g. by a scheduled task once every hour, or at the time of loading the study by the user.
  • the system can first apply user defined rules to determine additional studies to be displayed together with the primary Study 1105 .
  • additional studies can be prior examinations that are relevant for the diagnosis of the current Study, or additional current studies.
  • a PET examination will often be looked at together with a CT examination acquired at the same time.
  • the set of rules are constructed as follows:
  • Each rule consists of a matching criterion for the primary Study 1105 (primary condition), as well as matching criteria for additional studies (secondary condition).
  • the matching criterion is an expression consisting of operators that allow evaluating the parameters of the Study and comparing them to defined values.
  • the parameters of the Study can be any parameters defined by the DICOM standard, such as Study Description, Study Date, Modality, Patient Age, as well as any other parameters that can be derived from the DICOM parameters or from the Study itself, such as number of images, or number of image series as well as Image Content Based Parameters.
  • the operators are numeric or string based operators, such as equals, greater than, less than, contains, etc. Expressions can be combined using Boolean operators such as AND, OR, NOT. Operators can also contain more complex expressions, including user defined functions defined in an appropriate programming language, such as JavaScript or VisualBasic.
  • This rule will automatically load prior Chest X-Rays or prior Chest CT if the primary Study 1105 is a Chest X-RAY.
  • the rule is expressed in pseudo-code with the primary condition specified in the IF-clause and the secondary condition expressed in the SELECT-clause.
  • Parameter3 can be but need not be equal to Parameter1 and Parameter4 can be but need not be equal to Parameter2.
  • Parameter5 can be but need not be equal to Parameter1 and/or Parameter3
  • Parameter6 can be but need not be equal to Parameter2 and/or Parameter4.
  • Parameter5 is Primary.Vertebrae and Parameter6 is Other.Vertebrae the Parameter5 and Parameter6 denote the set of vertebrae in the primary and the other study respectively, and INTERSECTION( . . . ) NOT EMPTY selects only those prior studies for comparison that actually show at least parts of the same anatomy. This way the rule would not select a cervical spine scan for comparison when the current study is a lumbar spine, but it would select a prior thoracic spine scan for comparison, if that scan did have an overlap with the current scan of the lumbar spine.
  • the rules can normalize DICOM parameters.
  • a Study Selection Rule can contain arbitrary DICOM parameters.
  • the DICOM standard specifies if a particular parameter is defined on a patient, Study, series, or image level. For example, a Study-level parameter should have the same value in all images of a Study, while a series-level parameter should have the same value in all images of a series.
  • a Study-level tag should have the same value for all images of a Study this is not always true.
  • some parameters are defined on a series- or image-level (e.g. modality is a series-level parameter) and therefore can be unavailable. In both cases it can be unclear what value is to be used when evaluating the rule.
  • the invention described here provides different solutions to this problem.
  • a first approach is to choose a reference image and to read the value of a particular DICOM parameter from the reference image.
  • the reference image can be: (i) the image that was inserted into the system first, (ii) the image with the oldest image content date, (iii) the image that was inserted into the system last, or (iv) the image with the earliest image content date.
  • the choice of which image is to be chosen as the reference image can be configured for each parameter separately.
  • a second approach is to only allow original images to be chosen as the reference image.
  • Non-viewable DICOM objects like structured reports, key objects, or presentation states are disregarded, as well as derived images such as secondary capture images or reformatted images.
  • a third approach is to provide a list of all distinct values that a particular DICOM parameter has in the images of a Study.
  • a Study Selection Rule one can then check if that list contains a particular value. The example above can then read as follows:
  • the Study Selection Rules 1115 contain other derived parameters such as Abstract Tags and/or Image Content Based Parameters that characterize a Study in addition to or instead of DICOM parameters.
  • Abstract tags that are useful within Study Selection Rules 1115 include the following:
  • a rule that applies to a Mammogram Study and that selects at maximum three prior Mammogram studies no older than five years can read as follows.
  • a suitable display protocol can be selected. This is done using matching rules.
  • Each matching rule consists of conditions that are applied to the primary and other studies to be loaded.
  • protocol selection rules may contain DICOM parameters (either taken from a reference image or provided as a list of distinct values gathered from all images of a study), as well as Abstract Tags and/or Image Content Based Parameters and user-defined functions.
  • Each matching rule has a score and an associated display protocol.
  • all matching rules are evaluated and the display protocol of the matching rule that evaluates to true can be selected. If multiple matching rules evaluate to true, the one with the highest score can be selected.
  • the following example rule illustrates a matching rule that can apply for PET/CT studies of the abdomen to select a protocol named “StandardPetCTProtocol1” with a score of 10.
  • the rule is expressed in pseudo-code with the matching condition specified in the IF-clause and the chosen protocol specified by the SELECT.
  • the next step comprises creation of so-called image sets.
  • An image set consists of images that are logically grouped together. Usually, an image set is represented by a single preview icon in the application. It is an image set that is loaded into a viewer or tiled viewer.
  • DICOM series also represent a logical grouping of images. However, often DICOM series are not well suited for hanging of images and viewing. For example, in Mammography a single DICOM series may contain images of both left and right breast, in MRI it may contain both T1 and T2 images, or in CT it may contain both a localizer image and a 3D image stack. In all these cases the DICOM series can be split into different logical image sets. On the other hand, multiple DICOM series may represent the phases of a single 4D cardiac data set. In this case all those series can be joined into a single logical image set.
  • the logical image set is a 4D cardiac image set.
  • the individual images of the 4D cardiac image set are sorted in time to represent the temporal ordering of the cardiac cycle.
  • CNN is applied to the images of the 4D cardiac image in order to determine the presence of pathological indicators, e.g., myocardial infarction.
  • evidence of pathological indicator is memorialized in one or more Image Content Based Parameters.
  • the creation of image sets based on rules is a key component of the rule-based display system, specifically for the more advanced rendering techniques.
  • the rules-based display system is used to create image sets that are very similar to the rules described above in Study Selection Rules 1115 and Protocol Selection Rules 1135 sections.
  • a rule is a Boolean expression that can contain DICOM parameters, abstract tags, Image Content Based Parameters, or used-defined functions that are based on the DICOM parameters, abstract tags, Image Content Based Parameters.
  • Image set rules however, are applied to all images of a study that was selected for loading (and not to the study itself). Image-level parameters thus represent no problem and do not need to be normalized or otherwise treated specially. All images that match an image-set rule are grouped into a respective image set.
  • the following rule (expressed in pseudo-code) collects all images of a current CT study.
  • the resulting image sets can be assigned IDs or names that allow for referencing the image sets later in layout and display set rules.
  • ID 1 an image set with ID 1 was defined. If no image matches an image set rule, no such corresponding image set will be created.
  • the order of images within an image set is an important aspect. It determines how images are shown when the user browses through the image set or how images are distributed into the tiles of a tiled viewer.
  • the image set rules can contain an ordered list of sorting criteria. All images that are matched by a rule are sorted according to those criteria.
  • the following rule collects all images of a current CT study and sorts them according to DICOM series number at first and DICOM instance/image number at second.
  • sorting criteria can be extended by a split flag.
  • split flag it is possible to create multiple image sets from a single image set rule.
  • the value of a sorting criterion with split flag set to true changes, sub-sequent images are automatically inserted into a new image set. The resulting image sets are automatically enumerated by a sub-level ID.
  • the following rule essentially creates image sets that correspond to DICOM series, because all images with different series number will be split into different sets.
  • Condition.CTSoftTissueKernel denotes a user-defined Boolean condition that tests whether an image has a CT soft-tissue kernel.
  • the actual implementation of this condition can for example evaluate the manufacturer (which is encoded in a DICOM parameter). Depending on its value the rule can evaluate further parameters to find out if an image was reconstructed using a soft-tissue kernel or not. Since this Boolean condition was used as a sorting criterion with the split flag set to true, all non-soft-kernel images can be put into an image set with ID 1.1 and all soft-kernel images can be put into an image set with ID 1.2 (unless the image set is further split and IDs like 1.3 or 1.4 are created).
  • additional Abstract Tags and/or Image Content Based Parameters are used in image set rules.
  • One example is a tag that identifies whether an image has already been put into an image set. In principle, a single image can be put into multiple image sets, but sometimes this should be avoided. This can be achieved by evaluating image set rules in a pre-defined order and introducing an abstract tag AlreadyReferenced.
  • the image set rules are defined as follows.
  • Condition.IsLocalizer is a user-defined condition that returns true if an image is a localizer image, and false otherwise.
  • Rule 1 is applied first. Therefore the localizer image is put into a separate image set with ID 1.
  • the creation of the image sets based on rules is a key component of the efficient rules based display, specifically for the more advanced rendering techniques.
  • rules can be used to identify sets of 2D images that together form a 3D volume.
  • a display protocol defines multiple viewers, each with one or more tiles, i.e., viewports. To each viewer one or more image sets can be assigned based on Viewer Assignment Rules that are similar to the protocol section rules described herein. Viewer Assignment Rules are defined in the display protocol. The rules determine which image set shall be initially shown in a viewer. In case multiple image sets are assigned to a viewer, the one with the highest score is chosen. Afterwards users may cycle quickly through the remaining image sets using dedicated tools (Previous/Next Image Set), or pick another image set from a special image set menu.
  • Viewer Assignment Rules contain Boolean expressions of DICOM parameters, Abstract Tags, Image Content Based Parameters, or user-defined conditions based on DICOM parameters, Image Content Based Parameters, or abstract tags.
  • image set ID is simply set as a separate abstract tag.
  • the two rules load image sets with the IDs 1 and 2 into a viewer, but assign ID 1 a higher score so that this image set is initially visible (provided such an image set exists).
  • viewer assignment rules are applied to image sets.
  • This conflict is resolved in the same way as described herein in the Normalization of DICOM Parameters section. This means that values of DICOM parameters, Image Content Based Parameters, but also Abstract Tags, are automatically taken from some reference image.
  • a list of distinct values occurring in all images of the image set can be used in an assignment rule.
  • VRT volume rendering style display
  • MIP maximum intensity projection
  • the system uses a global, ordered list of style rules that is evaluated independently for each viewer and each image set loaded into a viewer.
  • An abstract tag DisplaySetID is provided that allows formulating style rules for a specific viewer or group of viewers.
  • Style Rules Parameters driven by Style Rules include the following:
  • the following is an example of a style rule that activates inverse 3D MIP rendering in all viewers with a DisplaySetID between 101 and 104, provided a PET data set is loaded into those viewers (modality PT, i.e., positron emission tomography). Also, an automatic window/level setting is used that is computed from the histogram of the image set (the 2% lowest values are all mapped to white, and the 2% highest values are all mapped to black):
  • the following is another example of a different style rule that always causes the image set with image set ID 200 to be displayed in MPR mode using 20 mm thick slices, with a window/level as specified in the DICOM parameters, and with a zoom factor so that the whole viewer window is filled out.
  • the rule is as follows.
  • Table I summarizes all types of rules that are applied in the rule-base display system:
  • FIG. 2 An example of how these aspects can be combined is shown in FIG. 2 .
  • the user has selected a CT examination of the abdomen.
  • the following rules have been used to determine that a recent X-Ray of the chest is relevant and shall be displayed as well:
  • a hanging protocol can be selected.
  • the protocol selection rules determine that the CT study is a thin slice CT study (i.e. that it has image series that form a 3D volume with sufficient resolution in all directions to display volume rendering or non-axial slices in a meaningful way). Furthermore the example rule determines that this is a study with enhanced vasculature, by looking for the key words “contrast” or “angio” in the study description.
  • This rule will actually form sets from images that contain images that are part of a ThinSliceVolume and that have been reconstructed with a “soft tissue” kernel.
  • the images will first be sorted by the size of the volume of which they are part (Abstract.NumberOfSlicesInVolume), then by DICOM series.
  • the split parameter in this case will ensure that an image set contains images from on series only.
  • a DICOM series can sometimes contain multiple non-consecutive volumes.
  • the abstract tag VolumeIndex will then indicate for each image, which of those volumes it is part of. If a series contains only one volume, then this will be “1” for all images in the series.
  • the images are ordered by slice position, but not split. This way we end up with one image set for each soft kernel thin slice volume, the largest volume being the first image set (ID 1.1). This ID will be used further in subsequent rules.
  • Image Set Rule 5 and 6 will be used to collect the remaining images of the primary Study 1105 .
  • the remaining images are not shown in the layout depicted in the example FIG. 2 .
  • the Display Protocol 1145 contains multiple layouts. The one shown in FIG. 2 is defined as follows:
  • the geometry is defined in a coordinate system having the origin in the upper left corner of the screen with the x axis pointing to the right and the y axis pointing down.
  • parameters of the viewers can be set in the layout definition. Parameters can also be set or overridden in the assignment and style rules, as will be explained next.
  • viewer assignment and style rules are as follows:
  • the rule to select the layout is rather simple: It is shown if the two image sets used exist. This is because the criteria to construct these images sets have been rather specific. As will be appreciated, the proposed system gives this type of flexibility.
  • Some aspects of this invention include methods of displaying one or more Sets of Images comprising the steps of:
  • Additional aspects include methods one or more Display Parameter are selected from the group consisting of Image Set Selection Parameters and View and Viewport Selection Parameters.
  • Display Parameters are selected from the group consisting of Image Set Selection Rules, View and Viewport Selection Rules, and Display Protocol Selection Rules.
  • Yet further aspects include methods where the step of identifying one or more Image Set Selection Rules is based on the one or more Image Set Selection Parameters.
  • Still further aspects include methods where the step of selecting one or more Viewpoint Selection Rules is based on one or more View and Viewport Selection Parameters.
  • Other aspects include methods where the step of displaying the one or more Sets of Images is based on one or more Display Protocol Selection Rules, one or more Image Set Selection Rules, and one or more View and Viewport Selection Rules.
  • Still other aspects include methods where one or more of the Study Selection Parameters are selected from the group consisting of DICOM parameters, Image Content Based Parameters, and Abstract Tags.
  • Display Protocol Selection Parameters are selected from the group consisting of DICOM parameters, Image Content Based Parameters, and Abstract Tags.
  • Additional aspects include methods where one or more of the Image Set Selection Parameters are selected from the group consisting of DICOM parameters, Image Content Based Parameters, and Abstract Tags.
  • View and Viewport Selection Parameters are selected from the group consisting of DICOM parameters, Image Content Based Parameters, and Abstract Tags.
  • More aspects include methods where one or more Study Selection Parameters are derived from a single reference image.
  • Still more aspects include methods where one or more Study Selection Parameters are derived from a single reference image DICOM Parameters.
  • Yet other aspects include methods where one or more Display Protocol Selection Parameters are derived using a list of all values of a DICOM parameter occurring in any of the one or more Sets of Images.
  • Alternative aspects include methods where the one or more View and Viewport Selection Rules contain protocols for one or more Viewports displaying images as 2D.
  • Other alternative aspects include methods where the one or more View and Viewport Selection Rules contain protocols for one or more Viewports displaying images in a 3D rendering mode.
  • Further alternative aspects include methods where one or more Study Selection Parameters include one or more Abstract Tags selected from the group consisting of RelativeStudyAge, PriorIndex. NumImages, NumSeries, Num3DVolumes, Num4DSequences and HasReport.
  • this invention includes methods where one or more View and Viewport Selection Rules include one or more Abstract Tags selected from the group consisting of Image Sets to be displayed, Rendering Style, Additional image sets for image fusion, Image Alignment, Colormap/Transfer Function, Slice Thickness, Zoom Factor, Camera position, Camera orientation and Labels/Overlay elements or one or more Image Content Based Parameters.
  • this invention includes methods further comprising the steps of:
  • this invention includes methods of displaying one or more Sets of Images comprising the steps of:
  • Another aspect of this invention includes methods of displaying images comprising the steps of:
  • this invention includes methods where the Study Selection Rule is:
  • this invention includes methods where the Study Selection Rule is:
  • this invention includes methods where the Protocol Selection Rule is:
  • methods include an Image Set Rule:
  • Image Set Rule is:
  • Still other aspects include methods where the Image Set Rule is:
  • Display Parameters include Viewer Assignment Rule:
  • Display Parameters include a Viewer Assignment Rule:
  • Additional aspects include methods further comprising a Viewer Assignment Rule:
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic in the primary study and a ParameterB as Modality in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParameterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract
  • the one or more ImageContentBased parameters identified are present in the primary study. In an embodiment, the one or more ImageContentBased parameters are vertebrae.
  • the vertebrae in the primary study are selected from the group consisting of L1, L2, L3, L4, and L5, and at least one of the vertebra in the secondary study is a vertebra present in the primary study.
  • the vertebrae in the primary study are selected from the group consisting of C1, C2, C3, C4, C5, C6, and C7, and at least one of the vertebra in the secondary study is a vertebra present in the primary study.
  • the vertebrae in the primary study are selected from the group consisting of Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12, and at least one of the vertebra in the secondary study is a vertebra present in the primary study.
  • the CNN is pretrained with the plurality of studies. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more of the ImageContentBased parameters.
  • the method further includes the CNN selecting based on psueudo code:
  • the one or more Study Selection Rules restrict the final list to studies of the patient.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA from the group consisting of an anatomical characteristic and a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary
  • the method further includes selecting a ParameterB, where the plurality of secondary studies exclude one or more based on ParameterB.
  • ParameterB is Modality.
  • ParameterB is selected from the group consisting of Computer Tomography (CT), then the Modality in the two or more secondary studies inserted is selected from the group consisting of Computed Radiography (CR), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CT Computer Tomography
  • CR Computed Radiography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • ParameterB in the primary study is equal to the ParameterB in the secondary study.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • ParameterA is BodyPartExamined.
  • ParameterA is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULA, KNEE, TIBULA, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the one or more Study Selection Rules restrict the final list to studies of the patient.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the ImageContentBased parameter is stored as a private DICOM tag.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease-based characteristic in the primary study and selecting as a ParameterB a Modality in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParamaterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study
  • the one or more ImageContentBased parameters identified are present in the primary study.
  • the one or more ImageContentBased parameters are vertebrae.
  • the vertebrae in the primary study are selected from the group consisting of L1, L2, L3, L4, and L5, and at least one of the vertebrae in the secondary study is a vertebra present in the primary study.
  • the vertebrae in the primary study are selected from the group consisting of C1, C2, C3, C4, C5, C6, and C7, and at least one of the vertebrae in the secondary study is a vertebra present in the primary study.
  • the vertebrae in the primary study are selected from the group consisting of Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12, and at least one of the vertebrae in the secondary study is a vertebra present in the primary study.
  • the CNN is pretrained with the plurality of studies.
  • the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the disease-based characteristic in the primary study.
  • the CNN selects one or more secondary studies which show the same anatomy.
  • the one or more Study Selection Rules restrict the final list to studies of the patient.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic in the primary study and selecting as a ParameterB a Modality in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParamaterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one
  • the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the method further includes the Study Selection Rule in step (c)(i)
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • the one or more ImageContentBased parameters are vertebrae.
  • the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12.
  • the CNN is pretrained with the plurality of studies.
  • the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study.
  • the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality
  • the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules requires
  • the method further includes where a ParameterB is a Modality in the primary study, where in step (c) the one or more Study Selection Rules restrict to studies where the Modality in the final list is equal to ParameterB.
  • a ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • the method further includes wherein a Study Selection Rule of the one or more Study Selection Rules requires
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters;
  • CNN Convolutional Neural Networks
  • the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules requires
  • the method further includes where a ParameterB is a Modality in the primary study, where in step (c) the one or more Study Selection Rules restrict to studies where the Modality in the final list is equal to ParameterB.
  • a ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • the method further includes the Study Selection Rule
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterB a Modality in the primary study and selecting as a ParameterA from the group consisting of an anatomical characteristic and a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParamaterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consist
  • the one or more ImageContentBased parameters are vertebrae.
  • the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12.
  • the CNN is pretrained with the plurality of studies.
  • the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study.
  • the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study.
  • the one or more ImageContentBased parameters are vertebrae.
  • the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12.
  • the CNN is pretrained with the plurality of studies.
  • the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study. In an embodiment, the one or more ImageContentBased parameters are vertebrae. In an embodiment, the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12.
  • the CNN is pretrained with the plurality of studies. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an AnatomicalCharacteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and one or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the AnatomicalCharacteristic in the primary study; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one
  • the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the one or more parameters include a ParameterB a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (c) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
  • the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the Modality in the two or more of the plurality of secondary studies is selected from the group consisting of Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the AnatomicalFeature is a medical condition.
  • the medical condition is selected from the group consisting of a fracture and a bleeding.
  • the AnatomicalFeature is a notation stored as meta data.
  • the AnatomicalFeature includes whether an organ is present in a study.
  • the ParameterA is selected from meta data stored in the primary Study.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and one or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the disease based characteristic in the primary study; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules
  • the one or more parameters include a ParameterB a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (c) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the AnatomicalFeature is a medical condition.
  • the medical condition is selected from the group consisting of a fracture and a bleeding.
  • the AnatomicalFeature is a notation stored as meta data.
  • the AnatomicalFeature includes whether an organ is present in a study.
  • the ParameterA is selected from meta data stored in the primary Study.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA from the primary study, where the ParameterA is selected from the group consisting of an anatomical characteristic, an anatomical region imaged, and a disease based characteristic; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and one or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the ParameterA selected;
  • step (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (e) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
  • ParameterA is the anatomical region imaged selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the one or more parameters include a ParameterB a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (c) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the AnatomicalFeature is a medical condition.
  • the medical condition is selected from the group consisting of a fracture and a bleeding.
  • the AnatomicalFeature is a notation stored as meta data.
  • the AnatomicalFeature includes whether an organ is present in a study.
  • the ParameterA is selected from meta data stored in the primary Study.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
  • the AnatomicalFeature is a medical condition.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA from the primary study, where the ParameterA is selected from the group consisting of an anatomical characteristic; (c) selecting ParameterB selected from the group consisting of one or more of regions of the anatomical characteristic, parts of a skeletal system of the anatomical characteristic, and organs of the anatomical characteristic; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from
  • the anatomical characteristic is selected from the group consisting of HEAD, EYE, EAR, NOSE, NOSTRIL, MOUTH, LIP, PHILTRUM, JAW, MANDIBLE, GINGIVA, TOOTH, TONGUE, THROAT, LARYNGEAL PROMINENCE, VERTEBRAL COLUMN, SCAPULA, HUMERUS, ELBOW, RADIUS, ULNA, CARPUS, METACARPUS, PHALANGES, THUMB, NAILS, THORAX, BREAST, ABDOMEN, PENIS, SCROTUM, VULVA, LEG, FEMUR, KNEE, PATELLA, TIBIA, SURA, TALOCRURAL REGION, METATARSUS, PHALANGES PROXIMALES, PHALANGES MEDIAE, AND PHALANGES DISTALES.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • one or both the ParameterA and the ParameterB are selected from meta data stored in the primary Study.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies, where the primary Study is an image measured after the patient has a contrast agent administered; (b) selecting a ParameterA an AnatomicalCharacteristic from the primary study; (c) selecting ParameterB a Modality from the primary study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient before the contrast agent was administered selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters
  • the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • one or both the ParameterA and the ParameterB are selected from meta data stored in the primary Study.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA an AnatomicalCharacteristic from the primary study; (c) selecting ParameterB selected from the group consisting of one or more anatomical characteristics of the AnatomicalCharacteristic, parts of a skeletal system of the AnatomicalCharacteristic, organs of the AnatomicalCharacteristic and Modality of the primary Study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one
  • the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • the ParameterA is selected from meta data stored in the primary Study.
  • a including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA from the primary study, where the ParameterA is selected from the group consisting of an anatomical characteristic, and a disease based characteristic; (c) selecting ParameterB from the group consisting of one or more anatomical features of the ParameterA, parts of a skeletal system of the ParameterA, organs of the ParameterA, and a Modality of the primary Study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters
  • the anatomical characteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the ParameterB is a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (d) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • the ParameterB is a Modality selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic imaged; (c) selecting ParameterB from the group consisting of one or more anatomical features of the ParameterA, parts of a skeletal system of the ParameterA, organs of the ParameterA, and a Modality of the primary Study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract
  • the anatomical characteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the ParameterB is a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (d) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • the ParameterB is a Modality selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • the method further includes the Study Selection Rule
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease based characteristic; (c) selecting ParameterB from the group consisting of one or more anatomical features of the ParameterA, parts of a skeletal system of the ParameterA, organs of the ParameterA, and a Modality of the primary Study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tag
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality
  • the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules requires
  • the method further includes where a ParameterB is a Modality in the primary study, where in step (c) the one or more Study Selection Rules restrict to studies where the Modality in the final list is equal to ParameterB.
  • a ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) requires
  • the one or more ImageContentBased parameters are vertebrae.
  • the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12.
  • the CNN is pretrained with the plurality of studies.
  • the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study.
  • the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study.
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the pluralit
  • the method further including where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • the method further includes where a ParameterB is a Modality in the primary study, where in step (c) the one or more Study Selection Rules restrict to studies where the Modality in the final list is equal to ParameterB.
  • a ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
  • CR Computed Radiography
  • CT Computer Tomography
  • DX Digital Radiography
  • MG Mammography
  • MR Magnetic Resonance
  • OP Opthalmic Photography
  • PT Positron Emission Tomography
  • RF Radio Fluoroscopy
  • XA X-Ray Angiography
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) require
  • the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
  • a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterB a Modality in the primary study and selecting as a ParameterA from the group consisting of an anatomical characteristic and a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParamaterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consist

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Medical Informatics (AREA)
  • Physics & Mathematics (AREA)
  • Public Health (AREA)
  • General Health & Medical Sciences (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Pathology (AREA)
  • Theoretical Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Veterinary Medicine (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • High Energy & Nuclear Physics (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Epidemiology (AREA)
  • Human Computer Interaction (AREA)
  • Primary Health Care (AREA)
  • Computer Graphics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Architecture (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Pulmonology (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)

Abstract

The invention provides, in some aspects, a system for implementing a rule derived basis to display image sets. In various embodiments of the invention, the selection of the images to be displayed, the layout of the images, as well as the rendering parameters and styles can be determined using a rule derived basis. The rules are based on meta data of the examination as well as image content that is being analyzed by neuronal networks. In an embodiment of the present invention, the user is presented with images displayed based on their preferences without having to first manually adjust parameters.

Description

PRIORITY CLAIM
This application is a continuation of (1) U.S. application Ser. No. 17/574,975 filed Jan. 13, 2022, which is a continuation of (2) U.S. application Ser. No. 17/060,432 filed Oct. 1, 2020, which issued as U.S. Pat. No. 11,244,494 on Feb. 8, 2022, which is a continuation of (3) U.S. application Ser. No. 16/531,436 filed Aug. 5, 2019, which issued as U.S. Pat. No. 10,832,467 on Nov. 10, 2020, which is a continuation (4) U.S. application Ser. No. 16/052,571 filed Aug. 1, 2018, which issued as U.S. Pat. No. 10,540,803 on Jan. 21, 2020, which is a continuation in part of (5) U.S. application Ser. No. 15/870,681 filed Jan. 12, 2018 which issued as U.S. Pat. No. 10,373,368 on Aug. 6, 2019, which is a continuation of (6) U.S. application Ser. No. 15/380,848 filed Dec. 15, 2016 which issued as U.S. Pat. No. 9,898,855 on Feb. 20, 2018, which is a continuation of (7) U.S. application Ser. No. 14/611,163 filed Jan. 30, 2015 which issued as U.S. Pat. No. 9,524,577 on Dec. 20, 2016, which is a continuation of (8) U.S. application Ser. No. 13/831,975 filed Mar. 15, 2013 which issued as U.S. Pat. No. 8,976,190 on Mar. 10, 2015. U.S. application Ser. No. 16/052,571 (see (4) above) filed Aug. 1, 2018, also claims the benefit of priority of (9) U.S. Provisional Patent Application Ser. No. 62/562,460, filed Sep. 24, 2017, (10) U.S. Provisional Patent Application Ser. No. 62/576,587, filed Oct. 24, 2017, and (11) U.S. Provisional Patent Application Ser. No. 62/712,912, filed Jul. 31, 2018. The teachings of (1)-(11) are herein incorporated by reference in their entireties and for all purposes.
BACKGROUND OF THE INVENTION
In order to diagnose a traditional X-Ray examination, the images printed on films would be ‘hung’ in front of a light box. For multi-image examinations, as well as for comparison with priors, the ‘hanging’ would often follow a specific protocol. For example, a particular organization or doctor may choose for a two-view chest X-Ray with a two-view prior exam, that the films be hung from left to right as follows: Frontal view of current examination, lateral view of current examination, frontal view of prior examination, lateral view of prior examination. In contrast, the doctor may hang mammography exams with the corresponding views of current and prior next to each other, if that was more appropriate for the diagnostic workflow in that case. Thus, the organization or doctor developed a traditional ‘Hanging Protocol’. Currently, the film and the light box are often being replaced by computer systems, called PACS (Picture Archiving and Communication System). PACS systems can mimic the Hanging Protocols.
Traditional X-Ray examinations typically produce one or a small number of single two dimensional (2D) images. In contrast, the more advanced imaging modalities such as Computer Tomography (CT), Magnetic Resonance Imaging (MRI) or Positron Emission Tomography (PET) can produce dozens of series, each consisting of a hundred or more images. It is possible and not uncommon to review images from these advanced modalities in the same manner as traditional X-Ray images, i.e., by hanging the individual images side-by-side, either on a light-box or using a PACS system.
SUMMARY OF THE INVENTION
The invention pertains to digital data processing and, more particularly, by way of example, to the visualization of image data. Three dimensional (3D) and four dimensional (4D) image data is routinely acquired with CT, MRI, PET, confocal microscopes, 3D ultrasound devices, and other imaging devices. The medical imaging market is just one example of a market that uses these devices. The visualization of image data market is growing rapidly, with new CT scanners collecting larger amounts of data more quickly than previous generation CT scanners. The invention has application to areas including medical imaging, atmospheric studies, astrophysics and geophysics.
With the rapid increase in the amounts and types of information that can be acquired using imaging technology, we have identified a substantial problem with integrating different types of image-based information into a form that can be used by a diagnostician, for example a physician. Namely, although there may be many different types of image data, the forms, formats, integration, and display of relevant information is extremely difficult for a person to carry out without sophisticated computer processing. Embodiments of this invention therefore provide a computer-based analytic framework whereby image-based information from a variety of different sources can be integrated to provide increased ability to display relevant information, e.g., to display information for a physician to diagnose and evaluate a patient's condition. We have identified another substantial problem in the art, namely the increased likelihood of confusion of image-based information from different problems, e.g., a physician can incur increased likelihood of confusion of image-based information from different patients. In such situations, a diagnostician (e.g., physician) may be presented with image-based information from different patients. Such inadvertent conflation can produce misdiagnosis or mistaken non-diagnosis. In each case, the outcome can be serious, e.g., misdiagnoses of a patient can result in increased chance of morbidity and/or mortality.
In general aspects of this invention, a First Study is first selected for review by a physician or diagnostician. Selection of a Study will generally be based on some particular characteristic. Such characteristic can be anatomical, disease-based, or both. Once a First Study is selected, an Additional Candidate Study can be selected based on the anatomical location of the First Study. Therefore, if the First Study is a Chest X-Ray, an Additional Candidate Study can be a Chest CT scan, MRI, positron-emission tomography (PET) scan, or other image of the chest. Alternatively, if a First Study is an X-Ray image of the gastrointestinal tract, an Additional Candidate Study could be a series of X-Ray images taken after ingestion of a contrast agent (such as barium). It can be appreciated that such anatomically selected Additional Candidate Studies can be applied to any organ, organ system, or tissue.
Alternatively, Additional Candidate Studies can be selected based on the type of disorder or disease being evaluated. For example, in a case in which a patient has had a diagnosis of cancer of one organ (e.g., lung), it can be desirable for Additional Candidate Studies to be targeted to identification of metastases in another organ. Thus, if a First Study is a Chest X-Ray, an Additional Candidate Study can be of the lymphatic system, head and neck, or various abdominal quadrants. Such Additional Candidate Studies may be X-ray, CT scans, MRI scans, PET scans, vascular visualizations (e.g., with injected contrast media) or histological images taken during a biopsy. Because the degree of detail (i.e., “granularity”) obtained using different imaging techniques may vary widely it can be desirable to have a Rule Based process whereby the granularity of an Additional Candidate Study is increased over that of the First Study.
For example, a Chest X-Ray is a two-dimensional image in which the entirety of the chest and lungs is represented as a flat image. An Additional Candidate Study could be a CT scan, where “2-dimensional” images are acquired at a series of different “depths” (e.g., “slices”) through the organ. If the 2-dimensional images are of sufficient quality to produce a 3-dimensional image of the organ with desirable degree of granularity, then the Additional Candidate Study can be depicted and displayed along with the image of the First Study.
    • General Rule 1 for selecting an Additional Candidate Study therefore can be: IF (Primary.Dicom.BodyPartExamined is “ANATOMICAL REGION 1”, and Primary.Dicom.Modality=IMAGE TYPE 1”)
    • THEN SELECT other studies for loading, WHERE (Other.Dicom.BodyPart Examined=ANATOMICAL REGION 1” and Other.Dicom.Modality=“IMAGE TYPE 2”).
If desired, in General Rule 1, Additional Candidate Studies can target “Other.Dicom.Modality=“IMAGE TYPE 2”).
It can be appreciated that any number of Additional Candidate Studies can be integrated using the computer-based processes of this invention.
Alternatively, General Rule 2 for selecting an Additional Candidate Study therefore can be:
    • IF (Primary.Dicom.Disease is “DISEASE 1”, and Primary.Dicom.Modality=IMAGE TYPE 1”)
    • THEN SELECT other studies for loading, WHERE (Other.Dicom.Disease=“DISEASE 1” and Primary.Dicom.Modality=“IMAGE TYPE 2”).
It can be readily appreciated that application of General Rule 2 can integrate other Anatomical Regions and a number of different Image Types.
In an embodiment of the present invention, a method or system uses a rule derived basis to display image sets. In various embodiments of the present invention, the selection of the images to be displayed, the layout of the images, i.e., the hanging, as well as the rendering parameters and styles can be determined using a rule derived basis. In an embodiment of the present invention, the user is presented with images displayed based on their preferences without having to first manually adjust parameters. Accordingly, there is a time saving in not displaying images initially in a non-rule derived basis.
The parameters used in the rules can be derived from meta data stored in the data files, such as the DICOM parameters, but they can also be derived from the image content using one or more Convolutional Neural Networks (CNN). Each CNN is pre-trained to derive relevant aspects about the image. At the time of data ingestion, the CNN is applied to the images of the Study, and the output of the CNN is used to define Image Content based parameters. Examples for such Image Content Based Parameters are (i) finer granular anatomic information, e.g. whether or not a particular organ is covered by a particular study, or (ii) whether or not a particular medical condition is present, such as a fracture or bleeding.
These and other aspects of the invention are evident in the drawings and in the description that follows.
BRIEF DESCRIPTION OF THE DRAWINGS
This invention is described with respect to specific embodiments thereof. Additional features can be appreciated from the Figures in which:
FIG. 1 depicts a flow chart showing the steps of applying various rules to the selected Study, according to an embodiment of the invention;
FIG. 2 depicts the resulting display for an example study, according to an embodiment of the invention; and
FIG. 3 shows an example of a user interface to specify rules including a dialog box to configure Study Selection rules, according to an embodiment of the invention.
FIG. 4 depicts a line drawing of an illustration of the human spine, with the vertebrae labeled according to standard terminology in human anatomy.
FIG. 5A is a line drawing of an illustration of Study S1 of a different part of the spine that may have been taken at different point in time to FIGS. 5B-5D. Study S1 is a scan of the lumbar spine containing all lumbar vertebrae L1, L2, L3, L4, and L5.
FIG. 5B is a line drawing of an illustration of Study S2 of a different part of the spine that may have been taken at different point in time to FIGS. 5A, 5C, and 5D. Study S2 is a scan of the cervical spine and does not contain any lumbar vertebrae.
FIG. 5C is a line drawing of an illustration of Study S3 of a different part of the spine that may have been taken at different point in time to FIGS. 5A, 5B, and 5D. Study S3 is a scan of vertebrae extending from lumbar to thoracic spine and also contains all five lumbar vertebrae (L1, L2, L3, L4, and L5).
FIG. 5D is a line drawing of an illustration of Study S4 of a different part of the spine that may have been taken at different point in time to FIGS. 5A-5C. Study S4 is a scan of the thoracic spine but it also contains lumbar vertebrae L1 and L2.
FIG. 6A is a line drawing corresponding to FIG. 5A and depicts the result of applying a particular neuronal network to Study S1 taken from one patient at a specific time, depicting different sections of the spine. The Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow, according to an embodiment of the present invention.
FIG. 6B is a line drawing corresponding to FIG. 5B and depicts the result of applying a particular neuronal network to Study S2 taken from one patient at a specific time, depicting different sections of the spine. The Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow, according to an embodiment of the present invention.
FIG. 6C is a line drawing corresponding to FIG. 5C and depicts the result of applying a particular neuronal network to Study S3 taken from one patient at a specific time, depicting different sections of the spine. The Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow, according to an embodiment of the present invention.
FIG. 6D is a line drawing corresponding to FIG. 5D and depicts the result of applying a particular neuronal network to Study S4 taken from one patient at a specific time, depicting different sections of the spine. The Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow, according to an embodiment of the present invention.
FIG. 7 depicts a subset of the DICOM tags and Image Content Based Parameters extracted from the Studies S1, S2, S3, S4 shown in FIG. 8 , namely Modality, BodyPartExamined, and Vertebrae, according to an embodiment of the present invention.
FIG. 8 depicts an example for a Study Selection Rule according to an embodiment of the present invention. The Rule uses the Image Content Based Parameter Vertebrae. The table shows the result of the selection if Study S1 was loaded by a user as primary study. Studies S3 and S4 would be selected for comparison, because they have common anatomy with the primary study, and Study S2 would not be selected. As is obvious in this example, this could not be achieved using a rule based on the DICOM tag BodyPartExamined alone.
FIG. 9A is a line drawing of a current study which has been loaded by a user into an exemplary hanging protocol showing FIGS. 9A-9D.
FIG. 9B is a line drawing of a current study which has been loaded by a user into an exemplary hanging protocol showing FIGS. 9A-9D.
FIG. 9C is a line drawing of a prior study which has been loaded into an exemplary hanging protocol showing FIGS. 9A-9D, where FIG. 9C is one of two series of relevant prior studies identified by a Study Selection Rule as containing different but overlapping parts of the anatomy of the same patient, according to an embodiment of the present invention.
FIG. 9D is a line drawing of a prior study which has been loaded into an exemplary hanging protocol showing FIGS. 9A-9D, where FIG. 9D is one of two series of relevant prior studies identified by a Study Selection Rule as containing different but overlapping parts of the anatomy of the same patient, according to an embodiment of the present invention.
DETAILED DESCRIPTION OF THE INVENTION Definitions
The transitional term “comprising” is synonymous with “including,” “containing,” or “characterized by,” is inclusive or open-ended and does not exclude additional, unrecited elements or method steps.
The transitional phrase “consisting of” excludes any element, step, or ingredient not specified in the claim, but does not exclude additional components or steps that are unrelated to the invention such as impurities ordinarily associated with a composition.
The transitional phrase “consisting essentially of” limits the scope of a claim to the specified materials or steps and those that do not materially affect the basic and novel characteristic(s) of the claimed invention.
The term “Study” will be used to refer to the set of images produced by an examination. A Study consists of one or more images. The images can be grouped into one or more image series. Each image, each series, and the whole Study can have different parameters attached. For medical images these can be defined by the Digital Imaging and Communication in Medicine (DICOM) standard.
Some or all of the images in a Study can form one or more three dimensional “Volumes.” For 3D modalities, such as CT or MRI, often each individual image in the series corresponds to a volume, but that is not a requirement. For example a cardiac CT may contain multiple 3D volumes covering the heart, each corresponding to a different point in the cardiac cycle, and all of the images belonging to all of these volumes being grouped into the same series.
The term “Hanging Protocol” will be used to refer to specific conventions how X-Ray films are arranged (hung) at a light box.
The term “Display Protocol” will be used to refer to the way images are displayed in a computer system, specifically the selection of the images to be displayed, the layout of the images, as well as the rendering parameters and styles.
The term “View” will be used to refer to data corresponding to a digital image view of a Set of Images rendered with a given set of rendering parameters and rendering modes.
The term “Viewport” will be used to refer to the logical part of the screen on the client computer in which a particular View is displayed, for example the user interface on the client computer can contain four rectangular Viewports 1160 of which three show a frontal, left, and bottom view respectively of a particular data, while the fourth viewer might show a 2D cross section through the same or a different data set.
The term “Sets of Images” or “Image Set” will be used to refer to one or more images, selected based on the rules.
The term “Study Selection Rules” will be used to refer to the rules used to select the studies to be displayed.
The term “Protocol Selection Rules” will be used to refer to the rules used to select the layout of the images to be displayed.
The term “Image Set Rules” will be used to refer to the rules used to form Image Sets 1165 from the images of one or more Study by applying selection, sorting, and breaking rules.
The term “Style Rules” will be used to refer to the rules to determine which rendering type, rendering style, and rendering parameters are used for a particular Image Set 1165 in a particular viewer.
The term “Volume Rendering” will be used to refer to Volume Rendering techniques including shaded Volume Rendering techniques, maximum intensity projection (MIP), oblique slicing or multi-planar reformats (MPR), axial/sagittal and coronal slice display, and thick slices (also called slabs). In medical imaging, for example, Volume Rendering is used to display 3D images from 3D image data sets, where a typical 3D image data set is a large number of 2D slice images acquired by a CT or MRI scanner and stored in a data structure.
The term “anatomical characteristic” will be selected from the group consisting of one or more of spine, chest, abdomen, breast, shoulder, trapezius, arm, elbow, wrist, finger, pelvis, hip, fibula, knee, tibula, ankle, foot, neck, head, temporomandibular junction, face, brain, dentition, sinus, adrenals, retina, pituitary, and prostate. The anatomical characteristic can include the Body Part Examined. An anatomical characteristic can be either natural or pathologic. A natural anatomical characteristic of a patient would be the presence of seven cervical vertebrae. A pathologic anatomical characteristic of a patient would be the presence of only six cervical vertebrae.
The term “anatomical feature” refers to a medical condition, e.g., whether a fracture or bleeding is present in a given image or volume. An anatomical feature can be a fractured fibula, a herniated disc, urethral bleeding, e.g. bleeding with benign prostate hyperplasia, lacerated breast, Gun Shot Wound (GSW) to the chest, infection by Treponema pertenue giving rise to YAWS lesion in left distal leg. An anatomical feature is pathologic. In an embodiment of the invention, if ParameterA is fibula then the anatomical feature can be a fractured fibula.
The term “disease based characteristic” can be selected from the type of disorder or disease being evaluated, e.g., a diagnosis of lung cancer. The disease based characteristic can include the Body Part Examined. A disease based characteristic is pathologic.
The phrase “carried out using Convolutional Neural Networks” means that CNN is used to select or identify based on an anatomical characteristic ParameterZ. For example, when ParameterZ is SPINE, other secondary studies with the same anatomical characteristic can be selected and CNN can generate one or more ImageContentBased parameters from one or more of these secondary studies. The presence of the ImageContentBased parameters can be used to generate a final list for display. This way the rule would not select a cervical spine scan for comparison when the current study is a lumbar spine, but it could select a prior thoracic spine scan for comparison, if that scan did have an overlap with the current scan of the lumbar spine.
Overview
Often, the traditional ‘Hanging Protocol’ is either not intuitive, cannot display the information in a manner in which it can be reviewed or is not the most efficient way to display images. Alternative ways of rendering the acquired images can be more efficient or more appropriate for displaying the information. Examples include Volume Rendering techniques or maximum intensity projections of stacks of cross-sectional images, rendering of oblique slices, rendering of thick slices or slabs, or rendering of fused images (e.g. in PET/CT). Specialized diagnostic workstations that are often specific to a clinical application area are used to provide appropriate rendering of the acquired images. As organizations and doctors require better and faster visualization methods that allow users to interact with the image data in real-time, the requirements and demands for displaying the data will increase.
FIG. 2 depicts an example study where the rules have created two Sets of Images. One Set of Images consists of a series of CT images forming a 3D volume, which is depicted in a volume rendered style in the Viewport 1160 in the upper left and in three orthogonal cross sections in the three other viewports in the left half of the screen. The second Set of Images consist of one chest X-Ray, assigned to a single Viewport 1160 covering the right half of the screen and rendering the X-Ray in 2D style. Appropriate data windows have been chosen by the rules to highlight the vasculature in the 3D rendering, as this is a study with contrast, as the rules can determine by the StudyDescription containing the word ‘contrast’.
FIG. 1 is a flow chart showing how the rules are used to create the two Sets of Images shown in FIG. 2 . As shown in FIG. 1 , a primary Study 1105 which can be manually selected by a user. In step (i) 1120, based on Study Selection Rules 1115 which interrogate parameters in the primary Study 1105 such as DICOM Parameters and Abstract Tags of both the primary Study 1105 and the candidate studies 1110, the Study Selection Rules 1115 can identify additional candidate studies 1110. The second set of studies 1125 which includes the candidate studies 1110 and the primary Study 1105 are available to be loaded into Viewports 1160. In step (ii) 1140, the Protocol Selection Rules 1135 select a Display Protocol 1145 from the Available Display Protocols 1130 based on DICOM Parameters and Abstract Tags present in the second studies 1125. In step (iii) 1155, Image Set Rules 1150 are used to define a plurality of Image Sets 1165 from the second studies 1125. The one or more Viewports 1160 are defined in the Display Protocol 1145. In step (iv) 1175, Viewport Assignment Rules 1170 assign one or more Image Sets 1165 to one or more Viewports 1160. In step (v) 1185, Style Rules 1180 define a rendering style and rendering parameters. In an embodiment of the invention steps (i) through (v) are performed by a server processor running a render server program with an interface shown in FIG. 3 in which the rules (Study Selection Rules 1115, Protocol Selection Rules 1135, Image Set Rules 1150, Viewport Assignment Rules 1170, and the one or more Style Rules 1180) are used to automatically select and display the Image Sets 1165 in the Viewports 1160.
A render server program is described in U.S. application Ser. No. 13/831,967, entitled “Multi-User Mult-GPU Render Server Apparatus and Methods”, inventors M. Westerhoff et al., which was filed Mar. 15, 2013, is herein expressly incorporated by reference in its entirety. A rule based render server program is described in U.S. application Ser. No. 13/831,982, entitled “Method and System for Transferring Data to Improve Responsiveness when Sending Large Data Sets”, inventors D Stalling et al., which was filed Mar. 15, 2013, is herein incorporated by reference in its entirety.
The system can be connected to a network, e.g. in a hospital, with data being sent to the system from Imaging Modalities, such as CT Scanners or an X-Ray machine, from other computer systems, such as an image archive or PACS system, e.g. using the DICOM network protocol and file format or other suitable network protocols, such as HTTP, HTTPS, SMB and other suitable file formats, such as TIFF, PNG, JPEG. Data can also be inserted into the system by using a CD or DVD, or a USB Memory Stick or other portable media. The system can also query other systems, such as an image archive, and retrieve data, using suitable network protocols and file formats, such as DICOM, or WADO.
We refer to the process of a new imaging study being sent to or retrieved by the system as “Study Insertion” in the following.
At the time of Study Insertion for each Study, the images and the volumes of the Study are being processed individually by one or more Convolutional Neural Network (CNN). Separate CNNs can be used for images and volumes respectively, and pre-selection rules can be used to determine which images or volumes to process with which CNN. For example, the DICOM tag Modality can be used to process CT images with a different CNN than MRI images. The term “Study Selection Parameters” will be used to refer to one or more parameters chosen from the group of DICOM Parameters, Abstract Tags, and Image Content Based Parameters.
The term “Convolutional Neural Network,” “CNN,” or the like refer, in the usual and customary sense, to a class of deep, feed-forward artificial neural networks that has successfully been applied to analyzing e.g., visual imagery. Exemplary references disclosing methods and systems for CNN include: Alex Krizhevsky et al., ImageNet Classification with Deep Convolutional Neural Networks, In: ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 25 (Eds. F. Pereira, C. J. C. Burges, L. Bottou and K. Q. Weinberger), Curran Associates, Inc., 2012, pp. 1097-1105; and Christian Szegedy et al, Going Deeper with Convolutions, In: COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2015, each of which is incorporated herein by reference and for all purposes.
The parameters used in the rules can be derived from an image content using one or more CNN. In an embodiment of the present invention, CNN can be used in the context of localization and object detection. In an embodiment of the present invention, a CNN consists of an input layer, one or more hidden layers and an output layer. In an embodiment of the present invention, optimizing the performance of a CNN can be accomplished by increasing the depth or the number of levels of the network and its width or the number of units at each level. In an embodiment of the present invention, the width defines the region of space within which visual stimuli affect the firing of a single neuron or the receptive field. Given the availability of a large amount of labeled training data it is possible to train higher quality models. However, increased layers and/or widths typically means a larger number of parameters, which makes the enlarged CNN prone to overfitting, and increased use of computational resources. In an embodiment of the present invention, the depth and width of the CNN can be maximized, while constraining the computational requirement. In an embodiment of the present invention, an additional 1×1 convolutional layers can be added to the receptive field. In an alternative embodiment of the present invention, an additional n×n convolutional layers can be added to the receptive field. In an embodiment of the present invention, filters can be used to reduce the dimension and thereby constrain computational demands. In an alternative embodiment of the present invention, the outputs of multiple nodes at one layer can be combined into a single node in the next layer to constrain the computational demands. In an embodiment of the present invention, a resulting matrix of the CNN would include sparse clustering between regions of dense clustering. In an embodiment of the present invention, reconfiguring a matrix containing sparse clustering between regions of dense clustering into two or more relatively dense submatrices can be used to constrain the computational demands. In an embodiment of the present invention, max-pooling in which a matrix is partitioned into a set of non-overlapping submatrices and the maximum for each submatrix is output can be used to constrain the computational demands. In an embodiment of the present invention, filtering is followed by rectified linear activation. In an embodiment of the present invention, if the probability distribution of the data-set is representable by a large, very sparse CNN, then the optimal network topology can be constructed layer by layer by analyzing the correlation statistics of the activations of the previous layer and clustering neurons with highly correlated outputs and the institution of multi-scale processing. In an embodiment of the present invention, each CNN is pre-trained to produce one or more output channels that represent relevant aspects of the input images or volumes. These output channels of the CNNs are referred to as “Image Content Based Parameters” in the following. At the time of data ingestion, the CNN is applied to the images of the Study. In an embodiment of the present invention, the output of the CNN is used to define Image Content based parameters. In an embodiment of the present invention, an Image Content Based Parameter includes finer granular anatomic information. In an embodiment of the present invention, an Image Content Based Parameter includes whether a particular organ is covered by a particular study. In an embodiment of the present invention, an Image Content Based Parameter includes whether a particular medical condition is present. In an embodiment of the present invention, an Image Content Based Parameter includes a fracture. In an embodiment of the present invention, an Image Content Based Parameter includes a fracture of a specific bone. In an embodiment of the present invention, an Image Content Based Parameter includes a fracture of a tibia. In an embodiment of the present invention, an Image Content Based Parameter includes bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes arterial bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes arterial bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes external venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes internal venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes external venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes internal venal bleeding. In an embodiment of the present invention, an Image Content Based Parameter includes varicose internal venal bleeding. As can be appreciated by a person of ordinary skill Image Content Based Parameter can cover a variety of medical conditions and their anatomic locations.
FIG. 4 depicts an illustration of the human spine, with the vertebrae (C1 130, C2 135, C3 140, C4 145, C5 150, C6 152, C7 154, Th1 156, Th2 158, Th3 160, Th4 162, Th5 164, Th6 166, Th7 168, Th8 170, Th9 172, Th10 174, Th11 1176, Th12 178, L1 180, L2 182, L3 184, L4 186, L5 188, Os sacrum 190 and coccyx 191) labeled according to standard terminology in human anatomy. Note that any labels and any grayscale coding (405 corresponds with cervical vertebrae, 410 corresponds with thoracic vertebrae, 415 correspond with lumbar vertebrae, 420 corresponds with sacrum and 425 corresponds with the coccyx) in FIG. 4 , and FIGS. 5A-5D are schematic representations and are not present in any original images. FIG. 5A depicts Study S1, FIG. 5B depicts Study S2, FIG. 5C depicts Study S3, and FIG. 5D depicts Study S4 showing different parts of the spine that may have been taken at different time points. FIG. 5A depicts an illustration of Study S1, a scan of the thoracic-sacrum region of the spine containing a portion of a thoracic vertebra 178, all lumbar vertebrae L1 180, L2 182, L3 184, L4 186, and L5 188, and a portion of the sacrum 190. FIG. 5B depicts an illustration of Study S2, a scan of the cervical spine showing cervical vertebrae C1 (Atlas) 130, C2 (Axis) 135, C3 140, C4 145, C5 150, C6 152, C7 154, Th1 156, and a portion of Th2 158. FIG. 5B does not contain any lumbar vertebrae. FIG. 5C depicts an illustration of Study S3, a scan of vertebrae extending from lumbar to sacral regions of the spine including a portion of C7 154, Th1 156, Th2 158, Th3 160, Th4 162, Th5 164, Th6 166, Th7 168, Th8 170, Th9 172, Th10 174, Th11 176, Th12 178, L1 180, L2 182, L3 184, L4 186, L5 188, and portion of Os sacrum 190. FIG. 5C contains all five lumbar vertebrae (L1 180, L2 182, L3 184, L4 186, and L5 188). FIG. 5D depicts an illustration of Study S4, a scan of the thoracic and lumbar regions of the spine including a portion of C7 154, Th1 156, Th2 158, Th3 160, Th4 162, Th5 164, Th6 166, Th7 168, Th8 170, Th9 172, Th10 174, Th11 176, Th12 178, L1 180, L2 182 and a portion of L3 184. In an embodiment of the invention, applying a Convolutional Neuronal Network (CNN) to Study S1 results in FIG. 6A which corresponds to FIG. 5A which recognizes L1 180, L2 182, L3 184, L4 186, and L5 188 in Study S1 and outputs the Image Content Based Parameters {L1, L2, L3, L4, L5}. In an embodiment of the invention, applying a Convolutional Neuronal Network (CNN) to Study S2 results in FIG. 6B which corresponds to FIG. 5B which recognizes C1 (Atlas) 130, C2 (Axis) 135, C3 140, C4 145, C5 150, C6 152, C7 154, Th1 156 in Study S2 and outputs the Image Content Based Parameters {C1, C2, C3, C4 C5, C6, C7, T1}. In an embodiment of the invention, applying a Convolutional Neuronal Network (CNN) to Study S3 results in FIG. 6C which corresponds to FIG. 5C which recognizes Th1 156, Th2 158, Th3 160, Th4 162, Th5 164, Th6 166, Th7 168, Th8 170, Th9 172, Th10 174, Th11 176, Th12 178, L1 180, L2 182, L3 184, L4 186, L5 188 in Study S3 and outputs the Image Content Based Parameters {T1, T2, T3, T4 T5, T6, T7, T8, T9, T10, T11, T12, L1, L2, L3, L4 L5}. In an embodiment of the invention, applying a Convolutional Neuronal Network (CNN) to Study S4 results in FIG. 6D which corresponds to FIG. 5D which recognizes Th1 156, Th2 158, Th3 160, Th4 162, Th5 164, Th6 166, Th7 168, Th8 170, Th9 172, Th10 174, Th11 176, Th12 178, L1 180, L2 182 in Study S4 and outputs the Image Content Based Parameters {T1, T2, T3, T4 T5, T6, T7, T8, T9, T10, T11, T12, L1, L2}. That is, based on the image the CNN recognizes vertebrae and outputs the Image Content Based Parameters. In an embodiment of the invention, these Image Content Based Parameters can then be used to select which of Studies S2, S3 and S4 can help a medical practitioner who has measured Study S1 make appropriate comparisons and diagnoses. For example, since the CNN analysis of Study S1 was able to recognize L1 180, L2 182, L3 184, L4 186, and L5 188 in Study S1 and output Image Content Based Parameters {L1, L2, L3, L4, L5} the medical professional can be interested in displaying other studies that display the lumber vertebra L1, L2, L3, L4, L5. As summarized in FIG. 7 , the CNN analysis of Study S2 did not recognize lumber vertebra L1, L2, L3, L4, L5. In contrast, the CNN analysis of Study S3 did recognize lumber vertebra L1, L2, L3, L4, L5 and the CNN analysis of Study S4 did recognize lumber vertebra L1, L2. As such the CNN analysis identifies Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, to be used in the Study Selection Rule:
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND INTERSECTION(ParameterC,ParameterD) NOT EMPTY), which rule is fulfilled for primary study S1, by study S3 and study S4, but not study S2, provided that the primary study anatomical characteristic is the same in study S3 and study S4.
Note that any labels and any grayscale coding (405 corresponds with cervical vertebrae, 410 corresponds with thoracic vertebrae, 415 correspond with lumbar vertebrae, 420 corresponds with sacrum and 425 corresponds with the coccyx) in FIG. 6 are not present in any original images. FIG. 6A shows that of the labeled vertebra 178, 180, 182, 184, 186, 188, and 190 all lumbar vertebrae 180, 182, 184, 186, and 188 appear as labeled by arrow. FIG. 6B shows that all of the labeled vertebra 130, 135, 140, 145, 150, 152, 154, 156, and 158 appear as labeled by arrow. FIG. 6C shows that of the labeled vertebra 154, 156, 158, 160, 162, 164, 166, 168, 170, 172, 174, 176, 178, 180, 182, 184, 186, 188, and 190 appearing as labeled by arrow are 156, 158, 160, 162, 164, 166, 168, 170, 172, 174, 176, 178, 180, 182, 184, 186, and 188. FIG. 6D shows that of the labeled vertebra 154, 156, 158, 160, 162, 164, 166, 168, 170, 172, 174, 176, 178, 180, 182, and 184 appearing as labeled by arrow are 156, 158, 160, 162, 164, 166, 168, 170, 172, 174, 176, 178, 180, and 182. The Image Content Based Parameter computed by the neuronal network in this example is the set (list) of vertebrae shown below the arrow. FIG. 8 depicts an example for a Study Selection Rule according to an embodiment of the present invention. The Study Selection Rule is as follows:
    • IF (Primary.Dicom.BodyPartExamined=“SPINE” and Primary.Dicom.Modality=“MR”)
    • THEN SELECT other studies for loading WHERE (Other.Dicom.BodyPartExamined=“SPINE” and (Other.Dicom.Modality=“MR”) AND
    • INTERSECTION(Primary.Vertebrae,Other.Vertebrae) NOT EMPTY).
      The Rule uses the Image Content Based Parameter ‘Vertebrae’. FIG. 8 shows the result of the selection if Study S1 was loaded by a user as primary study. Studies S3 and S4 would be selected for comparison, because they have common anatomy with the primary study, and Study S2 would not be selected. As is obvious in this example, this could not be achieved using a rule based on the DICOM tag BodyPartExamined alone (e.g., LSPINE, SPINE and TSPINE). FIG. 7 depicts a subset of the DICOM tags and Image Content Based Parameters extracted from the Studies S1, S2, S3, S4 shown in FIG. 8 , namely Modality, BodyPartExamined, and Vertebrae, according to an embodiment of the present invention. FIGS. 9A-D depict an exemplary hanging protocol where the client view is shown at 910, a menu including the functions ‘File’, ‘View’, ‘Protocol’, ‘Tools’ and ‘Help’ is shown at 915, as series of self-explanatory icons related to specific functions is shown at 920, the number of series including ‘1 Series’, ‘2 Series’, ‘4 Series’, ‘6 Series’, ‘8 Series’, ‘12 Series’, ‘Compare 1+1’, ‘Compare 2+2’, ‘Compare 4+4’, ‘Compare 6+6’, ‘Multiplanar (act. series)’, ‘Multiplanar+3D MIP (act. series)’ to be viewed is selected at 925, where the identification of the image displayed is shown at 930, the date of analysis of the image displayed is shown at 940, the series identification and image identification of the image displayed is shown at 950, and the anatomical location descriptor (L=lateral) of the image displayed is shown at 960, and the scale and magnification of the view of the image displayed is shown at 970. FIGS. 9A and 9B show two series of a current study that have been loaded by the user and where a Study Selection Rule has identified a relevant prior study containing a different but overlapping part of the anatomy of the same patient and displays FIGS. 9C and 9D, two series of the prior study for comparison (592 includes diaphragm, 593 includes disks, 594 includes the posterior spinal cord, 596 includes anterior spinal cord and vertebrae, 597 includes diaphragm and vertebrae and 598 includes the epithelial layer), according to an embodiment of the present invention.
In embodiments, each CNN is pre-trained to produce one or more output channels that represent relevant aspects of the input images or volumes. These output channels of the CNNs are referred to herein as “Image Content Based Parameters”. Image Content Based Parameters can be anatomical parameters. For example, they can be more fine granular than the information stored in DICOM parameters. For example the DICOM parameter BodyPartExamined may specify “SPINE”, whilst an Image Content Based Parameters can be defined for each vertebrae, such as L1, L2, L3 and so forth for the first, second, and third vertebrae in the lumbar spine. Image Content Based Parameters can also be defined for medical conditions, such as whether a fracture or bleeding is present in a given image or volume.
The Image Content Based Parameters computed for a Study are stored in an appropriate form, e.g. in a database, a text file, or as private DICOM tags.
As is easily appreciated, Image Content Based Parameters can also be computed at a later time than Study Insertion, e.g. by a scheduled task once every hour, or at the time of loading the study by the user.
Study Selection Rules 1115
In an embodiment of the present invention, based on the Study that the user selects for display (primary Study 1105), the system can first apply user defined rules to determine additional studies to be displayed together with the primary Study 1105. Such additional studies can be prior examinations that are relevant for the diagnosis of the current Study, or additional current studies. For example, a PET examination will often be looked at together with a CT examination acquired at the same time. The set of rules are constructed as follows:
Each rule consists of a matching criterion for the primary Study 1105 (primary condition), as well as matching criteria for additional studies (secondary condition). The matching criterion is an expression consisting of operators that allow evaluating the parameters of the Study and comparing them to defined values. The parameters of the Study can be any parameters defined by the DICOM standard, such as Study Description, Study Date, Modality, Patient Age, as well as any other parameters that can be derived from the DICOM parameters or from the Study itself, such as number of images, or number of image series as well as Image Content Based Parameters. The operators are numeric or string based operators, such as equals, greater than, less than, contains, etc. Expressions can be combined using Boolean operators such as AND, OR, NOT. Operators can also contain more complex expressions, including user defined functions defined in an appropriate programming language, such as JavaScript or VisualBasic.
Once a primary Study 1105 has been selected for display, the primary condition of each rule is evaluated. Those rules that match, i.e., evaluate to “true” for the given primary Study 1105, will then be applied to all other studies that are available for the same patient. Those other studies for which the secondary condition matches will be added to the list of studies to be displayed.
The following rule illustrates the concept. This rule will automatically load prior Chest X-Rays or prior Chest CT if the primary Study 1105 is a Chest X-RAY.
Study Selection Rule 1:
    • IF (Primary.Dicom.BodyPartExamined=“CHEST” and Primary.Dicom.Modality=“CR”)
    • THEN SELECT other studies for loading
    • WHERE (Other.Dicom.BodyPartExamined=“CHEST” and (Other.Dicom.Modality=“CR” or Other.Dicom.Modality=“CT”)).
The rule is expressed in pseudo-code with the primary condition specified in the IF-clause and the secondary condition expressed in the SELECT-clause.
Study Selection Rule 2A:
    • IF (Primary.Dicom.BodyPartExamined=“SPINE” and Primary.Dicom.Modality=“MR”)
    • THEN SELECT other studies for loading
    • WHERE (Other.Dicom.BodyPartExamined=“SPINE” and (Other.Dicom.Modality=“MR” AND INTERSECTION(Primary.Vertebrae,Other.Vertebrae) NOT EMPTY).
Study Selection Rule 2B:
    • IF (Primary.Dicom.BodyPartExamined=“SPINE” and Primary.Dicom.Modality=“CT”) THEN SELECT other studies for loading
    • WHERE (Other.Dicom.BodyPartExamined=“SPINE” and Other.Dicom.Modality=“CT” AND INTERSECTION(Spinous.process,Cobb.angle) NOT EMPTY).
Study Selection Rule 3A:
    • IF (Primary.Dicom.BodyPartExamined=“CHEST” and Primary.Dicom.Modality=“MR”)
    • THEN SELECT other studies for loading
    • WHERE (Other.Dicom.BodyPartExamined=“CHEST” and Other.Dicom.Modality=“MR” AND INTERSECTION(Spinous.process,Thyroid) NOT EMPTY).
Study Selection Rule 3B:
    • IF (Primary.Dicom.BodyPartExamined=“CHEST” and Primary.Dicom.Modality=“CT”)
    • THEN SELECT other studies for loading
    • WHERE (Other.Dicom.BodyPartExamined=“CHEST” and Other.Dicom.Modality=“CT” AND INTERSECTION(Spinous.process,Thyroid) NOT EMPTY).
Study Selection Rule 4A:
    • IF (Primary.Dicom.BodyPartExamined=“BREAST” and Primary.Dicom.Modality=“MR”)
    • THEN SELECT other studies for loading
    • WHERE (Other.Dicom.BodyPartExamined=“BREAST” and Other.Dicom.Modality=“MR” AND INTERSECTION(Implant,Tumor) NOT EMPTY).
Study Selection Rule 4B:
    • IF (Primary.Dicom.BodyPartExamined=“BREAST” and Primary.Dicom.Modality=“CT”)
    • THEN SELECT other studies for loading
    • WHERE (Other.Dicom.BodyPartExamined=“BREAST” and Other.Dicom.Modality=“CT” AND INTERSECTION(Implant,Tumor) NOT EMPTY).
Study Selection Rule 5A:
    • IF (Primary.Dicom.BodyPartExamined=“BODY_PART” and Primary.Dicom.Modality=“MR”)
    • THEN SELECT other studies for loading
    • WHERE (Other.Dicom.BodyPartExamined=“BODY_PART” and Other.Dicom.Modality=“MR” AND INTERSECTION(Parameter1,Parameter2) NOT EMPTY).
Study Selection Rule 5B:
    • IF (Primary.Dicom.BodyPartExamined=“BODY_PART” and Primary.Dicom.Modality=“CT”)
    • THEN SELECT other studies for loading
    • WHERE (Other.Dicom.BodyPartExamined=“BODY_PART” and
    • Other.Dicom.Modality=“CT” AND INTERSECTION(Parameter3,Parameter4) NOT EMPTY).
Where Parameter3 can be but need not be equal to Parameter1 and Parameter4 can be but need not be equal to Parameter2.
Study Selection Rule 5C:
    • IF (Primary.Dicom.BodyPartExamined=“BODY_PART” and Primary.Dicom.Modality=“MG”)
    • THEN SELECT other studies for loading WHERE (Other.Dicom.BodyPartExamined=“BODY_PART” and (Other.Dicom.Modality=“MG” AND
    • INTERSECTION(Parameter5,Parameter6) NOT EMPTY).
Where Parameter5 can be but need not be equal to Parameter1 and/or Parameter3, and Parameter6 can be but need not be equal to Parameter2 and/or Parameter4.
In this example if Parameter5 is Primary.Vertebrae and Parameter6 is Other.Vertebrae the Parameter5 and Parameter6 denote the set of vertebrae in the primary and the other study respectively, and INTERSECTION( . . . ) NOT EMPTY selects only those prior studies for comparison that actually show at least parts of the same anatomy. This way the rule would not select a cervical spine scan for comparison when the current study is a lumbar spine, but it would select a prior thoracic spine scan for comparison, if that scan did have an overlap with the current scan of the lumbar spine.
Study Selection Rules: Normalization of DICOM Parameters
In an embodiment of the present invention, the rules can normalize DICOM parameters. As described above, a Study Selection Rule can contain arbitrary DICOM parameters. The DICOM standard specifies if a particular parameter is defined on a patient, Study, series, or image level. For example, a Study-level parameter should have the same value in all images of a Study, while a series-level parameter should have the same value in all images of a series. There are two problems related to assuming that this statement is always the case. Firstly, although a Study-level tag should have the same value for all images of a Study this is not always true. Secondly, some parameters are defined on a series- or image-level (e.g. modality is a series-level parameter) and therefore can be unavailable. In both cases it can be unclear what value is to be used when evaluating the rule. The invention described here provides different solutions to this problem.
In an embodiment of the present invention, a first approach is to choose a reference image and to read the value of a particular DICOM parameter from the reference image. The reference image can be: (i) the image that was inserted into the system first, (ii) the image with the oldest image content date, (iii) the image that was inserted into the system last, or (iv) the image with the earliest image content date. The choice of which image is to be chosen as the reference image can be configured for each parameter separately.
In an embodiment of the present invention, a second approach is to only allow original images to be chosen as the reference image. Non-viewable DICOM objects like structured reports, key objects, or presentation states are disregarded, as well as derived images such as secondary capture images or reformatted images.
In an embodiment of the present invention, a third approach is to provide a list of all distinct values that a particular DICOM parameter has in the images of a Study. In a Study Selection Rule one can then check if that list contains a particular value. The example above can then read as follows:
Study Selection Rule 6:
    • IF (Primary.Dicom.BodyPartExamined=“CHEST” and Primary.DicomList.Modality contains “CR”)
    • THEN SELECT other studies for loading WHERE (Other.Dicom.BodyPartExamined=“CHEST” and (Other.DicomList.Modality contains “CR” or Other.DicomList.Modality contains “CT”)).
Study Selection Rules: Abstract Tags
In an embodiment of the present invention, the Study Selection Rules 1115 contain other derived parameters such as Abstract Tags and/or Image Content Based Parameters that characterize a Study in addition to or instead of DICOM parameters. Abstract tags that are useful within Study Selection Rules 1115 include the following:
    • (i) RelativeStudyAge indicates relative age of Study in days compared to primary Study 1105.
    • (ii) PriorIndex indicates an index that enumerates all other studies from youngest to oldest.
    • (iii) NumImages indicates number of images in Study.
    • (iv) NumSeries indicated number of image series in Study.
    • (v) Num3DVolumes indicates number of 3D volumes in Study.
    • (vi) Num4DSequences indicates number of 4D sequences in Study (e.g. Cardiac CT).
    • (vii) HasReport indicates a flag that indicates if a report is available for a Study.
    • (viii) HasThinSliceVolumes indicates whether the study has at least one set of images that form a true 3D volume, i.e. a sufficiently large number of equidistant slices (the exact number can be user configurable, e.g. 30 would be a common choice) and a sufficiently small spacing between two consecutive slices to guarantee an isotropic (or close to isotropic) (again, this parameter can be user defined, values between 1 mm and 3 mm are common thresholds for CT and MR examinations).
For example, a rule that applies to a Mammogram Study and that selects at maximum three prior Mammogram studies no older than five years can read as follows.
Study Selection Rule 7:
    • IF (Primary.Dicom.Modality=“MG”
    • THEN SELECT other studies for loading WHERE (Other.Dicom.Modality=“MG” and Other.Abstract.PriorIndex<=3 and Other.Abstract.RelativeStudyAge<5*365).
Protocol Selection Rules 1135
In an embodiment of the present invention, once the studies to be displayed are determined as described above, a suitable display protocol can be selected. This is done using matching rules. Each matching rule consists of conditions that are applied to the primary and other studies to be loaded. Like in Study Selection Rules 1115, protocol selection rules may contain DICOM parameters (either taken from a reference image or provided as a list of distinct values gathered from all images of a study), as well as Abstract Tags and/or Image Content Based Parameters and user-defined functions. Each matching rule has a score and an associated display protocol.
In an embodiment of the present invention, all matching rules are evaluated and the display protocol of the matching rule that evaluates to true can be selected. If multiple matching rules evaluate to true, the one with the highest score can be selected.
The following example rule illustrates a matching rule that can apply for PET/CT studies of the abdomen to select a protocol named “StandardPetCTProtocol1” with a score of 10.
Protocol Selection Rule 1:
    • IF (Primary.Dicom.BodyPartExamined=“ABDOMEN” and Primary.Dicom.Modality=“CT” and Exists(Other1) and Other1.Dicom.Modality=“PET”)
    • THEN SELECT “StandardPetCTProtocol1” with score=10.
In an embodiment of the present invention, the rule is expressed in pseudo-code with the matching condition specified in the IF-clause and the chosen protocol specified by the SELECT.
Image Set Rules 1150
In an embodiment of the present invention, once a display protocol is selected, further rules defined within the protocol are evaluated. The next step comprises creation of so-called image sets. An image set consists of images that are logically grouped together. Usually, an image set is represented by a single preview icon in the application. It is an image set that is loaded into a viewer or tiled viewer. Note that DICOM series also represent a logical grouping of images. However, often DICOM series are not well suited for hanging of images and viewing. For example, in Mammography a single DICOM series may contain images of both left and right breast, in MRI it may contain both T1 and T2 images, or in CT it may contain both a localizer image and a 3D image stack. In all these cases the DICOM series can be split into different logical image sets. On the other hand, multiple DICOM series may represent the phases of a single 4D cardiac data set. In this case all those series can be joined into a single logical image set.
In an embodiment, the logical image set is a 4D cardiac image set. In an embodiment, the individual images of the 4D cardiac image set are sorted in time to represent the temporal ordering of the cardiac cycle. In an embodiment, CNN is applied to the images of the 4D cardiac image in order to determine the presence of pathological indicators, e.g., myocardial infarction. In embodiments, evidence of pathological indicator is memorialized in one or more Image Content Based Parameters.
Thus the creation of image sets based on rules is a key component of the rule-based display system, specifically for the more advanced rendering techniques. For example, the rules-based display system is used to create image sets that are very similar to the rules described above in Study Selection Rules 1115 and Protocol Selection Rules 1135 sections. A rule is a Boolean expression that can contain DICOM parameters, abstract tags, Image Content Based Parameters, or used-defined functions that are based on the DICOM parameters, abstract tags, Image Content Based Parameters. Image set rules however, are applied to all images of a study that was selected for loading (and not to the study itself). Image-level parameters thus represent no problem and do not need to be normalized or otherwise treated specially. All images that match an image-set rule are grouped into a respective image set.
In an embodiment of the present invention, the following rule (expressed in pseudo-code) collects all images of a current CT study.
Image Set Rule 1:
    • IF (Dicom.Modality=“CT” and Abstract.PriorIndex=0)
    • THEN CREATE image set with ID 1.
In an embodiment of the present invention, the resulting image sets can be assigned IDs or names that allow for referencing the image sets later in layout and display set rules. In the above example an image set with ID 1 was defined. If no image matches an image set rule, no such corresponding image set will be created.
Image Set Rules: Sorting
In an embodiment of the present invention, the order of images within an image set is an important aspect. It determines how images are shown when the user browses through the image set or how images are distributed into the tiles of a tiled viewer. In one embodiment of the present invention, in order to specify image sorting, the image set rules can contain an ordered list of sorting criteria. All images that are matched by a rule are sorted according to those criteria.
For example, the following rule collects all images of a current CT study and sorts them according to DICOM series number at first and DICOM instance/image number at second.
Image Set Rule 2:
    • IF (Dicom.Modality=“CT” and Abstract.PriorIndex=0)
    • THEN CREATE image set with ID 1
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending.
Image Set Rules: Splitting
In an embodiment of the present invention, sorting criteria can be extended by a split flag. With the split flag it is possible to create multiple image sets from a single image set rule. When the value of a sorting criterion with split flag set to true changes, sub-sequent images are automatically inserted into a new image set. The resulting image sets are automatically enumerated by a sub-level ID.
For example, the following rule essentially creates image sets that correspond to DICOM series, because all images with different series number will be split into different sets.
Image Set Rule 3:
    • IF (Dicom.Modality=“CT” and Abstract.PriorIndex=0)
    • THEN CREATE image set with ID 1.x
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
In applications where a CT has been measured, it can happen that a study contains both a soft-kernel series and a hard kernel series and both series have the same series number. In order to separate the images into different image sets the above rule can be extended by the following:
Image Set Rule 4:
    • IF (Dicom.Modality=“CT” and Abstract.PriorIndex=0)
    • THEN CREATE image set with ID 1.x
      • SORTED BY Condition.CTSoftTisseKernel SPLIT:=true
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
Here, Condition.CTSoftTissueKernel denotes a user-defined Boolean condition that tests whether an image has a CT soft-tissue kernel. The actual implementation of this condition can for example evaluate the manufacturer (which is encoded in a DICOM parameter). Depending on its value the rule can evaluate further parameters to find out if an image was reconstructed using a soft-tissue kernel or not. Since this Boolean condition was used as a sorting criterion with the split flag set to true, all non-soft-kernel images can be put into an image set with ID 1.1 and all soft-kernel images can be put into an image set with ID 1.2 (unless the image set is further split and IDs like 1.3 or 1.4 are created).
Image Set Rules: More Abstract Tags
In an embodiment of the present invention, additional Abstract Tags and/or Image Content Based Parameters are used in image set rules. One example is a tag that identifies whether an image has already been put into an image set. In principle, a single image can be put into multiple image sets, but sometimes this should be avoided. This can be achieved by evaluating image set rules in a pre-defined order and introducing an abstract tag AlreadyReferenced.
For example, in CT study that has a localizer image and a 3D image stack both stored in one DICOM series, one may want to create an image set, one for the localizer and one for the 3D image stack. Accordingly, the image set rules are defined as follows.
Image Set Rule 5 (Localizer):
    • IF (Dicom.Modality=“CT” and Condition.IsLocalizer=true)
    • THEN CREATE image set with ID 1
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
Image Set Rule 6 (Images):
    • IF (Dicom.Modality=“CT” and Abstract.AlreadyReferenced=false)
    • THEN CREATE image set with ID 2
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
Here Condition.IsLocalizer is a user-defined condition that returns true if an image is a localizer image, and false otherwise. In an embodiment of the present invention, Rule 1 is applied first. Therefore the localizer image is put into a separate image set with ID 1. Next rule 2 is applied. This rule can match for all CT images including the localizer image. However, because AlreadyReferenced=false is specified, the localizer image is skipped and not placed into image set 2.
In an embodiment of the present invention, the creation of the image sets based on rules is a key component of the efficient rules based display, specifically for the more advanced rendering techniques. For example rules can be used to identify sets of 2D images that together form a 3D volume.
Viewer Assignment Rules
In another embodiment of the present invention, a display protocol defines multiple viewers, each with one or more tiles, i.e., viewports. To each viewer one or more image sets can be assigned based on Viewer Assignment Rules that are similar to the protocol section rules described herein. Viewer Assignment Rules are defined in the display protocol. The rules determine which image set shall be initially shown in a viewer. In case multiple image sets are assigned to a viewer, the one with the highest score is chosen. Afterwards users may cycle quickly through the remaining image sets using dedicated tools (Previous/Next Image Set), or pick another image set from a special image set menu.
Like the other rule types Viewer Assignment Rules contain Boolean expressions of DICOM parameters, Abstract Tags, Image Content Based Parameters, or user-defined conditions based on DICOM parameters, Image Content Based Parameters, or abstract tags. In many cases it is sufficient to specify the image sets to be assigned to a viewer by their image set ID instead of evaluating the underlying DICOM parameters, Image Content Based Parameters and/or Abstract Tags again. Therefore, the image set ID is simply set as a separate abstract tag. In the following example the two rules load image sets with the IDs 1 and 2 into a viewer, but assign ID 1 a higher score so that this image set is initially visible (provided such an image set exists).
Viewer Assignment Rule 1:
    • IF (EXISTS ImageSet[1])
    • THEN Viewport[0].AddImageSet(ID=1, score=10).
Viewer Assignment Rule 2:
    • IF (EXISTS ImageSet[2])
    • THEN Viewport[0].AddImageSet(ID=2, score=5).
In an embodiment of the present invention, viewer assignment rules are applied to image sets. Thus there is a possible conflict regarding ambiguous image-level and series-level tags. This conflict is resolved in the same way as described herein in the Normalization of DICOM Parameters section. This means that values of DICOM parameters, Image Content Based Parameters, but also Abstract Tags, are automatically taken from some reference image. Alternatively, for all DICOM parameters a list of distinct values occurring in all images of the image set can be used in an assignment rule.
Style Rules
In one embodiment of the present invention, there is a final set of rules that specify the rendering style and other rendering parameters to be used when showing a particular image set. For example, for a CT Angiogram study, often a volume rendering style display (VRT) is desired, whereas for a study looking for lung nodules a maximum intensity projection (MIP) of 20 mm slabs may be desired. Style rules, that can be user specific, allow driving that automatically. The rules can use the same parameters as discussed above, as well as the existence or absence of certain image sets.
In one embodiment of the present invention, the system uses a global, ordered list of style rules that is evaluated independently for each viewer and each image set loaded into a viewer. An abstract tag DisplaySetID is provided that allows formulating style rules for a specific viewer or group of viewers.
Parameters driven by Style Rules include the following:
    • Rendering style (can be 2D, oblique, curved, MIP slab, 3D MIP, VRT, shaded VRT, etc.);
    • Image alignment (left, right, top, bottom, centered);
    • Inverse display (black on white versus white on black);
    • Colormap or transfer function;
    • Window/level (data window);
    • Slice thickness;
    • Zoom factor;
    • Camera position and orientation; and
    • Labels/OverlayDisplay of labels, annotations and other overlay elements.
The following is an example of a style rule that activates inverse 3D MIP rendering in all viewers with a DisplaySetID between 101 and 104, provided a PET data set is loaded into those viewers (modality PT, i.e., positron emission tomography). Also, an automatic window/level setting is used that is computed from the histogram of the image set (the 2% lowest values are all mapped to white, and the 2% highest values are all mapped to black):
Style Rule 1:
    • IF (Abstract.DisplaySetID>100 and
      • Abstract.DisplaySetID<105 and
      • Dicom.Modality=“PT”)
    • THEN SET
      • RenderingStyle:=“3D MIP”
      • Inverse:=true
      • DataWindow:=“2% 98%”
The following is another example of a different style rule that always causes the image set with image set ID 200 to be displayed in MPR mode using 20 mm thick slices, with a window/level as specified in the DICOM parameters, and with a zoom factor so that the whole viewer window is filled out. The rule is as follows.
Style Rule 2:
    • IF (Abstract.ImageSetID=200)
    • THEN SET
      • RenderingStyle:=“MPR”
      • SliceThickness:=“20”
      • DataWindow:=“DICOM1”
      • ZoomFactor:=“FitToWindow”
Summary of Rule Types
Table I summarizes all types of rules that are applied in the rule-base display system:
TABLE I
Normalized
Rule Type Applies to Parameters Defined where
Study Selection Rule Studies yes globally
Protocol Selection Rule Studies yes globally
Image Set Rule Images not required protocol
Viewer Assignment Rule Image Sets yes globally, protocol
Style Rule Image Sets yes globally, protocol
Described above are methods and systems for implementing a rule derived basis to display image sets. The foregoing description of embodiments of the methods, systems, and components of the present invention has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations will be apparent to one of ordinary skill in the relevant arts. For example, steps performed in the embodiments of the invention disclosed can be performed in alternate orders, certain steps can be omitted, and additional steps can be added. The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications that are suited to the particular used contemplated. Other embodiments are possible and are covered by the invention. Such embodiments will be apparent to persons skilled in the relevant art(s) based on the teachings contained herein. The breadth and scope of the present invention should not be limited by any of the above-described exemplary embodiments, but should be defined only in accordance with the following claims and their equivalents.
Example Shown in FIG. 2
An example of how these aspects can be combined is shown in FIG. 2 . In the example the user has selected a CT examination of the abdomen. The following rules have been used to determine that a recent X-Ray of the chest is relevant and shall be displayed as well:
    • IF (Primary.Dicom.BodyPartExamined=“ABDOMEN” and Primary.Dicom.Modality=“CT”)
    • THEN SELECT other studies for loading WHERE (Other.Dicom.BodyPartExamined=“ABDOMEN” OR Other.Dicom.BodyPartExamined=“CHEST”) and
    • (Other.Dicom.Modality=“CR” or Other.Dicom.Modality=“CT”) AND Other.RelativeStudyAge<“90 days”
From this rule, a hanging protocol can be selected. In the example the protocol selection rules determine that the CT study is a thin slice CT study (i.e. that it has image series that form a 3D volume with sufficient resolution in all directions to display volume rendering or non-axial slices in a meaningful way). Furthermore the example rule determines that this is a study with enhanced vasculature, by looking for the key words “contrast” or “angio” in the study description.
The display protocol selection rule that applies here and select the protocol CTThinSliceVesselWithPrior can read:
    • IF (Primary.Dicom.BodyPartExamined=“ABDOMEN” and Primary.Dicom.Modality=“CT” and Primary.Abstract.HasThinSliceVolumes and (Primary.Dicom.StudyDescription containsAnyOf “contrast, angio” and exists Other1 THEN SELECT “CTThinSliceVesselWithPrior” with score=10.
From this image sets are generated using Image Set Rules:
    • IF (Dicom.Modality=“CT” and Abstract.PriorIndex=0 and Condition.IsPartOfThinSliceVolume and Condition.CTSoftTisseKernel)
    • THEN CREATE image set with ID 1.x
      • SORTED BY Abstract.NumberOfSlicesInVolume ORDER:=descending SPLIT:=true
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.Abstract.VolumeIndex ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.Abstract.SlicePosition ORDER:=ascending SPLIT:=false.
This rule will actually form sets from images that contain images that are part of a ThinSliceVolume and that have been reconstructed with a “soft tissue” kernel. Given the protocol selection rule has specifically matched for just CT studies, the conditions Dicom.Modality=“CT” and Abstract.PriorIndex=0 are actually redundant, but could be useful if a different selection rule was used.
The images will first be sorted by the size of the volume of which they are part (Abstract.NumberOfSlicesInVolume), then by DICOM series. The split parameter in this case will ensure that an image set contains images from on series only. A DICOM series can sometimes contain multiple non-consecutive volumes. The abstract tag VolumeIndex will then indicate for each image, which of those volumes it is part of. If a series contains only one volume, then this will be “1” for all images in the series. The split=true in this part of the rule would result in a separate image set for each of those volumes. Finally, within each volume, the images are ordered by slice position, but not split. This way we end up with one image set for each soft kernel thin slice volume, the largest volume being the first image set (ID 1.1). This ID will be used further in subsequent rules.
The rule to form an image set from any CR prior study in this example is much simpler:
    • IF (Dicom.Modality=“CR” and Abstract.PriorIndex=1)
    • THEN CREATE image set with ID 10
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=false
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
This creates one image set with ID=10 containing all images for the first prior study, if that is a CR.
In practice, additional rules, such as Image Set Rule 5 and 6 (see above) will be used to collect the remaining images of the primary Study 1105. The remaining images are not shown in the layout depicted in the example FIG. 2 .
The Display Protocol 1145 contains multiple layouts. The one shown in FIG. 2 is defined as follows:
DEFINE Layout {
ID=“Layout5”;
NAME=”+PlainFilm”
Viewports {
{ ID=50, Geometry=”(0,0)-(0.25,0.5)”},
{ ID=51, Geometry=”(0.25,0)-(0.5,0.5)”},
{ ID=52, Geometry=”(0,0.5)-(0.25,1)”},
{ ID=53, Geometry=”(0.25,0.5)-(0. 5,0.5)”},
{ ID=54, Geometry=”(0.5,0)-(1,1)”, Style=”2D”}
}
}
In this example the geometry is defined in a coordinate system having the origin in the upper left corner of the screen with the x axis pointing to the right and the y axis pointing down. Please note how parameters of the viewers can be set in the layout definition. Parameters can also be set or overridden in the assignment and style rules, as will be explained next.
In this example, viewer assignment and style rules are as follows:
IF ImageSetExists (1.1) and ImageSetExists(10) THEN
SHOW_LAYOUT Layout5 WITH
 Viewport[0].AddImageSet(1.1)
 Viewport [0].Style=”VRT(diffuse)”
 Viewport [0].Colormap=”CTAngioSoftTissue”
 Viewport [1,2,3].AddImageSet(1.1)
 Viewport [1,2,3].Style=”MPR”
 Viewport [1,2,3].DataWindow=”DICOM1”
 Viewport [1].oriantation=”axial”
 Viewport [2].oriantation=”sagittal”
 Viewport [3].oriantation=”coronal”
 Viewport [4].AddImageSet(10)
 IF (ImageSet[10].Dicom.Columns > 1024) THEN
 Viewport[4].Zoom=”FitToWindow”
ELSE
 Viewport[4].Zoom=”1:1”
In this particular example, the rule to select the layout is rather simple: It is shown if the two image sets used exist. This is because the criteria to construct these images sets have been rather specific. As will be appreciated, the proposed system gives this type of flexibility.
Aspects of the Invention
Some aspects of this invention include methods of displaying one or more Sets of Images comprising the steps of:
    • a. selecting a primary Study;
    • b. selecting one or more Study Selection Parameters based on the primary Study;
    • c. selecting one or more Study Selection Rules based on the one or more Study Selection Parameters;
    • d. selecting one or more Sets of Images from a plurality of images based on the one or more Study Selection Rules;
    • e. selecting one or more Display Protocol Selection Parameters based on the one or more Sets of Images selected;
    • f. selecting one or more Display Protocol Selection Rules based on the one or more Display Protocol Selection Parameters;
    • g. selecting one or more Display Parameters using the one or more Display Protocol Selection Rules; and
    • h. displaying the one or more Sets of Images according to the Display Parameters.
Additional aspects include methods one or more Display Parameter are selected from the group consisting of Image Set Selection Parameters and View and Viewport Selection Parameters.
Further aspects include methods where the one or more Display Parameters are selected from the group consisting of Image Set Selection Rules, View and Viewport Selection Rules, and Display Protocol Selection Rules.
Yet further aspects include methods where the step of identifying one or more Image Set Selection Rules is based on the one or more Image Set Selection Parameters.
Still further aspects include methods where the step of selecting one or more Viewpoint Selection Rules is based on one or more View and Viewport Selection Parameters.
Other aspects include methods where the step of displaying the one or more Sets of Images is based on one or more Display Protocol Selection Rules, one or more Image Set Selection Rules, and one or more View and Viewport Selection Rules.
Still other aspects include methods where one or more of the Study Selection Parameters are selected from the group consisting of DICOM parameters, Image Content Based Parameters, and Abstract Tags.
Other aspects include methods where one or more of the Display Protocol Selection Parameters are selected from the group consisting of DICOM parameters, Image Content Based Parameters, and Abstract Tags.
Additional aspects include methods where one or more of the Image Set Selection Parameters are selected from the group consisting of DICOM parameters, Image Content Based Parameters, and Abstract Tags.
Further aspects include methods where one or more of the View and Viewport Selection Parameters are selected from the group consisting of DICOM parameters, Image Content Based Parameters, and Abstract Tags.
More aspects include methods where one or more Study Selection Parameters are derived from a single reference image.
Still more aspects include methods where one or more Study Selection Parameters are derived from a single reference image DICOM Parameters.
Yet other aspects include methods where one or more Display Protocol Selection Parameters are derived using a list of all values of a DICOM parameter occurring in any of the one or more Sets of Images.
Alternative aspects include methods where the one or more View and Viewport Selection Rules contain protocols for one or more Viewports displaying images as 2D.
Other alternative aspects include methods where the one or more View and Viewport Selection Rules contain protocols for one or more Viewports displaying images in a 3D rendering mode.
Further alternative aspects include methods where one or more Study Selection Parameters include one or more Abstract Tags selected from the group consisting of RelativeStudyAge, PriorIndex. NumImages, NumSeries, Num3DVolumes, Num4DSequences and HasReport.
In other aspects, this invention includes methods where one or more View and Viewport Selection Rules include one or more Abstract Tags selected from the group consisting of Image Sets to be displayed, Rendering Style, Additional image sets for image fusion, Image Alignment, Colormap/Transfer Function, Slice Thickness, Zoom Factor, Camera position, Camera orientation and Labels/Overlay elements or one or more Image Content Based Parameters.
In still other aspects, this invention includes methods further comprising the steps of:
    • receiving one or more Sets of Images based on the Study Selection Rules;
    • selecting one or more Image Set Selection Parameters;
    • selecting one or more Image Set Selection Rules based on the one or more Image Set Selection Parameters; and
    • displaying the one or more Sets of Images based on the Display Protocol Selection Rules and the Image Set Selection Rules.
In another aspect, this invention includes methods of displaying one or more Sets of Images comprising the steps of:
    • selecting one or more Study Selection Parameters;
    • selecting or more Study Selection Rules based on the one or more Study Selection Parameters;
    • receiving one or more Sets of Images based on the Study Selection Rules;
    • selecting one or more Display Protocol Selection Parameters based on the one or more Sets of Images selected;
    • selecting one or more Display Protocol Selection Rules based on the one or more Display Protocol Selection Parameters; and
    • displaying the one or more Sets of Images based on the Display Protocol Selection Rules.
Another aspect of this invention includes methods of displaying images comprising the steps of:
    • a. selecting one or more Study Selection Parameters;
    • b. selecting Study Selection Rules based on the one or more Study Selection Parameters;
    • c. receiving one or more images based on the Study Selection Rules;
    • d. selecting one or more Display Protocol Selection Parameters based on the one or more images selected;
    • e. selecting Display Protocol Selection Rules based on the one or more Display Protocol Selection Parameters;
    • f. selecting one or more Image Set Selection Parameters;
    • g. selecting Image Set Selection Rules based on the one or more Image Set Selection Parameters;
    • h. selecting one or more View and Viewport Selection Parameters;
    • i. selecting View and Viewport Selection Rules based on the one or more View and Viewport Selection Parameters; and
    • j. displaying the one or more images based on the Display Protocol Selection Rules, the Image Set Selection Rules and the View and Viewport Selection Rules.
Other aspects of the invention include methods where the Study Selection Rule is:
    • IF (Primary.Dicom.BodyPartExamined=“CHEST” and Primary.Dicom.Modality=“CR”)
    • THEN SELECT other studies for loading WHERE (Other.Dicom.BodyPartExamined=“CHEST” and (Other.Dicom.Modality=“CR” or Other.Dicom.Modality=“CT”)).
In another aspect, this invention includes methods where the Study Selection Rule is:
    • IF (Primary.Dicom.BodyPartExamined=“CHEST” and Primary.DicomList.Modality contains “CR”)
    • THEN SELECT other studies for loading WHERE (Other.Dicom.BodyPartExamined=“CHEST” and (Other.DicomList.Modality contains “CR” or Other.DicomList.Modality contains “CT”)).
In other aspects, this invention includes methods where the Study Selection Rule is:
    • IF (Primary.Dicom.Modality=“MG”
    • THEN SELECT other studies for loading WHERE (Other.Dicom.Modality=“MG” and Other.Abstract.PriorIndex<=3 and Other.Abstract.RelativeStudyAge<5*365).
In yet another aspect, this invention includes methods where the Protocol Selection Rule is:
    • IF (Primary.Dicom.BodyPartExamined=“ABDOMEN” and Primary.Dicom.Modality=“CT” and Exists(Other1) and Other1.Dicom.Modality=“PET”)
    • THEN SELECT “StandardPetCTProtocol1” with score=10.
In aspects of the invention, methods include an Image Set Rule:
    • IF (Dicom.Modality=“CT” and Abstract.PriorIndex=0)
    • THEN CREATE image set with ID 1.
Additionally, other aspects include methods where the Image Set Rule is:
    • IF (Dicom.Modality=“CT” and Abstract.PriorIndex=0)
    • THEN CREATE image set with ID 1
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending.
Still other aspects include methods where the Image Set Rule is:
    • IF (Dicom.Modality=“CT” and Abstract.PriorIndex=0)
    • THEN CREATE image set with ID 1.x
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
Moreover, other aspects include methods where the Image Set Rule is:
    • IF (Dicom.Modality=“CT” and Abstract.PriorIndex=0)
    • THEN CREATE image set with ID 1.x
      • SORTED BY Condition.CTSoftTisseKernel SPLIT:=true
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
Yet other aspects include methods where the Image Set Rule (Localizer) is:
    • IF (Dicom.Modality=“CT” and Condition.IsLocalizer=true)
    • THEN CREATE image set with ID 1
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
Other aspects of the methods of this invention include an Image Set Rule (Images):
    • IF (Dicom.Modality=“CT” and Abstract.AlreadyReferenced=false)
    • THEN CREATE image set with ID 2
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
Yet other aspects of the methods of this invention include using Image Set Rule (Images):
    • IF (Dicom.Modality=“CT” and Abstract.AlreadyReferenced=false)
    • THEN CREATE image set with ID 2
      • SORTED BY Dicom.SeriesNumber ORDER:=ascending SPLIT:=true
      • SORTED BY Dicom.InstanceNumber ORDER:=ascending SPLIT:=false.
Additionally, other aspects include methods where the Display Parameters include Viewer Assignment Rule:
    • IF (Abstract.ImageSetID=1)
    • THEN SELECT image set with score=10.
Yet further aspects include methods where the Display Parameters include a Viewer Assignment Rule:
    • IF (Abstract.ImageSetID=2)
    • THEN SELECT image set with score=5.
Additional aspects include methods further comprising a Viewer Assignment Rule:
    • IF (Abstract.ImageSetID=2)
    • THEN SELECT image set with score=5.
Further Embodiments
Further embodiments contemplated herein include the following.
In an aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic in the primary study and a ParameterB as Modality in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParameterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the one or more ImageContentBased parameters identified are present in the primary study. In an embodiment, the one or more ImageContentBased parameters are vertebrae.
In an embodiment, the vertebrae in the primary study are selected from the group consisting of L1, L2, L3, L4, and L5, and at least one of the vertebra in the secondary study is a vertebra present in the primary study. In an embodiment, the vertebrae in the primary study are selected from the group consisting of C1, C2, C3, C4, C5, C6, and C7, and at least one of the vertebra in the secondary study is a vertebra present in the primary study. In an embodiment, the vertebrae in the primary study are selected from the group consisting of Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12, and at least one of the vertebra in the secondary study is a vertebra present in the primary study. In an embodiment, the CNN is pretrained with the plurality of studies. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more of the ImageContentBased parameters.
In an embodiment, the method further includes the CNN selecting based on psueudo code:
    • IF (Primary.Dicom.AnatomicalCharacteristic=“SPINE” and
    • Primary.Dicom.Modality=“MR”)
      THEN SELECT other studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“SPINE” and Other.Dicom.Modality=“MR” AND INTERSECTION(Primary.Vertebrae,Other.Vertebrae) NOT EMPTY).
In an embodiment, the one or more Study Selection Rules restrict the final list to studies of the patient.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA from the group consisting of an anatomical characteristic and a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the method further includes selecting a ParameterB, where the plurality of secondary studies exclude one or more based on ParameterB. In an embodiment, ParameterB is Modality. In an embodiment, ParameterB is selected from the group consisting of Computer Tomography (CT), then the Modality in the two or more secondary studies inserted is selected from the group consisting of Computed Radiography (CR), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA). In an embodiment, ParameterB in the primary study is equal to the ParameterB in the secondary study.
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • SELECT other studies for loading, WHERE
      INTERSECTION(Primary.Dicom.Modality,Other.Dicom.Modality) NOT EMPTY.
In an embodiment, ParameterA is BodyPartExamined. In an embodiment, ParameterA is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULA, KNEE, TIBULA, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
In an embodiment, the one or more Study Selection Rules restrict the final list to studies of the patient. In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • SELECT other studies for loading, WHERE
      INTERSECTION(Primary.Dicom.Modality,Other.Dicom.Modality) NOT EMPTY.
In an embodiment, the ImageContentBased parameter is stored as a private DICOM tag.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease-based characteristic in the primary study and selecting as a ParameterB a Modality in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParamaterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the one or more ImageContentBased parameters identified are present in the primary study. In an embodiment, the one or more ImageContentBased parameters are vertebrae. In an embodiment, the vertebrae in the primary study are selected from the group consisting of L1, L2, L3, L4, and L5, and at least one of the vertebrae in the secondary study is a vertebra present in the primary study. In an embodiment, the vertebrae in the primary study are selected from the group consisting of C1, C2, C3, C4, C5, C6, and C7, and at least one of the vertebrae in the secondary study is a vertebra present in the primary study. In an embodiment, the vertebrae in the primary study are selected from the group consisting of Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12, and at least one of the vertebrae in the secondary study is a vertebra present in the primary study. In an embodiment, the CNN is pretrained with the plurality of studies. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the disease-based characteristic in the primary study. In an embodiment, the CNN selects one or more secondary studies which show the same anatomy. In an embodiment, the one or more Study Selection Rules restrict the final list to studies of the patient.
In another aspect there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic in the primary study and selecting as a ParameterB a Modality in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParamaterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
In an embodiment, the method further includes the Study Selection Rule in step (c)(i)
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB”).
In an embodiment, the method further includes where in step (c)(ii) the CNN generates Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, and the Study Selection Rule in step (c)(i) and in step (c)(iii)
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(ParameterC,ParameterD) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
      • THEN SELECT other studies for loading
        WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality=“ParameterB” AND
        INTERSECTION(Primary.ImageContentBasedParameter,Other.ImageContentBasedParameter) NOT EMPTY).
In an embodiment, the one or more ImageContentBased parameters are vertebrae. In an embodiment, the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12. In an embodiment, the CNN is pretrained with the plurality of studies. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA”).
In an embodiment, the method further includes where in step (c)(ii) the CNN generates Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, and a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND INTERSECTION(ParameterC,ParameterD) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
      • THEN SELECT other studies for loading
        WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND
        INTERSECTION(Primary.ImageContentBasedParameter,Other.ImageContentBasedParameter) NOT EMPTY).
In an embodiment, the method further includes where a ParameterB is a Modality in the primary study, where in step (c) the one or more Study Selection Rules restrict to studies where the Modality in the final list is equal to ParameterB.
In an embodiment, a ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB”).
In an embodiment, the method further includes where in step (c)(ii) the CNN generates Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, and a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) require
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(ParameterC,ParameterD) NOT EMPTY).
In an embodiment, the method further includes wherein a Study Selection Rule of the one or more Study Selection Rules requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.ImageContentBasedParameter,Other.ImageContentBasedParameter) NOT EMPTY).
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters;
(d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA”),
    • THEN SELECT other studies for loading WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA”).
In an embodiment, the method further includes where in step (c)(ii) the CNN generates Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, and a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) require
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA”),
    • THEN SELECT other studies for loading WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” AND INTERSECTION(ParameterC,ParameterD) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
    • THEN SELECT other studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND
      INTERSECTION(Primary.ImageContentBasedParameter,Other.ImageContentBasedParameter) NOT EMPTY).
In an embodiment, the method further includes where a ParameterB is a Modality in the primary study, where in step (c) the one or more Study Selection Rules restrict to studies where the Modality in the final list is equal to ParameterB.
In an embodiment, a ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other studies for loading WHERE
      (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB”).
In an embodiment, the method further includes where in step (c)(ii) the CNN generates Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, and a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) require
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other studies for loading WHERE
      (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality=“ParameterB” AND INTERSECTION(ParameterC,ParameterD) NOT EMPTY).
In an embodiment, the method further includes the Study Selection Rule
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality==“ParameterB” AND
      INTERSECTION(Primary.ImageContentBasedParameter,Other.ImageContentBasedParameter) NOT EMPTY).
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterB a Modality in the primary study and selecting as a ParameterA from the group consisting of an anatomical characteristic and a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParamaterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the one or more ImageContentBased parameters are vertebrae. In an embodiment, the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12. In an embodiment, the CNN is pretrained with the plurality of studies. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study. In an embodiment, the one or more ImageContentBased parameters are vertebrae. In an embodiment, the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12. In an embodiment, the CNN is pretrained with the plurality of studies. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study. In an embodiment, the one or more ImageContentBased parameters are vertebrae. In an embodiment, the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12. In an embodiment, the CNN is pretrained with the plurality of studies. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an AnatomicalCharacteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and one or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the AnatomicalCharacteristic in the primary study; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (e) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE. In an embodiment, the one or more parameters include a ParameterB a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (c) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
In an embodiment, the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA). In an embodiment, when the ParameterB is Computed Radiography (CR) then the Modality in the two or more of the plurality of secondary studies is selected from the group consisting of Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“CR”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and (Other.Dicom.Modality=“CR”) or Other.Dicom.Modality=“CT”).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterZ”) AND INTERSECTION (ParameterB, ParameterZ) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterY”) AND INTERSECTION (ParameterB, ParameterY) EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and (Other.Dicom.Modality=NOT “ParameterB” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In an embodiment, the AnatomicalFeature is a medical condition. In an embodiment, the medical condition is selected from the group consisting of a fracture and a bleeding. In an embodiment, the AnatomicalFeature is a notation stored as meta data. In an embodiment, the AnatomicalFeature includes whether an organ is present in a study. In an embodiment, the ParameterA is selected from meta data stored in the primary Study.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and one or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the disease based characteristic in the primary study; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (e) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the one or more parameters include a ParameterB a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (c) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • SELECT other secondary studies for loading
      WHERE (Primary.Dicom.Modality=“ParameterB” and (Other.Dicom.Modality=“ParameterY” AND INTERSECTION(ParameterB,ParameterY) NOT EMPTY).
In an embodiment, the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA” and Primary.Dicom.Modality=“CR”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” and (=“CR” or Other.Dicom.Modality=“CT”)).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterB” or Other.Dicom.Modality=“ParameterZ”)).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterY”)
      WHERE Intersection (ParameterB, ParameterY) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“DISEASE 1” and
    • Primary.Dicom.Modality=“IMAGE TYPE 1”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.DiseaseCharacteristic=“DISEASE 1” and Other.Dicom.Modality=“IMAGE TYPE 2”).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“SpinalDegradation” and Primary.Dicom.Modality=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“SpinalDegradation” and Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In an embodiment, the AnatomicalFeature is a medical condition. In an embodiment, the medical condition is selected from the group consisting of a fracture and a bleeding. In an embodiment, the AnatomicalFeature is a notation stored as meta data. In an embodiment, the AnatomicalFeature includes whether an organ is present in a study. In an embodiment, the ParameterA is selected from meta data stored in the primary Study.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA from the primary study, where the ParameterA is selected from the group consisting of an anatomical characteristic, an anatomical region imaged, and a disease based characteristic; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and one or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the ParameterA selected;
(d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (e) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, ParameterA is the anatomical region imaged selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE. In an embodiment, the one or more parameters include a ParameterB a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (c) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • SELECT other secondary studies for loading
      WHERE (Primary.Dicom.Modality=“ParameterB” and (Other.Dicom.Modality=“ParameterY”
      AND INTERSECTION(ParameterB,ParameterY) NOT EMPTY).
In an embodiment, the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“CR”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“CR” or (Other.Dicom.Modality=“CT”)).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterB” or Other.Dicom.Modality=“ParameterZ”) WHERE Intersection (ParameterB, ParameterZ) EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterY”) AND INTERSECTION (ParameterB, ParameterY) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“Spine” and Primary.Dicom.Modality=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“Spine” and Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In an embodiment, the AnatomicalFeature is a medical condition. In an embodiment, the medical condition is selected from the group consisting of a fracture and a bleeding. In an embodiment, the AnatomicalFeature is a notation stored as meta data. In an embodiment, the AnatomicalFeature includes whether an organ is present in a study. In an embodiment, the ParameterA is selected from meta data stored in the primary Study.
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“CR”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” and Other.Dicom.Modality=“CR” or Other.Dicom.Modality=“CT”).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterB” or Other.Dicom.Modality=“ParameterZ”) WHERE Intersection (ParameterB, ParameterZ) EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterY” WHERE Intersection (ParameterB, ParameterY) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA” and
    • Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB”).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“SpinalDegradation” and Primary.Dicom.Modality=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“SpinalDegradation” and Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In an embodiment, the AnatomicalFeature is a medical condition.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA from the primary study, where the ParameterA is selected from the group consisting of an anatomical characteristic; (c) selecting ParameterB selected from the group consisting of one or more of regions of the anatomical characteristic, parts of a skeletal system of the anatomical characteristic, and organs of the anatomical characteristic; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (f) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (e).
In an embodiment, the anatomical characteristic is selected from the group consisting of HEAD, EYE, EAR, NOSE, NOSTRIL, MOUTH, LIP, PHILTRUM, JAW, MANDIBLE, GINGIVA, TOOTH, TONGUE, THROAT, LARYNGEAL PROMINENCE, VERTEBRAL COLUMN, SCAPULA, HUMERUS, ELBOW, RADIUS, ULNA, CARPUS, METACARPUS, PHALANGES, THUMB, NAILS, THORAX, BREAST, ABDOMEN, PENIS, SCROTUM, VULVA, LEG, FEMUR, KNEE, PATELLA, TIBIA, SURA, TALOCRURAL REGION, METATARSUS, PHALANGES PROXIMALES, PHALANGES MEDIAE, AND PHALANGES DISTALES.
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.RegionAnatomicalCharacteristic=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.OrganAnatomicalCharacteristic=“ParameterC”).
In an embodiment, one or both the ParameterA and the ParameterB are selected from meta data stored in the primary Study.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies, where the primary Study is an image measured after the patient has a contrast agent administered; (b) selecting a ParameterA an AnatomicalCharacteristic from the primary study; (c) selecting ParameterB a Modality from the primary study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient before the contrast agent was administered selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (f) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (e).
In an embodiment, the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE. In an embodiment, the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“CR”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“CR” or Other.Dicom.Modality=“CT”).
In an embodiment, one or both the ParameterA and the ParameterB are selected from meta data stored in the primary Study.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA an AnatomicalCharacteristic from the primary study; (c) selecting ParameterB selected from the group consisting of one or more anatomical characteristics of the AnatomicalCharacteristic, parts of a skeletal system of the AnatomicalCharacteristic, organs of the AnatomicalCharacteristic and Modality of the primary Study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (f) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (e).
In an embodiment, the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE. In an embodiment, the ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“SPINE” and Primary.Dicom.AnatomicalFeature=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“SPINE” and Other.Dicom.AnatomicalFeature=“ParameterB” AND
      INTERSECTION(Primary.AnatomicalFeature,Other.AnatomicalFeature) NOT EMPTY).
In an embodiment, the ParameterA is selected from meta data stored in the primary Study.
In another aspect, there is provided a including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting a ParameterA from the primary study, where the ParameterA is selected from the group consisting of an anatomical characteristic, and a disease based characteristic; (c) selecting ParameterB from the group consisting of one or more anatomical features of the ParameterA, parts of a skeletal system of the ParameterA, organs of the ParameterA, and a Modality of the primary Study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (f) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (e).
In an embodiment, the anatomical characteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE. In an embodiment, the ParameterB is a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (d) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • SELECT other secondary studies for loading
      WHERE (Primary.Dicom.Modality=“ParameterB” and (Other.Dicom.Modality=“ParameterY” AND INTERSECTION(ParameterB,ParameterY) NOT EMPTY).
In an embodiment, the ParameterB is a Modality selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB”).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterY” AND INTERSECTION (ParameterB, ParameterY) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic imaged; (c) selecting ParameterB from the group consisting of one or more anatomical features of the ParameterA, parts of a skeletal system of the ParameterA, organs of the ParameterA, and a Modality of the primary Study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (f) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (e).
In an embodiment, the anatomical characteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE. In an embodiment, the ParameterB is a Modality in the primary study and a ParameterY a Modality in a secondary study, where in step (d) the one or more Study Selection Rules restrict to studies where ParameterB is equal to ParameterY.
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • SELECT other secondary studies for loading
      WHERE (Primary.Dicom.Modality=“ParameterB” and (Other.Dicom.Modality=“ParameterY” AND INTERSECTION(ParameterB,ParameterY) NOT EMPTY).
In an embodiment, the ParameterB is a Modality selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB”).
In an embodiment, the method further includes the Study Selection Rule
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterY” AND INTERSECTION (ParameterB, ParameterY) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and (Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease based characteristic; (c) selecting ParameterB from the group consisting of one or more anatomical features of the ParameterA, parts of a skeletal system of the ParameterA, organs of the ParameterA, and a Modality of the primary Study; (d) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to generate a list of a plurality of secondary studies based on the ParameterA and the ParameterB, where the one or more Study Selection Rules restrict the plurality of secondary studies to studies of the patient selected from the plurality of studies; (e) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies and one or more Abstract Tags from the plurality of secondary studies; and (f) displaying two or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (e).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (d) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”)
    • THEN SELECT other secondary studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND
      INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA an anatomical characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the AnatomicalCharacteristic is selected from the group consisting of SPINE, CHEST, ABDOMEN, BREAST, SHOULDER, TRAPEZIUS, ARM, ELBOW, WRIST, FINGER, PELVIS, HIP, FIBULAR, KNEE, TIBULAR, ANKLE, FOOT, NECK, HEAD, TEMPOROMANDIBULAR JUNCTION, FACE, BRAIN, DENTITION, SINUS, ADRENALS, RETINA, PITUITARY, and PROSTATE.
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA”).
In an embodiment, the method further includes where in step (c)(ii) the CNN generates Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, and a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) require
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND INTERSECTION(ParameterC,ParameterD) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
    • THEN SELECT other studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND
      INTERSECTION(Primary.ImageContentBasedParameter,Other.ImageContentBasedParameter) NOT EMPTY).
In an embodiment, the method further includes where a ParameterB is a Modality in the primary study, where in step (c) the one or more Study Selection Rules restrict to studies where the Modality in the final list is equal to ParameterB.
In an embodiment, a ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB”).
In an embodiment, the method further includes where in step (c)(ii) the CNN generates Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, and a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) require
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
      THEN SELECT other studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(ParameterC,ParameterD) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.ImageContentBasedParameter,Other.ImageContentBasedParameter) NOT EMPTY).
In an embodiment, the one or more ImageContentBased parameters are vertebrae. In an embodiment, the CNN identifies vertebrae in the primary study selected from the group consisting of L1, L2, L3, L4, L5, C1, C2, C3, C4, C5, C6, C7, Th1, Th2, Th3, Th4, Th5, Th6, Th7, Th8, Th9, Th10, Th11, and Th12. In an embodiment, the CNN is pretrained with the plurality of studies. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on the anatomical characteristic in the primary study. In an embodiment, the CNN is pretrained with a first plurality of studies where the first plurality of studies is selected based on one or more ImageContentBased parameters identified in the primary study.
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterA a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).
In an embodiment, the method further including where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA”),
    • THEN SELECT other studies for loading WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA”).
In an embodiment, the method further includes where in step (c)(ii) the CNN generates Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, and a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) require
    • IF (Primary.Dicom.DiseaseCharacteristic=“ParameterA”),
    • THEN SELECT other studies for loading WHERE (Other.Dicom.DiseaseCharacteristic=“ParameterA” AND INTERSECTION(ParameterC,ParameterD) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
    • THEN SELECT other studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND
      INTERSECTION(Primary.ImageContentBasedParameter,Other.ImageContentBasedParameter) NOT EMPTY).
In an embodiment, the method further includes where a ParameterB is a Modality in the primary study, where in step (c) the one or more Study Selection Rules restrict to studies where the Modality in the final list is equal to ParameterB.
In an embodiment, a ParameterB is selected from the group consisting of Computed Radiography (CR), Computer Tomography (CT), Digital Radiography (DX), Mammography (MG), Magnetic Resonance (MR), Opthalmic Photography (OP), Positron Emission Tomography (PT), Radio Fluoroscopy (RF), and X-Ray Angiography (XA).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) require
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other studies for loading WHERE
      (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterB”).
In an embodiment, the method further including where in step (c)(ii) the CNN generates Primary.ImageContentBasedParameter=“ParameterC” and Other.ImageContentBasedParameter=“ParameterD”, and a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) and in step (c)(iii) require
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality=“ParameterB” AND INTERSECTION(ParameterC,ParameterD) NOT EMPTY).
In an embodiment, the method further includes where a Study Selection Rule of the one or more Study Selection Rules in step (c)(i) requires
    • IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterB”),
    • THEN SELECT other studies for loading
      WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” AND Other.Dicom.Modality=“ParameterB” AND
      INTERSECTION(Primary.ImageContentBasedParameter,Other.ImageContentBasedParameter) NOT EMPTY).
In another aspect, there is provided a method including: (a) selecting a primary Study of a patient selected from a plurality of studies; (b) selecting as a ParameterB a Modality in the primary study and selecting as a ParameterA from the group consisting of an anatomical characteristic and a disease based characteristic in the primary study; (c) executing on a server digital data processor a render server program which applies one or more Study Selection Rules to: (i) generate a list of a plurality of secondary studies based on ParameterA and ParamaterB; (ii) generate from the list of the plurality of secondary studies one or more ImageContentBased parameters using Convolutional Neural Networks (CNN); (iii) select from the list of the plurality of secondary studies a final list based on the one or more ImageContentBased parameters; (d) executing on the server digital data processor the render server program which applies one or more Protocol Selection Rules to select a Display Protocol, where the one or more Protocol Selection Rules are based on two or more parameters selected from the group consisting of one or more DICOM parameters from the primary Study, one or more Abstract Tags from the primary Study, one or more DICOM parameters from the plurality of secondary studies, one or more Abstract Tags from the plurality of secondary studies and one or more ImageContentBased parameters; and (e) displaying the primary study and one or more of the plurality of secondary studies selected from the list based on the Display Protocol selected in step (d).

Claims (20)

What is claimed is:
1. A method comprising:
(A) receiving at a server a ParameterA and a ParameterB from a client computer, where the ParameterA is an anatomical characteristic of a patient, where the ParameterB is a patient name, the server comprising a server digital data processor, and the client computer comprising a client digital data processor;
(B) executing a render server program on the server which applies one or more Study Selection Rules to generate one or more primary studies based on the ParameterA and the ParameterB, where the one or more primary studies each comprise one or more images, where a Study Selection Rule of the one or more Study Selection Rules requires
IF (Primary.Dicom. AnatomicalCharacteristic =“ParameterA”),
THEN SELECT other studies for loading
WHERE (Other.Dicom.AnatomicalCharacteristic =“ParameterA”;
(C) applying convolutional neural networks to the one or more images to identify one or more image content based parameters;
(D) executing the render server program on the server which applies the one or more image content based parameters to generate a list of a plurality of secondary studies; and
(E) sending from the server to the client computer for display on the client computer the list generated in step (D).
2. The method of claim 1, further comprising in step (E) sending the one or more image content based parameters identified.
3. The method of claim 1, where the ParameterA is selected from the group consisting of ABDOMEN, ADRENALS, ANKLE, ARM, BRAIN, BREAST, CERVICAL SPINE, CERVICAL VERTEBRAE, CHEST, DENTITION, ELBOW, FACE, FIBULAR, FINGER, FOOT, HEAD, HIP, KNEE, LUMBAR SPINE, LUMBAR VERTEBRAE, NECK, PELVIS, PITUITARY, PROSTATE, RADIUS, RETINA, SACRAL SPINE, SACRUM, SINUS, SHOULDER, SPINE, TEMPOROMANDIBULAR JUNCTION, TIBIA, THORACIC SPINE, THORACIC VERTEBRAE, TRAPEZIUS, ULNA, URETHRA, VERTEBRAE, and WRIST.
4. The method of claim 1, further comprising where ParameterA=“SpinalDegradation”.
5. The method of claim 1, further comprising receiving at the server a ParameterC from the client computer where the ParameterC is a Modality selected from the group consisting of Computed Radiography, Computer Tomography, Confocal Microscopy, Digital Radiography, histological image Mammography, Magnetic Resonance Imaging, Opthalmic Photography, Positron Emission Tomography, Radio Fluoroscopy, Ultrasound, 3-D Ultrasound, Vascular Visualization, X-Ray Angiography, and X-Ray Image.
6. The method of claim 1, where the one or more image content based parameters identify finer granular anatomic information.
7. A method comprising:
(A) receiving at a server a ParameterA and a ParameterB and a ParameterC from a client computer, where the ParameterA is an anatomical characteristic of a patient, where the ParameterB is a patient name, the server comprising a server digital data processor, and the client computer comprising a client digital data processor, where the ParameterC is a Modality selected from the group consisting of Computed Radiography, Computer Tomography, Confocal Microscopy, Digital Radiography, histological image Mammography, Magnetic Resonance Imaging, Opthalmic Photography, Positron Emission Tomography, Radio Fluoroscopy, Ultrasound, 3-D Ultrasound, Vascular Visualization, X-Ray Angiography, and X-Ray Image;
(B) executing a render server program on the server which applies one or more Study Selection Rules to generate one or more primary studies based on the ParameterA and the ParameterB, where the one or more primary studies each comprise one or more images;
(C) applying convolutional neural networks to the one or more images to identify one or more image content based parameters;
(D) executing the render server program on the server which applies the one or more image content based parameters to generate a list of a plurality of secondary studies; and
(E) sending from the server to the client computer for display on the client computer the list generated in step (D).
8. The method of claim 7, further comprising in step (E) sending the one or more image content based parameters identified.
9. The method of claim 7, where the ParameterA is selected from the group consisting of ABDOMEN, ADRENALS, ANKLE, ARM, BRAIN, BREAST, CERVICAL SPINE, CERVICAL VERTEBRAE, CHEST, DENTITION, ELBOW, FACE, FIBULAR, FINGER, FOOT, HEAD, HIP, KNEE, LUMBAR SPINE, LUMBAR VERTEBRAE, NECK, PELVIS, PITUITARY, PROSTATE, RADIUS, RETINA, SACRAL SPINE, SPINAL DEGRADATION, SACRUM, SINUS, SHOULDER, SPINE, TEMPOROMANDIBULAR JUNCTION, TIBIA, THORACIC SPINE, THORACIC VERTEBRAE, TRAPEZIUS, ULNA, URETHRA, VERTEBRAE, and WRIST.
10. The method of claim 7, further comprising where a Study Selection Rule of the one or more Study Selection Rules in step (B) requires
IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA”),
THEN SELECT other studies for loading
WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA”.
11. The method of claim 7, further comprising where a Study Selection Rule of the one or more Study Selection Rules in step (B) requires
SELECT other secondary studies for loading
WHERE (Primary.Dicom.Modality=“ParameterC” and (Other.Dicom.Modality=“ParameterC”.
12. The method of claim 7, further comprising where a Study Selection Rule of the one or more Study Selection Rules in step (B) requires
IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterC”),
THEN SELECT other secondary studies for loading
WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterC”).
13. The method of claim 7, further comprising where a Study Selection Rule of the one or more Study Selection Rules in step (B) requires
IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterC”)
THEN SELECT other secondary studies for loading WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterC” AND INTERSECTION(Primary.Dicom.AnatomicalFeature,Other.Dicom.AnatomicalFeature) NOT EMPTY).
14. The method of claim 7, where the one or more image content based parameters identify finer granular anatomic information.
15. A method comprising:
(A) receiving at a server a ParameterA and a ParameterB and a ParameterC from a client computer, where the ParameterA is an anatomical characteristic of a patient, where the ParameterB is a patient name, the server comprising a server digital data processor, and the client computer comprising a client digital data processor;
(B) executing a render server program on the server which applies one or more Study Selection Rules to generate one or more primary studies based on the ParameterA and the ParameterB, where the one or more primary studies each comprise one or more images, where a Study Selection Rule of the one or more Study Selection Rules requires
IF (Primary.Dicom.AnatomicalCharacteristic=“ParameterA” and Primary.Dicom.Modality=“ParameterC”),
THEN SELECT other secondary studies for loading
WHERE (Other.Dicom.AnatomicalCharacteristic=“ParameterA” and Other.Dicom.Modality=“ParameterC”);
(C) applying convolutional neural networks to the one or more images to identify one or more image content based parameters;
(D) executing the render server program on the server which applies the one or more image content based parameters to generate a list of a plurality of secondary studies; and
(E) sending from the server to the client computer for display on the client computer the list generated in step (D).
16. The method of claim 15, further comprising in step (E) sending the one or more image content based parameters identified.
17. The method of claim 15, where the ParameterA is selected from the group consisting of ABDOMEN, ADRENALS, ANKLE, ARM, BRAIN, BREAST, CERVICAL SPINE, CERVICAL VERTEBRAE, CHEST, DENTITION, ELBOW, FACE, FIBULAR, FINGER, FOOT, HEAD, HIP, KNEE, LUMBAR SPINE, LUMBAR VERTEBRAE, NECK, PELVIS, PITUITARY, PROSTATE, RADIUS, RETINA, SACRAL SPINE, SACRUM, SINUS, SHOULDER, SPINE, SPINAL DEGRADATION, TEMPOROMANDIBULAR JUNCTION, TIBIA, THORACIC SPINE, THORACIC VERTEBRAE, TRAPEZIUS, ULNA, URETHRA, VERTEBRAE, and WRIST.
18. The method of claim 15, further comprising where the INTERSECTION(ParameterB, ParameterC) NOT EMPTY).
19. The method of claim 15, where the one or more image content based parameters identify finer granular anatomic information.
20. The method of claim 15, where ParameterC is selected from the group consisting of Computed Radiography, Computer Tomography, Confocal Microscopy, Digital Radiography, histological image Mammography, Magnetic Resonance Imaging, Opthalmic Photography, Positron Emission Tomography, Radio Fluoroscopy, Ultrasound, 3-D Ultrasound, Vascular Visualization, X-Ray Angiography, and X-Ray Image.
US18/229,101 2013-03-15 2023-08-01 Method and system for rule based display of sets of images using image content derived parameters Active US12136164B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/229,101 US12136164B2 (en) 2013-03-15 2023-08-01 Method and system for rule based display of sets of images using image content derived parameters

Applications Claiming Priority (12)

Application Number Priority Date Filing Date Title
US13/831,975 US8976190B1 (en) 2013-03-15 2013-03-15 Method and system for rule based display of sets of images
US14/611,163 US9524577B1 (en) 2013-03-15 2015-01-30 Method and system for rule based display of sets of images
US15/380,848 US9898855B2 (en) 2013-03-15 2016-12-15 Method and system for rule based display of sets of images
US201762562460P 2017-09-24 2017-09-24
US201762576587P 2017-10-24 2017-10-24
US15/870,681 US10373368B2 (en) 2013-03-15 2018-01-12 Method and system for rule-based display of sets of images
US201862712912P 2018-07-31 2018-07-31
US16/052,571 US10540803B2 (en) 2013-03-15 2018-08-01 Method and system for rule-based display of sets of images
US16/531,436 US10832467B2 (en) 2013-03-15 2019-08-05 Method and system for rule based display of sets of images using image content derived parameters
US17/060,432 US11244495B2 (en) 2013-03-15 2020-10-01 Method and system for rule based display of sets of images using image content derived parameters
US17/574,975 US11763516B2 (en) 2013-03-15 2022-01-13 Method and system for rule based display of sets of images using image content derived parameters
US18/229,101 US12136164B2 (en) 2013-03-15 2023-08-01 Method and system for rule based display of sets of images using image content derived parameters

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US17/574,975 Continuation US11763516B2 (en) 2013-03-15 2022-01-13 Method and system for rule based display of sets of images using image content derived parameters

Publications (2)

Publication Number Publication Date
US20230386125A1 US20230386125A1 (en) 2023-11-30
US12136164B2 true US12136164B2 (en) 2024-11-05

Family

ID=80080596

Family Applications (3)

Application Number Title Priority Date Filing Date
US17/060,432 Active US11244495B2 (en) 2013-03-15 2020-10-01 Method and system for rule based display of sets of images using image content derived parameters
US17/574,975 Active US11763516B2 (en) 2013-03-15 2022-01-13 Method and system for rule based display of sets of images using image content derived parameters
US18/229,101 Active US12136164B2 (en) 2013-03-15 2023-08-01 Method and system for rule based display of sets of images using image content derived parameters

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US17/060,432 Active US11244495B2 (en) 2013-03-15 2020-10-01 Method and system for rule based display of sets of images using image content derived parameters
US17/574,975 Active US11763516B2 (en) 2013-03-15 2022-01-13 Method and system for rule based display of sets of images using image content derived parameters

Country Status (1)

Country Link
US (3) US11244495B2 (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8976190B1 (en) * 2013-03-15 2015-03-10 Pme Ip Australia Pty Ltd Method and system for rule based display of sets of images
US11183292B2 (en) 2013-03-15 2021-11-23 PME IP Pty Ltd Method and system for rule-based anonymized display and data export
US11244495B2 (en) * 2013-03-15 2022-02-08 PME IP Pty Ltd Method and system for rule based display of sets of images using image content derived parameters
JP7172796B2 (en) * 2019-03-28 2022-11-16 コニカミノルタ株式会社 Display system, display control device and display control method

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6765570B1 (en) 1998-07-21 2004-07-20 Magic Earth, Inc. System and method for analyzing and imaging three-dimensional volume data sets using a three-dimensional sampling probe
US20060149850A1 (en) 2005-01-05 2006-07-06 Control4 Corporation Method and apparatus for synchronizing playback of streaming media in multiple output devices
US20080052126A1 (en) 2006-08-25 2008-02-28 Konica Minolta Medical & Graphic, Inc. Database system, program, image retrieving method, and report retrieving method
US20080123917A1 (en) * 2006-11-29 2008-05-29 Siemens Medical Solutions Usa, Inc. An Imaging Study Completion Processing System
US20080166070A1 (en) * 2007-01-04 2008-07-10 General Electric Company Method for providing adaptive hanging protocols for image reading
US20130089248A1 (en) * 2011-10-05 2013-04-11 Cireca Theranostics, Llc Method and system for analyzing biological specimens by spectral imaging
US20140173287A1 (en) 2011-07-11 2014-06-19 Takeshi Mizunuma Identifier management method and system
US20170032546A1 (en) 2015-07-28 2017-02-02 PME IP Pty Ltd Apparatus and method for visualizing digital breast tomosynthesis and other volumetric images
US10540803B2 (en) * 2013-03-15 2020-01-21 PME IP Pty Ltd Method and system for rule-based display of sets of images
US10762872B2 (en) 2007-11-23 2020-09-01 PME IP Pty Ltd Client-server visualization system with hybrid data processing
US20200327669A1 (en) 2007-11-23 2020-10-15 PME IP Pty Ltd Automatic image segmentation methods and analysis
US20200366614A1 (en) 2013-03-15 2020-11-19 Pme Ip Pty Ltd. Method and system fpor transferring data to improve responsiveness when sending large data sets
US10909679B2 (en) * 2017-09-24 2021-02-02 PME IP Pty Ltd Method and system for rule based display of sets of images using image content derived parameters
US20210352133A1 (en) 2007-08-27 2021-11-11 PME IP Pty Ltd Fast file server methods and systems
US11244495B2 (en) * 2013-03-15 2022-02-08 PME IP Pty Ltd Method and system for rule based display of sets of images using image content derived parameters
US11599672B2 (en) 2015-07-31 2023-03-07 PME IP Pty Ltd Method and apparatus for anonymized display and data export

Family Cites Families (308)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US2658310A (en) 1950-12-22 1953-11-10 Carnegie Inst Of Washington Apparatus and process for the production of photosynthetic microorganisms, particularly algae
US3431200A (en) 1967-06-13 1969-03-04 North American Rockwell Flocculation of suspensions
US3645040A (en) 1967-10-18 1972-02-29 Era Inc Unbalanced culture method of algae production
US4137868A (en) 1976-09-29 1979-02-06 Pryor Taylor A Method and apparatus for growing seafood in commercially significant quantities on land
JPS54160696A (en) 1978-05-29 1979-12-19 Atomenergi Ab Fish breeding pond apparatus
JPS5561736A (en) 1978-10-28 1980-05-09 Nippon Carbide Kogyo Kk Laver breeding method and material
US4320594A (en) 1978-12-28 1982-03-23 Battelle Memorial Institute Mass algal culture system
US4267038A (en) 1979-11-20 1981-05-12 Thompson Worthington J Controlled natural purification system for advanced wastewater treatment and protein conversion and recovery
DE3145046C2 (en) 1980-11-15 1985-08-29 Rolls-Royce Ltd., London Method and device for generating a representation of the spatial distribution of a liquid flow within a hollow body structure
US4562139A (en) 1985-01-02 1985-12-31 Eastman Kodak Company Photographic products employing novel nondiffusible metal complexes of azo dyes
US4910912A (en) 1985-12-24 1990-03-27 Lowrey Iii O Preston Aquaculture in nonconvective solar ponds
US4928250A (en) 1986-07-02 1990-05-22 Hewlett-Packard Company System for deriving radiation images
US4958460A (en) 1988-05-09 1990-09-25 Algae Farms Method of growing and harvesting microorganisms
US4905148A (en) 1988-08-04 1990-02-27 General Electric Company Three-dimensional surface representation using connectivity method without leaks
US5091960A (en) 1988-09-26 1992-02-25 Visual Information Technologies, Inc. High-speed image rendering method using look-ahead images
FR2641099B1 (en) 1988-12-22 1991-02-22 Gen Electric Cgr
FR2642198B1 (en) 1989-01-20 1991-04-05 Gen Electric Cgr METHOD FOR CALCULATING AND USING THE CONICAL PROJECTION IMAGE, FOR EXAMPLE WITH X-RAY MEANS, OF A SAMPLE THREE-DIMENSIONAL OBJECT, AND METHOD FOR THREE-DIMENSIONAL RECONSTRUCTION OF A STUDY OBJECT USING THIS CALCULATION PROCESS
FR2644590B1 (en) 1989-03-20 1994-08-19 General Electric Cgr Sa PROCESS FOR ACQUIRING RADIOLOGICAL DATA AND RECONSTRUCTING STRUCTURES CORRESPONDING TO THIS BODY
US5078952A (en) 1989-06-16 1992-01-07 Science Applications International Corporation Multi-sensor explosive detection system
US5274759A (en) 1989-07-11 1993-12-28 Kabushiki Kaisha Toshiba Ultrasonic diagnostic apparatus capable of multi-frame representation
US5128864A (en) 1989-08-09 1992-07-07 W. L. Systems, Inc. Method for computing tomographic scans
US5163131A (en) 1989-09-08 1992-11-10 Auspex Systems, Inc. Parallel i/o network file server architecture
FR2656129B1 (en) 1989-12-20 1992-03-13 Gen Electric Cgr METHOD FOR MULTI-SCALE RECONSTRUCTION OF THE IMAGE OF THE STRUCTURE OF A BODY.
US5031117A (en) 1990-02-13 1991-07-09 International Business Machines Corporation Prioritization scheme for enhancing the display of ray traced images
JPH0454682A (en) 1990-06-22 1992-02-21 Toshiba Corp Method and device for three-dimensional picture processing
US5253171A (en) 1990-09-21 1993-10-12 General Electric Company Parallel processing method and apparatus based on the algebra reconstruction technique for reconstructing a three-dimensional computerized tomography (CT) image from cone beam projection data
US5633999A (en) 1990-11-07 1997-05-27 Nonstop Networks Limited Workstation-implemented data storage re-routing for server fault-tolerance on computer networks
US5293313A (en) 1990-11-21 1994-03-08 Picker International, Inc. Real time physician view box
JP3083606B2 (en) 1990-11-22 2000-09-04 株式会社東芝 Medical diagnosis support system
CA2056528A1 (en) 1990-12-21 1992-06-22 Kwok C. Tam Parallel processing method and apparatus for reconstructing a three-dimensional computerized tomography (ct) image of an object from cone beam projection data or from planar integrals
US5121708A (en) 1991-02-14 1992-06-16 Nuttle David A Hydroculture crop production system
CA2131705C (en) 1992-03-09 2008-10-21 Aaron G. Filler Image neurography and diffusion anisotropy imaging
US5375156A (en) 1992-03-31 1994-12-20 Siemens Medical Systems, Inc. Method and apparatus for 3-D computer tomography
CA2133825C (en) 1992-04-13 2002-12-31 Alan Benn Image analysis for meat
US5280428A (en) 1992-07-14 1994-01-18 General Electric Company Method and apparatus for projecting diagnostic images from volumed diagnostic data accessed in data tubes
EP0602730B1 (en) 1992-12-18 2002-06-19 Koninklijke Philips Electronics N.V. Registration of Volumetric images which are relatively elastically deformed by matching surfaces
US5452416A (en) 1992-12-30 1995-09-19 Dominator Radiology, Inc. Automated system and a method for organizing, presenting, and manipulating medical images
FR2701135B1 (en) 1993-01-29 1995-03-10 Commissariat Energie Atomique Method for reconstructing three-dimensional images of an evolving object.
US5412703A (en) 1993-02-04 1995-05-02 Institute For Radiological Image Science, Inc. Reduced partial volume artifacts in image reconstruction, with application to X-ray computed tomography
WO1996042022A1 (en) 1993-03-31 1996-12-27 Eg & G Astrophysics Research Corporation Three-dimensional reconstruction based on a limited number of x-ray projections
US5442672A (en) 1993-03-31 1995-08-15 Bjorkholm; Paul J. Three-dimensional reconstruction based on a limited number of X-ray projections
US5368033A (en) 1993-04-20 1994-11-29 North American Philips Corporation Magnetic resonance angiography method and apparatus employing an integration projection
US5488700A (en) 1993-07-30 1996-01-30 Xerox Corporation Image rendering system with local, adaptive estimation of incident diffuse energy
US6304771B1 (en) 1993-10-29 2001-10-16 The Trustees Of The University Of Pennsylvania Systems and methods for imaging fluorophores
US5666291A (en) 1994-06-02 1997-09-09 Sony Corporation Device for interfacing a CD-ROM player to an entertainment or information network and a network including such device
US6473793B1 (en) 1994-06-08 2002-10-29 Hughes Electronics Corporation Method and apparatus for selectively allocating and enforcing bandwidth usage requirements on network users
US5594842A (en) 1994-09-06 1997-01-14 The Research Foundation Of State University Of New York Apparatus and method for real-time volume visualization
US5640436A (en) 1995-01-26 1997-06-17 Hitachi Medical Corporation Method and apparatus for X-ray computed tomography
US5813988A (en) 1995-02-03 1998-09-29 Research Foundation Time-resolved diffusion tomographic imaging in highly scattering turbid media
FR2736455B1 (en) 1995-07-03 1997-08-08 Commissariat Energie Atomique METHOD FOR RECONSTRUCTING A 3D IMAGE WITH IMPROVED CONTRAST AND RESOLUTION AND APPLICATION OF THIS PROCESS TO PROVIDING AN ATTENUATION MAPPING OF AN OBJECT
JP2914891B2 (en) 1995-07-05 1999-07-05 株式会社東芝 X-ray computed tomography apparatus
US5671265A (en) 1995-07-14 1997-09-23 Siemens Corporate Research, Inc. Evidential reconstruction of vessel trees from X-ray angiograms with a dynamic contrast bolus
US5793374A (en) 1995-07-28 1998-08-11 Microsoft Corporation Specialized shaders for shading objects in computer generated images
US5744802A (en) 1995-10-25 1998-04-28 Adac Laboratories Image generation from limited projections in positron emission tomography using multi-slice rebinning
US6018562A (en) 1995-11-13 2000-01-25 The United States Of America As Represented By The Secretary Of The Army Apparatus and method for automatic recognition of concealed objects using multiple energy computed tomography
US7720672B1 (en) 1995-12-29 2010-05-18 Wyse Technology Inc. Method and apparatus for display of windowing application programs on a terminal
US5838756A (en) 1996-01-08 1998-11-17 Kabushiki Kaisha Toshiba Radiation computed tomography apparatus
US5821541A (en) 1996-02-02 1998-10-13 Tuemer; Tuemay O. Method and apparatus for radiation detection
US6108576A (en) 1996-03-18 2000-08-22 The Research Foundation Of City College Of New York Time-resolved diffusion tomographic 2D and 3D imaging in highly scattering turbid media
US5931789A (en) 1996-03-18 1999-08-03 The Research Foundation City College Of New York Time-resolved diffusion tomographic 2D and 3D imaging in highly scattering turbid media
US5602892A (en) 1996-03-21 1997-02-11 Llacer; Jorge Method for optimization of radiation therapy planning
AU4169497A (en) 1996-08-29 1998-04-14 David T. Borup Apparatus and method for imaging with wavefields using inverse scattering techniques
FR2752975B1 (en) 1996-09-04 1998-12-04 Ge Medical Syst Sa METHOD FOR RECONSTRUCTING A THREE-DIMENSIONAL IMAGE OF AN OBJECT, IN PARTICULAR A THREE-DIMENSIONAL ANGIOGRAPHIC IMAGE
US6175655B1 (en) 1996-09-19 2001-01-16 Integrated Medical Systems, Inc. Medical imaging system for displaying, manipulating and analyzing three-dimensional images
US6123733A (en) 1996-11-27 2000-09-26 Voxel, Inc. Method and apparatus for rapidly evaluating digital data processing parameters
US5841140A (en) 1997-01-08 1998-11-24 Smv America, Inc. Gamma camera for pet and spect studies
US5963658A (en) 1997-01-27 1999-10-05 University Of North Carolina Method and apparatus for detecting an abnormality within a host medium
US5774519A (en) 1997-01-30 1998-06-30 Analogic Corporation Method of and apparatus for calibration of CT scanners
US6226005B1 (en) 1997-01-31 2001-05-01 LAFERRIèRE ALAIN M Method and system for determining and/or using illumination maps in rendering images
US5859891A (en) 1997-03-07 1999-01-12 Hibbard; Lyn Autosegmentation/autocontouring system and method for use with three-dimensional radiation therapy treatment planning
US6032264A (en) 1997-04-22 2000-02-29 Micron Technology, Inc. Apparatus and method implementing repairs on a memory device
US5960056A (en) 1997-07-01 1999-09-28 Analogic Corporation Method and apparatus for reconstructing volumetric images in a helical scanning computed tomography system with multiple rows of detectors
US6008813A (en) 1997-08-01 1999-12-28 Mitsubishi Electric Information Technology Center America, Inc. (Ita) Real-time PC based volume rendering system
US6557102B1 (en) 1997-09-05 2003-04-29 Koninklijke Philips Electronics N.V. Digital trust center for medical image authentication
US6105029A (en) 1997-09-17 2000-08-15 International Business Machines Corporation Retrieving network files through parallel channels
US5909476A (en) 1997-09-22 1999-06-01 University Of Iowa Research Foundation Iterative process for reconstructing cone-beam tomographic images
US6745070B2 (en) 1997-10-03 2004-06-01 Tasc Ltd. High definition electrical impedance tomography
US6108007A (en) 1997-10-09 2000-08-22 Silicon Graphics, Inc. Method, system, and computer program product for increasing interpolation precision using multi-channel texture mapping
US6807581B1 (en) 2000-09-29 2004-10-19 Alacritech, Inc. Intelligent network storage interface system
US6360279B1 (en) 1997-10-14 2002-03-19 Bea Systems, Inc. True parallel client server system and method
US7185266B2 (en) 2003-02-12 2007-02-27 Alacritech, Inc. Network interface device for error detection using partial CRCS of variable length message portions
US6658480B2 (en) 1997-10-14 2003-12-02 Alacritech, Inc. Intelligent network interface system and method for accelerated protocol processing
US6049390A (en) 1997-11-05 2000-04-11 Barco Graphics Nv Compressed merging of raster images for high speed digital printing
US6377266B1 (en) 1997-11-26 2002-04-23 3Dlabs Inc., Ltd. Bit BLT with multiple graphics processors
US5963613A (en) 1997-12-31 1999-10-05 Siemens Corporate Research, Inc. C-arm calibration method for 3D reconstruction in an imaging system
US7010532B1 (en) 1997-12-31 2006-03-07 International Business Machines Corporation Low overhead methods and apparatus for shared access storage devices
US5950203A (en) 1997-12-31 1999-09-07 Mercury Computer Systems, Inc. Method and apparatus for high-speed access to and sharing of storage devices on a networked digital data processing system
US6289235B1 (en) 1998-03-05 2001-09-11 Wake Forest University Method and system for creating three-dimensional images using tomosynthetic computed tomography
US6205120B1 (en) 1998-03-13 2001-03-20 Packeteer, Inc. Method for transparently determining and setting an optimal minimum required TCP window size
US6091422A (en) 1998-04-03 2000-07-18 Avid Technology, Inc. System for editing complex visual data providing a continuously updated rendering
US6002739A (en) 1998-04-28 1999-12-14 Hewlett Packard Company Computed tomography with iterative reconstruction of thin cross-sectional planes
US6088423A (en) 1998-06-05 2000-07-11 Vivid Technologies, Inc. Multiview x-ray based system for detecting contraband such as in baggage
FR2779853B1 (en) 1998-06-11 2000-08-11 Ge Medical Syst Sa PROCESS FOR RECONSTRUCTING A THREE-DIMENSIONAL IMAGE OF AN OBJECT, IN PARTICULAR AN ANGIOGRAPHIC THREE-DIMENSIONAL IMAGE
US6268846B1 (en) 1998-06-22 2001-07-31 Adobe Systems Incorporated 3D graphics based on images and morphing
US20020034817A1 (en) 1998-06-26 2002-03-21 Henry Eric C. Process and apparatus for isolating and continuosly cultivating, harvesting, and processing of a substantially pure form of a desired species of algae
US6674430B1 (en) 1998-07-16 2004-01-06 The Research Foundation Of State University Of New York Apparatus and method for real-time volume processing and universal 3D rendering
US6119244A (en) 1998-08-25 2000-09-12 Network Appliance, Inc. Coordinating persistent status information with multiple file servers
US6327050B1 (en) 1999-04-23 2001-12-04 Electronics For Imaging, Inc. Printing method and apparatus having multiple raster image processors
US7664883B2 (en) 1998-08-28 2010-02-16 Alacritech, Inc. Network interface device that fast-path processes solicited session layer read commands
US6591004B1 (en) 1998-09-21 2003-07-08 Washington University Sure-fit: an automated method for modeling the shape of cerebral cortex and other complex structures using customized filters and transformations
DE19843812A1 (en) 1998-09-24 2000-03-30 Philips Corp Intellectual Pty Computed tomography procedure with a conical beam
US6526305B1 (en) 1998-11-25 2003-02-25 The Johns Hopkins University Method of fiber reconstruction employing data acquired by magnetic resonance imaging
IL127359A0 (en) 1998-12-01 1999-10-28 Yeda Res & Dev Computerized adaptive imaging
US6278460B1 (en) 1998-12-15 2001-08-21 Point Cloud, Inc. Creating a three-dimensional model from two-dimensional images
US20030158786A1 (en) 1999-02-26 2003-08-21 Skyline Software Systems, Inc. Sending three-dimensional images over a network
US7010554B2 (en) 2002-04-04 2006-03-07 Emc Corporation Delegation of metadata management in a storage system by leasing of free file system blocks and i-nodes from a file system owner
US6852225B1 (en) 1999-04-20 2005-02-08 The Regents Of The University Of California Method and apparatus to establish and optimize sedimentation and methane fermentation in primary wastewater ponds
US6264610B1 (en) 1999-05-05 2001-07-24 The University Of Connecticut Combined ultrasound and near infrared diffused light imaging system
US6236704B1 (en) 1999-06-30 2001-05-22 Siemens Corporate Research, Inc. Method and apparatus using a virtual detector for three-dimensional reconstruction from x-ray images
DE60015675T3 (en) 1999-09-06 2010-01-14 Komori Corporation Color management method and apparatus for a printing press
WO2001020546A2 (en) 1999-09-14 2001-03-22 The Research Foundation Of State University Of New York Imaging of scattering media using relative detector values
US6798417B1 (en) 1999-09-23 2004-09-28 International Business Machines Corporation Just in time graphics dispatching
FR2799028B1 (en) 1999-09-27 2002-05-03 Ge Medical Syst Sa METHOD FOR RECONSTRUCTING A THREE-DIMENSIONAL IMAGE OF ELEMENTS OF STRONG CONTRAST
US6654012B1 (en) 1999-10-01 2003-11-25 Terarecon, Inc. Early ray termination in a parallel pipelined volume rendering system
US6384821B1 (en) 1999-10-04 2002-05-07 International Business Machines Corporation Method and apparatus for delivering 3D graphics in a networked environment using transparent video
US6377257B1 (en) 1999-10-04 2002-04-23 International Business Machines Corporation Methods and apparatus for delivering 3D graphics in a networked environment
WO2001034027A1 (en) 1999-11-10 2001-05-17 The University Of Toledo System and method for skin lesion examination using multi-spectral, multi-source transillumination
US6324241B1 (en) 1999-12-30 2001-11-27 Ge Medical Systems Global Technology Company, Llc Method and apparatus for CT reconstruction
US6535821B2 (en) 2000-02-11 2003-03-18 University Of Iowa Research Foundation System and method of bolus-chasing angiography with adaptive real-time computed tomography (CT)
AU2001239926A1 (en) 2000-02-25 2001-09-03 The Research Foundation Of State University Of New York Apparatus and method for volume processing and rendering
DE10009285A1 (en) 2000-02-28 2001-08-30 Philips Corp Intellectual Pty Computer tomograph for determining the pulse transfer spectrum in an examination area
US7003547B1 (en) 2000-03-30 2006-02-21 United Devices, Inc. Distributed parallel processing system having capability-based incentives and associated method
WO2001074238A1 (en) 2000-03-31 2001-10-11 Koninklijke Philips Electronics N.V. Method and device for localizing a deviant region in a turbid medium
US6747654B1 (en) 2000-04-20 2004-06-08 Ati International Srl Multiple device frame synchronization method and apparatus
US6633688B1 (en) 2000-04-28 2003-10-14 Earth Resource Mapping, Inc. Method system and apparatus for providing image data in client/server systems
US20030065268A1 (en) 2000-05-05 2003-04-03 Massachusetts Institute Of Technology Optical computed tomography in a turbid media
EP1278454A2 (en) 2000-05-05 2003-01-29 Massachusetts Institute Of Technology Optical computed tomography in a turbid media
US20020089587A1 (en) 2000-05-18 2002-07-11 Imove Inc. Intelligent buffering and reporting in a multiple camera data streaming video system
US6664963B1 (en) 2000-05-31 2003-12-16 Nvidia Corporation System, method and computer program product for programmable shading using pixel shaders
US7274368B1 (en) 2000-07-31 2007-09-25 Silicon Graphics, Inc. System method and computer program product for remote graphics processing
FR2812741B1 (en) 2000-08-02 2003-01-17 Ge Med Sys Global Tech Co Llc METHOD AND DEVICE FOR RECONSTRUCTING A DYNAMIC THREE-DIMENSIONAL IMAGE OF AN OBJECT TRAVELED BY A CONTRAST PRODUCT
US7392291B2 (en) 2000-08-11 2008-06-24 Applied Micro Circuits Corporation Architecture for providing block-level storage access over a computer network
US7034828B1 (en) 2000-08-23 2006-04-25 Nintendo Co., Ltd. Recirculating shade tree blender for a graphics system
US7047309B2 (en) 2000-08-23 2006-05-16 International Business Machines Corporation Load balancing and dynamic control of multiple data streams in a network
US6785409B1 (en) 2000-10-24 2004-08-31 Koninklijke Philips Electronics, N.V. Segmentation method and apparatus for medical images using diffusion propagation, pixel classification, and mathematical morphology
US20020080143A1 (en) 2000-11-08 2002-06-27 Morgan David L. Rendering non-interactive three-dimensional content
US6615063B1 (en) 2000-11-27 2003-09-02 The General Hospital Corporation Fluorescence-mediated molecular tomography
US7383076B2 (en) 2000-11-27 2008-06-03 The General Hospital Corporation Fluorescence-mediated molecular tomography
US6475150B2 (en) 2000-12-01 2002-11-05 The Regents Of The University Of California System and method for ultrasonic tomography
US6470070B2 (en) 2000-12-20 2002-10-22 Cedara Software Corp. Image reconstruction using multiple X-ray projections
US7937470B2 (en) 2000-12-21 2011-05-03 Oracle International Corp. Methods of determining communications protocol latency
US6415013B1 (en) 2000-12-28 2002-07-02 Ge Medical Systems Global Technology Company, Llc Backprojection methods and apparatus for computed tomography imaging systems
US7788335B2 (en) 2001-01-11 2010-08-31 F5 Networks, Inc. Aggregated opportunistic lock and aggregated implicit lock management for locking aggregated files in a switched file system
WO2002067201A1 (en) 2001-02-15 2002-08-29 The Regents Of The University Of Michigan Statistically reconstructing an x-ray computed tomography image with beam hardening corrector
US6507633B1 (en) 2001-02-15 2003-01-14 The Regents Of The University Of Michigan Method for statistically reconstructing a polyenergetic X-ray computed tomography image and image reconstructor apparatus utilizing the method
US6778127B2 (en) 2001-03-28 2004-08-17 Larry G. Stolarczyk Drillstring radar
US6519355B2 (en) 2001-03-28 2003-02-11 Alan C. Nelson Optical projection imaging system and method for automatically detecting cells having nuclear and cytoplasmic densitometric features associated with disease
US7170521B2 (en) 2001-04-03 2007-01-30 Ultravisual Medical Systems Corporation Method of and system for storing, communicating, and displaying image data
DE60222768T2 (en) 2001-04-03 2008-07-17 Koninklijke Philips Electronics N.V. CT Scanner
US6914610B2 (en) 2001-05-18 2005-07-05 Sun Microsystems, Inc. Graphics primitive size estimation and subdivision for use with a texture accumulation buffer
US7054852B1 (en) 2001-05-23 2006-05-30 Ncr Corporation Performance of join operations in parallel database systems
US6687733B2 (en) 2001-06-01 2004-02-03 Intergenix Method and system for automatically configuring a client-server network
US6697064B1 (en) 2001-06-08 2004-02-24 Nvidia Corporation System, method and computer program product for matrix tracking during vertex processing in a graphics pipeline
US7006101B1 (en) 2001-06-08 2006-02-28 Nvidia Corporation Graphics API with branching capabilities
WO2002101497A2 (en) 2001-06-08 2002-12-19 Nvidia Corporation System, method and computer program product for programmable fragment processing in a graphics pipeline
US6577752B2 (en) 2001-06-15 2003-06-10 Arch Development Corporation Automated method and system for the delineation of the chest wall in computed tomography scans for the assessment of pleural disease
US6956570B2 (en) 2001-06-27 2005-10-18 Believe, Inc. Object visibility control for ray tracing
US6636623B2 (en) 2001-08-10 2003-10-21 Visiongate, Inc. Optical projection imaging system and method for automatically detecting cells with molecular marker compartmentalization associated with malignancy and disease
US6741730B2 (en) 2001-08-10 2004-05-25 Visiongate, Inc. Method and apparatus for three-dimensional imaging in the fourier domain
US6771733B2 (en) 2001-08-16 2004-08-03 University Of Central Florida Method of reconstructing images for spiral and non-spiral computer tomography
US7039723B2 (en) 2001-08-31 2006-05-02 Hinnovation, Inc. On-line image processing and communication system
US6947047B1 (en) 2001-09-20 2005-09-20 Nvidia Corporation Method and system for programmable pipelined graphics processing with branching instructions
US6917899B2 (en) 2001-09-21 2005-07-12 Microsoft Corporation System and methods for providing histogram computation in a high precision rasterization data pipeline
US7558611B2 (en) 2001-11-24 2009-07-07 Image Analysis, Inc. Automatic detection and quantification of coronary and aortic calcium
FR2833100B1 (en) 2001-11-30 2004-03-12 Ge Med Sys Global Tech Co Llc METHOD FOR RECONSTRUCTING AN IMAGE OF AN ORGAN
US6879715B2 (en) 2001-12-05 2005-04-12 General Electric Company Iterative X-ray scatter correction method and apparatus
US7650412B2 (en) 2001-12-21 2010-01-19 Netapp, Inc. Systems and method of implementing disk ownership in networked storage
US6809736B1 (en) 2002-01-08 2004-10-26 Apple Computer, Inc. Virtualization of graphics resources
US6744253B2 (en) 2002-01-15 2004-06-01 Larry G. Stolarczyk Synchronous radio-imaging of underground structures
AU2002236414A1 (en) 2002-01-18 2003-07-30 Kent Ridge Digital Labs Method and apparatus for determining symmetry in 2d and 3d images
US7262770B2 (en) 2002-03-21 2007-08-28 Microsoft Corporation Graphics image rendering with radiance self-transfer for low-frequency lighting environments
US6878115B2 (en) 2002-03-28 2005-04-12 Ultrasound Detection Systems, Llc Three-dimensional ultrasound computed tomography imaging system
US6877059B2 (en) 2002-03-29 2005-04-05 Emc Corporation Communications architecture for a high throughput storage processor
AU2003220677A1 (en) 2002-04-06 2003-10-27 Randall L Barbour Modification of the normalized difference method for real-time optical tomography
US6707878B2 (en) 2002-04-15 2004-03-16 General Electric Company Generalized filtered back-projection reconstruction in digital tomosynthesis
US6973479B2 (en) 2002-05-01 2005-12-06 Thales Avionics, Inc. Method and system for configuration and download in a restricted architecture network
US6740232B1 (en) 2002-05-01 2004-05-25 Aquascape Designs, Inc. Constructed wetlands system, treatment apparatus and method
US6697508B2 (en) 2002-05-10 2004-02-24 Visiongate, Inc. Tomographic reconstruction of small objects using a priori knowledge
US6770893B2 (en) 2002-05-13 2004-08-03 Visiongate, Inc. Method and apparatus for emission computed tomography using temporal signatures
US20050239182A1 (en) 2002-05-13 2005-10-27 Isaac Berzin Synthetic and biologically-derived products produced using biomass produced by photobioreactors configured for mitigation of pollutants in flue gases
US7050953B2 (en) 2002-05-22 2006-05-23 Bigwood Technology Incorporated Dynamical methods for solving large-scale discrete and continuous optimization problems
US20040078238A1 (en) 2002-05-31 2004-04-22 Carson Thomas Anonymizing tool for medical data
US6825840B2 (en) 2002-06-06 2004-11-30 Nvidia Corporation System and method of adjusting ray origins when shading vertices with rays
US7324116B2 (en) 2002-06-20 2008-01-29 Microsoft Corporation Systems and methods for providing controllable texture sampling
US6978206B1 (en) 2002-06-21 2005-12-20 Infogation Corporation Distributed navigation system
US6825843B2 (en) 2002-07-18 2004-11-30 Nvidia Corporation Method and apparatus for loop and branch instructions in a programmable graphics pipeline
US6765981B2 (en) 2002-07-31 2004-07-20 Agilent Technologies, Inc. Computed tomography
US6754299B2 (en) 2002-08-02 2004-06-22 Ge Medical Systems Global Technology Company, Llc Methods and apparatus for weighting of computed tomography data
US6658080B1 (en) 2002-08-05 2003-12-02 Voxar Limited Displaying image data using automatic presets
DE10237347B4 (en) 2002-08-14 2006-05-11 Siemens Ag Test method for a group of two-dimensional images of a three-dimensional object to meet a termination criterion and objects corresponding thereto
FI116750B (en) 2002-08-28 2006-02-15 Instrumentarium Corp Procedure and arrangement of medical x-rays
US20060122498A1 (en) 2002-08-30 2006-06-08 Sharpe James A Optical projection tomography
DE60303613T2 (en) 2002-08-30 2006-08-17 Medical Research Council OPTICAL PROJECTION TOMOGRAPHY
JP4467267B2 (en) 2002-09-06 2010-05-26 株式会社ソニー・コンピュータエンタテインメント Image processing method, image processing apparatus, and image processing system
US7475124B2 (en) 2002-09-25 2009-01-06 Emc Corporation Network block services for client access of network-attached data storage in an IP network
AU2003300646A1 (en) 2002-10-04 2004-05-04 Core Laboratories Lp Method and system for distributed tomographic velocity analysis using dense p-maps
US7058644B2 (en) 2002-10-07 2006-06-06 Click Commerce, Inc. Parallel tree searches for matching multiple, hierarchical data structures
DE10252662A1 (en) 2002-11-11 2004-05-27 Philips Intellectual Property & Standards Gmbh Medical computer tomography procedure has source rotated around object and back projection image reconstruction using square law distance and cosine beam angle weighting
US7185003B2 (en) 2002-11-14 2007-02-27 Seisint, Inc. Query scheduling in a parallel-processing database system
US7123760B2 (en) 2002-11-21 2006-10-17 General Electric Company Method and apparatus for removing obstructing structures in CT imaging
US7272429B2 (en) 2002-11-27 2007-09-18 Ge Medical Systems Global Technology Company, Llc Methods and apparatus for facilitating a reduction in artifacts
US7366199B1 (en) 2002-12-10 2008-04-29 Apple Inc. Method and apparatus measuring bandwidth
US6768782B1 (en) 2002-12-16 2004-07-27 University Of Notre Dame Du Lac Iterative method for region-of-interest reconstruction
US7552192B2 (en) 2002-12-18 2009-06-23 Ronnie Gerome Carmichael Massively parallel computer network-utilizing MPACT and multipoint parallel server (MPAS) technologies
WO2004066215A1 (en) 2003-01-21 2004-08-05 Philips Intellectual Property & Standards Gmbh Computed tomography method with coherent scattered rays, and computed tomograph
US7098907B2 (en) 2003-01-30 2006-08-29 Frantic Films Corporation Method for converting explicitly represented geometric surfaces into accurate level sets
US7647091B2 (en) 2003-02-05 2010-01-12 The General Hospital Corporation Method and system for free space optical tomography of diffuse media
US7542036B2 (en) 2003-02-19 2009-06-02 California Institute Of Technology Level set surface editing operators
WO2004077211A2 (en) 2003-02-28 2004-09-10 Tilmon Systems Ltd. Method and apparatus for increasing file server performance by offloading data path processing
US7519591B2 (en) 2003-03-12 2009-04-14 Siemens Medical Solutions Usa, Inc. Systems and methods for encryption-based de-identification of protected health information
US20040193901A1 (en) 2003-03-27 2004-09-30 Ge Medical Systems Global Company, Llc Dynamic configuration of patient tags and masking types while de-identifying patient data during image export from PACS diagnostic workstation
US7047337B2 (en) 2003-04-24 2006-05-16 International Business Machines Corporation Concurrent access of shared resources utilizing tracking of request reception and completion order
US7154985B2 (en) 2003-05-13 2006-12-26 Medical Insight A/S Method and system for simulating X-ray images
CA2527053C (en) 2003-05-30 2013-12-10 Karl Johann Schmidt Shading computer-generated objects using generalized shading regions
WO2004109603A1 (en) 2003-06-11 2004-12-16 Koninklijke Philips Electronics, N.V. User control of 3d volume plane crop
US20050012753A1 (en) 2003-07-18 2005-01-20 Microsoft Corporation Systems and methods for compositing graphics overlays without altering the primary display image and presenting them to the display on-demand
US7076735B2 (en) 2003-07-21 2006-07-11 Landmark Graphics Corporation System and method for network transmission of graphical data through a distributed application
US7167176B2 (en) 2003-08-15 2007-01-23 Microsoft Corporation Clustered principal components for precomputed radiance transfer
US7865485B2 (en) 2003-09-23 2011-01-04 Emc Corporation Multi-threaded write interface and methods for increasing the single file read and write throughput of a file server
US7948490B2 (en) 2003-10-22 2011-05-24 Microsoft Corporation Hardware-accelerated computation of radiance transfer coefficients in computer graphics
WO2005055137A2 (en) 2003-11-26 2005-06-16 Viatronix Incorporated Vessel segmentation using vesselness and edgeness
US7219085B2 (en) 2003-12-09 2007-05-15 Microsoft Corporation System and method for accelerating and optimizing the processing of machine learning techniques using a graphics processing unit
EP1728213B1 (en) 2003-12-12 2008-02-20 Agency for Science, Technology and Research Method and apparatus for identifying pathology in brain images
US7693318B1 (en) 2004-01-12 2010-04-06 Pme Ip Australia Pty Ltd Method and apparatus for reconstruction of 3D image volumes from projection images
US7120283B2 (en) 2004-01-12 2006-10-10 Mercury Computer Systems, Inc. Methods and apparatus for back-projection and forward-projection
US7907759B2 (en) 2006-02-02 2011-03-15 Wake Forest University Health Sciences Cardiac visualization systems for displaying 3-D images of cardiac voxel intensity distributions with optional physician interactive boundary tracing tools
US7593762B2 (en) 2004-04-09 2009-09-22 Siemens Medical Solutions Usa, Inc. System and method for automatically segmenting bones in computed tomography angiography data
US7385607B2 (en) 2004-04-12 2008-06-10 Nvidia Corporation Scalable shader architecture
US7154500B2 (en) 2004-04-20 2006-12-26 The Chinese University Of Hong Kong Block-based fragment filtration with feasible multi-GPU acceleration for real-time volume rendering on conventional personal computer
US7706633B2 (en) 2004-04-21 2010-04-27 Siemens Corporation GPU-based image manipulation method for registration applications
CA2564313A1 (en) 2004-05-05 2005-11-17 Ims Health Incorporated Data encryption applications for multi-source longitudinal patient-level data integration
US20050270298A1 (en) 2004-05-14 2005-12-08 Mercury Computer Systems, Inc. Daughter card approach to employing multiple graphics cards within a system
US20070005798A1 (en) 2004-05-17 2007-01-04 Adrian Gropper System and method for virtual radiology and patient document flow
US7899516B2 (en) 2004-06-23 2011-03-01 M2S, Inc. Method and apparatus for determining the risk of rupture of a blood vessel using the contiguous element defined area
US7242401B2 (en) 2004-06-25 2007-07-10 Siemens Medical Solutions Usa, Inc. System and method for fast volume rendering
WO2006014480A2 (en) 2004-07-08 2006-02-09 Actuality Systems, Inc. Architecture for rendering graphics on output devices over diverse connections
US7339585B2 (en) 2004-07-19 2008-03-04 Pie Medical Imaging B.V. Method and apparatus for visualization of biological structures with use of 3D position information from segmentation results
US7315926B2 (en) 2004-09-21 2008-01-01 Emc Corporation Lock management for concurrent access to a single file from multiple data mover computers
US20060066609A1 (en) 2004-09-28 2006-03-30 Iodice Arthur P Methods and systems for viewing geometry of an object model generated by a CAD tool
US7848231B2 (en) 2004-10-29 2010-12-07 Nippon Telegraph And Telephone Corporation Packet communication network and packet communication method
US8189002B1 (en) 2004-10-29 2012-05-29 PME IP Australia Pty, Ltd. Method and apparatus for visualizing three-dimensional and higher-dimensional image data sets
US7778392B1 (en) 2004-11-02 2010-08-17 Pme Ip Australia Pty Ltd Method of reconstructing computed tomography (CT) volumes suitable for execution on commodity central processing units (CPUs) and graphics processors, and apparatus operating in accord with those methods (rotational X-ray on GPUs)
US20090005693A1 (en) 2004-12-22 2009-01-01 Biotree Systems, Inc. Medical Imaging Methods and Apparatus for Diagnosis and Monitoring of Diseases and Uses Therefor
US7609884B1 (en) 2004-12-23 2009-10-27 Pme Ip Australia Pty Ltd Mutual information based registration of 3D-image volumes on GPU using novel accelerated methods of histogram computation
US20060239540A1 (en) 2005-03-09 2006-10-26 Bracco Imaging, S.P.A. Methods and systems for creating 4D images using multiple 2D images acquired in real-time ("4D ultrasound")
US7385614B2 (en) 2005-03-28 2008-06-10 Silicon Graphics, Inc. Compositing images using logically divided object space
US8041093B2 (en) 2005-04-22 2011-10-18 General Electric Company System and method for definition of DICOM header values
US20070038939A1 (en) 2005-07-11 2007-02-15 Challen Richard F Display servers and systems and methods of graphical display
US7483939B2 (en) 2005-08-25 2009-01-27 General Electric Company Medical processing system allocating resources for processing 3D to form 2D image data based on report of monitor data
US20070092864A1 (en) 2005-09-30 2007-04-26 The University Of Iowa Research Foundation Treatment planning methods, devices and systems
CA2626040A1 (en) 2005-10-17 2007-04-26 The General Hospital Corporation Structure-analysis system, method, software arrangement and computer-accessible medium for digital cleansing of computed tomography colonography images
WO2007058997A2 (en) 2005-11-11 2007-05-24 The University Of Houston System Scoring method for imaging-based detection of vulnerable patients
JP4762693B2 (en) 2005-11-22 2011-08-31 株式会社日立製作所 File server, file server log management system, and file server log management method
US20070165917A1 (en) 2005-11-26 2007-07-19 Zhujiang Cao Fully automatic vessel tree segmentation
US7756316B2 (en) 2005-12-05 2010-07-13 Siemens Medicals Solutions USA, Inc. Method and system for automatic lung segmentation
CA2640802A1 (en) 2005-12-20 2007-09-13 University Of Maryland, Baltimore Method and apparatus for accelerated elastic registration of multiple scans of internal properties of a body
US20070185879A1 (en) 2005-12-23 2007-08-09 Metacommunications, Inc. Systems and methods for archiving and retrieving digital assets
DE102005062419B4 (en) 2005-12-27 2008-01-17 Vega Grieshaber Kg Circuit arrangement for a field device
US20070156955A1 (en) 2005-12-30 2007-07-05 Royer Robert J Jr Method and apparatus for queuing disk drive access requests
CN101410874A (en) 2006-01-13 2009-04-15 卓斯企业有限责任公司 Computer network-based 3D rendering system
US7995824B2 (en) 2006-02-10 2011-08-09 University Of Medicine And Dentistry Of New Jersey Precision subtraction computed tomographic angiography
US20070226314A1 (en) 2006-03-22 2007-09-27 Sss Research Inc. Server-based systems and methods for enabling interactive, collabortive thin- and no-client image-based applications
US20070255704A1 (en) 2006-04-26 2007-11-01 Baek Ock K Method and system of de-identification of a record
JP5105786B2 (en) 2006-07-07 2012-12-26 株式会社東芝 Magnetic resonance imaging apparatus and image processing apparatus
US20080009055A1 (en) 2006-07-10 2008-01-10 Greenfuel Technologies Corp. Integrated photobioreactor-based pollution mitigation and oil extraction processes and systems
US7711797B1 (en) 2006-07-31 2010-05-04 Juniper Networks, Inc. Optimizing batch size for prefetching data over wide area networks
US8878833B2 (en) 2006-08-16 2014-11-04 Barco, Inc. Systems, methods, and apparatus for recording of graphical display
US20080086557A1 (en) 2006-10-06 2008-04-10 Ace*Comm Network service provider platform for supporting usage sensitive billing and operation services
US20080115139A1 (en) 2006-10-27 2008-05-15 Todd Alan Inglett Barrier-based access to a shared resource in a massively parallel computer system
US9355273B2 (en) 2006-12-18 2016-05-31 Bank Of America, N.A., As Collateral Agent System and method for the protection and de-identification of health care data
WO2008083351A2 (en) 2006-12-29 2008-07-10 Genifuel Corporation Controlled growth environments for algae cultivation
US8130245B2 (en) 2007-01-18 2012-03-06 Harris Corporation System and method for processing map images
US20080208961A1 (en) 2007-02-23 2008-08-28 Hostway Corporation Parallel retrieval system
WO2008115859A2 (en) 2007-03-16 2008-09-25 The General Hospital Corporation System and method for displaying medical imaging spectral data as hysometric maps
SE532218C2 (en) 2007-03-28 2009-11-17 Agency 9 Ab Systems, method, computer programs and computer-readable media for graphics processing
US20090077097A1 (en) 2007-04-16 2009-03-19 Attune Systems, Inc. File Aggregation in a Switched File System
WO2008138008A1 (en) 2007-05-08 2008-11-13 Riverbed Technology, Inc A hybrid segment-oriented file server and wan accelerator
US20090012382A1 (en) 2007-07-02 2009-01-08 General Electric Company Method and system for detection of obstructions in vasculature
US7673011B2 (en) 2007-08-10 2010-03-02 International Business Machines Corporation Configuring compute nodes of a parallel computer in an operational group into a plurality of independent non-overlapping collective networks
US10311541B2 (en) 2007-11-23 2019-06-04 PME IP Pty Ltd Multi-user multi-GPU render server apparatus and methods
US9904969B1 (en) 2007-11-23 2018-02-27 PME IP Pty Ltd Multi-user multi-GPU render server apparatus and methods
WO2011065929A1 (en) 2007-11-23 2011-06-03 Mercury Computer Systems, Inc. Multi-user multi-gpu render server apparatus and methods
EP2098994A1 (en) 2008-03-04 2009-09-09 Agfa HealthCare NV System for real-time volume rendering on thin clients via a render server
US8416196B2 (en) 2008-03-04 2013-04-09 Apple Inc. Touch event model programming interface
US8369600B2 (en) 2008-03-25 2013-02-05 General Electric Company Method and apparatus for detecting irregularities in tissue microarrays
US8781197B2 (en) 2008-04-28 2014-07-15 Cornell University Tool for accurate quantification in molecular MRI
US20090313170A1 (en) 2008-06-16 2009-12-17 Agmednet, Inc. Agent for Medical Image Transmission
US8229193B2 (en) 2008-09-03 2012-07-24 General Electric Company System and methods for applying image presentation context functions to image sub-regions
US8386560B2 (en) 2008-09-08 2013-02-26 Microsoft Corporation Pipeline for network based server-side 3D image rendering
US8432413B2 (en) 2008-11-17 2013-04-30 Xrfiles, Inc. System and method for the display of extended bit depth high resolution images
US8542136B1 (en) 2009-04-02 2013-09-24 Comtech Ef Data Corp. Data compression system and related methods
EP2486503A2 (en) 2009-10-07 2012-08-15 Hologic, Inc. Processing and displaying computer-aided detection information associated with breast x-ray images
US20110112862A1 (en) 2009-11-06 2011-05-12 Yi-Cheng Yu System and Method for Securely Managing and Storing Individually Identifiable Information in Web-Based and Alliance-Based Networks
US8296359B2 (en) 2010-07-12 2012-10-23 Opus Medicus, Inc. Systems and methods for networked, in-context, high resolution image viewing
US20120078088A1 (en) 2010-09-28 2012-03-29 Point of Contact, LLC. Medical image projection and tracking system
US8745413B2 (en) 2011-03-02 2014-06-03 Appature, Inc. Protected health care data marketing system and method
US20120233153A1 (en) 2011-03-11 2012-09-13 International Business Machines Corporation Hierarchical browsing operations on a directory attribute
DE102012201169A1 (en) 2012-01-27 2013-08-01 Siemens Aktiengesellschaft Automatic registration of image pairs of medical image data sets
US11183292B2 (en) * 2013-03-15 2021-11-23 PME IP Pty Ltd Method and system for rule-based anonymized display and data export
US8976190B1 (en) 2013-03-15 2015-03-10 Pme Ip Australia Pty Ltd Method and system for rule based display of sets of images
US10070839B2 (en) * 2013-03-15 2018-09-11 PME IP Pty Ltd Apparatus and system for rule based visualization of digital breast tomosynthesis and other volumetric images
US10448911B2 (en) 2013-10-30 2019-10-22 Koninklijke Philips N.V. Method and device for displaying medical images
US9582680B2 (en) 2014-01-30 2017-02-28 Microsoft Technology Licensing, Llc Scrubbe to remove personally identifiable information
US20160012181A1 (en) 2014-07-11 2016-01-14 Elevated Capital Group Llc Method for assigning a qualitative importance of relevant genetic phenotypes to the use of specific drugs for individual patients based on genetic test results
US10452813B2 (en) * 2016-11-17 2019-10-22 Terarecon, Inc. Medical image identification and interpretation
WO2018205232A1 (en) 2017-05-11 2018-11-15 上海联影医疗科技有限公司 Method for automatically and accurately positioning reference line according to spliced result
US11969280B2 (en) * 2020-01-07 2024-04-30 Cleerly, Inc. Systems, methods, and devices for medical image analysis, diagnosis, risk stratification, decision making and/or disease tracking
AU2021205821A1 (en) * 2020-01-07 2022-07-21 Cleerly, Inc. Systems, methods, and devices for medical image analysis, diagnosis, risk stratification, decision making and/or disease tracking

Patent Citations (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6765570B1 (en) 1998-07-21 2004-07-20 Magic Earth, Inc. System and method for analyzing and imaging three-dimensional volume data sets using a three-dimensional sampling probe
US20060149850A1 (en) 2005-01-05 2006-07-06 Control4 Corporation Method and apparatus for synchronizing playback of streaming media in multiple output devices
US20080052126A1 (en) 2006-08-25 2008-02-28 Konica Minolta Medical & Graphic, Inc. Database system, program, image retrieving method, and report retrieving method
US20080123917A1 (en) * 2006-11-29 2008-05-29 Siemens Medical Solutions Usa, Inc. An Imaging Study Completion Processing System
US20080166070A1 (en) * 2007-01-04 2008-07-10 General Electric Company Method for providing adaptive hanging protocols for image reading
US20210352133A1 (en) 2007-08-27 2021-11-11 PME IP Pty Ltd Fast file server methods and systems
US20230260478A1 (en) 2007-11-23 2023-08-17 PME IP Pty Ltd Client-server visualization system with hybrid data processing
US11640809B2 (en) 2007-11-23 2023-05-02 PME IP Pty Ltd Client-server visualization system with hybrid data processing
US11514572B2 (en) 2007-11-23 2022-11-29 PME IP Pty Ltd Automatic image segmentation methods and analysis
US10762872B2 (en) 2007-11-23 2020-09-01 PME IP Pty Ltd Client-server visualization system with hybrid data processing
US20200327669A1 (en) 2007-11-23 2020-10-15 PME IP Pty Ltd Automatic image segmentation methods and analysis
US20220165231A1 (en) 2007-11-23 2022-05-26 PME IP Pty Ltd Client-server visualization system with hybrid data processing
US20140173287A1 (en) 2011-07-11 2014-06-19 Takeshi Mizunuma Identifier management method and system
US20130089248A1 (en) * 2011-10-05 2013-04-11 Cireca Theranostics, Llc Method and system for analyzing biological specimens by spectral imaging
US20200366614A1 (en) 2013-03-15 2020-11-19 Pme Ip Pty Ltd. Method and system fpor transferring data to improve responsiveness when sending large data sets
US11244495B2 (en) * 2013-03-15 2022-02-08 PME IP Pty Ltd Method and system for rule based display of sets of images using image content derived parameters
US10832467B2 (en) * 2013-03-15 2020-11-10 PME IP Pty Ltd Method and system for rule based display of sets of images using image content derived parameters
US10540803B2 (en) * 2013-03-15 2020-01-21 PME IP Pty Ltd Method and system for rule-based display of sets of images
US11763516B2 (en) * 2013-03-15 2023-09-19 PME IP Pty Ltd Method and system for rule based display of sets of images using image content derived parameters
US20210256742A1 (en) 2015-07-28 2021-08-19 PME IP Pty Ltd Apparatus and method for visualizing digital breast tomosynthesis and other volumetric images
US11620773B2 (en) 2015-07-28 2023-04-04 PME IP Pty Ltd Apparatus and method for visualizing digital breast tomosynthesis and other volumetric images
US20170032546A1 (en) 2015-07-28 2017-02-02 PME IP Pty Ltd Apparatus and method for visualizing digital breast tomosynthesis and other volumetric images
US11599672B2 (en) 2015-07-31 2023-03-07 PME IP Pty Ltd Method and apparatus for anonymized display and data export
US20230195937A1 (en) 2015-07-31 2023-06-22 PME IP Pty Ltd Method and apparatus for anonymized display and data export
US11669969B2 (en) * 2017-09-24 2023-06-06 PME IP Pty Ltd Method and system for rule based display of sets of images using image content derived parameters
US10909679B2 (en) * 2017-09-24 2021-02-02 PME IP Pty Ltd Method and system for rule based display of sets of images using image content derived parameters

Non-Patent Citations (9)

* Cited by examiner, † Cited by third party
Title
AU2018335370, Office Action, dated May 15, 2023, 4 pages.
Au2022200601, Office Action, dated Dec. 8, 2022, 4 pages.
CA2991378, Office Action, dated Oct. 20, 2023, 8 pages.
CA2991378, Office Action, dated Oct. 24, 2022, 7 pages.
EP202303856_OA_August_23_2023, 7 pages.
EP3329405, Summons, dated Jul. 7, 2023, 14 pages.
Gou et al. Persuit of medical image diagnostic system and method, Jul. 24, 2009 (Year: 2009). *
Higgins et al., Distributed System for Processing 3D Medical Images, Compat Biol Med (1997) 27, pp. 97-115.
JP2022-191827, Office Action, dated Oct. 18, 2023, 2 pages (& English translation).

Also Published As

Publication number Publication date
US20210019933A1 (en) 2021-01-21
US20220139025A1 (en) 2022-05-05
US20230386125A1 (en) 2023-11-30
US11244495B2 (en) 2022-02-08
US11763516B2 (en) 2023-09-19

Similar Documents

Publication Publication Date Title
US10832467B2 (en) Method and system for rule based display of sets of images using image content derived parameters
US11669969B2 (en) Method and system for rule based display of sets of images using image content derived parameters
US11701064B2 (en) Method and system for rule based display of sets of images
US12136164B2 (en) Method and system for rule based display of sets of images using image content derived parameters
US8369585B2 (en) Automatic classification of information in images
US20100231605A1 (en) Medical image processing device and medical image processing program
US8786601B2 (en) Generating views of medical images
JP6945474B2 (en) Learning data creation support device, learning data creation support method, and learning data creation support program
JP2006034585A (en) Picture display device and picture display method, and program thereof
US10860894B2 (en) Learning data generation support apparatus, operation method of learning data generation support apparatus, and learning data generation support program
JP2022058397A (en) Medical image processing device, medical image processing method and medical image processing program
US20040024292A1 (en) System and method for assigning a computer aided detection application to a digital image
JP2004041490A (en) Diagnostic imaging support system
US20240370995A1 (en) Method and system for rule based display of sets of images using image content derived parameters
Suganthi et al. Interactive Visualization for Understanding and Analyzing Medical Data
Chalida Aphinives et al. Artificial Intelligence Development for Detecting Microcalcification within Mammography

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE