[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

US20090274271A1 - System and method for selecting a guidance mode for performing a percutaneous procedure - Google Patents

System and method for selecting a guidance mode for performing a percutaneous procedure Download PDF

Info

Publication number
US20090274271A1
US20090274271A1 US12/431,518 US43151809A US2009274271A1 US 20090274271 A1 US20090274271 A1 US 20090274271A1 US 43151809 A US43151809 A US 43151809A US 2009274271 A1 US2009274271 A1 US 2009274271A1
Authority
US
United States
Prior art keywords
ray
instrument
trajectory
planned
instrument trajectory
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US12/431,518
Other versions
US8165660B2 (en
Inventor
Marcus Pfister
Norbert Strobel
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Siemens Healthineers AG
Siemens Medical Solutions USA Inc
Original Assignee
Siemens Medical Solutions USA Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Siemens Medical Solutions USA Inc filed Critical Siemens Medical Solutions USA Inc
Priority to US12/431,518 priority Critical patent/US8165660B2/en
Assigned to SIEMENS MEDICAL SOLUTIONS USA, INC. reassignment SIEMENS MEDICAL SOLUTIONS USA, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PFISTER, MARCUS, STROBEL, NORBERT
Publication of US20090274271A1 publication Critical patent/US20090274271A1/en
Assigned to SIEMENS AKTIENGESELLSCHAFT reassignment SIEMENS AKTIENGESELLSCHAFT ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PFISTER, MARCUS, STROBEL, NORBERT
Application granted granted Critical
Publication of US8165660B2 publication Critical patent/US8165660B2/en
Assigned to SIEMENS HEALTHCARE GMBH reassignment SIEMENS HEALTHCARE GMBH ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS AKTIENGESELLSCHAFT
Assigned to Siemens Healthineers Ag reassignment Siemens Healthineers Ag ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIEMENS HEALTHCARE GMBH
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/12Arrangements for detecting or locating foreign bodies
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/44Constructional features of apparatus for radiation diagnosis
    • A61B6/4429Constructional features of apparatus for radiation diagnosis related to the mounting of source units and detector units
    • A61B6/4435Constructional features of apparatus for radiation diagnosis related to the mounting of source units and detector units the source unit and the detector unit being coupled by a rigid structure
    • A61B6/4441Constructional features of apparatus for radiation diagnosis related to the mounting of source units and detector units the source unit and the detector unit being coupled by a rigid structure the rigid structure being a C-arm or U-arm
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • A61B6/5211Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
    • A61B6/5229Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/10Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
    • A61B90/11Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis with guides for needles or instruments, e.g. arcuate slides or ball joints
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/10Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
    • A61B90/11Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis with guides for needles or instruments, e.g. arcuate slides or ball joints
    • A61B90/13Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis with guides for needles or instruments, e.g. arcuate slides or ball joints guided by light, e.g. laser pointers
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/107Visualisation of planned trajectories or target regions
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B2090/364Correlation of different images or relation of image positions in respect to the body
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/37Surgical systems with images on a monitor during operation
    • A61B2090/376Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61NELECTROTHERAPY; MAGNETOTHERAPY; RADIATION THERAPY; ULTRASOUND THERAPY
    • A61N5/00Radiation therapy
    • A61N5/10X-ray therapy; Gamma-ray therapy; Particle-irradiation therapy
    • A61N5/1048Monitoring, verifying, controlling systems and methods
    • A61N5/1049Monitoring, verifying, controlling systems and methods for verifying the position of the patient with respect to the radiation beam
    • A61N2005/1061Monitoring, verifying, controlling systems and methods for verifying the position of the patient with respect to the radiation beam using an x-ray imaging system having a separate imaging source

Definitions

  • target point selection may be done by selecting a desired MPR view, such as by a right-click of a mouse pointer on an appropriate soft key 3 A in a pop-up window in the display. In the illustrated embodiment, this results in desired MPR views being displayed in the upper left quadrant 3 B, and upper right quadrant 3 C.
  • the target point is “selected” by clicking with a mouse pointer at the target position 3 D in the lower left quadrant MPR display 3 E.
  • the skin entry point may be selected (“clicked”) in the same manner.

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Medical Informatics (AREA)
  • Surgery (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Pathology (AREA)
  • Public Health (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Veterinary Medicine (AREA)
  • Animal Behavior & Ethology (AREA)
  • Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • High Energy & Nuclear Physics (AREA)
  • Radiology & Medical Imaging (AREA)
  • Biophysics (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Apparatus For Radiation Diagnosis (AREA)

Abstract

A system and method are disclosed for planning a percutaneous procedure and for guiding an instrument to engage a target within a patient's body. A patient 3-dimensional image data set is provided, within which a user defines an instrument trajectory, for example, by selecting a skin entry point and a target point. A line, or “planned path,” is generated between the points. The system determines whether the path can be targeted so the optical axis of the movable arm coincides with the path so that a laser can be used for instrument guidance. Alternatively, the user can determine whether planned path can be targeted so that the C-arm can be made to coincide with the path so that the extension of the path is projected onto the detector, using x-ray radiation. If neither laser guidance or x-ray guidance can be used, the physician is prompted to re-plan the path so that at least x-ray radiation guidance can be used. Once the desired guidance method is selected, the puncture instrument is guided along the planned path to engage the target.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This is a U.S. non-provisional application of U.S. provisional patent application Ser. No. 61/049,803, filed May 2, 2008, by Marcus Pfister, et al., titled “Workflow for Percutaneous Procedures,” the entirety of which application is incorporated herein by reference.
  • FIELD OF THE DISCLOSURE
  • The disclosure relates generally to methods for planning and performing percutaneous procedures, and more particularly to methods for automatically aligning medical devices to perform percutaneous procedures that reduce overall radiation exposure to patient and user.
  • BACKGROUND
  • With modem angiography systems, it is possible to obtain not only 2-dimensional (2D) fluoroscopic images but, by rotation of a C-arm around the patient, 3-dimensional (3D) CT-like images. For the purpose of intervention planning or navigation, these 3D datasets can be co-registered with the 2D fluoroscopic image. This registration can be used to plan percutaneous puncture procedures, such as needle biopsies. Once registered, 2D fluoroscopic overlay images can be rendered under a given C-arm view angle to provide 3D anatomical information as well as a graphical representation of the interventional plan that can be combined with the live x-ray fluoroscopy image.
  • Methods are known for utilizing a laser fixed to the C-arm and orientated along the planned path of a needle puncture to orient the actual needle along a desired puncture path so that it intersects a targeted location within the patent. Proper orientation of the laser can require manual adjustment of the patient table and/or a C-arm to ensure that the actual needle path coincides with the “virtual” (i.e., theoretical) laser path. This is because not all theoretical paths may be achievable with a C-arm, since the C-arm may be physically limited in angulation and also may be limited so that it does not collide with the patient or physician. The advantage of laser-based guidance methods, however, is that a biopsy needle can be oriented without exposing the patient or physician to x-ray radiation.
  • Other methods are known for positioning a needle along a virtual path without the use of a laser fixed to a specific position with respect to the C-arm. Such methods instead utilize x-ray radiation to orient the needle position. To adjust the C-arm so that a needle is correctly oriented along a desired path, the puncture target and the skin entry point—which together define a “path of puncture”—are imaged on top of each other on an x-ray detector to provide a “bull's eye view” that enables appropriate adjustment of the C-arm to achieve a desired puncture path.
  • There is a need for a system and method that combine the benefits of both types of procedures in a way that minimizes overall radiation exposure to patients and practitioners.
  • SUMMARY OF THE DISCLOSURE
  • A method is disclosed for planning a percutaneous procedure for use in a system comprising an imaging system having a movable arm, an x-ray source and an x-ray detector and a display and a system controller connected to and in communication with the imaging system and display, comprising: providing a three-dimensional image data set of a patient tissue region; obtaining an x-ray image of the patient tissue region using the x-ray source and the x-ray detector; co-registering the three-dimensional image data set to the x-ray image acquired using the imaging system; obtaining a planned instrument trajectory based on target point data representative of a target object within the patient tissue region and another point along the planned instrument trajectory, which may be a skin entry point data representative of a skin entry point; and aligning an instrument along the planned instrument trajectory using one of a plurality of instrument guidance procedures, the one of a plurality of instrument guidance procedures being selected on the basis of the planned instrument trajectory and a position of the movable arm; wherein the plurality of instrument guidance procedures comprises a laser-guided procedure, an x-ray guided procedure, and a procedure that replans the instrument trajectory.
  • A system is disclosed for planning a percutaneous procedure for use in a system comprising an imaging system having a movable arm, an x-ray source and an x-ray detector and a display and a system controller connected to and in communication with the imaging system and display. The system further comprises a machine-readable storage medium encoded with a computer program code such that, when the computer program code is executed by a processor, the processor performs a method comprising: providing a three-dimensional image data set of a patient tissue region; obtaining an x-ray image of the patient tissue region using the x-ray source and the x-ray detector; co-registering the three-dimensional image data set to the x-ray image acquired using the imaging system; obtaining a planned instrument trajectory based on target point data representative of a target object within the patient tissue region and a skin entry point data representative of a skin entry point; and aligning an instrument along the planned instrument trajectory using one of a plurality of instrument guidance procedures, the one of a plurality of instrument guidance procedures being selected on the basis of the planned instrument trajectory and a position of the movable arm; wherein the plurality of instrument guidance procedures comprises a laser-guided procedure, an x-ray guided procedure, and a procedure that replans the instrument trajectory.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings illustrate preferred embodiments of the disclosure so far devised for the practical application of the principles thereof, and in which:
  • FIG. 1 is a is a schematic diagram showing an exemplary x-ray acquisition system for performing the disclosed method;
  • FIG. 2 is a display view from an x-ray guidance system showing a three-dimensional rendering of a test phantom showing objects internal to the phantom that are representative of different types of patient tissue;
  • FIG. 3 is a display view showing the selection of a target point on the first and second fluoroscopic views and the MPR view;
  • FIG. 4 is a display view showing the selection of a skin entry point on the first and second fluoroscopic views and the MPR view;
  • FIG. 5 is a side view of an exemplary arrangement for supporting a laser pointer for use with the disclosed methods;
  • FIG. 6 is a top view of the arrangement of FIG. 5;
  • FIG. 7 is an isometric view of a flat x-ray detector in combination with a laser pointer;
  • FIGS. 8 and 9 are schematic views of the interaction of the flat x-ray detector and laser pointer of FIG. 7 with the support arrangement of FIG. 6;
  • FIG. 10 is an isometric view of a flat x-ray detector in combination with a laser pointer according to another aspect of the disclosure; and
  • FIG. 11 is a display view from an x-ray guidance system showing a collimated fluoroscopic view of the target and skin entry points, as well as an oblique fluoroscopic view showing a planned path trajectory intersecting the target and skin entry points;
  • FIG. 12 is an enlarged view of the collimated fluoroscopic view of FIG. 11;
  • FIG. 13 is an enlarged view of the collimated fluoroscopic view of FIG. 11 showing an instrument inserted at the skin entry point;
  • FIG. 14 is a collimated fluoroscopic view taken oblique to the view of FIG. 13 showing the position of the instrument relative to the graphical overlay of the planned instrument trajectory;
  • FIG. 15 is a collimated fluoroscopic view taken oblique to the views of FIGS. 13 and 14 showing the position of the instrument relative to the graphical overlay of the planned instrument trajectory;
  • FIG. 16 is a collimated fluoroscopic view showing the position of the instrument as it intersects the target;
  • FIG. 17 is a C-arm CT (DynaCT) scan of the completed instrument insertion;
  • FIGS. 18A-18E are flowcharts for using the disclosed method to select between a laser guidance technique and an x-ray guidance technique for planning and performing a percutaneous procedure;
  • FIG. 19 is a flow chart showing a sequence of steps of a method for aligning an instrument using laser guidance; and
  • FIGS. 20 is a flow chart describing a sequence of steps of an exemplary x-ray alignment method.
  • DETAILED DESCRIPTION Definitions
  • An “imaging system” is a system that includes at least a movable arm, an x-ray source, an x-ray detector, a display and a system controller. A “patient 3-dimensional image data set” is a three dimensional numerical array whose elements hold the values of specific physical properties at points in space inside the patient's body. A “multiplanar reformation image (MPR)” is a planar cross-section of the patient 3-dimensional image data set generated by cutting through the three-dimensional data set at some orientation (e.g., axial, coronal, sagittal, or oblique). A “fluoroscopic image” is a two-dimensional x-ray projection image showing internal tissues of a region of the body. A “live fluoroscopic image” is a sequence of x-ray images taken successively showing live movement of internal tissues of a region of the body. A “combined image” is an image in which an x-ray image is combined with a two-dimensional rendering of a three-dimensional data set computed under the same orientation as the x-ray image. A two-dimensional rendering of a three-dimensional data set is also referred to as a fluoroscopic overlay image or overlay image. “Co-registering” means aligning an x-ray image with a patient 3-dimensional image data set such that associated features within the x-ray image and a two-dimensional overlay image generated from the patient 3-dimensional image data set appear at the same location on a display in which the x-ray image and the overlay image are shown together. Co-registration can be point-based or gray-level based. In point-based co-registration, a transform is applied to the 3-dimensional image data set such that points in the resulting overlay image line up with their counterparts in the x-ray image as closely as possible. Gray-level based co-registration techniques determine the transform not by minimizing the distance between associated points in the overlay image and x-ray image, but by minimizing an error metric based on the resulting overlay image's gray levels and the x-ray image's gray levels. Co-registration can be static or elastic. In the latter case, the 3D data set may not only be rotated and shifted but warped as well. “Instrument” refers to any object which may pierce tissue of a patient, a non-limiting listing of which include needles and other biopsy devices, screws, implants, cannula, endoscopes, and anything else that can be inserted into a patient's body either percutaneously or intravascularly. A “skin entry point” is the position on a patient's skin at which an instrument is inserted. “Skin entry point data” is data representative of the skin entry point within the patient 3-dimensional image data set or within two x-ray views taken under different view orientations using a triangulation technique. A “target” or “target point” is a point within the body of a patient that is the subject of a percutaneous procedure. “Target point data” is data representative of the skin entry point within the patient 3-dimensional image data set or within two x-ray views taken under different view orientations using a triangulation technique. A “planned path” is a line generated between the skin entry point and the target point. “Instrument trajectory” is a desired trajectory of the instrument defined by the planned path. A “Bull's Eye View” is an x-ray view under which a target point and another point along the instrument trajectory are projected onto each other. The other point along the instrument trajectory may be the skin entry point. The movable arm view direction can be visualized using a graphical overlay in which the target point and skin entry point, forward-projected from 3-dimensions to 2-dimensions, are displayed as individual circles. If the Bull's Eye View has been reached, these two circles are projected at the same 2-dimensional position (i.e., they appear concentrically aligned). A “progression view” is an x-ray image taken at an oblique angle with respect to a line joining the skin entry point and the target. The term movable arm tomographic reconstruction refers to a technique in which multiple x-ray images taken along a particular image acquisition trajectory of the movable arm system, e.g., a partial circle. The multiple x-ray images are used to construct a patient 3-dimensional image data set. An “optical axis,” or “central ray” is an imaginary straight line passing through the midpoint of an x-ray source and the iso-center of the imaging device. In an ideal system geometry, the optical axis is perpendicular to the x-ray detector, intersecting the detector at its midpoint.
  • Referring to FIG. 1, an exemplary system 1 is shown for performing a percutaneous procedure. The system 1 may comprise an x-ray tube or source 2 and associated support and filtering components. The x-ray source may be affixed to a support, such as a movable arm 4 to allow the x-ray source to be moved within a constrained region. In one embodiment, the movable arm 4 is a C-arm. The constrained region may be arcuate or otherwise three dimensional, depending on the nature of the support structure. A collimator may also be included, which defines the size and shape of x-ray beam 6 emerging from the source. An x-ray exposure controller 8 and system controller 10 may also be included. System controller 10 may be a personal computer or any known controller capable of receiving and transmitting control signals to/from the above-described x-ray system components via a hardware interface 12. System controller 10 may include a user input device 14, such as a trackball, mouse, joystick, and/or computer keyboard to provide for user input in carrying out various system functions, such as mode selection, linearity control, x-ray dose control, data storage, etc. The system controller 10 may include a processor 16 executing instructions for performing one or more steps of the disclosed method.
  • In the illustrated embodiment, a patient 18 is shown on patient-support table 20 such that an x-ray beam 6 generated by the x-ray source passes through him/her onto a detector 22. In one embodiment the detector 22 is a flat panel detector that acquires digital image frames directly, which are transferred to an image processor 24. A display/record device 26 records and/displays the processed image(s). The display/record device 26 may include a display for displaying the displayed image output, as well as a separate device for archiving. The image is arranged for storage in an archive such as a network storage device. The x-ray source 2 is controlled by the system controller 10 via exposure controller 8 and x-ray generator 28. The position of the x-ray source 2 may be adjusted via a drive system associated with the movable arm 4. The movable arm 4, x-ray source 2, x-ray detector 22, display 26 and system controller 10 may together be referred to as an imaging system.
  • As previously noted, laser-guided and radiation-guided systems are known for use in guiding instruments during percutaneous procedures. Although two specific laser guidance and x-ray guidance systems and methods will be described herein, it will be appreciated that the specifics of each system are not critical, and that any of a variety of other laser guidance and/or x-ray guidance systems and methods may be used with the disclosed system and method.
    • Workflow Steps
  • Initially, a 3-dimensional data set of a targeted region within the patient will be obtained using one of a variety of techniques, such as computed tomography (CT), magnetic resonance imaging (MRI), positron emission tomography (PET), C-arm tomographic imaging (syngo DynaCT, Siemens AG, Forchheim, Germany) or the like. Referring to FIG. 2, a 3-dimensional rendering of such a 3-dimensional data set is shown. In the illustrated embodiment, the 3-dimensional dataset has been loaded into an appropriate 3D rendering program, such as the Siemens InSpace system, for viewing (the figures show 3-dimensional images representative of test phantoms that have a plurality of objects placed inside to simulate vessels, landmarks and targets).
  • The x-ray views (fluoroscopic images) obtained using the movable arm 4, source 2 and detector 22, need to be appropriately “registered” with the overlay images derived from the 3-dimensional data set for the region of interest of the patient. Data registration may be performed manually, automatically or semi-automatically (i.e., computer assisted). This registration step ensures that the fluoroscopic (x-ray) images of the patient obtained using the source 2 and detector 22 match the images of the patient constructed from the 3-dimensional data set. This enables instrument positioning using information on target position obtained from the 3-dimensional data set. In one embodiment, the co-registration step is performed by minimizing an error metric based on gray levels of a resulting overlay image and the x-ray image. In another embodiment, the co-registration step is performed by applying a transform to the 3-dimensional image data set such that points in a resulting overlay image align with counterpart points in the x-ray image.
  • Once the 3-dimensional data set has been appropriately registered to the 2-dimensional x-ray geometry, the instrument trajectory may be planned. To this end, the user may select a target point, Xt, and a skin entry point, Xe within the overlay images by visualizing the areas within a particular image and clicking on the point(s) using a selector such as a mouse button. Alternatively, the instrument trajectory may be planned before the registration step. In either case, registration ensures that the planned trajectory is mapped to the corresponding anatomical environment.
  • As shown in FIG. 3, target point selection may be done by selecting a desired MPR view, such as by a right-click of a mouse pointer on an appropriate soft key 3A in a pop-up window in the display. In the illustrated embodiment, this results in desired MPR views being displayed in the upper left quadrant 3B, and upper right quadrant 3C. The target point is “selected” by clicking with a mouse pointer at the target position 3D in the lower left quadrant MPR display 3E. The skin entry point may be selected (“clicked”) in the same manner.
  • Based on where the click points are made in the MPR view, the system obtains data representative of the target and skin entry points using data from the 3-dimensional patient data set. Using the target point data and skin entry point data, the system generates a graphical overlay showing a line which represents the planned instrument trajectory. Such a graphical overlay is applied to each of the images shown on the user display (as seen as line 4F in FIG. 4). This graphical overly may consist of the target and skin entry points, as well as a line intersecting the two, and may be overlaid onto one or more of the display views (e.g., x-ray, MPR, 3-dimensional rendering) as desired by the user. Since the x-ray views and the patient 3-dimensional image data set are registered with one another at this point in the procedure, the system can map the exact location of the target point Xt, and the skin entry point Xe (as well as the connecting vector “n”) at their precise locations on each of the display views. As will be described in greater detail later, the displayed line represents the planned instrument trajectory.
  • As shown in the four display quadrants 4A, 4B, 4C and 4D of FIG. 4, the skin entry point Xe and the target point Xt define a line 4F (shown in 4 places) in 3-dimensional space having a path vector (n=Xe−Xt). It will be appreciated that, as an alternative to defining the path vector using two points in space, the trajectory could instead be defined using xt and the path vector “n”. This may be appropriate, for example, in the case where the patient is large and the skin entry point cannot be seen in the movable arm CT (i.e., it is outside the physical range of the movable arm CT). In such a case, the user may not need to see the exact location of the skin entry point if there are no organs in the immediate area (e.g., where there is only fat tissue). Whether this is indeed the case or not can be checked if the registered 3-dimensional patient data set comprises the complete volume of the patient or if there is another 3-dimensional data set (e.g., CT, MRI) that provides similar information in outer patient regions. In such a case, the physician may simply “click” the desired target point xt in the appropriate display view(s), and may “click” a second point that is visible in the movable arm CT to define a desired path vector n without specifying an actual skin entry point Xe. A desired instrument trajectory is a straight line (path vector) “n” that originates outside the patient's body and passes through the skin entry point xe and the target point xt without crossing bones or vital organs.
  • As shown in FIG. 4, the graphical overlay (i.e., the one in which points Xt and Xe are shown along with the line connecting them) may be combined with an anatomical image (i.e., an MPR view). In addition, the graphical overlay may be combined with both the forward projected anatomical image (overlay image) and live x-ray views. In the final movable arm position (again, the Bull's Eye View position), the graphical overlay may also adjust to different x-ray zoom conditions so that the user may confirm final positioning by revealing small deviations from the optimal view orientation. This resizing is automatically achieved through the use of a calculated conversion factor determined e.g., using a “similar triangles” technique.
  • Once the planned instrument trajectory is obtained, a determination is made regarding whether a laser guidance procedure can be performed, or whether an x-ray guidance procedure can be performed, or whether the path needs to be replanned. Laser guidance can be used if the planned instrument trajectory can be targeted in such a manner that the optical axis of the x-ray source on the C-arm coincides with the planned trajectory. Since the C-arm is calibrated to the 3D image data set, the system can automatically determine whether such an orientation is possible.
  • If the system determines that laser guidance cannot be employed, a determination is made regarding whether the C-arm can be made to coincide with the planned instrument trajectory so that the extension of the trajectory is projected onto the detector (is in the “Bull's Eye View”), in which case an x-ray guidance procedure can be used for instrument guidance.
  • If the planned instrument trajectory does not satisfy either of the aforementioned criteria, the instrument trajectory is replanned until at least an x-ray guidance procedure can be used. Of course, the user can re-plan the instrument trajectory more than once (e.g., if a laser-guided procedure is preferred, re-planning can be performed as many times as desired in order to achieve a planned instrument trajectory that meets the criteria for utilizing laser guidance).
  • Once a final instrument trajectory is obtained, the instrument is guided to the target using the selected guidance technique, as will be explained in greater detail below with reference to the respective procedure).
  • Exemplary Laser Guidance System and Method
  • Referring now to FIGS. 5-10, an exemplary laser guidance system 29 for use with the disclosed method will be described in detail. Referring to FIGS. 5 and 6, a movable stand 30 is attached to a patient table 20. The stand 30 has a swivel arm 32 which is rotatable according to the arrow 34 in FIG. 6. A fastening device 36 is provided on the swivel arm 32, on which a laser pointer 38 is mounted. The fastening device 36 is designed in such a way that 3-dimensional movement of the laser pointer 38 is possible, in particular according to the motion arrows 40, 42 and 44.
  • The stand 32 may have a position sensor 46 through which the position of the laser pointer 38 is measurable relative to the patient table 20. The position sensor 46 may be used to define a coordinate system which can be recorded with the previously obtained 3-dimensional data set.
  • In one embodiment, motors may be used to achieve the automated movements according to the motion arrows 34, 40, 42 and 44. Such motors are not required, and the laser guidance system 29 can instead be moved by hand.
  • In an alternative embodiment, the laser pointer 38 may be affixed to the x-ray detector 22. FIG. 7 shows the x-ray detector 22 to which two braces 48 are attached, which run parallel to the surface 50 of the x-ray detector 22. A sheath 52 is positioned perpendicular to the braces 48 and is used to attach the laser pointer 38. With this embodiment, to move the laser guidance system the entire x-ray detector 22 is moved, in particular the C-arm 4 is swiveled or tilted, to which the x-ray detector 22 and laser pointer 38 are attached.
  • In the event that no additional fluoroscopic images are deemed necessary during a percutaneous procedure, the laser pointer 38 can remain in the sheath 52. Where braces 48 are used to attach the laser pointer 38 to the x-ray detector 22, the braces and pointer may be separated from the x-ray detector 22 after reaching a desired instrument alignment.
  • Referring now to FIG. 8, a stand 54, similar to stand 32 in FIGS. 5 and 6, may be used to guide the laser pointer 38. The laser pointer 38 may be fixed to the stand 54, separate from the x-ray detector 22. The C-arm 4 may be turned, e.g., into the position shown in FIG. 9. The laser pointer 38, since it is attached to the stand 54, remains aligned to the patient 18, and thus an instrument can be aligned along the laser beam 56 generated by the laser pointer 38. During the percutaneous procedure, fluoroscopic images can be taken using the x-ray source 2, x-ray detector 22, and C-arm 4 oriented at an oblique angle with respect to the instrument trajectory (as defined by the laser beam 56). As will be appreciated, these fluoroscopic images may aid the physician to guide the instrument precisely and accurately to the target.
  • In a further alternative embodiment, a pair of laser sources 58 may be arranged on or near the edges of the x-ray detector 22 so that they are not visible in the images produced using the detector 22. Such an arrangement is shown in FIG. 10. The laser sources 58 may generate fan-shaped light in which a fan angle a is defined and along which the fan extends, such that produced laser beams are on a single level. The two levels intersect in a central straight beam line 60 such that this straight beam line 60 can readily be recognized by the user. The extension of the straight line 60 is the skin entry point Xe and the target point Xt, and thus represents the instrument trajectory. The instrument is aligned using this beam line 60. In one embodiment, a laser pointer 38 is aligned along straight beam line 60, and the x-ray detector 22 may be moved as desired to obtain one or more fluoroscope images, while the laser pointer 38 is used as part of the guidance system.
    • Exemplary X-ray Guidance System and Method
  • Referring now to FIGS. 11-17, an exemplary x-ray guidance system and method will be described in greater detail. As previously described, the system generates a line “n” between the skin entry point Xe and the target point Xt. This line is used to align the movable arm 4 to achieve a “Bull's Eye View,” in which the two points are superimposed to show only a single point to the user. The instrument is placed at the skin entry point and aligned using the Bull's Eye View, shown in display quadrant 11A of FIG. 11, to orient the instrument along the planned instrument trajectory (i.e., one that hits both points). Initial alignment is verified using a fluoroscopic image of the oriented instrument. After the initial alignment is verified, the user inserts the instrument a short distance into the patient. One or more progression x-ray views are used to verify that the instrument is on the planned path between the skin entry point and the target point. The user may employ an iterative approach of inserting the instrument a small distance followed by a verification of the instrument's position using progression x-ray views to guide the instrument to the target. When the instrument reaches the target, a desired additional procedure may be performed, such as a biopsy, a drainage procedure, a radiofrequency ablation, or other medical interventional procedure.
  • Once an acceptable instrument trajectory has been planned, the movable arm 4 may be moved into the Bull's Eye View position. As previously noted, the Bull's Eye View orientation is one in which the skin entry point and the target point (Xe and Xt) overlie each other on the same detector positions (see FIG. 4). Adjustment of the movable arm 4 to achieve this positioning can either be performed manually or automatically.
  • In practice, positioning an instrument at the skin entry point Xe may be a difficult task, and thus a positioning aid may be used, such as a biopsy grid, a biopsy mesh, or other appropriate device. Once the appropriate instrument positioning has been achieved, collimation may be set around the Bull's Eye View to limit radiation exposure to the user and patient. Collimation may be either manual or automatic.
  • To facilitate fine adjustment of the instrument position, the Bull's Eye View may be isolated and enlarged, as shown in FIG. 12, to reveal slight deviations from the desired instrument positioning and orientation. Thus, FIG. 12 shows a switch from the four-quadrant view of FIG. 11 to a full-window view with an increased zoom level to reveal deviations from the ideal Bull's Eye View. As can be seen, the zoomed view of FIG. 12 shows concentric overlapping circles 62 (in black), 64 (in white) indicating that the Bull's Eye View has been achieved. In the illustrated embodiment, a SeeStar device (manufactured by Radi Medical Devices, Uppsala, Sweden) has been used to aid instrument positioning. The SeeStar device is an instrument guide that produces an elongated artifact in an x-ray image. This elongated artifact indicates the trajectory of an instrument inserted through the SeeStar, and thus it can be determined whether the selected trajectory will intersect the target as desired. The SeeStar shows up as a circle 66 (i.e., a black tube-like shadow in the figure) in the center of the displayed circles, which indicates that it is in the desired orientation (i.e., one that is in alignment with a trajectory that passes through the skin entry point and the target point). If the SeeStar were to show up as a line, its position/orientation would be adjusted, followed by re-verification of the new position/orientation by subsequent x-ray views.
  • As previously noted, in lieu of a SeeStar device, the user could instead use a hollow instrument guide to verify instrument placement. The hollow instrument guide may be configured so that it shows up as a point under fluoroscopy in the Bull's Eye View when a desired alignment is achieved. The hollow instrument guide may be clamped in position during fluoroscopy to limit radiation to the user, and its position may be adjusted and verified in a manner similar to that described in relation to the SeeStar device.
  • Once the desired instrument alignment is achieved, the instrument is pushed forward by a small amount into the patient tissue to stabilize the instrument's orientation. This insertion is performed under the Bull's Eye View. As shown in FIG. 13, the user can see straight down the instrument guide as well. The large circle represents the instrument body and instrument tip. In the illustrated embodiment they are exactly aligned, which is why only one large circle is visible in the figure. The black “bulb” in the center is the instrument (in the illustrated case, a needle). It appears in this way because it is almost (but not perfectly) aligned with the viewing direction. If the instrument were perfectly aligned, it would be shown as a circle in this view.
  • Instrument alignment may again be verified at this early stage of insertion. Such verification can be performed using x-ray “progression views,” which are oblique x-ray views (i.e., non-Bull's Eye Views) obtained using the source 2 and detector 22. It will be appreciated that the user may also return to the Bull's Eye View at any time during the procedure to obtain additional information regarding instrument alignment. If a bi-plane x-ray device is available with the B-plane providing a progression, it is possible to check if the instrument remains aligned with the associated graphical overlay (shown as line 68 in FIG. 14) while the instrument is being pushed forward into the tissue. In the illustrated embodiment, the instrument appears as a thin diagonal line starting from the bottom left of the image, just above the graphical overlay line 68.
  • During the procedure, the movable arm 4 may be moved between the first and second progression views to enable the user to control the actual instrument movement from two oblique angles until the instrument has reached the target. When the target has been almost reached in one progression view, the user can return to the other progression view to confirm that the instrument has indeed been placed correctly before making the final push or releasing a spring-loaded biopsy device if one is used. The user can also return to the Bull's Eye View to obtain additional orientation information.
  • Referring to FIG. 15, the movable arm is moved into the second progression view to check on instrument placement. If instrument 70 and graphical trajectory 72 align, the instrument 70 can be moved into the target. In the illustrated embodiment, a small degree of bend is shown in the instrument 70, which can occur when the instrument is small/thin and the target is dense.
  • Referring to FIG. 16, a return to the first progression view is performed to confirm instrument placement at the target. It will be appreciated that if a bi-plane fluoroscopic device is available, there is no need to rotate the movable arm's A-plane back and forth between two progression views. Instead, the A-plane can be put at the first progression view while the B-plane is positioned under the second progression view, and both may be viewed simultaneously As an alternative to the use of progression views to verify instrument positioning during insertion, movable arm CT (DynaCT) acquisitions can be performed throughout the workflow to verify the instrument position and orientation at each stage of insertion. Such a movable arm CT acquisition can be performed at one or more stages of the procedure, as desired by the user. It is noted, however, that the movable arm CT procedure can take up to several minutes and increases the overall amount of radiation exposure to the patient. Progression views, by contrast, are relatively fast (almost instantaneous). The user simply rotates the movable arm (if required) to the desired progression view location, and releases the x-rays. The x-ray image shows up on the display in a few seconds.
  • FIG. 17 shows a movable arm CT (DynaCT) scan of the completed instrument insertion position. Although not required, performing such a verification ensures that the positioning is correct prior to completing the procedure. As can be seen in the MPR views shown in the upper and lower left quadrants 17A, 17B of FIG. 17, the instrument 70 has been appropriately engaged with the target 72. The upper right quadrant view (which shows the Bull's Eye View), however, reveals that the instrument 70 just made it into the target 72.
  • Referring now to FIGS. 18A-18E, the disclosed method for selecting a guidance technique for performing a percutaneous procedure will be described in greater detail. At step 1800 (FIG. 18A), a patient 18 is placed on a table 20 in proximity to an imaging system having a movable arm 4, an x-ray source 2, an x-ray detector 22, a system controller 10 and a display 26. At step 1810, a 3D patient image data set of a patient tissue region is provided. At step 1820, an x-ray image of a targeted patient tissue region is acquired using the x-ray source and x-ray detector. In one embodiment (see FIG. 18B), at step 1822 obtaining an x-ray image of the patient tissue region using the x-ray source 2 and x-ray detector 22 comprises obtaining a plurality of x-ray images, displaying the plurality of x-ray images on the display 26 along with a three-dimensional rendering of the three-dimensional image data set, and an overlay image representing a 2-dimensional rendering of the 3-dimensional image data set. At step 1824, the skin entry point, the target point and the planned instrument trajectory are graphically displayed in their respective positions on the plurality of displayed x-ray images, the three-dimensional rendering and the overlay image. At step 1830 (FIG. 18A), the 3D patient data set is co-registered to the x-ray image acquired using the imaging system. In one embodiment (see FIG. 18C), at step 1832 the co-registering step comprises applying a transform to the 3D patient image data set such that points in a resulting overlay image align with counterpart points in the x-ray image. At step 1840 (FIG. 18A), the user plans an instrument trajectory in the 3D volume by selecting target point data representative of a target point within the patient tissue region, and skin entry point data representative of a skin entry point. The target point data and skin entry point data may be obtained from the co-registered 3D image data set. As a result of co-registration, it may also be possible to plan the instrument path immediately after step 1800. This may be beneficial if planning is performed on a pre-procedural data set acquired on an imaging device that is different from the C-arm imaging system.
  • At step 1850 instrument alignment is performed using one of a plurality of instrument guidance procedures, where the instrument guidance procedure is selected on the basis of the planned instrument trajectory, and a position of the movable arm 4. The plurality of instrument guidance procedures may include a laser guided procedure, an x-ray guided procedure, or a procedure that replans the instrument trajectory. If the planned instrument trajectory can be targeted in such a manner that the optical axis (central ray) of the x-ray source 2 on the C-arm 4 coincides with the trajectory, at step 1860 a laser guidance procedure (such as the one described previously in relation to FIGS. 5-10, and discussed in more detail below) is selected for instrument guidance. This determination is based on information relating to calibration of the C-arm with the 3D volume and a collision control feature of the system. The step of aligning the instrument may comprise aligning the instrument along a laser beam path generated by a laser pointer, the laser beam path intersecting the skin entry point Xe and the target point Xt (step 1852, FIG. 18D). At step 1854, the laser beam path may comprise the intersection of at least two laser fan patterns.
  • If the central ray of the x-ray source does not coincide with the planned instrument trajectory, and an extension of the planned instrument trajectory intersects with the x-ray detector 22, at step 1870 an x-ray guidance procedure (such as the one described previously in relation to FIGS. 11-17, and discussed in more detail below) is selected for instrument guidance. If a central ray of the x-ray source does not coincide with the planned instrument trajectory, and an extension of the planned instrument trajectory does not intersect with the x-ray detector 22, the instrument trajectory is replanned at step 1840. This replanning step may involve repositioning the patient by moving the patient table. At step 1890, a warning signal may be provided to the user if the central ray of the x-ray source does not coincide with the planned instrument trajectory, and an extension of the planned instrument trajectory does not intersect with the x-ray detector 22.
  • If laser guidance can be employed, at step 1900 (see FIG. 19), the 3D image data set is represented from the perspective of the target guidance system. The target guidance system defines a vector in a view direction to the patient. The patient is assigned to the 3D image data set so that the target guidance system defines a view direction on the 3D image data set. This may occur so that the target guidance system is represented in the central point of the monitor as an X, and the target point should also be represented as an X. The skin entry point Xe and target point Xt may also be represented. The goal is for the target guidance system to precisely show these points superimposed on each other. It will be appreciated that the target may be represented as an X, while the skin entry point may be represented using a circle, or that circles may be used to represent both the target and skin entry points. Either graphical representation can be combined with a fluoroscopic overlay image representing the patient's anatomical structures.
  • The user may move the laser guidance system 29 or the patient 18 at step 1910 to achieve an appropriate position so that the points are superimposed. Because this movement changes the view of the laser guidance system to the patient and with it the 3D image data set, the representation is adjusted at step 1920. At step 1930, a check is performed to determine whether the target representation has been reached (i.e., whether all of the points align with each other). If the target representation is not achieved, steps 1910, 1920 and 1930 are repeated until the points are superimposed. Once the desired target representation is reached, at step 1940 the instrument is guided using the target guidance system. If the laser beam exactly aligns with the central x-ray of the x-ray source, the laser may be used by itself to provide instrument guidance. It will be appreciated that for instances in which the laser alone is used for instrument guidance, that the practitioner still may verify the instrument position within the patient using x-ray imaging.
  • In one embodiment, automatic positioning of the patient table 20 may be achieved at step 1910. The goal is to align the central x-ray of the x-ray source with the planned instrument trajectory. Thus, the position of the patient table 20 is automatically adjusted using positioning motors system until the planned instrument trajectory intersects the iso-center of the C-arm 4. The C-arm 4 is rotated such that its view orientation matches the orientation of the instrument trajectory.
  • Referring again to FIG. 18A, if, at step 1850 it is determined that laser guidance cannot be employed, but the planned instrument trajectory still can be targeted in such a manner that a central ray of the x-ray source can be made to coincide with the planned instrument trajectory so that the extension of the trajectory intersects with the detector 22 (the “Bull's Eye View”), at step 1860 an x-ray guidance procedure (such as the one described in relation to FIGS. 11-17) is selected for instrument guidance.
  • If x-ray guidance is used, at step 2010 (see FIG. 20), the movable arm 4 is adjusted to a position at which an x-ray image taken using the x-ray source 2 and the x-ray detector 6 results in the target point Xt and the skin entry point Xe being superimposed on top of each other. In one embodiment, the step of adjusting the movable arm may comprise determining a spatial orientation within the three-dimensional image data set at which the target point and skin entry point are superimposed on each other, and automatically moving the movable arm so that a further x-ray image obtained using the x-ray source 2 and detector 22 images the target and skin entry points onto the same pixels of the x-ray detector.
  • At step 2020, alignment of an instrument positioned between the x-ray source 2 and the skin entry point Xe is verified as an acceptable position with respect to the planned instrument trajectory when the instrument appears on the display as a point overlying the target point Xt and the skin entry point in a verification x-ray image taken using the x-ray source 2 and detector 22. In one embodiment, acceptable position with respect to the planned instrument trajectory is verified by taking multiple x-ray images using the x-ray source 2 and detector 22 at movable arm positions oblique to the position of the movable arm 4 used to obtain the verification x-ray image.
  • In further steps, the user may insert the instrument into the patient at the skin entry point. One or more progression x-ray views may be taken to ensure that the instrument remains aligned with the projected instrument path. It will be appreciated that the user may also return to the Bull's Eye View to gain additional insights regarding instrument orientation. The user may press the instrument further into the patient toward the target while making adjustments to ensure the instrument remains aligned with the projected instrument path. The pressing and progression x-ray steps may be repeated as desired by the user to guide the instrument in an incremental manner to intersect the target.
  • Referring again to FIG. 18A, if neither laser guidance nor x-ray guidance can be used (i.e., the central ray of the x-ray source does not coincide with the planned instrument trajectory and an extension of the planned instrument trajectory does not intersect with the x-ray detector, at step 1880 it is determined that the planned instrument trajectory needs to be replanned, and the process returns to step 1840 and the instrument trajectory is replanned. Replanning is performed that at least the instrument trajectory can be planned using a laser or x-ray guidance techniques, as determined again in steps 1850, 1860 and 1870. Where the system determines that the planned instrument trajectory cannot be targeted using either laser or x-ray guidance techniques (i.e., when a central ray of the x-ray source does not coincide with the planned instrument trajectory and an extension of the planned instrument trajectory does not intersect with the x-ray detector, an alert such as a warning or other signal may be provided to indicate to the user that the trajectory must be re-planned.
  • In one embodiment, shown in FIG. 18E, a suggested re-planned instrument trajectory may be automatically formulated and presented to a user so that either laser guidance or x-ray radiation guidance techniques can be used. Thus, at step 1842 re-planning may be carried out manually (using a graphical user interface (GUI)), or by means of a “tracked” instrument (as will be described in greater detail below). Alternatively, at step 1844 re-planning can be carried out automatically. Once the instrument is reoriented along the re-planned trajectory, the instrument may be advanced and controlled as described previously in relation to the laser-guidance and x-ray-guidance techniques.
  • In one alternative embodiment, skin entry point replanning may be facilitated by the use of an optical or electromagnetic position sensor attached to the instrument. The position sensor may automatically identify both the location and the direction of movement of the instrument so that when the position sensor is moved on the surface of the skin of a patient, the skin entry point Xe can be newly determined such that, given a stationary target point (but a variable “tracked” starting point), the resultant path of the instrument can automatically be entered into the 3D dataset (and rendered as desired). Once an appropriate instrument trajectory has been selected, the skin entry point Xe is automatically determined and a marking (e.g., color mark, optionally combined with the injection of a contrast agent) can be made on the skin. In one exemplary embodiment, the instrument may comprise a biopsy localization device (e.g., SeeStar) with a position sensor.
  • In another embodiment, a flexible mounting arm may be provided with a holder for an instrument guide (e.g., guide sleeve, instrument mounting or clamping device). The flexible mounting arm can be connected to structure such as the patient table or other adjacent equipment and may be used to hold an instrument guide during the procedure. The flexible mounting arm may be automatically robotically controllable to enable the user to determine the instrument trajectory automatically or semi-automatically, without touching the patient. Once the instrument trajectory has been finalized, the instrument may be precisely and accurately inserted into the patient. This technique may also facilitate the rapid percutaneous insertion of an instrument (e.g., by means of a biopsy cannula).
  • In further embodiments in which the instrument comprises a biopsy cannula, a “tracked” head (e.g., ultrasound) may be used to enable real time display of the instrument in the 3D volume (i.e., rendered or otherwise displayed on the display 26, providing a “live image” of the instrument as additional piece of information to the user.
  • Furthermore, when replanning of the instrument trajectory is performed, a forward-looking (i.e., “endoscopic view”) may be provided to the user to enable the user to “look” into the direction of the target (see FIG. 12). Similarly, other “context-based” visualization methods may be employed, such as those which eliminate soft tissue from the 3D dataset, while retaining other elements of interest such as bones or vessels. This may enable the physician to see whether obstructions exist along the path of a planned instrument trajectory, and, furthermore, whether the path is safe (i.e., does not intersect with blood vessels). Such a 3D dataset could alternatively or additionally be acquired by means of an ultrasound head which can be movably attached on a mounting arm and “tracked.”
  • Alternatively, a “tracked” 2D ultrasound transducer may be employed to allow the user to determine both the target and the direction of a percutaneous intervention. If the ultrasound transducer and the C-arm are registered with each other, the ultrasound path can be imported into the display, and used to orient the C-arm.
  • It will be appreciated that if the x-ray guidance technique is used to guide the instrument, the projection of the “skin entry point” and the “target point” does not impinge on the center of the detector 22 but rather on an outer region of the detector which cannot be shielded by a central collimator (if one is used). To alleviate this, the table 20 may be moved in such a manner that the “bull's eye view” passes through the central collimating opening.
  • The system may additionally employ a limitation on potential C-arm trajectories to ensure that the C-arm will not collide with the user. For example, when using the x-ray guidance procedure, the user may move from the “bull's eye view” into first and second progression views to check the position of the inserted instrument within the patient (see FIG. 14). In such cases, it may not be possible to determine whether moving the C-arm to obtain those progression views will cause a collision between the C-arm and the user. This uncertainty factor can be avoided by limiting the direction of motion of the C-arm during interventions in such a manner that rotations around the patient are carried out along the median plane (i.e., at a cranial/caudal (CRAN/CAUD) angle of 0 degree). Thus, when moving the C-arm from one position to another, the angulation of the C-arm would first be retracted (in CRAN/CAUD direction), a new angle of rotation would be targeted (in left anterior oblique/right anterior oblique (LAO/RAO) direction), and the new angulation set. This eliminates the risk that a C-arm could collide with the leg of the user since he would know what movement to expect.
  • In a further embodiment, feedback-controlled laser orientation technique can also be employed. In this embodiment, an additional signal (in addition to the optical signal on the screen, given by the superimposition of certain markers) is provided to the system to provide feedback as to how accurately a planned instrument trajectory coincides with the optical axis of the system (i.e., the direction of the laser).
  • Other user-feedback techniques may also be used, such as displaying on the display 26 offsets (in degree or distance) between the planned instrument trajectory and the optical axis of the system. Alternatively, a lamp may be provided, the color of which changes as the two are more closely aligned. Acoustic signals and haptic force feedback are other examples of acceptable feedback techniques. If such a feedback is available, the workflow could be as follows: at the beginning of the procedure, the C-arm is moved into a “generalized bull's eye view.” A “generalized bull's eye view” is an x-ray view projecting target and skin entry points onto each other along an x-ray beam direction that does not coincide with the central ray of the x-ray cone. If work with the laser is desired, the table can be manually moved until the patient is in such a position that the C-arm can be moved into a viewing direction in which the axis of the optical system intersects the target point. Only after the patient and the C-arm are moved into the correct positions does the laser turn on. In all other cases, it will not. If the C-arm is moved manually, “force feedback” can be used to ensure that the positions can be approached in a controlled manner, without overshooting them. Alternatively, positioning motors may be provided to enable the table to be automatically moved into a position suitable for laser operation.
  • The method described herein may be automated by, for example, tangibly embodying a program of instructions upon a computer readable storage media capable of being read by machine capable of executing the instructions. A general purpose computer is one example of such a machine. A non-limiting exemplary list of appropriate storage media well known in the art would include such devices as a readable or writeable CD, flash memory chips (e.g., thumb drives), various magnetic storage media, and the like.
  • The features of the method have been disclosed, and further variations will be apparent to persons skilled in the art. All such variations are considered to be within the scope of the appended claims. Reference should be made to the appended claims, rather than the foregoing specification, as indicating the true scope of the disclosed method.
  • The functions and process steps herein may be performed automatically or wholly or partially in response to user command. An activity (including a step) performed automatically is performed in response to executable instruction or device operation without user direct initiation of the activity.
  • The systems and processes of FIGS. 1-20 are not exclusive. Other systems, processes and menus may be derived in accordance with the principles of the invention to accomplish the same objectives. Although this invention has been described with reference to particular embodiments, it is to be understood that the embodiments and variations shown and described herein are for illustration purposes only. Modifications to the current design may be implemented by those skilled in the art, without departing from the scope of the invention. The processes and applications may, in alternative embodiments, be located on one or more (e.g., distributed) processing devices accessing a network linking the elements of FIGS. 1 and 5-10. Further, any of the functions and steps provided in FIGS. 18A-20 may be implemented in hardware, software or a combination of both and may reside on one or more processing devices located at any location of a network linking the elements of FIGS. 1 and 5-10 or another linked network, including the Internet.

Claims (24)

1. A method for planning a percutaneous procedure for use in a system comprising an imaging system having a movable arm, an x-ray source and an x-ray detector and a display and a system controller connected to and in communication with the imaging system and display, comprising:
providing a three-dimensional image data set of a patient tissue region;
obtaining an x-ray image of the patient tissue region using the x-ray source and the x-ray detector;
co-registering the three-dimensional image data set to the x-ray image acquired using the imaging system;
obtaining a planned instrument trajectory based on target point data representative of a target object within the patient tissue region and a skin entry point data representative of a skin entry point; and
aligning an instrument along the planned instrument trajectory using one of a plurality of instrument guidance procedures, the one of a plurality instrument guidance procedures being selected on the basis of the planned instrument trajectory and a position of the movable arm;
wherein the plurality of instrument guidance procedures comprises a laser guided procedure, an x-ray guided procedure, and a procedure that replans the instrument trajectory.
2. The method of claim 1, wherein the laser-guided procedure is selected if a central ray of the x-ray source coincides with the planned instrument trajectory.
3. The method of claim 1, wherein the x-ray guided procedure is selected if a central ray of the x-ray source does not coincide with the planned instrument trajectory, and an extension of the planned instrument trajectory intersects with the x-ray detector.
4. The method of claim 1, wherein the instrument trajectory is replanned if a central ray of the x-ray source does not coincide with the planned instrument trajectory and an extension of the planned instrument trajectory does not intersect with the x-ray detector.
5. The method of claim 4, further comprising a warning signal to a user when a central ray of the x-ray source does not coincide with the planned instrument trajectory and an extension of the planned instrument trajectory does not intersect with the x-ray detector.
6. The method of claim 1, wherein replanning the instrument trajectory comprises manually replanning the instrument trajectory using a graphical user interface.
7. The method of claim 1, wherein replanning the instrument trajectory comprises automatically replanning the instrument trajectory.
8. The method of claim 1, wherein the co-registering step comprises applying a transform to the 3-dimensional image data set such that points in a resulting overlay image align with counterpart points in the x-ray image.
9. The method of claim 1, wherein the step of aligning an instrument comprises aligning the instrument along a laser beam path generated by a laser pointer, the laser beam path intersecting the skin entry point and the target object.
10. The method of claim 9, wherein the laser beam path comprises the intersection of at least two laser fan patterns.
11. The method of claim 1, wherein the step of obtaining an x-ray image of the patient tissue region using the x-ray source and the x-ray detector comprises obtaining a plurality of x-ray images, displaying the plurality of x-ray images on the display along with a three-dimensional rendering of the three-dimensional image data set, and an overlay image representing a 2-dimensional rendering of the 3-dimensional image data set.
12. The method of claim 11, wherein the skin entry point, the target point and the planned instrument trajectory are graphically displayed in their respective positions on the plurality of displayed x-ray images, the three-dimensional rendering and the overlay image.
13. A system for planning a percutaneous procedure for use in a system comprising an imaging system having a movable arm, an x-ray source and an x-ray detector and a display and a system controller connected to and in communication with the imaging system and display, and a machine-readable storage medium encoded with a computer program code such that, when the computer program code is executed by a processor, the processor performs a method comprising:
providing a three-dimensional image data set of a patient tissue region;
obtaining an x-ray image of the patient tissue region using the x-ray source and the x-ray detector;
co-registering the three-dimensional image data set to the x-ray image acquired using the imaging system;
obtaining a planned instrument trajectory based on target point data representative of a target object within the patient tissue region and a skin entry point data representative of a skin entry point; and
aligning an instrument along the planned instrument trajectory using one of a plurality of instrument guidance procedures, the one of a plurality of instrument guidance procedures being selected on the basis of the planned instrument trajectory and a position of the movable arm;
wherein the plurality of instrument guidance procedures comprises a laser-guided procedure, an x-ray guided procedure, and a procedure that replans the instrument trajectory.
13. The system of claim 12, wherein the laser-guided procedure is selected if a central ray of the x-ray source coincides with the planned instrument trajectory.
14. The system of claim 12, wherein the x-ray guided procedure is selected if a central ray of the x-ray source does not coincide with the planned instrument trajectory, and an extension of the planned instrument trajectory intersects with the x-ray detector.
15. The system of claim 12, wherein the instrument trajectory is replanned if a central ray of the x-ray source does not coincide with the planned instrument trajectory and an extension of the planned instrument trajectory does not intersect with the x-ray detector.
16. The system of claim 15, further comprising a warning signal to a user when a central ray of the x-ray source does not coincide with the planned instrument trajectory and an extension of the planned instrument trajectory does not intersect with the x-ray detector.
17. The system of claim 12, wherein replanning the instrument trajectory comprises manually replanning the instrument trajectory using a graphical user interface.
18. The system of claim 12, wherein replanning the instrument trajectory comprises automatically replanning the instrument trajectory.
19. The system of claim 12, wherein the co-registering step comprises applying a transform to the 3-dimensional image data set such that points in a resulting overlay image align with counterpart points in the x-ray image.
20. The system of claim 12, wherein the step of aligning an instrument comprises aligning the instrument along a laser beam path generated by a laser pointer, the laser beam path intersecting the skin entry point and the target object.
21. The system of claim 20, wherein the laser beam path comprises the intersection of at least two laser fan patterns.
22. The system of claim 12, wherein the step of obtaining an x-ray image of the patient tissue region using the x-ray source and the x-ray detector comprises obtaining a plurality of x-ray images, displaying the plurality of x-ray images on the display along with a three-dimensional rendering of the three-dimensional image data set, and an overlay image representing a 2-dimensional rendering of the 3-dimensional image data set.
23. The system of claim 22, wherein the skin entry point, the target point and the planned instrument trajectory are graphically displayed in their respective positions on the plurality of displayed x-ray images, the three-dimensional rendering and the overlay image.
US12/431,518 2008-05-02 2009-04-28 System and method for selecting a guidance mode for performing a percutaneous procedure Active 2030-06-07 US8165660B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/431,518 US8165660B2 (en) 2008-05-02 2009-04-28 System and method for selecting a guidance mode for performing a percutaneous procedure

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US4980308P 2008-05-02 2008-05-02
US12/431,518 US8165660B2 (en) 2008-05-02 2009-04-28 System and method for selecting a guidance mode for performing a percutaneous procedure

Publications (2)

Publication Number Publication Date
US20090274271A1 true US20090274271A1 (en) 2009-11-05
US8165660B2 US8165660B2 (en) 2012-04-24

Family

ID=41257080

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/429,546 Abandoned US20090281452A1 (en) 2008-05-02 2009-04-24 System and method for a medical procedure using computed tomography
US12/431,518 Active 2030-06-07 US8165660B2 (en) 2008-05-02 2009-04-28 System and method for selecting a guidance mode for performing a percutaneous procedure

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/429,546 Abandoned US20090281452A1 (en) 2008-05-02 2009-04-24 System and method for a medical procedure using computed tomography

Country Status (1)

Country Link
US (2) US20090281452A1 (en)

Cited By (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120022508A1 (en) * 2010-07-22 2012-01-26 Patrick Gross Method and orientation device for marking a predetermined guide path of a medical instrument
US20120108952A1 (en) * 2010-10-29 2012-05-03 Eli Stern Methods and systems for patient alignment for nuclear medicine imaging
US20120209123A1 (en) * 2011-02-10 2012-08-16 Timothy King Surgeon's Aid for Medical Display
US20120281812A1 (en) * 2011-04-22 2012-11-08 Toshiba Medical Systems Corporation X-ray image diagnosing apparatus
US20120300910A1 (en) * 2011-05-26 2012-11-29 Sebastien Gorges X-ray imaging apparatus having a variable distance between an x-ray source and an object to be imaged
WO2012168749A1 (en) * 2011-06-06 2012-12-13 Sarr Souleymane Removable guide device for radiofluoroscopic infiltration having an image intensifier
US20130083894A1 (en) * 2011-09-30 2013-04-04 Siemens Aktiengesellschaft Method for controlling the movement of an x-ray apparatus and x-ray system
US20130089180A1 (en) * 2011-10-11 2013-04-11 Siemens Aktiengesellschaft Method and apparatus for producing an x-ray projection image in a desired direction
US20130116550A1 (en) * 2011-07-06 2013-05-09 Hideaki Ishii Medical diagnostic imaging apparatus
US20130136332A1 (en) * 2011-11-29 2013-05-30 Hisayuki Uehara X-ray image diagnosis apparatus
JP2013172951A (en) * 2012-01-27 2013-09-05 Toshiba Corp X-ray ct apparatus and x-ray ct system
WO2014125789A1 (en) * 2013-02-14 2014-08-21 Seiko Epson Corporation Head mounted display and control method for head mounted display
FR3007962A1 (en) * 2013-07-04 2015-01-09 X Nov Ip GRAPHICAL SELECTION OF BONE ANCHOR PROSTHESIS
US20150080740A1 (en) * 2013-09-18 2015-03-19 iMIRGE Medical INC. Optical targeting and visualization of trajectories
EP2526869A3 (en) * 2011-05-23 2015-10-07 Fujifilm Corporation Image processing device, radiographic image capture system, image processing method and image processing program storage medium
US20160147939A1 (en) * 2014-11-26 2016-05-26 Koninklijke Philips N.V. Efficient management of visible light still images and/or video
EP3028675A1 (en) * 2014-12-05 2016-06-08 X.Nov IP Sarl Graphical selection of a prosthesis with bone anchor
WO2016119053A1 (en) * 2015-01-27 2016-08-04 iMIRGE Medical INC. Optical targeting and visualization of trajectories
US20160345923A1 (en) * 2015-06-01 2016-12-01 Toshiba Medical Systems Corporation Medical image processing apparatus and x-ray diagnostic apparatus
US20170071672A1 (en) * 2014-03-04 2017-03-16 Xact Robotics Ltd. Dynamic planning method for needle insertion
WO2017055990A1 (en) * 2015-09-28 2017-04-06 Koninklijke Philips N.V. Optical registration of a remote center of motion robot
WO2017070124A1 (en) * 2015-10-19 2017-04-27 New York University Electronic position guidance device with real-time auditory and visual feedback
US20170156800A1 (en) * 2014-03-17 2017-06-08 Roy Anthony Brown Surgical Targeting Systems and Methods
US20170193160A1 (en) * 2016-01-06 2017-07-06 Boston Scientific Scimed, Inc. Systems and methods for planning medical procedures
US20170245951A1 (en) * 2012-06-21 2017-08-31 Globus Medical, Inc. Surgical robot platform
US20170290559A1 (en) * 2014-09-25 2017-10-12 Siemens Healthcare Gmbh Mobile x-ray machine
NL2016800A (en) * 2016-05-19 2017-11-23 Umc Utrecht Holding Bv Method of positioning an interventional device.
EP3254627A1 (en) * 2016-06-08 2017-12-13 General Electric Company Fluoroscopic guidance system with offset light source and method of use
US20180008217A1 (en) * 2016-07-08 2018-01-11 Alexander Gemmel Motion controllers for mobile x-ray devices
US20180035984A1 (en) * 2016-08-03 2018-02-08 Siemens Healthcare Gmbh Biopsy unit
US20180087900A1 (en) * 2016-09-28 2018-03-29 Kla-Tencor Corporation Optical Near-Field Metrology
US20180199901A1 (en) * 2015-07-16 2018-07-19 Koninklijke Philips N.V. Device for remote fluoroscopy, nearby fluoroscopy and radiology
US20180214007A1 (en) * 2017-01-30 2018-08-02 Seiko Epson Corporation Endoscope operation support system
US10130388B2 (en) 2014-03-13 2018-11-20 New York University Position guidance device with bubble level
US20180333208A1 (en) * 2017-05-17 2018-11-22 General Electric Company Guidance system for needle procedures
CN109481018A (en) * 2018-12-29 2019-03-19 上海联影医疗科技有限公司 A kind of navigation equipment and method applied in medical care precess
WO2019109211A1 (en) * 2017-12-04 2019-06-13 Covidien Lp Automatic ablation antenna segmentation from ct image
JP2019122767A (en) * 2018-01-15 2019-07-25 キヤノンメディカルシステムズ株式会社 Medical information processing apparatus, x-ray diagnostic system and medical information processing program
US10376179B2 (en) 2011-04-21 2019-08-13 Koninklijke Philips N.V. MPR slice selection for visualization of catheter in three-dimensional ultrasound
CN110680481A (en) * 2019-11-15 2020-01-14 元亨同基医疗器械(北京)有限公司 Method for adjusting launching position of guide holder of puncture positioning instrument
WO2020105049A1 (en) * 2018-11-22 2020-05-28 Vuze Medical Ltd. Apparatus and methods for use with image-guided skeletal procedures
US10716631B2 (en) 2016-03-13 2020-07-21 Vuze Medical Ltd. Apparatus and methods for use with skeletal procedures
DE102019108753A1 (en) * 2019-04-03 2020-10-08 Amedo Smart Tracking Solutions Gmbh 3D laser-based positioning system
WO2020201956A1 (en) * 2019-04-05 2020-10-08 Medos International Sarl Systems, devices, and methods for providing surgical trajectory guidance
US20200315736A1 (en) * 2019-04-05 2020-10-08 Siemens Healthcare Gmbh Guiding a needle using a radiopaque contour
US11006906B2 (en) * 2017-02-28 2021-05-18 Medicaroid Corporation Robotic operating table, and robotic operating table operation device
CN113164134A (en) * 2018-10-31 2021-07-23 皇家飞利浦有限公司 X-ray imaging guidance
CN113367779A (en) * 2021-06-16 2021-09-10 张涛 Puncture system and method based on C-arm CT and semiconductor laser
US11224483B2 (en) 2017-07-08 2022-01-18 Vuze Medical Ltd. Apparatus and methods for use with image-guided skeletal procedures
CN114271838A (en) * 2021-12-01 2022-04-05 赛诺威盛科技(北京)股份有限公司 Multifunctional scanning method, multifunctional scanning support, electronic equipment and storage medium
CN116096322A (en) * 2020-09-04 2023-05-09 上海联影医疗科技股份有限公司 Systems and methods for assisting in placement of a surgical instrument into a subject
US20230270501A1 (en) * 2022-02-25 2023-08-31 Siemens Healthcare Gmbh Trajectory planning for a medical robot system

Families Citing this family (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2429400B1 (en) * 2009-05-14 2014-01-01 University Health Network Quantitative endoscopy
JP5317933B2 (en) * 2009-11-17 2013-10-16 富士フイルム株式会社 Image display device and program thereof
WO2011073824A1 (en) * 2009-12-18 2011-06-23 Koninklijke Philips Electronics N.V. Multi-section alignment of imaging data
JP4956635B2 (en) * 2010-02-24 2012-06-20 財団法人仙台市医療センター Percutaneous puncture support system
DE102010012621A1 (en) * 2010-03-24 2011-09-29 Siemens Aktiengesellschaft Method and device for automatically adapting a reference image
US9282944B2 (en) * 2010-06-22 2016-03-15 Queen's University At Kingston C-arm pose estimation using intensity-based registration of imaging modalities
US20120065498A1 (en) * 2010-09-14 2012-03-15 Thomas Redel Apparatus and method for minimally invasive therapy of mitral regurgitation
US8848874B2 (en) 2010-09-20 2014-09-30 Siemens Medical Solutions Usa, Inc. System for recovering from collision of components of an X-ray imaging unit
DE102011075435A1 (en) * 2011-05-06 2012-11-08 Siemens Aktiengesellschaft display unit
US8900131B2 (en) * 2011-05-13 2014-12-02 Intuitive Surgical Operations, Inc. Medical system providing dynamic registration of a model of an anatomical structure for image-guided surgery
US9510771B1 (en) 2011-10-28 2016-12-06 Nuvasive, Inc. Systems and methods for performing spine surgery
DE102011086941B4 (en) * 2011-11-23 2016-01-21 Kuka Roboter Gmbh industrial robots
US20130218003A1 (en) * 2012-02-21 2013-08-22 Siemens Aktiengesellschaft Rapid entry point localization for percutaneous interventions
DE102012208037A1 (en) * 2012-05-14 2013-11-14 Siemens Aktiengesellschaft Patient storage device, a medical imaging device with the patient support device and a method for marking a maximum resident area
US20130345554A1 (en) * 2012-06-08 2013-12-26 Baylis Medical Company Inc. Methods for recanalization of vessels
BR112014032112A2 (en) * 2012-06-28 2017-06-27 Koninklijke Philips Nv image acquisition system; and method for multimodal image acquisition
CN104902821B (en) 2012-10-05 2018-11-09 皇家飞利浦有限公司 ROI is drawn
JP6222835B2 (en) * 2013-02-14 2017-11-01 東芝メディカルシステムズ株式会社 X-ray diagnostic equipment
US20140270067A1 (en) * 2013-03-14 2014-09-18 Stephen Knecht Clark Radiographic marker
DE102013208793B4 (en) * 2013-05-14 2015-04-23 Siemens Aktiengesellschaft A method for generating a 3D image data set from a volume to be examined as a basis for display image data
US9848922B2 (en) 2013-10-09 2017-12-26 Nuvasive, Inc. Systems and methods for performing spine surgery
DE102013226975A1 (en) * 2013-12-20 2015-06-25 Siemens Aktiengesellschaft Method for image monitoring during interventional intervention, X-ray device, computer program and data carrier
AU2015296645A1 (en) 2014-07-28 2017-02-16 Econolite Group, Inc. Self-configuring traffic signal controller
KR20160033546A (en) * 2014-09-18 2016-03-28 삼성전자주식회사 X-ray apparatus and method for scanning thereof
US10722200B2 (en) * 2015-06-04 2020-07-28 Siemens Healthcare Gmbh Apparatus and methods for a projection display device on X-ray imaging devices
DE102015215820A1 (en) * 2015-08-19 2017-02-23 Siemens Healthcare Gmbh Controlling a medical device
KR101798939B1 (en) * 2015-09-08 2017-11-17 삼성전자주식회사 X-ray image apparatus and control method for the same
RU2746151C2 (en) * 2016-03-16 2021-04-07 Конинклейке Филипс Н.В. System for supporting the performance of interventional procedures
US11576746B2 (en) * 2016-09-20 2023-02-14 Kornerstone Devices Pvt. Ltd. Light and shadow guided needle positioning system and method
EP3518807A4 (en) * 2016-09-30 2020-05-27 Intuitive Surgical Operations Inc. Systems and methods for entry point localization
DE102017207908B4 (en) * 2017-05-10 2023-06-15 Siemens Healthcare Gmbh X-ray imaging system and method for taking X-ray images
AU2018265160B2 (en) 2017-05-10 2023-11-16 Mako Surgical Corp. Robotic spine surgery system and methods
US11627924B2 (en) 2019-09-24 2023-04-18 Covidien Lp Systems and methods for image-guided navigation of percutaneously-inserted devices
DE102019135782A1 (en) * 2019-12-23 2021-06-24 Otto-Von-Guericke-Universität Magdeburg Computed tomography system
US11490872B2 (en) 2020-08-21 2022-11-08 GE Precision Healthcare LLC C-arm imaging system and method
US12004284B2 (en) 2021-10-21 2024-06-04 GE Precision Healthcare LLC Methods and systems for x-ray imaging

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5782842A (en) * 1995-01-16 1998-07-21 Daum Gmbh Medical instrument guidance apparatus and method
US6195577B1 (en) * 1998-10-08 2001-02-27 Regents Of The University Of Minnesota Method and apparatus for positioning a device in a body
US6811313B2 (en) * 2001-09-25 2004-11-02 Siemens Aktiengesellschaft C-arm X-ray system with adjustable detector positioning
US20060039537A1 (en) * 2004-05-28 2006-02-23 Strobel Norbert K C-arm device with adjustable detector offset for cone beam imaging involving partial circle scan trajectories
US20060120507A1 (en) * 2004-11-26 2006-06-08 Thomas Brunner Angiographic x-ray diagnostic device for rotation angiography
US20070255292A1 (en) * 2006-04-28 2007-11-01 Siemens Aktiengesellschaft X-ray C-arm system, frame for a flat-panel x-ray detector of an x-ray C-arm system and method for technical support of targeting during puncturing of a patient
US20080089467A1 (en) * 2006-08-10 2008-04-17 Siemens Aktiengesellschaft Method for recording x-ray images of an area lying outside a center of rotation of a C-arm system and the associated C-arm system
US20080194945A1 (en) * 2007-02-13 2008-08-14 Siemens Medical Solutions Usa, Inc. Apparatus and Method for Aligning a Light Pointer With a Medical Interventional Device Trajectory
US20080200876A1 (en) * 2007-02-20 2008-08-21 Siemens Medical Solutions Usa, Inc. Needle Guidance With a Dual-Headed Laser

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6064904A (en) * 1997-11-28 2000-05-16 Picker International, Inc. Frameless stereotactic CT scanner with virtual needle display for planning image guided interventional procedures
US6470207B1 (en) * 1999-03-23 2002-10-22 Surgical Navigation Technologies, Inc. Navigational guidance via computer-assisted fluoroscopic imaging
DE102005023165A1 (en) * 2005-05-19 2006-11-23 Siemens Ag Medical imaging system with a patient traversable part and collision protection method
DE102005030609A1 (en) * 2005-06-30 2007-01-04 Siemens Ag Method or X-ray device for creating a series recording of medical X-ray images of a possibly moving patient during the series recording
DE102005041602A1 (en) * 2005-09-01 2007-04-05 Siemens Ag Method for displaying a medical implant in an image and medical imaging system

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5782842A (en) * 1995-01-16 1998-07-21 Daum Gmbh Medical instrument guidance apparatus and method
US6195577B1 (en) * 1998-10-08 2001-02-27 Regents Of The University Of Minnesota Method and apparatus for positioning a device in a body
US6811313B2 (en) * 2001-09-25 2004-11-02 Siemens Aktiengesellschaft C-arm X-ray system with adjustable detector positioning
US20060039537A1 (en) * 2004-05-28 2006-02-23 Strobel Norbert K C-arm device with adjustable detector offset for cone beam imaging involving partial circle scan trajectories
US20060120507A1 (en) * 2004-11-26 2006-06-08 Thomas Brunner Angiographic x-ray diagnostic device for rotation angiography
US20070255292A1 (en) * 2006-04-28 2007-11-01 Siemens Aktiengesellschaft X-ray C-arm system, frame for a flat-panel x-ray detector of an x-ray C-arm system and method for technical support of targeting during puncturing of a patient
US20080089467A1 (en) * 2006-08-10 2008-04-17 Siemens Aktiengesellschaft Method for recording x-ray images of an area lying outside a center of rotation of a C-arm system and the associated C-arm system
US20080194945A1 (en) * 2007-02-13 2008-08-14 Siemens Medical Solutions Usa, Inc. Apparatus and Method for Aligning a Light Pointer With a Medical Interventional Device Trajectory
US20080200876A1 (en) * 2007-02-20 2008-08-21 Siemens Medical Solutions Usa, Inc. Needle Guidance With a Dual-Headed Laser

Cited By (101)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102342861A (en) * 2010-07-22 2012-02-08 西门子公司 Method and orientation device for marking a predetermined guide path of a medical instrument
US20120022508A1 (en) * 2010-07-22 2012-01-26 Patrick Gross Method and orientation device for marking a predetermined guide path of a medical instrument
US20120108952A1 (en) * 2010-10-29 2012-05-03 Eli Stern Methods and systems for patient alignment for nuclear medicine imaging
US9078618B2 (en) * 2010-10-29 2015-07-14 General Electric Company Methods and systems for patient alignment for nuclear medicine imaging
US20120209123A1 (en) * 2011-02-10 2012-08-16 Timothy King Surgeon's Aid for Medical Display
US10631712B2 (en) * 2011-02-10 2020-04-28 Karl Storz Imaging, Inc. Surgeon's aid for medical display
US10376179B2 (en) 2011-04-21 2019-08-13 Koninklijke Philips N.V. MPR slice selection for visualization of catheter in three-dimensional ultrasound
US8781068B2 (en) * 2011-04-22 2014-07-15 Kabushiki Kaisha Toshiba X-ray image diagnosing apparatus
US20120281812A1 (en) * 2011-04-22 2012-11-08 Toshiba Medical Systems Corporation X-ray image diagnosing apparatus
EP2526869A3 (en) * 2011-05-23 2015-10-07 Fujifilm Corporation Image processing device, radiographic image capture system, image processing method and image processing program storage medium
US20120300910A1 (en) * 2011-05-26 2012-11-29 Sebastien Gorges X-ray imaging apparatus having a variable distance between an x-ray source and an object to be imaged
WO2012168749A1 (en) * 2011-06-06 2012-12-13 Sarr Souleymane Removable guide device for radiofluoroscopic infiltration having an image intensifier
US20130116550A1 (en) * 2011-07-06 2013-05-09 Hideaki Ishii Medical diagnostic imaging apparatus
US9445773B2 (en) * 2011-07-06 2016-09-20 Toshiba Medical Systems Corporation Medical diagnostic imaging apparatus
CN103027699A (en) * 2011-09-30 2013-04-10 西门子公司 Method for controlling the movement of an x-ray apparatus and x-ray system
US9642584B2 (en) * 2011-09-30 2017-05-09 Siemens Aktiengesellschaft Method for controlling the movement of an x-ray apparatus and x-ray system
US20130083894A1 (en) * 2011-09-30 2013-04-04 Siemens Aktiengesellschaft Method for controlling the movement of an x-ray apparatus and x-ray system
US20130089180A1 (en) * 2011-10-11 2013-04-11 Siemens Aktiengesellschaft Method and apparatus for producing an x-ray projection image in a desired direction
US9039283B2 (en) * 2011-10-11 2015-05-26 Siemens Aktiengesellschaft Method and apparatus for producing an X-ray projection image in a desired direction
US20130136332A1 (en) * 2011-11-29 2013-05-30 Hisayuki Uehara X-ray image diagnosis apparatus
US9384545B2 (en) * 2011-11-29 2016-07-05 Kabushiki Kaisha Toshiba X-ray image diagnosis apparatus
JP2013172951A (en) * 2012-01-27 2013-09-05 Toshiba Corp X-ray ct apparatus and x-ray ct system
US10485617B2 (en) * 2012-06-21 2019-11-26 Globus Medical, Inc. Surgical robot platform
US20170245951A1 (en) * 2012-06-21 2017-08-31 Globus Medical, Inc. Surgical robot platform
US9916691B2 (en) 2013-02-14 2018-03-13 Seiko Epson Corporation Head mounted display and control method for head mounted display
WO2014125789A1 (en) * 2013-02-14 2014-08-21 Seiko Epson Corporation Head mounted display and control method for head mounted display
US10169925B2 (en) 2013-02-14 2019-01-01 Seiko Epson Corporation Head mounted display and control method for head mounted display
FR3007962A1 (en) * 2013-07-04 2015-01-09 X Nov Ip GRAPHICAL SELECTION OF BONE ANCHOR PROSTHESIS
US9877795B2 (en) * 2013-09-18 2018-01-30 Imirge Medical Inc Optical targeting and visualization of trajectories
WO2015039246A1 (en) * 2013-09-18 2015-03-26 iMIRGE Medical INC. Optical targeting and visualization of trajectories
US20150080740A1 (en) * 2013-09-18 2015-03-19 iMIRGE Medical INC. Optical targeting and visualization of trajectories
CN106413621A (en) * 2013-09-18 2017-02-15 伊美格医药公司 Surgical assemblies for housing force transmitting members
US10433911B2 (en) * 2013-09-18 2019-10-08 iMIRGE Medical INC. Optical targeting and visualization of trajectories
US10245110B2 (en) * 2014-03-04 2019-04-02 Xact Robotics Ltd. Dynamic planning method for needle insertion
US20170071672A1 (en) * 2014-03-04 2017-03-16 Xact Robotics Ltd. Dynamic planning method for needle insertion
US10130388B2 (en) 2014-03-13 2018-11-20 New York University Position guidance device with bubble level
US20170156800A1 (en) * 2014-03-17 2017-06-08 Roy Anthony Brown Surgical Targeting Systems and Methods
US20170296273A9 (en) * 2014-03-17 2017-10-19 Roy Anthony Brown Surgical Targeting Systems and Methods
US20170290559A1 (en) * 2014-09-25 2017-10-12 Siemens Healthcare Gmbh Mobile x-ray machine
US10653386B2 (en) * 2014-09-25 2020-05-19 Siemens Healthcare Gmbh Mobile x-ray machine
US20160147939A1 (en) * 2014-11-26 2016-05-26 Koninklijke Philips N.V. Efficient management of visible light still images and/or video
US10162935B2 (en) * 2014-11-26 2018-12-25 Koninklijke Philips N.V. Efficient management of visible light still images and/or video
EP3028675A1 (en) * 2014-12-05 2016-06-08 X.Nov IP Sarl Graphical selection of a prosthesis with bone anchor
WO2016119053A1 (en) * 2015-01-27 2016-08-04 iMIRGE Medical INC. Optical targeting and visualization of trajectories
US20160345923A1 (en) * 2015-06-01 2016-12-01 Toshiba Medical Systems Corporation Medical image processing apparatus and x-ray diagnostic apparatus
US9936928B2 (en) * 2015-06-01 2018-04-10 Toshiba Medical Systems Corporation Medical image processing apparatus and X-ray diagnostic apparatus
US20180199901A1 (en) * 2015-07-16 2018-07-19 Koninklijke Philips N.V. Device for remote fluoroscopy, nearby fluoroscopy and radiology
US10993685B2 (en) * 2015-07-16 2021-05-04 Koninklijke Philips N.V. Device for remote fluoroscopy, nearby fluoroscopy and radiology
WO2017055990A1 (en) * 2015-09-28 2017-04-06 Koninklijke Philips N.V. Optical registration of a remote center of motion robot
WO2017070124A1 (en) * 2015-10-19 2017-04-27 New York University Electronic position guidance device with real-time auditory and visual feedback
AU2017206019B2 (en) * 2016-01-06 2021-08-12 Boston Scientific Scimed, Inc. Systems and methods for planning medical procedures
US11830604B2 (en) 2016-01-06 2023-11-28 Boston Scientific Scimed, Inc. Systems and methods for planning medical procedures
US11302435B2 (en) 2016-01-06 2022-04-12 Boston Scientific Scimed, Inc. Systems and methods for planning medical procedures
US20170193160A1 (en) * 2016-01-06 2017-07-06 Boston Scientific Scimed, Inc. Systems and methods for planning medical procedures
WO2017120369A1 (en) * 2016-01-06 2017-07-13 Boston Scientific Scimed, Inc. Systems and methods for planning medical procedures
EP3795108A1 (en) * 2016-01-06 2021-03-24 Boston Scientific Scimed, Inc. Systems and methods for planning medical procedures
US11490967B2 (en) 2016-03-13 2022-11-08 Vuze Medical Ltd. Apparatus and methods for use with skeletal procedures
US10716631B2 (en) 2016-03-13 2020-07-21 Vuze Medical Ltd. Apparatus and methods for use with skeletal procedures
US11452570B2 (en) 2016-03-13 2022-09-27 Vuze Medical Ltd. Apparatus and methods for use with skeletal procedures
US11911118B2 (en) 2016-03-13 2024-02-27 Vuze Medical Ltd. Apparatus and methods for use with skeletal procedures
US11723616B2 (en) * 2016-05-19 2023-08-15 Cart-Tech B.V Method of verifying a position of an interventional device
NL2016800A (en) * 2016-05-19 2017-11-23 Umc Utrecht Holding Bv Method of positioning an interventional device.
US20190125289A1 (en) * 2016-05-19 2019-05-02 Umc Utrecht Holding B.V. Method of verifying a position of an interventional device
WO2017200385A1 (en) * 2016-05-19 2017-11-23 Umc Utrecht Holding B.V. Method of verifying a position of an interventional device
EP3254627A1 (en) * 2016-06-08 2017-12-13 General Electric Company Fluoroscopic guidance system with offset light source and method of use
US20170354387A1 (en) * 2016-06-08 2017-12-14 General Electric Company Fluoroscopic Guidance System With Offset Light Source and Method Of Use
CN107468336A (en) * 2016-06-08 2017-12-15 通用电气公司 With the fluoroscopic guidance system and its application method for deviateing light source
JP2017217448A (en) * 2016-06-08 2017-12-14 ゼネラル・エレクトリック・カンパニイ Fluoroscopic guidance system with offset light source and method of use
US20180008217A1 (en) * 2016-07-08 2018-01-11 Alexander Gemmel Motion controllers for mobile x-ray devices
US10751010B2 (en) * 2016-07-08 2020-08-25 Siemens Healthcare Gmbh Motion controllers for mobile X-ray devices
US20180035984A1 (en) * 2016-08-03 2018-02-08 Siemens Healthcare Gmbh Biopsy unit
US11815347B2 (en) * 2016-09-28 2023-11-14 Kla-Tencor Corporation Optical near-field metrology
US20180087900A1 (en) * 2016-09-28 2018-03-29 Kla-Tencor Corporation Optical Near-Field Metrology
US20180214007A1 (en) * 2017-01-30 2018-08-02 Seiko Epson Corporation Endoscope operation support system
US10842349B2 (en) * 2017-01-30 2020-11-24 Seiko Epson Corporation Endoscope operation support system
US11006906B2 (en) * 2017-02-28 2021-05-18 Medicaroid Corporation Robotic operating table, and robotic operating table operation device
US10667869B2 (en) * 2017-05-17 2020-06-02 General Electric Company Guidance system for needle procedures
US20180333208A1 (en) * 2017-05-17 2018-11-22 General Electric Company Guidance system for needle procedures
US11806183B2 (en) 2017-07-08 2023-11-07 Vuze Medical Ltd. Apparatus and methods for use with image-guided skeletal procedures
US11406338B2 (en) 2017-07-08 2022-08-09 Vuze Medical Ltd. Apparatus and methods for use with image-guided skeletal procedures
US11224483B2 (en) 2017-07-08 2022-01-18 Vuze Medical Ltd. Apparatus and methods for use with image-guided skeletal procedures
WO2019109211A1 (en) * 2017-12-04 2019-06-13 Covidien Lp Automatic ablation antenna segmentation from ct image
US11638612B2 (en) 2017-12-04 2023-05-02 Covidien Lp Automatic ablation antenna segmentation from CT image
JP7271183B2 (en) 2018-01-15 2023-05-11 キヤノンメディカルシステムズ株式会社 Medical information processing device, X-ray diagnostic system and medical information processing program
JP2019122767A (en) * 2018-01-15 2019-07-25 キヤノンメディカルシステムズ株式会社 Medical information processing apparatus, x-ray diagnostic system and medical information processing program
CN113164134A (en) * 2018-10-31 2021-07-23 皇家飞利浦有限公司 X-ray imaging guidance
WO2020105049A1 (en) * 2018-11-22 2020-05-28 Vuze Medical Ltd. Apparatus and methods for use with image-guided skeletal procedures
EP3886749A4 (en) * 2018-12-29 2022-01-26 Shanghai United Imaging Healthcare Co., Ltd. System and method for placing surgical instrument in subject
CN109481018A (en) * 2018-12-29 2019-03-19 上海联影医疗科技有限公司 A kind of navigation equipment and method applied in medical care precess
WO2020135769A1 (en) 2018-12-29 2020-07-02 Shanghai United Imaging Healthcare Co., Ltd. System and method for placing surgical instrument in subject
DE102019108753A1 (en) * 2019-04-03 2020-10-08 Amedo Smart Tracking Solutions Gmbh 3D laser-based positioning system
DE102019204920A1 (en) * 2019-04-05 2020-10-08 Siemens Healthcare Gmbh Guide unit for guiding a needle with an X-ray-opaque contour; Management system; Investment; Computer program product and method
WO2020201956A1 (en) * 2019-04-05 2020-10-08 Medos International Sarl Systems, devices, and methods for providing surgical trajectory guidance
US20200315736A1 (en) * 2019-04-05 2020-10-08 Siemens Healthcare Gmbh Guiding a needle using a radiopaque contour
US11813026B2 (en) 2019-04-05 2023-11-14 Medos International Sarl Systems, devices, and methods for providing surgical trajectory guidance
CN110680481A (en) * 2019-11-15 2020-01-14 元亨同基医疗器械(北京)有限公司 Method for adjusting launching position of guide holder of puncture positioning instrument
CN116096322A (en) * 2020-09-04 2023-05-09 上海联影医疗科技股份有限公司 Systems and methods for assisting in placement of a surgical instrument into a subject
CN113367779A (en) * 2021-06-16 2021-09-10 张涛 Puncture system and method based on C-arm CT and semiconductor laser
CN114271838A (en) * 2021-12-01 2022-04-05 赛诺威盛科技(北京)股份有限公司 Multifunctional scanning method, multifunctional scanning support, electronic equipment and storage medium
US20230270501A1 (en) * 2022-02-25 2023-08-31 Siemens Healthcare Gmbh Trajectory planning for a medical robot system
US11877805B2 (en) * 2022-02-25 2024-01-23 Siemens Healthcare Gmbh Trajectory planning for a medical robot system

Also Published As

Publication number Publication date
US8165660B2 (en) 2012-04-24
US20090281452A1 (en) 2009-11-12

Similar Documents

Publication Publication Date Title
US8165660B2 (en) System and method for selecting a guidance mode for performing a percutaneous procedure
US11989338B2 (en) Using optical codes with augmented reality displays
US20100111389A1 (en) System and method for planning and guiding percutaneous procedures
US11992349B2 (en) System and method for local three dimensional volume reconstruction using a standard fluoroscope
US20220313190A1 (en) System and method for pose estimation of an imaging device and for determining the location of a medical device with respect to a target
US9202387B2 (en) Methods for planning and performing percutaneous needle procedures
CN110741414B (en) Systems and methods for identifying, marking, and navigating to a target using real-time two-dimensional fluoroscopic data
EP2004071B1 (en) Targeting device, computer readable medium and program element
US20100292565A1 (en) Medical imaging medical device navigation from at least two 2d projections from different angles
JP6713563B2 (en) System and method for performing local three-dimensional volume reconstruction using standard fluoroscopy equipment
EP3254627A1 (en) Fluoroscopic guidance system with offset light source and method of use
EP1727471A1 (en) System for guiding a medical instrument in a patient body
US20220409290A1 (en) Method and system for reproducing an insertion point for a medical instrument
US8731643B2 (en) Imaging system and methods for medical needle procedures
US11576557B2 (en) Method for supporting a user, computer program product, data medium and imaging system
WO2019075074A1 (en) System and method for identifying and marking a target in a fluoroscopic three-dimensional reconstruction
EP4163874A1 (en) Method and device for an improved display in a manual or a robot assisted intervention in a region of interest of the patient
Chen et al. Accuracy and efficiency of an infrared based positioning and tracking system for image-guided intervention

Legal Events

Date Code Title Description
AS Assignment

Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC., PENNSYLVANIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PFISTER, MARCUS;STROBEL, NORBERT;REEL/FRAME:022938/0107

Effective date: 20090629

AS Assignment

Owner name: SIEMENS AKTIENGESELLSCHAFT, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PFISTER, MARCUS;STROBEL, NORBERT;REEL/FRAME:027432/0102

Effective date: 20111213

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

AS Assignment

Owner name: SIEMENS HEALTHCARE GMBH, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS AKTIENGESELLSCHAFT;REEL/FRAME:039011/0400

Effective date: 20160610

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 12

AS Assignment

Owner name: SIEMENS HEALTHINEERS AG, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS HEALTHCARE GMBH;REEL/FRAME:066088/0256

Effective date: 20231219