US20090274271A1 - System and method for selecting a guidance mode for performing a percutaneous procedure - Google Patents
System and method for selecting a guidance mode for performing a percutaneous procedure Download PDFInfo
- Publication number
- US20090274271A1 US20090274271A1 US12/431,518 US43151809A US2009274271A1 US 20090274271 A1 US20090274271 A1 US 20090274271A1 US 43151809 A US43151809 A US 43151809A US 2009274271 A1 US2009274271 A1 US 2009274271A1
- Authority
- US
- United States
- Prior art keywords
- ray
- instrument
- trajectory
- planned
- instrument trajectory
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 137
- 238000003384 imaging method Methods 0.000 claims description 21
- 238000009877 rendering Methods 0.000 claims description 14
- 238000003860 storage Methods 0.000 claims description 7
- 238000004891 communication Methods 0.000 claims description 4
- 238000004590 computer program Methods 0.000 claims description 4
- 230000003287 optical effect Effects 0.000 abstract description 10
- 230000005855 radiation Effects 0.000 abstract description 6
- 244000309464 bull Species 0.000 description 25
- 210000001519 tissue Anatomy 0.000 description 17
- 238000002591 computed tomography Methods 0.000 description 11
- 230000033001 locomotion Effects 0.000 description 10
- 238000001574 biopsy Methods 0.000 description 9
- 238000003780 insertion Methods 0.000 description 7
- 230000037431 insertion Effects 0.000 description 7
- 238000012795 verification Methods 0.000 description 6
- 238000002604 ultrasonography Methods 0.000 description 5
- 230000008569 process Effects 0.000 description 4
- 238000002594 fluoroscopy Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 210000000988 bone and bone Anatomy 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000002595 magnetic resonance imaging Methods 0.000 description 2
- 210000000056 organ Anatomy 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 210000003484 anatomy Anatomy 0.000 description 1
- 238000002583 angiography Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 210000004204 blood vessel Anatomy 0.000 description 1
- 239000003795 chemical substances by application Substances 0.000 description 1
- 239000002872 contrast media Substances 0.000 description 1
- 238000005520 cutting process Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000012279 drainage procedure Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 239000007943 implant Substances 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 238000002347 injection Methods 0.000 description 1
- 239000007924 injection Substances 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000013152 interventional procedure Methods 0.000 description 1
- 238000005304 joining Methods 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000013188 needle biopsy Methods 0.000 description 1
- 230000000704 physical effect Effects 0.000 description 1
- 238000002600 positron emission tomography Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 238000007674 radiofrequency ablation Methods 0.000 description 1
- 210000004872 soft tissue Anatomy 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 210000003813 thumb Anatomy 0.000 description 1
- 238000007794 visualization technique Methods 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/12—Arrangements for detecting or locating foreign bodies
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/44—Constructional features of apparatus for radiation diagnosis
- A61B6/4429—Constructional features of apparatus for radiation diagnosis related to the mounting of source units and detector units
- A61B6/4435—Constructional features of apparatus for radiation diagnosis related to the mounting of source units and detector units the source unit and the detector unit being coupled by a rigid structure
- A61B6/4441—Constructional features of apparatus for radiation diagnosis related to the mounting of source units and detector units the source unit and the detector unit being coupled by a rigid structure the rigid structure being a C-arm or U-arm
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5211—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
- A61B6/5229—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/10—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
- A61B90/11—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis with guides for needles or instruments, e.g. arcuate slides or ball joints
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/10—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
- A61B90/11—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis with guides for needles or instruments, e.g. arcuate slides or ball joints
- A61B90/13—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis with guides for needles or instruments, e.g. arcuate slides or ball joints guided by light, e.g. laser pointers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/107—Visualisation of planned trajectories or target regions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B2090/364—Correlation of different images or relation of image positions in respect to the body
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/376—Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61N—ELECTROTHERAPY; MAGNETOTHERAPY; RADIATION THERAPY; ULTRASOUND THERAPY
- A61N5/00—Radiation therapy
- A61N5/10—X-ray therapy; Gamma-ray therapy; Particle-irradiation therapy
- A61N5/1048—Monitoring, verifying, controlling systems and methods
- A61N5/1049—Monitoring, verifying, controlling systems and methods for verifying the position of the patient with respect to the radiation beam
- A61N2005/1061—Monitoring, verifying, controlling systems and methods for verifying the position of the patient with respect to the radiation beam using an x-ray imaging system having a separate imaging source
Definitions
- target point selection may be done by selecting a desired MPR view, such as by a right-click of a mouse pointer on an appropriate soft key 3 A in a pop-up window in the display. In the illustrated embodiment, this results in desired MPR views being displayed in the upper left quadrant 3 B, and upper right quadrant 3 C.
- the target point is “selected” by clicking with a mouse pointer at the target position 3 D in the lower left quadrant MPR display 3 E.
- the skin entry point may be selected (“clicked”) in the same manner.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Medical Informatics (AREA)
- Surgery (AREA)
- Heart & Thoracic Surgery (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Pathology (AREA)
- Public Health (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Veterinary Medicine (AREA)
- Animal Behavior & Ethology (AREA)
- Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- High Energy & Nuclear Physics (AREA)
- Radiology & Medical Imaging (AREA)
- Biophysics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Description
- This is a U.S. non-provisional application of U.S. provisional patent application Ser. No. 61/049,803, filed May 2, 2008, by Marcus Pfister, et al., titled “Workflow for Percutaneous Procedures,” the entirety of which application is incorporated herein by reference.
- The disclosure relates generally to methods for planning and performing percutaneous procedures, and more particularly to methods for automatically aligning medical devices to perform percutaneous procedures that reduce overall radiation exposure to patient and user.
- With modem angiography systems, it is possible to obtain not only 2-dimensional (2D) fluoroscopic images but, by rotation of a C-arm around the patient, 3-dimensional (3D) CT-like images. For the purpose of intervention planning or navigation, these 3D datasets can be co-registered with the 2D fluoroscopic image. This registration can be used to plan percutaneous puncture procedures, such as needle biopsies. Once registered, 2D fluoroscopic overlay images can be rendered under a given C-arm view angle to provide 3D anatomical information as well as a graphical representation of the interventional plan that can be combined with the live x-ray fluoroscopy image.
- Methods are known for utilizing a laser fixed to the C-arm and orientated along the planned path of a needle puncture to orient the actual needle along a desired puncture path so that it intersects a targeted location within the patent. Proper orientation of the laser can require manual adjustment of the patient table and/or a C-arm to ensure that the actual needle path coincides with the “virtual” (i.e., theoretical) laser path. This is because not all theoretical paths may be achievable with a C-arm, since the C-arm may be physically limited in angulation and also may be limited so that it does not collide with the patient or physician. The advantage of laser-based guidance methods, however, is that a biopsy needle can be oriented without exposing the patient or physician to x-ray radiation.
- Other methods are known for positioning a needle along a virtual path without the use of a laser fixed to a specific position with respect to the C-arm. Such methods instead utilize x-ray radiation to orient the needle position. To adjust the C-arm so that a needle is correctly oriented along a desired path, the puncture target and the skin entry point—which together define a “path of puncture”—are imaged on top of each other on an x-ray detector to provide a “bull's eye view” that enables appropriate adjustment of the C-arm to achieve a desired puncture path.
- There is a need for a system and method that combine the benefits of both types of procedures in a way that minimizes overall radiation exposure to patients and practitioners.
- A method is disclosed for planning a percutaneous procedure for use in a system comprising an imaging system having a movable arm, an x-ray source and an x-ray detector and a display and a system controller connected to and in communication with the imaging system and display, comprising: providing a three-dimensional image data set of a patient tissue region; obtaining an x-ray image of the patient tissue region using the x-ray source and the x-ray detector; co-registering the three-dimensional image data set to the x-ray image acquired using the imaging system; obtaining a planned instrument trajectory based on target point data representative of a target object within the patient tissue region and another point along the planned instrument trajectory, which may be a skin entry point data representative of a skin entry point; and aligning an instrument along the planned instrument trajectory using one of a plurality of instrument guidance procedures, the one of a plurality of instrument guidance procedures being selected on the basis of the planned instrument trajectory and a position of the movable arm; wherein the plurality of instrument guidance procedures comprises a laser-guided procedure, an x-ray guided procedure, and a procedure that replans the instrument trajectory.
- A system is disclosed for planning a percutaneous procedure for use in a system comprising an imaging system having a movable arm, an x-ray source and an x-ray detector and a display and a system controller connected to and in communication with the imaging system and display. The system further comprises a machine-readable storage medium encoded with a computer program code such that, when the computer program code is executed by a processor, the processor performs a method comprising: providing a three-dimensional image data set of a patient tissue region; obtaining an x-ray image of the patient tissue region using the x-ray source and the x-ray detector; co-registering the three-dimensional image data set to the x-ray image acquired using the imaging system; obtaining a planned instrument trajectory based on target point data representative of a target object within the patient tissue region and a skin entry point data representative of a skin entry point; and aligning an instrument along the planned instrument trajectory using one of a plurality of instrument guidance procedures, the one of a plurality of instrument guidance procedures being selected on the basis of the planned instrument trajectory and a position of the movable arm; wherein the plurality of instrument guidance procedures comprises a laser-guided procedure, an x-ray guided procedure, and a procedure that replans the instrument trajectory.
- The accompanying drawings illustrate preferred embodiments of the disclosure so far devised for the practical application of the principles thereof, and in which:
-
FIG. 1 is a is a schematic diagram showing an exemplary x-ray acquisition system for performing the disclosed method; -
FIG. 2 is a display view from an x-ray guidance system showing a three-dimensional rendering of a test phantom showing objects internal to the phantom that are representative of different types of patient tissue; -
FIG. 3 is a display view showing the selection of a target point on the first and second fluoroscopic views and the MPR view; -
FIG. 4 is a display view showing the selection of a skin entry point on the first and second fluoroscopic views and the MPR view; -
FIG. 5 is a side view of an exemplary arrangement for supporting a laser pointer for use with the disclosed methods; -
FIG. 6 is a top view of the arrangement ofFIG. 5 ; -
FIG. 7 is an isometric view of a flat x-ray detector in combination with a laser pointer; -
FIGS. 8 and 9 are schematic views of the interaction of the flat x-ray detector and laser pointer ofFIG. 7 with the support arrangement ofFIG. 6 ; -
FIG. 10 is an isometric view of a flat x-ray detector in combination with a laser pointer according to another aspect of the disclosure; and -
FIG. 11 is a display view from an x-ray guidance system showing a collimated fluoroscopic view of the target and skin entry points, as well as an oblique fluoroscopic view showing a planned path trajectory intersecting the target and skin entry points; -
FIG. 12 is an enlarged view of the collimated fluoroscopic view ofFIG. 11 ; -
FIG. 13 is an enlarged view of the collimated fluoroscopic view ofFIG. 11 showing an instrument inserted at the skin entry point; -
FIG. 14 is a collimated fluoroscopic view taken oblique to the view ofFIG. 13 showing the position of the instrument relative to the graphical overlay of the planned instrument trajectory; -
FIG. 15 is a collimated fluoroscopic view taken oblique to the views ofFIGS. 13 and 14 showing the position of the instrument relative to the graphical overlay of the planned instrument trajectory; -
FIG. 16 is a collimated fluoroscopic view showing the position of the instrument as it intersects the target; -
FIG. 17 is a C-arm CT (DynaCT) scan of the completed instrument insertion; -
FIGS. 18A-18E are flowcharts for using the disclosed method to select between a laser guidance technique and an x-ray guidance technique for planning and performing a percutaneous procedure; -
FIG. 19 is a flow chart showing a sequence of steps of a method for aligning an instrument using laser guidance; and -
FIGS. 20 is a flow chart describing a sequence of steps of an exemplary x-ray alignment method. - An “imaging system” is a system that includes at least a movable arm, an x-ray source, an x-ray detector, a display and a system controller. A “patient 3-dimensional image data set” is a three dimensional numerical array whose elements hold the values of specific physical properties at points in space inside the patient's body. A “multiplanar reformation image (MPR)” is a planar cross-section of the patient 3-dimensional image data set generated by cutting through the three-dimensional data set at some orientation (e.g., axial, coronal, sagittal, or oblique). A “fluoroscopic image” is a two-dimensional x-ray projection image showing internal tissues of a region of the body. A “live fluoroscopic image” is a sequence of x-ray images taken successively showing live movement of internal tissues of a region of the body. A “combined image” is an image in which an x-ray image is combined with a two-dimensional rendering of a three-dimensional data set computed under the same orientation as the x-ray image. A two-dimensional rendering of a three-dimensional data set is also referred to as a fluoroscopic overlay image or overlay image. “Co-registering” means aligning an x-ray image with a patient 3-dimensional image data set such that associated features within the x-ray image and a two-dimensional overlay image generated from the patient 3-dimensional image data set appear at the same location on a display in which the x-ray image and the overlay image are shown together. Co-registration can be point-based or gray-level based. In point-based co-registration, a transform is applied to the 3-dimensional image data set such that points in the resulting overlay image line up with their counterparts in the x-ray image as closely as possible. Gray-level based co-registration techniques determine the transform not by minimizing the distance between associated points in the overlay image and x-ray image, but by minimizing an error metric based on the resulting overlay image's gray levels and the x-ray image's gray levels. Co-registration can be static or elastic. In the latter case, the 3D data set may not only be rotated and shifted but warped as well. “Instrument” refers to any object which may pierce tissue of a patient, a non-limiting listing of which include needles and other biopsy devices, screws, implants, cannula, endoscopes, and anything else that can be inserted into a patient's body either percutaneously or intravascularly. A “skin entry point” is the position on a patient's skin at which an instrument is inserted. “Skin entry point data” is data representative of the skin entry point within the patient 3-dimensional image data set or within two x-ray views taken under different view orientations using a triangulation technique. A “target” or “target point” is a point within the body of a patient that is the subject of a percutaneous procedure. “Target point data” is data representative of the skin entry point within the patient 3-dimensional image data set or within two x-ray views taken under different view orientations using a triangulation technique. A “planned path” is a line generated between the skin entry point and the target point. “Instrument trajectory” is a desired trajectory of the instrument defined by the planned path. A “Bull's Eye View” is an x-ray view under which a target point and another point along the instrument trajectory are projected onto each other. The other point along the instrument trajectory may be the skin entry point. The movable arm view direction can be visualized using a graphical overlay in which the target point and skin entry point, forward-projected from 3-dimensions to 2-dimensions, are displayed as individual circles. If the Bull's Eye View has been reached, these two circles are projected at the same 2-dimensional position (i.e., they appear concentrically aligned). A “progression view” is an x-ray image taken at an oblique angle with respect to a line joining the skin entry point and the target. The term movable arm tomographic reconstruction refers to a technique in which multiple x-ray images taken along a particular image acquisition trajectory of the movable arm system, e.g., a partial circle. The multiple x-ray images are used to construct a patient 3-dimensional image data set. An “optical axis,” or “central ray” is an imaginary straight line passing through the midpoint of an x-ray source and the iso-center of the imaging device. In an ideal system geometry, the optical axis is perpendicular to the x-ray detector, intersecting the detector at its midpoint.
- Referring to
FIG. 1 , anexemplary system 1 is shown for performing a percutaneous procedure. Thesystem 1 may comprise an x-ray tube orsource 2 and associated support and filtering components. The x-ray source may be affixed to a support, such as amovable arm 4 to allow the x-ray source to be moved within a constrained region. In one embodiment, themovable arm 4 is a C-arm. The constrained region may be arcuate or otherwise three dimensional, depending on the nature of the support structure. A collimator may also be included, which defines the size and shape ofx-ray beam 6 emerging from the source. Anx-ray exposure controller 8 andsystem controller 10 may also be included.System controller 10 may be a personal computer or any known controller capable of receiving and transmitting control signals to/from the above-described x-ray system components via ahardware interface 12.System controller 10 may include auser input device 14, such as a trackball, mouse, joystick, and/or computer keyboard to provide for user input in carrying out various system functions, such as mode selection, linearity control, x-ray dose control, data storage, etc. Thesystem controller 10 may include aprocessor 16 executing instructions for performing one or more steps of the disclosed method. - In the illustrated embodiment, a
patient 18 is shown on patient-support table 20 such that anx-ray beam 6 generated by the x-ray source passes through him/her onto adetector 22. In one embodiment thedetector 22 is a flat panel detector that acquires digital image frames directly, which are transferred to animage processor 24. A display/record device 26 records and/displays the processed image(s). The display/record device 26 may include a display for displaying the displayed image output, as well as a separate device for archiving. The image is arranged for storage in an archive such as a network storage device. Thex-ray source 2 is controlled by thesystem controller 10 viaexposure controller 8 andx-ray generator 28. The position of thex-ray source 2 may be adjusted via a drive system associated with themovable arm 4. Themovable arm 4,x-ray source 2,x-ray detector 22,display 26 andsystem controller 10 may together be referred to as an imaging system. - As previously noted, laser-guided and radiation-guided systems are known for use in guiding instruments during percutaneous procedures. Although two specific laser guidance and x-ray guidance systems and methods will be described herein, it will be appreciated that the specifics of each system are not critical, and that any of a variety of other laser guidance and/or x-ray guidance systems and methods may be used with the disclosed system and method.
- Workflow Steps
- Initially, a 3-dimensional data set of a targeted region within the patient will be obtained using one of a variety of techniques, such as computed tomography (CT), magnetic resonance imaging (MRI), positron emission tomography (PET), C-arm tomographic imaging (syngo DynaCT, Siemens AG, Forchheim, Germany) or the like. Referring to
FIG. 2 , a 3-dimensional rendering of such a 3-dimensional data set is shown. In the illustrated embodiment, the 3-dimensional dataset has been loaded into an appropriate 3D rendering program, such as the Siemens InSpace system, for viewing (the figures show 3-dimensional images representative of test phantoms that have a plurality of objects placed inside to simulate vessels, landmarks and targets). - The x-ray views (fluoroscopic images) obtained using the
movable arm 4,source 2 anddetector 22, need to be appropriately “registered” with the overlay images derived from the 3-dimensional data set for the region of interest of the patient. Data registration may be performed manually, automatically or semi-automatically (i.e., computer assisted). This registration step ensures that the fluoroscopic (x-ray) images of the patient obtained using thesource 2 anddetector 22 match the images of the patient constructed from the 3-dimensional data set. This enables instrument positioning using information on target position obtained from the 3-dimensional data set. In one embodiment, the co-registration step is performed by minimizing an error metric based on gray levels of a resulting overlay image and the x-ray image. In another embodiment, the co-registration step is performed by applying a transform to the 3-dimensional image data set such that points in a resulting overlay image align with counterpart points in the x-ray image. - Once the 3-dimensional data set has been appropriately registered to the 2-dimensional x-ray geometry, the instrument trajectory may be planned. To this end, the user may select a target point, Xt, and a skin entry point, Xe within the overlay images by visualizing the areas within a particular image and clicking on the point(s) using a selector such as a mouse button. Alternatively, the instrument trajectory may be planned before the registration step. In either case, registration ensures that the planned trajectory is mapped to the corresponding anatomical environment.
- As shown in
FIG. 3 , target point selection may be done by selecting a desired MPR view, such as by a right-click of a mouse pointer on an appropriate soft key 3A in a pop-up window in the display. In the illustrated embodiment, this results in desired MPR views being displayed in the upperleft quadrant 3B, and upperright quadrant 3C. The target point is “selected” by clicking with a mouse pointer at thetarget position 3D in the lower leftquadrant MPR display 3E. The skin entry point may be selected (“clicked”) in the same manner. - Based on where the click points are made in the MPR view, the system obtains data representative of the target and skin entry points using data from the 3-dimensional patient data set. Using the target point data and skin entry point data, the system generates a graphical overlay showing a line which represents the planned instrument trajectory. Such a graphical overlay is applied to each of the images shown on the user display (as seen as
line 4F inFIG. 4 ). This graphical overly may consist of the target and skin entry points, as well as a line intersecting the two, and may be overlaid onto one or more of the display views (e.g., x-ray, MPR, 3-dimensional rendering) as desired by the user. Since the x-ray views and the patient 3-dimensional image data set are registered with one another at this point in the procedure, the system can map the exact location of the target point Xt, and the skin entry point Xe (as well as the connecting vector “n”) at their precise locations on each of the display views. As will be described in greater detail later, the displayed line represents the planned instrument trajectory. - As shown in the four
display quadrants FIG. 4 , the skin entry point Xe and the target point Xt define aline 4F (shown in 4 places) in 3-dimensional space having a path vector (n=Xe−Xt). It will be appreciated that, as an alternative to defining the path vector using two points in space, the trajectory could instead be defined using xt and the path vector “n”. This may be appropriate, for example, in the case where the patient is large and the skin entry point cannot be seen in the movable arm CT (i.e., it is outside the physical range of the movable arm CT). In such a case, the user may not need to see the exact location of the skin entry point if there are no organs in the immediate area (e.g., where there is only fat tissue). Whether this is indeed the case or not can be checked if the registered 3-dimensional patient data set comprises the complete volume of the patient or if there is another 3-dimensional data set (e.g., CT, MRI) that provides similar information in outer patient regions. In such a case, the physician may simply “click” the desired target point xt in the appropriate display view(s), and may “click” a second point that is visible in the movable arm CT to define a desired path vector n without specifying an actual skin entry point Xe. A desired instrument trajectory is a straight line (path vector) “n” that originates outside the patient's body and passes through the skin entry point xe and the target point xt without crossing bones or vital organs. - As shown in
FIG. 4 , the graphical overlay (i.e., the one in which points Xt and Xe are shown along with the line connecting them) may be combined with an anatomical image (i.e., an MPR view). In addition, the graphical overlay may be combined with both the forward projected anatomical image (overlay image) and live x-ray views. In the final movable arm position (again, the Bull's Eye View position), the graphical overlay may also adjust to different x-ray zoom conditions so that the user may confirm final positioning by revealing small deviations from the optimal view orientation. This resizing is automatically achieved through the use of a calculated conversion factor determined e.g., using a “similar triangles” technique. - Once the planned instrument trajectory is obtained, a determination is made regarding whether a laser guidance procedure can be performed, or whether an x-ray guidance procedure can be performed, or whether the path needs to be replanned. Laser guidance can be used if the planned instrument trajectory can be targeted in such a manner that the optical axis of the x-ray source on the C-arm coincides with the planned trajectory. Since the C-arm is calibrated to the 3D image data set, the system can automatically determine whether such an orientation is possible.
- If the system determines that laser guidance cannot be employed, a determination is made regarding whether the C-arm can be made to coincide with the planned instrument trajectory so that the extension of the trajectory is projected onto the detector (is in the “Bull's Eye View”), in which case an x-ray guidance procedure can be used for instrument guidance.
- If the planned instrument trajectory does not satisfy either of the aforementioned criteria, the instrument trajectory is replanned until at least an x-ray guidance procedure can be used. Of course, the user can re-plan the instrument trajectory more than once (e.g., if a laser-guided procedure is preferred, re-planning can be performed as many times as desired in order to achieve a planned instrument trajectory that meets the criteria for utilizing laser guidance).
- Once a final instrument trajectory is obtained, the instrument is guided to the target using the selected guidance technique, as will be explained in greater detail below with reference to the respective procedure).
- Referring now to
FIGS. 5-10 , an exemplarylaser guidance system 29 for use with the disclosed method will be described in detail. Referring toFIGS. 5 and 6 , amovable stand 30 is attached to a patient table 20. Thestand 30 has aswivel arm 32 which is rotatable according to thearrow 34 inFIG. 6 . Afastening device 36 is provided on theswivel arm 32, on which alaser pointer 38 is mounted. Thefastening device 36 is designed in such a way that 3-dimensional movement of thelaser pointer 38 is possible, in particular according to themotion arrows - The
stand 32 may have aposition sensor 46 through which the position of thelaser pointer 38 is measurable relative to the patient table 20. Theposition sensor 46 may be used to define a coordinate system which can be recorded with the previously obtained 3-dimensional data set. - In one embodiment, motors may be used to achieve the automated movements according to the
motion arrows laser guidance system 29 can instead be moved by hand. - In an alternative embodiment, the
laser pointer 38 may be affixed to thex-ray detector 22.FIG. 7 shows thex-ray detector 22 to which twobraces 48 are attached, which run parallel to thesurface 50 of thex-ray detector 22. Asheath 52 is positioned perpendicular to thebraces 48 and is used to attach thelaser pointer 38. With this embodiment, to move the laser guidance system theentire x-ray detector 22 is moved, in particular the C-arm 4 is swiveled or tilted, to which thex-ray detector 22 andlaser pointer 38 are attached. - In the event that no additional fluoroscopic images are deemed necessary during a percutaneous procedure, the
laser pointer 38 can remain in thesheath 52. Where braces 48 are used to attach thelaser pointer 38 to thex-ray detector 22, the braces and pointer may be separated from thex-ray detector 22 after reaching a desired instrument alignment. - Referring now to
FIG. 8 , astand 54, similar to stand 32 inFIGS. 5 and 6 , may be used to guide thelaser pointer 38. Thelaser pointer 38 may be fixed to thestand 54, separate from thex-ray detector 22. The C-arm 4 may be turned, e.g., into the position shown inFIG. 9 . Thelaser pointer 38, since it is attached to thestand 54, remains aligned to thepatient 18, and thus an instrument can be aligned along thelaser beam 56 generated by thelaser pointer 38. During the percutaneous procedure, fluoroscopic images can be taken using thex-ray source 2,x-ray detector 22, and C-arm 4 oriented at an oblique angle with respect to the instrument trajectory (as defined by the laser beam 56). As will be appreciated, these fluoroscopic images may aid the physician to guide the instrument precisely and accurately to the target. - In a further alternative embodiment, a pair of
laser sources 58 may be arranged on or near the edges of thex-ray detector 22 so that they are not visible in the images produced using thedetector 22. Such an arrangement is shown inFIG. 10 . The laser sources 58 may generate fan-shaped light in which a fan angle a is defined and along which the fan extends, such that produced laser beams are on a single level. The two levels intersect in a centralstraight beam line 60 such that thisstraight beam line 60 can readily be recognized by the user. The extension of thestraight line 60 is the skin entry point Xe and the target point Xt, and thus represents the instrument trajectory. The instrument is aligned using thisbeam line 60. In one embodiment, alaser pointer 38 is aligned alongstraight beam line 60, and thex-ray detector 22 may be moved as desired to obtain one or more fluoroscope images, while thelaser pointer 38 is used as part of the guidance system. - Exemplary X-ray Guidance System and Method
- Referring now to
FIGS. 11-17 , an exemplary x-ray guidance system and method will be described in greater detail. As previously described, the system generates a line “n” between the skin entry point Xe and the target point Xt. This line is used to align themovable arm 4 to achieve a “Bull's Eye View,” in which the two points are superimposed to show only a single point to the user. The instrument is placed at the skin entry point and aligned using the Bull's Eye View, shown indisplay quadrant 11A ofFIG. 11 , to orient the instrument along the planned instrument trajectory (i.e., one that hits both points). Initial alignment is verified using a fluoroscopic image of the oriented instrument. After the initial alignment is verified, the user inserts the instrument a short distance into the patient. One or more progression x-ray views are used to verify that the instrument is on the planned path between the skin entry point and the target point. The user may employ an iterative approach of inserting the instrument a small distance followed by a verification of the instrument's position using progression x-ray views to guide the instrument to the target. When the instrument reaches the target, a desired additional procedure may be performed, such as a biopsy, a drainage procedure, a radiofrequency ablation, or other medical interventional procedure. - Once an acceptable instrument trajectory has been planned, the
movable arm 4 may be moved into the Bull's Eye View position. As previously noted, the Bull's Eye View orientation is one in which the skin entry point and the target point (Xe and Xt) overlie each other on the same detector positions (seeFIG. 4 ). Adjustment of themovable arm 4 to achieve this positioning can either be performed manually or automatically. - In practice, positioning an instrument at the skin entry point Xe may be a difficult task, and thus a positioning aid may be used, such as a biopsy grid, a biopsy mesh, or other appropriate device. Once the appropriate instrument positioning has been achieved, collimation may be set around the Bull's Eye View to limit radiation exposure to the user and patient. Collimation may be either manual or automatic.
- To facilitate fine adjustment of the instrument position, the Bull's Eye View may be isolated and enlarged, as shown in
FIG. 12 , to reveal slight deviations from the desired instrument positioning and orientation. Thus,FIG. 12 shows a switch from the four-quadrant view ofFIG. 11 to a full-window view with an increased zoom level to reveal deviations from the ideal Bull's Eye View. As can be seen, the zoomed view ofFIG. 12 shows concentric overlapping circles 62 (in black), 64 (in white) indicating that the Bull's Eye View has been achieved. In the illustrated embodiment, a SeeStar device (manufactured by Radi Medical Devices, Uppsala, Sweden) has been used to aid instrument positioning. The SeeStar device is an instrument guide that produces an elongated artifact in an x-ray image. This elongated artifact indicates the trajectory of an instrument inserted through the SeeStar, and thus it can be determined whether the selected trajectory will intersect the target as desired. The SeeStar shows up as a circle 66 (i.e., a black tube-like shadow in the figure) in the center of the displayed circles, which indicates that it is in the desired orientation (i.e., one that is in alignment with a trajectory that passes through the skin entry point and the target point). If the SeeStar were to show up as a line, its position/orientation would be adjusted, followed by re-verification of the new position/orientation by subsequent x-ray views. - As previously noted, in lieu of a SeeStar device, the user could instead use a hollow instrument guide to verify instrument placement. The hollow instrument guide may be configured so that it shows up as a point under fluoroscopy in the Bull's Eye View when a desired alignment is achieved. The hollow instrument guide may be clamped in position during fluoroscopy to limit radiation to the user, and its position may be adjusted and verified in a manner similar to that described in relation to the SeeStar device.
- Once the desired instrument alignment is achieved, the instrument is pushed forward by a small amount into the patient tissue to stabilize the instrument's orientation. This insertion is performed under the Bull's Eye View. As shown in
FIG. 13 , the user can see straight down the instrument guide as well. The large circle represents the instrument body and instrument tip. In the illustrated embodiment they are exactly aligned, which is why only one large circle is visible in the figure. The black “bulb” in the center is the instrument (in the illustrated case, a needle). It appears in this way because it is almost (but not perfectly) aligned with the viewing direction. If the instrument were perfectly aligned, it would be shown as a circle in this view. - Instrument alignment may again be verified at this early stage of insertion. Such verification can be performed using x-ray “progression views,” which are oblique x-ray views (i.e., non-Bull's Eye Views) obtained using the
source 2 anddetector 22. It will be appreciated that the user may also return to the Bull's Eye View at any time during the procedure to obtain additional information regarding instrument alignment. If a bi-plane x-ray device is available with the B-plane providing a progression, it is possible to check if the instrument remains aligned with the associated graphical overlay (shown asline 68 inFIG. 14 ) while the instrument is being pushed forward into the tissue. In the illustrated embodiment, the instrument appears as a thin diagonal line starting from the bottom left of the image, just above thegraphical overlay line 68. - During the procedure, the
movable arm 4 may be moved between the first and second progression views to enable the user to control the actual instrument movement from two oblique angles until the instrument has reached the target. When the target has been almost reached in one progression view, the user can return to the other progression view to confirm that the instrument has indeed been placed correctly before making the final push or releasing a spring-loaded biopsy device if one is used. The user can also return to the Bull's Eye View to obtain additional orientation information. - Referring to
FIG. 15 , the movable arm is moved into the second progression view to check on instrument placement. Ifinstrument 70 andgraphical trajectory 72 align, theinstrument 70 can be moved into the target. In the illustrated embodiment, a small degree of bend is shown in theinstrument 70, which can occur when the instrument is small/thin and the target is dense. - Referring to
FIG. 16 , a return to the first progression view is performed to confirm instrument placement at the target. It will be appreciated that if a bi-plane fluoroscopic device is available, there is no need to rotate the movable arm's A-plane back and forth between two progression views. Instead, the A-plane can be put at the first progression view while the B-plane is positioned under the second progression view, and both may be viewed simultaneously As an alternative to the use of progression views to verify instrument positioning during insertion, movable arm CT (DynaCT) acquisitions can be performed throughout the workflow to verify the instrument position and orientation at each stage of insertion. Such a movable arm CT acquisition can be performed at one or more stages of the procedure, as desired by the user. It is noted, however, that the movable arm CT procedure can take up to several minutes and increases the overall amount of radiation exposure to the patient. Progression views, by contrast, are relatively fast (almost instantaneous). The user simply rotates the movable arm (if required) to the desired progression view location, and releases the x-rays. The x-ray image shows up on the display in a few seconds. -
FIG. 17 shows a movable arm CT (DynaCT) scan of the completed instrument insertion position. Although not required, performing such a verification ensures that the positioning is correct prior to completing the procedure. As can be seen in the MPR views shown in the upper and lowerleft quadrants FIG. 17 , theinstrument 70 has been appropriately engaged with thetarget 72. The upper right quadrant view (which shows the Bull's Eye View), however, reveals that theinstrument 70 just made it into thetarget 72. - Referring now to
FIGS. 18A-18E , the disclosed method for selecting a guidance technique for performing a percutaneous procedure will be described in greater detail. At step 1800 (FIG. 18A ), apatient 18 is placed on a table 20 in proximity to an imaging system having amovable arm 4, anx-ray source 2, anx-ray detector 22, asystem controller 10 and adisplay 26. Atstep 1810, a 3D patient image data set of a patient tissue region is provided. Atstep 1820, an x-ray image of a targeted patient tissue region is acquired using the x-ray source and x-ray detector. In one embodiment (seeFIG. 18B ), atstep 1822 obtaining an x-ray image of the patient tissue region using thex-ray source 2 andx-ray detector 22 comprises obtaining a plurality of x-ray images, displaying the plurality of x-ray images on thedisplay 26 along with a three-dimensional rendering of the three-dimensional image data set, and an overlay image representing a 2-dimensional rendering of the 3-dimensional image data set. Atstep 1824, the skin entry point, the target point and the planned instrument trajectory are graphically displayed in their respective positions on the plurality of displayed x-ray images, the three-dimensional rendering and the overlay image. At step 1830 (FIG. 18A ), the 3D patient data set is co-registered to the x-ray image acquired using the imaging system. In one embodiment (seeFIG. 18C ), atstep 1832 the co-registering step comprises applying a transform to the 3D patient image data set such that points in a resulting overlay image align with counterpart points in the x-ray image. At step 1840 (FIG. 18A ), the user plans an instrument trajectory in the 3D volume by selecting target point data representative of a target point within the patient tissue region, and skin entry point data representative of a skin entry point. The target point data and skin entry point data may be obtained from the co-registered 3D image data set. As a result of co-registration, it may also be possible to plan the instrument path immediately afterstep 1800. This may be beneficial if planning is performed on a pre-procedural data set acquired on an imaging device that is different from the C-arm imaging system. - At
step 1850 instrument alignment is performed using one of a plurality of instrument guidance procedures, where the instrument guidance procedure is selected on the basis of the planned instrument trajectory, and a position of themovable arm 4. The plurality of instrument guidance procedures may include a laser guided procedure, an x-ray guided procedure, or a procedure that replans the instrument trajectory. If the planned instrument trajectory can be targeted in such a manner that the optical axis (central ray) of thex-ray source 2 on the C-arm 4 coincides with the trajectory, at step 1860 a laser guidance procedure (such as the one described previously in relation toFIGS. 5-10 , and discussed in more detail below) is selected for instrument guidance. This determination is based on information relating to calibration of the C-arm with the 3D volume and a collision control feature of the system. The step of aligning the instrument may comprise aligning the instrument along a laser beam path generated by a laser pointer, the laser beam path intersecting the skin entry point Xe and the target point Xt (step 1852,FIG. 18D ). Atstep 1854, the laser beam path may comprise the intersection of at least two laser fan patterns. - If the central ray of the x-ray source does not coincide with the planned instrument trajectory, and an extension of the planned instrument trajectory intersects with the
x-ray detector 22, atstep 1870 an x-ray guidance procedure (such as the one described previously in relation toFIGS. 11-17 , and discussed in more detail below) is selected for instrument guidance. If a central ray of the x-ray source does not coincide with the planned instrument trajectory, and an extension of the planned instrument trajectory does not intersect with thex-ray detector 22, the instrument trajectory is replanned atstep 1840. This replanning step may involve repositioning the patient by moving the patient table. Atstep 1890, a warning signal may be provided to the user if the central ray of the x-ray source does not coincide with the planned instrument trajectory, and an extension of the planned instrument trajectory does not intersect with thex-ray detector 22. - If laser guidance can be employed, at step 1900 (see
FIG. 19 ), the 3D image data set is represented from the perspective of the target guidance system. The target guidance system defines a vector in a view direction to the patient. The patient is assigned to the 3D image data set so that the target guidance system defines a view direction on the 3D image data set. This may occur so that the target guidance system is represented in the central point of the monitor as an X, and the target point should also be represented as an X. The skin entry point Xe and target point Xt may also be represented. The goal is for the target guidance system to precisely show these points superimposed on each other. It will be appreciated that the target may be represented as an X, while the skin entry point may be represented using a circle, or that circles may be used to represent both the target and skin entry points. Either graphical representation can be combined with a fluoroscopic overlay image representing the patient's anatomical structures. - The user may move the
laser guidance system 29 or the patient 18 atstep 1910 to achieve an appropriate position so that the points are superimposed. Because this movement changes the view of the laser guidance system to the patient and with it the 3D image data set, the representation is adjusted at step 1920. Atstep 1930, a check is performed to determine whether the target representation has been reached (i.e., whether all of the points align with each other). If the target representation is not achieved,steps step 1940 the instrument is guided using the target guidance system. If the laser beam exactly aligns with the central x-ray of the x-ray source, the laser may be used by itself to provide instrument guidance. It will be appreciated that for instances in which the laser alone is used for instrument guidance, that the practitioner still may verify the instrument position within the patient using x-ray imaging. - In one embodiment, automatic positioning of the patient table 20 may be achieved at
step 1910. The goal is to align the central x-ray of the x-ray source with the planned instrument trajectory. Thus, the position of the patient table 20 is automatically adjusted using positioning motors system until the planned instrument trajectory intersects the iso-center of the C-arm 4. The C-arm 4 is rotated such that its view orientation matches the orientation of the instrument trajectory. - Referring again to
FIG. 18A , if, atstep 1850 it is determined that laser guidance cannot be employed, but the planned instrument trajectory still can be targeted in such a manner that a central ray of the x-ray source can be made to coincide with the planned instrument trajectory so that the extension of the trajectory intersects with the detector 22 (the “Bull's Eye View”), atstep 1860 an x-ray guidance procedure (such as the one described in relation toFIGS. 11-17 ) is selected for instrument guidance. - If x-ray guidance is used, at step 2010 (see
FIG. 20 ), themovable arm 4 is adjusted to a position at which an x-ray image taken using thex-ray source 2 and thex-ray detector 6 results in the target point Xt and the skin entry point Xe being superimposed on top of each other. In one embodiment, the step of adjusting the movable arm may comprise determining a spatial orientation within the three-dimensional image data set at which the target point and skin entry point are superimposed on each other, and automatically moving the movable arm so that a further x-ray image obtained using thex-ray source 2 anddetector 22 images the target and skin entry points onto the same pixels of the x-ray detector. - At
step 2020, alignment of an instrument positioned between thex-ray source 2 and the skin entry point Xe is verified as an acceptable position with respect to the planned instrument trajectory when the instrument appears on the display as a point overlying the target point Xt and the skin entry point in a verification x-ray image taken using thex-ray source 2 anddetector 22. In one embodiment, acceptable position with respect to the planned instrument trajectory is verified by taking multiple x-ray images using thex-ray source 2 anddetector 22 at movable arm positions oblique to the position of themovable arm 4 used to obtain the verification x-ray image. - In further steps, the user may insert the instrument into the patient at the skin entry point. One or more progression x-ray views may be taken to ensure that the instrument remains aligned with the projected instrument path. It will be appreciated that the user may also return to the Bull's Eye View to gain additional insights regarding instrument orientation. The user may press the instrument further into the patient toward the target while making adjustments to ensure the instrument remains aligned with the projected instrument path. The pressing and progression x-ray steps may be repeated as desired by the user to guide the instrument in an incremental manner to intersect the target.
- Referring again to
FIG. 18A , if neither laser guidance nor x-ray guidance can be used (i.e., the central ray of the x-ray source does not coincide with the planned instrument trajectory and an extension of the planned instrument trajectory does not intersect with the x-ray detector, atstep 1880 it is determined that the planned instrument trajectory needs to be replanned, and the process returns to step 1840 and the instrument trajectory is replanned. Replanning is performed that at least the instrument trajectory can be planned using a laser or x-ray guidance techniques, as determined again insteps - In one embodiment, shown in
FIG. 18E , a suggested re-planned instrument trajectory may be automatically formulated and presented to a user so that either laser guidance or x-ray radiation guidance techniques can be used. Thus, atstep 1842 re-planning may be carried out manually (using a graphical user interface (GUI)), or by means of a “tracked” instrument (as will be described in greater detail below). Alternatively, atstep 1844 re-planning can be carried out automatically. Once the instrument is reoriented along the re-planned trajectory, the instrument may be advanced and controlled as described previously in relation to the laser-guidance and x-ray-guidance techniques. - In one alternative embodiment, skin entry point replanning may be facilitated by the use of an optical or electromagnetic position sensor attached to the instrument. The position sensor may automatically identify both the location and the direction of movement of the instrument so that when the position sensor is moved on the surface of the skin of a patient, the skin entry point Xe can be newly determined such that, given a stationary target point (but a variable “tracked” starting point), the resultant path of the instrument can automatically be entered into the 3D dataset (and rendered as desired). Once an appropriate instrument trajectory has been selected, the skin entry point Xe is automatically determined and a marking (e.g., color mark, optionally combined with the injection of a contrast agent) can be made on the skin. In one exemplary embodiment, the instrument may comprise a biopsy localization device (e.g., SeeStar) with a position sensor.
- In another embodiment, a flexible mounting arm may be provided with a holder for an instrument guide (e.g., guide sleeve, instrument mounting or clamping device). The flexible mounting arm can be connected to structure such as the patient table or other adjacent equipment and may be used to hold an instrument guide during the procedure. The flexible mounting arm may be automatically robotically controllable to enable the user to determine the instrument trajectory automatically or semi-automatically, without touching the patient. Once the instrument trajectory has been finalized, the instrument may be precisely and accurately inserted into the patient. This technique may also facilitate the rapid percutaneous insertion of an instrument (e.g., by means of a biopsy cannula).
- In further embodiments in which the instrument comprises a biopsy cannula, a “tracked” head (e.g., ultrasound) may be used to enable real time display of the instrument in the 3D volume (i.e., rendered or otherwise displayed on the
display 26, providing a “live image” of the instrument as additional piece of information to the user. - Furthermore, when replanning of the instrument trajectory is performed, a forward-looking (i.e., “endoscopic view”) may be provided to the user to enable the user to “look” into the direction of the target (see
FIG. 12 ). Similarly, other “context-based” visualization methods may be employed, such as those which eliminate soft tissue from the 3D dataset, while retaining other elements of interest such as bones or vessels. This may enable the physician to see whether obstructions exist along the path of a planned instrument trajectory, and, furthermore, whether the path is safe (i.e., does not intersect with blood vessels). Such a 3D dataset could alternatively or additionally be acquired by means of an ultrasound head which can be movably attached on a mounting arm and “tracked.” - Alternatively, a “tracked” 2D ultrasound transducer may be employed to allow the user to determine both the target and the direction of a percutaneous intervention. If the ultrasound transducer and the C-arm are registered with each other, the ultrasound path can be imported into the display, and used to orient the C-arm.
- It will be appreciated that if the x-ray guidance technique is used to guide the instrument, the projection of the “skin entry point” and the “target point” does not impinge on the center of the
detector 22 but rather on an outer region of the detector which cannot be shielded by a central collimator (if one is used). To alleviate this, the table 20 may be moved in such a manner that the “bull's eye view” passes through the central collimating opening. - The system may additionally employ a limitation on potential C-arm trajectories to ensure that the C-arm will not collide with the user. For example, when using the x-ray guidance procedure, the user may move from the “bull's eye view” into first and second progression views to check the position of the inserted instrument within the patient (see
FIG. 14 ). In such cases, it may not be possible to determine whether moving the C-arm to obtain those progression views will cause a collision between the C-arm and the user. This uncertainty factor can be avoided by limiting the direction of motion of the C-arm during interventions in such a manner that rotations around the patient are carried out along the median plane (i.e., at a cranial/caudal (CRAN/CAUD) angle of 0 degree). Thus, when moving the C-arm from one position to another, the angulation of the C-arm would first be retracted (in CRAN/CAUD direction), a new angle of rotation would be targeted (in left anterior oblique/right anterior oblique (LAO/RAO) direction), and the new angulation set. This eliminates the risk that a C-arm could collide with the leg of the user since he would know what movement to expect. - In a further embodiment, feedback-controlled laser orientation technique can also be employed. In this embodiment, an additional signal (in addition to the optical signal on the screen, given by the superimposition of certain markers) is provided to the system to provide feedback as to how accurately a planned instrument trajectory coincides with the optical axis of the system (i.e., the direction of the laser).
- Other user-feedback techniques may also be used, such as displaying on the
display 26 offsets (in degree or distance) between the planned instrument trajectory and the optical axis of the system. Alternatively, a lamp may be provided, the color of which changes as the two are more closely aligned. Acoustic signals and haptic force feedback are other examples of acceptable feedback techniques. If such a feedback is available, the workflow could be as follows: at the beginning of the procedure, the C-arm is moved into a “generalized bull's eye view.” A “generalized bull's eye view” is an x-ray view projecting target and skin entry points onto each other along an x-ray beam direction that does not coincide with the central ray of the x-ray cone. If work with the laser is desired, the table can be manually moved until the patient is in such a position that the C-arm can be moved into a viewing direction in which the axis of the optical system intersects the target point. Only after the patient and the C-arm are moved into the correct positions does the laser turn on. In all other cases, it will not. If the C-arm is moved manually, “force feedback” can be used to ensure that the positions can be approached in a controlled manner, without overshooting them. Alternatively, positioning motors may be provided to enable the table to be automatically moved into a position suitable for laser operation. - The method described herein may be automated by, for example, tangibly embodying a program of instructions upon a computer readable storage media capable of being read by machine capable of executing the instructions. A general purpose computer is one example of such a machine. A non-limiting exemplary list of appropriate storage media well known in the art would include such devices as a readable or writeable CD, flash memory chips (e.g., thumb drives), various magnetic storage media, and the like.
- The features of the method have been disclosed, and further variations will be apparent to persons skilled in the art. All such variations are considered to be within the scope of the appended claims. Reference should be made to the appended claims, rather than the foregoing specification, as indicating the true scope of the disclosed method.
- The functions and process steps herein may be performed automatically or wholly or partially in response to user command. An activity (including a step) performed automatically is performed in response to executable instruction or device operation without user direct initiation of the activity.
- The systems and processes of
FIGS. 1-20 are not exclusive. Other systems, processes and menus may be derived in accordance with the principles of the invention to accomplish the same objectives. Although this invention has been described with reference to particular embodiments, it is to be understood that the embodiments and variations shown and described herein are for illustration purposes only. Modifications to the current design may be implemented by those skilled in the art, without departing from the scope of the invention. The processes and applications may, in alternative embodiments, be located on one or more (e.g., distributed) processing devices accessing a network linking the elements of FIGS. 1 and 5-10. Further, any of the functions and steps provided inFIGS. 18A-20 may be implemented in hardware, software or a combination of both and may reside on one or more processing devices located at any location of a network linking the elements of FIGS. 1 and 5-10 or another linked network, including the Internet.
Claims (24)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/431,518 US8165660B2 (en) | 2008-05-02 | 2009-04-28 | System and method for selecting a guidance mode for performing a percutaneous procedure |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US4980308P | 2008-05-02 | 2008-05-02 | |
US12/431,518 US8165660B2 (en) | 2008-05-02 | 2009-04-28 | System and method for selecting a guidance mode for performing a percutaneous procedure |
Publications (2)
Publication Number | Publication Date |
---|---|
US20090274271A1 true US20090274271A1 (en) | 2009-11-05 |
US8165660B2 US8165660B2 (en) | 2012-04-24 |
Family
ID=41257080
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/429,546 Abandoned US20090281452A1 (en) | 2008-05-02 | 2009-04-24 | System and method for a medical procedure using computed tomography |
US12/431,518 Active 2030-06-07 US8165660B2 (en) | 2008-05-02 | 2009-04-28 | System and method for selecting a guidance mode for performing a percutaneous procedure |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/429,546 Abandoned US20090281452A1 (en) | 2008-05-02 | 2009-04-24 | System and method for a medical procedure using computed tomography |
Country Status (1)
Country | Link |
---|---|
US (2) | US20090281452A1 (en) |
Cited By (52)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120022508A1 (en) * | 2010-07-22 | 2012-01-26 | Patrick Gross | Method and orientation device for marking a predetermined guide path of a medical instrument |
US20120108952A1 (en) * | 2010-10-29 | 2012-05-03 | Eli Stern | Methods and systems for patient alignment for nuclear medicine imaging |
US20120209123A1 (en) * | 2011-02-10 | 2012-08-16 | Timothy King | Surgeon's Aid for Medical Display |
US20120281812A1 (en) * | 2011-04-22 | 2012-11-08 | Toshiba Medical Systems Corporation | X-ray image diagnosing apparatus |
US20120300910A1 (en) * | 2011-05-26 | 2012-11-29 | Sebastien Gorges | X-ray imaging apparatus having a variable distance between an x-ray source and an object to be imaged |
WO2012168749A1 (en) * | 2011-06-06 | 2012-12-13 | Sarr Souleymane | Removable guide device for radiofluoroscopic infiltration having an image intensifier |
US20130083894A1 (en) * | 2011-09-30 | 2013-04-04 | Siemens Aktiengesellschaft | Method for controlling the movement of an x-ray apparatus and x-ray system |
US20130089180A1 (en) * | 2011-10-11 | 2013-04-11 | Siemens Aktiengesellschaft | Method and apparatus for producing an x-ray projection image in a desired direction |
US20130116550A1 (en) * | 2011-07-06 | 2013-05-09 | Hideaki Ishii | Medical diagnostic imaging apparatus |
US20130136332A1 (en) * | 2011-11-29 | 2013-05-30 | Hisayuki Uehara | X-ray image diagnosis apparatus |
JP2013172951A (en) * | 2012-01-27 | 2013-09-05 | Toshiba Corp | X-ray ct apparatus and x-ray ct system |
WO2014125789A1 (en) * | 2013-02-14 | 2014-08-21 | Seiko Epson Corporation | Head mounted display and control method for head mounted display |
FR3007962A1 (en) * | 2013-07-04 | 2015-01-09 | X Nov Ip | GRAPHICAL SELECTION OF BONE ANCHOR PROSTHESIS |
US20150080740A1 (en) * | 2013-09-18 | 2015-03-19 | iMIRGE Medical INC. | Optical targeting and visualization of trajectories |
EP2526869A3 (en) * | 2011-05-23 | 2015-10-07 | Fujifilm Corporation | Image processing device, radiographic image capture system, image processing method and image processing program storage medium |
US20160147939A1 (en) * | 2014-11-26 | 2016-05-26 | Koninklijke Philips N.V. | Efficient management of visible light still images and/or video |
EP3028675A1 (en) * | 2014-12-05 | 2016-06-08 | X.Nov IP Sarl | Graphical selection of a prosthesis with bone anchor |
WO2016119053A1 (en) * | 2015-01-27 | 2016-08-04 | iMIRGE Medical INC. | Optical targeting and visualization of trajectories |
US20160345923A1 (en) * | 2015-06-01 | 2016-12-01 | Toshiba Medical Systems Corporation | Medical image processing apparatus and x-ray diagnostic apparatus |
US20170071672A1 (en) * | 2014-03-04 | 2017-03-16 | Xact Robotics Ltd. | Dynamic planning method for needle insertion |
WO2017055990A1 (en) * | 2015-09-28 | 2017-04-06 | Koninklijke Philips N.V. | Optical registration of a remote center of motion robot |
WO2017070124A1 (en) * | 2015-10-19 | 2017-04-27 | New York University | Electronic position guidance device with real-time auditory and visual feedback |
US20170156800A1 (en) * | 2014-03-17 | 2017-06-08 | Roy Anthony Brown | Surgical Targeting Systems and Methods |
US20170193160A1 (en) * | 2016-01-06 | 2017-07-06 | Boston Scientific Scimed, Inc. | Systems and methods for planning medical procedures |
US20170245951A1 (en) * | 2012-06-21 | 2017-08-31 | Globus Medical, Inc. | Surgical robot platform |
US20170290559A1 (en) * | 2014-09-25 | 2017-10-12 | Siemens Healthcare Gmbh | Mobile x-ray machine |
NL2016800A (en) * | 2016-05-19 | 2017-11-23 | Umc Utrecht Holding Bv | Method of positioning an interventional device. |
EP3254627A1 (en) * | 2016-06-08 | 2017-12-13 | General Electric Company | Fluoroscopic guidance system with offset light source and method of use |
US20180008217A1 (en) * | 2016-07-08 | 2018-01-11 | Alexander Gemmel | Motion controllers for mobile x-ray devices |
US20180035984A1 (en) * | 2016-08-03 | 2018-02-08 | Siemens Healthcare Gmbh | Biopsy unit |
US20180087900A1 (en) * | 2016-09-28 | 2018-03-29 | Kla-Tencor Corporation | Optical Near-Field Metrology |
US20180199901A1 (en) * | 2015-07-16 | 2018-07-19 | Koninklijke Philips N.V. | Device for remote fluoroscopy, nearby fluoroscopy and radiology |
US20180214007A1 (en) * | 2017-01-30 | 2018-08-02 | Seiko Epson Corporation | Endoscope operation support system |
US10130388B2 (en) | 2014-03-13 | 2018-11-20 | New York University | Position guidance device with bubble level |
US20180333208A1 (en) * | 2017-05-17 | 2018-11-22 | General Electric Company | Guidance system for needle procedures |
CN109481018A (en) * | 2018-12-29 | 2019-03-19 | 上海联影医疗科技有限公司 | A kind of navigation equipment and method applied in medical care precess |
WO2019109211A1 (en) * | 2017-12-04 | 2019-06-13 | Covidien Lp | Automatic ablation antenna segmentation from ct image |
JP2019122767A (en) * | 2018-01-15 | 2019-07-25 | キヤノンメディカルシステムズ株式会社 | Medical information processing apparatus, x-ray diagnostic system and medical information processing program |
US10376179B2 (en) | 2011-04-21 | 2019-08-13 | Koninklijke Philips N.V. | MPR slice selection for visualization of catheter in three-dimensional ultrasound |
CN110680481A (en) * | 2019-11-15 | 2020-01-14 | 元亨同基医疗器械(北京)有限公司 | Method for adjusting launching position of guide holder of puncture positioning instrument |
WO2020105049A1 (en) * | 2018-11-22 | 2020-05-28 | Vuze Medical Ltd. | Apparatus and methods for use with image-guided skeletal procedures |
US10716631B2 (en) | 2016-03-13 | 2020-07-21 | Vuze Medical Ltd. | Apparatus and methods for use with skeletal procedures |
DE102019108753A1 (en) * | 2019-04-03 | 2020-10-08 | Amedo Smart Tracking Solutions Gmbh | 3D laser-based positioning system |
WO2020201956A1 (en) * | 2019-04-05 | 2020-10-08 | Medos International Sarl | Systems, devices, and methods for providing surgical trajectory guidance |
US20200315736A1 (en) * | 2019-04-05 | 2020-10-08 | Siemens Healthcare Gmbh | Guiding a needle using a radiopaque contour |
US11006906B2 (en) * | 2017-02-28 | 2021-05-18 | Medicaroid Corporation | Robotic operating table, and robotic operating table operation device |
CN113164134A (en) * | 2018-10-31 | 2021-07-23 | 皇家飞利浦有限公司 | X-ray imaging guidance |
CN113367779A (en) * | 2021-06-16 | 2021-09-10 | 张涛 | Puncture system and method based on C-arm CT and semiconductor laser |
US11224483B2 (en) | 2017-07-08 | 2022-01-18 | Vuze Medical Ltd. | Apparatus and methods for use with image-guided skeletal procedures |
CN114271838A (en) * | 2021-12-01 | 2022-04-05 | 赛诺威盛科技(北京)股份有限公司 | Multifunctional scanning method, multifunctional scanning support, electronic equipment and storage medium |
CN116096322A (en) * | 2020-09-04 | 2023-05-09 | 上海联影医疗科技股份有限公司 | Systems and methods for assisting in placement of a surgical instrument into a subject |
US20230270501A1 (en) * | 2022-02-25 | 2023-08-31 | Siemens Healthcare Gmbh | Trajectory planning for a medical robot system |
Families Citing this family (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2429400B1 (en) * | 2009-05-14 | 2014-01-01 | University Health Network | Quantitative endoscopy |
JP5317933B2 (en) * | 2009-11-17 | 2013-10-16 | 富士フイルム株式会社 | Image display device and program thereof |
WO2011073824A1 (en) * | 2009-12-18 | 2011-06-23 | Koninklijke Philips Electronics N.V. | Multi-section alignment of imaging data |
JP4956635B2 (en) * | 2010-02-24 | 2012-06-20 | 財団法人仙台市医療センター | Percutaneous puncture support system |
DE102010012621A1 (en) * | 2010-03-24 | 2011-09-29 | Siemens Aktiengesellschaft | Method and device for automatically adapting a reference image |
US9282944B2 (en) * | 2010-06-22 | 2016-03-15 | Queen's University At Kingston | C-arm pose estimation using intensity-based registration of imaging modalities |
US20120065498A1 (en) * | 2010-09-14 | 2012-03-15 | Thomas Redel | Apparatus and method for minimally invasive therapy of mitral regurgitation |
US8848874B2 (en) | 2010-09-20 | 2014-09-30 | Siemens Medical Solutions Usa, Inc. | System for recovering from collision of components of an X-ray imaging unit |
DE102011075435A1 (en) * | 2011-05-06 | 2012-11-08 | Siemens Aktiengesellschaft | display unit |
US8900131B2 (en) * | 2011-05-13 | 2014-12-02 | Intuitive Surgical Operations, Inc. | Medical system providing dynamic registration of a model of an anatomical structure for image-guided surgery |
US9510771B1 (en) | 2011-10-28 | 2016-12-06 | Nuvasive, Inc. | Systems and methods for performing spine surgery |
DE102011086941B4 (en) * | 2011-11-23 | 2016-01-21 | Kuka Roboter Gmbh | industrial robots |
US20130218003A1 (en) * | 2012-02-21 | 2013-08-22 | Siemens Aktiengesellschaft | Rapid entry point localization for percutaneous interventions |
DE102012208037A1 (en) * | 2012-05-14 | 2013-11-14 | Siemens Aktiengesellschaft | Patient storage device, a medical imaging device with the patient support device and a method for marking a maximum resident area |
US20130345554A1 (en) * | 2012-06-08 | 2013-12-26 | Baylis Medical Company Inc. | Methods for recanalization of vessels |
BR112014032112A2 (en) * | 2012-06-28 | 2017-06-27 | Koninklijke Philips Nv | image acquisition system; and method for multimodal image acquisition |
CN104902821B (en) | 2012-10-05 | 2018-11-09 | 皇家飞利浦有限公司 | ROI is drawn |
JP6222835B2 (en) * | 2013-02-14 | 2017-11-01 | 東芝メディカルシステムズ株式会社 | X-ray diagnostic equipment |
US20140270067A1 (en) * | 2013-03-14 | 2014-09-18 | Stephen Knecht Clark | Radiographic marker |
DE102013208793B4 (en) * | 2013-05-14 | 2015-04-23 | Siemens Aktiengesellschaft | A method for generating a 3D image data set from a volume to be examined as a basis for display image data |
US9848922B2 (en) | 2013-10-09 | 2017-12-26 | Nuvasive, Inc. | Systems and methods for performing spine surgery |
DE102013226975A1 (en) * | 2013-12-20 | 2015-06-25 | Siemens Aktiengesellschaft | Method for image monitoring during interventional intervention, X-ray device, computer program and data carrier |
AU2015296645A1 (en) | 2014-07-28 | 2017-02-16 | Econolite Group, Inc. | Self-configuring traffic signal controller |
KR20160033546A (en) * | 2014-09-18 | 2016-03-28 | 삼성전자주식회사 | X-ray apparatus and method for scanning thereof |
US10722200B2 (en) * | 2015-06-04 | 2020-07-28 | Siemens Healthcare Gmbh | Apparatus and methods for a projection display device on X-ray imaging devices |
DE102015215820A1 (en) * | 2015-08-19 | 2017-02-23 | Siemens Healthcare Gmbh | Controlling a medical device |
KR101798939B1 (en) * | 2015-09-08 | 2017-11-17 | 삼성전자주식회사 | X-ray image apparatus and control method for the same |
RU2746151C2 (en) * | 2016-03-16 | 2021-04-07 | Конинклейке Филипс Н.В. | System for supporting the performance of interventional procedures |
US11576746B2 (en) * | 2016-09-20 | 2023-02-14 | Kornerstone Devices Pvt. Ltd. | Light and shadow guided needle positioning system and method |
EP3518807A4 (en) * | 2016-09-30 | 2020-05-27 | Intuitive Surgical Operations Inc. | Systems and methods for entry point localization |
DE102017207908B4 (en) * | 2017-05-10 | 2023-06-15 | Siemens Healthcare Gmbh | X-ray imaging system and method for taking X-ray images |
AU2018265160B2 (en) | 2017-05-10 | 2023-11-16 | Mako Surgical Corp. | Robotic spine surgery system and methods |
US11627924B2 (en) | 2019-09-24 | 2023-04-18 | Covidien Lp | Systems and methods for image-guided navigation of percutaneously-inserted devices |
DE102019135782A1 (en) * | 2019-12-23 | 2021-06-24 | Otto-Von-Guericke-Universität Magdeburg | Computed tomography system |
US11490872B2 (en) | 2020-08-21 | 2022-11-08 | GE Precision Healthcare LLC | C-arm imaging system and method |
US12004284B2 (en) | 2021-10-21 | 2024-06-04 | GE Precision Healthcare LLC | Methods and systems for x-ray imaging |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5782842A (en) * | 1995-01-16 | 1998-07-21 | Daum Gmbh | Medical instrument guidance apparatus and method |
US6195577B1 (en) * | 1998-10-08 | 2001-02-27 | Regents Of The University Of Minnesota | Method and apparatus for positioning a device in a body |
US6811313B2 (en) * | 2001-09-25 | 2004-11-02 | Siemens Aktiengesellschaft | C-arm X-ray system with adjustable detector positioning |
US20060039537A1 (en) * | 2004-05-28 | 2006-02-23 | Strobel Norbert K | C-arm device with adjustable detector offset for cone beam imaging involving partial circle scan trajectories |
US20060120507A1 (en) * | 2004-11-26 | 2006-06-08 | Thomas Brunner | Angiographic x-ray diagnostic device for rotation angiography |
US20070255292A1 (en) * | 2006-04-28 | 2007-11-01 | Siemens Aktiengesellschaft | X-ray C-arm system, frame for a flat-panel x-ray detector of an x-ray C-arm system and method for technical support of targeting during puncturing of a patient |
US20080089467A1 (en) * | 2006-08-10 | 2008-04-17 | Siemens Aktiengesellschaft | Method for recording x-ray images of an area lying outside a center of rotation of a C-arm system and the associated C-arm system |
US20080194945A1 (en) * | 2007-02-13 | 2008-08-14 | Siemens Medical Solutions Usa, Inc. | Apparatus and Method for Aligning a Light Pointer With a Medical Interventional Device Trajectory |
US20080200876A1 (en) * | 2007-02-20 | 2008-08-21 | Siemens Medical Solutions Usa, Inc. | Needle Guidance With a Dual-Headed Laser |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6064904A (en) * | 1997-11-28 | 2000-05-16 | Picker International, Inc. | Frameless stereotactic CT scanner with virtual needle display for planning image guided interventional procedures |
US6470207B1 (en) * | 1999-03-23 | 2002-10-22 | Surgical Navigation Technologies, Inc. | Navigational guidance via computer-assisted fluoroscopic imaging |
DE102005023165A1 (en) * | 2005-05-19 | 2006-11-23 | Siemens Ag | Medical imaging system with a patient traversable part and collision protection method |
DE102005030609A1 (en) * | 2005-06-30 | 2007-01-04 | Siemens Ag | Method or X-ray device for creating a series recording of medical X-ray images of a possibly moving patient during the series recording |
DE102005041602A1 (en) * | 2005-09-01 | 2007-04-05 | Siemens Ag | Method for displaying a medical implant in an image and medical imaging system |
-
2009
- 2009-04-24 US US12/429,546 patent/US20090281452A1/en not_active Abandoned
- 2009-04-28 US US12/431,518 patent/US8165660B2/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5782842A (en) * | 1995-01-16 | 1998-07-21 | Daum Gmbh | Medical instrument guidance apparatus and method |
US6195577B1 (en) * | 1998-10-08 | 2001-02-27 | Regents Of The University Of Minnesota | Method and apparatus for positioning a device in a body |
US6811313B2 (en) * | 2001-09-25 | 2004-11-02 | Siemens Aktiengesellschaft | C-arm X-ray system with adjustable detector positioning |
US20060039537A1 (en) * | 2004-05-28 | 2006-02-23 | Strobel Norbert K | C-arm device with adjustable detector offset for cone beam imaging involving partial circle scan trajectories |
US20060120507A1 (en) * | 2004-11-26 | 2006-06-08 | Thomas Brunner | Angiographic x-ray diagnostic device for rotation angiography |
US20070255292A1 (en) * | 2006-04-28 | 2007-11-01 | Siemens Aktiengesellschaft | X-ray C-arm system, frame for a flat-panel x-ray detector of an x-ray C-arm system and method for technical support of targeting during puncturing of a patient |
US20080089467A1 (en) * | 2006-08-10 | 2008-04-17 | Siemens Aktiengesellschaft | Method for recording x-ray images of an area lying outside a center of rotation of a C-arm system and the associated C-arm system |
US20080194945A1 (en) * | 2007-02-13 | 2008-08-14 | Siemens Medical Solutions Usa, Inc. | Apparatus and Method for Aligning a Light Pointer With a Medical Interventional Device Trajectory |
US20080200876A1 (en) * | 2007-02-20 | 2008-08-21 | Siemens Medical Solutions Usa, Inc. | Needle Guidance With a Dual-Headed Laser |
Cited By (101)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102342861A (en) * | 2010-07-22 | 2012-02-08 | 西门子公司 | Method and orientation device for marking a predetermined guide path of a medical instrument |
US20120022508A1 (en) * | 2010-07-22 | 2012-01-26 | Patrick Gross | Method and orientation device for marking a predetermined guide path of a medical instrument |
US20120108952A1 (en) * | 2010-10-29 | 2012-05-03 | Eli Stern | Methods and systems for patient alignment for nuclear medicine imaging |
US9078618B2 (en) * | 2010-10-29 | 2015-07-14 | General Electric Company | Methods and systems for patient alignment for nuclear medicine imaging |
US20120209123A1 (en) * | 2011-02-10 | 2012-08-16 | Timothy King | Surgeon's Aid for Medical Display |
US10631712B2 (en) * | 2011-02-10 | 2020-04-28 | Karl Storz Imaging, Inc. | Surgeon's aid for medical display |
US10376179B2 (en) | 2011-04-21 | 2019-08-13 | Koninklijke Philips N.V. | MPR slice selection for visualization of catheter in three-dimensional ultrasound |
US8781068B2 (en) * | 2011-04-22 | 2014-07-15 | Kabushiki Kaisha Toshiba | X-ray image diagnosing apparatus |
US20120281812A1 (en) * | 2011-04-22 | 2012-11-08 | Toshiba Medical Systems Corporation | X-ray image diagnosing apparatus |
EP2526869A3 (en) * | 2011-05-23 | 2015-10-07 | Fujifilm Corporation | Image processing device, radiographic image capture system, image processing method and image processing program storage medium |
US20120300910A1 (en) * | 2011-05-26 | 2012-11-29 | Sebastien Gorges | X-ray imaging apparatus having a variable distance between an x-ray source and an object to be imaged |
WO2012168749A1 (en) * | 2011-06-06 | 2012-12-13 | Sarr Souleymane | Removable guide device for radiofluoroscopic infiltration having an image intensifier |
US20130116550A1 (en) * | 2011-07-06 | 2013-05-09 | Hideaki Ishii | Medical diagnostic imaging apparatus |
US9445773B2 (en) * | 2011-07-06 | 2016-09-20 | Toshiba Medical Systems Corporation | Medical diagnostic imaging apparatus |
CN103027699A (en) * | 2011-09-30 | 2013-04-10 | 西门子公司 | Method for controlling the movement of an x-ray apparatus and x-ray system |
US9642584B2 (en) * | 2011-09-30 | 2017-05-09 | Siemens Aktiengesellschaft | Method for controlling the movement of an x-ray apparatus and x-ray system |
US20130083894A1 (en) * | 2011-09-30 | 2013-04-04 | Siemens Aktiengesellschaft | Method for controlling the movement of an x-ray apparatus and x-ray system |
US20130089180A1 (en) * | 2011-10-11 | 2013-04-11 | Siemens Aktiengesellschaft | Method and apparatus for producing an x-ray projection image in a desired direction |
US9039283B2 (en) * | 2011-10-11 | 2015-05-26 | Siemens Aktiengesellschaft | Method and apparatus for producing an X-ray projection image in a desired direction |
US20130136332A1 (en) * | 2011-11-29 | 2013-05-30 | Hisayuki Uehara | X-ray image diagnosis apparatus |
US9384545B2 (en) * | 2011-11-29 | 2016-07-05 | Kabushiki Kaisha Toshiba | X-ray image diagnosis apparatus |
JP2013172951A (en) * | 2012-01-27 | 2013-09-05 | Toshiba Corp | X-ray ct apparatus and x-ray ct system |
US10485617B2 (en) * | 2012-06-21 | 2019-11-26 | Globus Medical, Inc. | Surgical robot platform |
US20170245951A1 (en) * | 2012-06-21 | 2017-08-31 | Globus Medical, Inc. | Surgical robot platform |
US9916691B2 (en) | 2013-02-14 | 2018-03-13 | Seiko Epson Corporation | Head mounted display and control method for head mounted display |
WO2014125789A1 (en) * | 2013-02-14 | 2014-08-21 | Seiko Epson Corporation | Head mounted display and control method for head mounted display |
US10169925B2 (en) | 2013-02-14 | 2019-01-01 | Seiko Epson Corporation | Head mounted display and control method for head mounted display |
FR3007962A1 (en) * | 2013-07-04 | 2015-01-09 | X Nov Ip | GRAPHICAL SELECTION OF BONE ANCHOR PROSTHESIS |
US9877795B2 (en) * | 2013-09-18 | 2018-01-30 | Imirge Medical Inc | Optical targeting and visualization of trajectories |
WO2015039246A1 (en) * | 2013-09-18 | 2015-03-26 | iMIRGE Medical INC. | Optical targeting and visualization of trajectories |
US20150080740A1 (en) * | 2013-09-18 | 2015-03-19 | iMIRGE Medical INC. | Optical targeting and visualization of trajectories |
CN106413621A (en) * | 2013-09-18 | 2017-02-15 | 伊美格医药公司 | Surgical assemblies for housing force transmitting members |
US10433911B2 (en) * | 2013-09-18 | 2019-10-08 | iMIRGE Medical INC. | Optical targeting and visualization of trajectories |
US10245110B2 (en) * | 2014-03-04 | 2019-04-02 | Xact Robotics Ltd. | Dynamic planning method for needle insertion |
US20170071672A1 (en) * | 2014-03-04 | 2017-03-16 | Xact Robotics Ltd. | Dynamic planning method for needle insertion |
US10130388B2 (en) | 2014-03-13 | 2018-11-20 | New York University | Position guidance device with bubble level |
US20170156800A1 (en) * | 2014-03-17 | 2017-06-08 | Roy Anthony Brown | Surgical Targeting Systems and Methods |
US20170296273A9 (en) * | 2014-03-17 | 2017-10-19 | Roy Anthony Brown | Surgical Targeting Systems and Methods |
US20170290559A1 (en) * | 2014-09-25 | 2017-10-12 | Siemens Healthcare Gmbh | Mobile x-ray machine |
US10653386B2 (en) * | 2014-09-25 | 2020-05-19 | Siemens Healthcare Gmbh | Mobile x-ray machine |
US20160147939A1 (en) * | 2014-11-26 | 2016-05-26 | Koninklijke Philips N.V. | Efficient management of visible light still images and/or video |
US10162935B2 (en) * | 2014-11-26 | 2018-12-25 | Koninklijke Philips N.V. | Efficient management of visible light still images and/or video |
EP3028675A1 (en) * | 2014-12-05 | 2016-06-08 | X.Nov IP Sarl | Graphical selection of a prosthesis with bone anchor |
WO2016119053A1 (en) * | 2015-01-27 | 2016-08-04 | iMIRGE Medical INC. | Optical targeting and visualization of trajectories |
US20160345923A1 (en) * | 2015-06-01 | 2016-12-01 | Toshiba Medical Systems Corporation | Medical image processing apparatus and x-ray diagnostic apparatus |
US9936928B2 (en) * | 2015-06-01 | 2018-04-10 | Toshiba Medical Systems Corporation | Medical image processing apparatus and X-ray diagnostic apparatus |
US20180199901A1 (en) * | 2015-07-16 | 2018-07-19 | Koninklijke Philips N.V. | Device for remote fluoroscopy, nearby fluoroscopy and radiology |
US10993685B2 (en) * | 2015-07-16 | 2021-05-04 | Koninklijke Philips N.V. | Device for remote fluoroscopy, nearby fluoroscopy and radiology |
WO2017055990A1 (en) * | 2015-09-28 | 2017-04-06 | Koninklijke Philips N.V. | Optical registration of a remote center of motion robot |
WO2017070124A1 (en) * | 2015-10-19 | 2017-04-27 | New York University | Electronic position guidance device with real-time auditory and visual feedback |
AU2017206019B2 (en) * | 2016-01-06 | 2021-08-12 | Boston Scientific Scimed, Inc. | Systems and methods for planning medical procedures |
US11830604B2 (en) | 2016-01-06 | 2023-11-28 | Boston Scientific Scimed, Inc. | Systems and methods for planning medical procedures |
US11302435B2 (en) | 2016-01-06 | 2022-04-12 | Boston Scientific Scimed, Inc. | Systems and methods for planning medical procedures |
US20170193160A1 (en) * | 2016-01-06 | 2017-07-06 | Boston Scientific Scimed, Inc. | Systems and methods for planning medical procedures |
WO2017120369A1 (en) * | 2016-01-06 | 2017-07-13 | Boston Scientific Scimed, Inc. | Systems and methods for planning medical procedures |
EP3795108A1 (en) * | 2016-01-06 | 2021-03-24 | Boston Scientific Scimed, Inc. | Systems and methods for planning medical procedures |
US11490967B2 (en) | 2016-03-13 | 2022-11-08 | Vuze Medical Ltd. | Apparatus and methods for use with skeletal procedures |
US10716631B2 (en) | 2016-03-13 | 2020-07-21 | Vuze Medical Ltd. | Apparatus and methods for use with skeletal procedures |
US11452570B2 (en) | 2016-03-13 | 2022-09-27 | Vuze Medical Ltd. | Apparatus and methods for use with skeletal procedures |
US11911118B2 (en) | 2016-03-13 | 2024-02-27 | Vuze Medical Ltd. | Apparatus and methods for use with skeletal procedures |
US11723616B2 (en) * | 2016-05-19 | 2023-08-15 | Cart-Tech B.V | Method of verifying a position of an interventional device |
NL2016800A (en) * | 2016-05-19 | 2017-11-23 | Umc Utrecht Holding Bv | Method of positioning an interventional device. |
US20190125289A1 (en) * | 2016-05-19 | 2019-05-02 | Umc Utrecht Holding B.V. | Method of verifying a position of an interventional device |
WO2017200385A1 (en) * | 2016-05-19 | 2017-11-23 | Umc Utrecht Holding B.V. | Method of verifying a position of an interventional device |
EP3254627A1 (en) * | 2016-06-08 | 2017-12-13 | General Electric Company | Fluoroscopic guidance system with offset light source and method of use |
US20170354387A1 (en) * | 2016-06-08 | 2017-12-14 | General Electric Company | Fluoroscopic Guidance System With Offset Light Source and Method Of Use |
CN107468336A (en) * | 2016-06-08 | 2017-12-15 | 通用电气公司 | With the fluoroscopic guidance system and its application method for deviateing light source |
JP2017217448A (en) * | 2016-06-08 | 2017-12-14 | ゼネラル・エレクトリック・カンパニイ | Fluoroscopic guidance system with offset light source and method of use |
US20180008217A1 (en) * | 2016-07-08 | 2018-01-11 | Alexander Gemmel | Motion controllers for mobile x-ray devices |
US10751010B2 (en) * | 2016-07-08 | 2020-08-25 | Siemens Healthcare Gmbh | Motion controllers for mobile X-ray devices |
US20180035984A1 (en) * | 2016-08-03 | 2018-02-08 | Siemens Healthcare Gmbh | Biopsy unit |
US11815347B2 (en) * | 2016-09-28 | 2023-11-14 | Kla-Tencor Corporation | Optical near-field metrology |
US20180087900A1 (en) * | 2016-09-28 | 2018-03-29 | Kla-Tencor Corporation | Optical Near-Field Metrology |
US20180214007A1 (en) * | 2017-01-30 | 2018-08-02 | Seiko Epson Corporation | Endoscope operation support system |
US10842349B2 (en) * | 2017-01-30 | 2020-11-24 | Seiko Epson Corporation | Endoscope operation support system |
US11006906B2 (en) * | 2017-02-28 | 2021-05-18 | Medicaroid Corporation | Robotic operating table, and robotic operating table operation device |
US10667869B2 (en) * | 2017-05-17 | 2020-06-02 | General Electric Company | Guidance system for needle procedures |
US20180333208A1 (en) * | 2017-05-17 | 2018-11-22 | General Electric Company | Guidance system for needle procedures |
US11806183B2 (en) | 2017-07-08 | 2023-11-07 | Vuze Medical Ltd. | Apparatus and methods for use with image-guided skeletal procedures |
US11406338B2 (en) | 2017-07-08 | 2022-08-09 | Vuze Medical Ltd. | Apparatus and methods for use with image-guided skeletal procedures |
US11224483B2 (en) | 2017-07-08 | 2022-01-18 | Vuze Medical Ltd. | Apparatus and methods for use with image-guided skeletal procedures |
WO2019109211A1 (en) * | 2017-12-04 | 2019-06-13 | Covidien Lp | Automatic ablation antenna segmentation from ct image |
US11638612B2 (en) | 2017-12-04 | 2023-05-02 | Covidien Lp | Automatic ablation antenna segmentation from CT image |
JP7271183B2 (en) | 2018-01-15 | 2023-05-11 | キヤノンメディカルシステムズ株式会社 | Medical information processing device, X-ray diagnostic system and medical information processing program |
JP2019122767A (en) * | 2018-01-15 | 2019-07-25 | キヤノンメディカルシステムズ株式会社 | Medical information processing apparatus, x-ray diagnostic system and medical information processing program |
CN113164134A (en) * | 2018-10-31 | 2021-07-23 | 皇家飞利浦有限公司 | X-ray imaging guidance |
WO2020105049A1 (en) * | 2018-11-22 | 2020-05-28 | Vuze Medical Ltd. | Apparatus and methods for use with image-guided skeletal procedures |
EP3886749A4 (en) * | 2018-12-29 | 2022-01-26 | Shanghai United Imaging Healthcare Co., Ltd. | System and method for placing surgical instrument in subject |
CN109481018A (en) * | 2018-12-29 | 2019-03-19 | 上海联影医疗科技有限公司 | A kind of navigation equipment and method applied in medical care precess |
WO2020135769A1 (en) | 2018-12-29 | 2020-07-02 | Shanghai United Imaging Healthcare Co., Ltd. | System and method for placing surgical instrument in subject |
DE102019108753A1 (en) * | 2019-04-03 | 2020-10-08 | Amedo Smart Tracking Solutions Gmbh | 3D laser-based positioning system |
DE102019204920A1 (en) * | 2019-04-05 | 2020-10-08 | Siemens Healthcare Gmbh | Guide unit for guiding a needle with an X-ray-opaque contour; Management system; Investment; Computer program product and method |
WO2020201956A1 (en) * | 2019-04-05 | 2020-10-08 | Medos International Sarl | Systems, devices, and methods for providing surgical trajectory guidance |
US20200315736A1 (en) * | 2019-04-05 | 2020-10-08 | Siemens Healthcare Gmbh | Guiding a needle using a radiopaque contour |
US11813026B2 (en) | 2019-04-05 | 2023-11-14 | Medos International Sarl | Systems, devices, and methods for providing surgical trajectory guidance |
CN110680481A (en) * | 2019-11-15 | 2020-01-14 | 元亨同基医疗器械(北京)有限公司 | Method for adjusting launching position of guide holder of puncture positioning instrument |
CN116096322A (en) * | 2020-09-04 | 2023-05-09 | 上海联影医疗科技股份有限公司 | Systems and methods for assisting in placement of a surgical instrument into a subject |
CN113367779A (en) * | 2021-06-16 | 2021-09-10 | 张涛 | Puncture system and method based on C-arm CT and semiconductor laser |
CN114271838A (en) * | 2021-12-01 | 2022-04-05 | 赛诺威盛科技(北京)股份有限公司 | Multifunctional scanning method, multifunctional scanning support, electronic equipment and storage medium |
US20230270501A1 (en) * | 2022-02-25 | 2023-08-31 | Siemens Healthcare Gmbh | Trajectory planning for a medical robot system |
US11877805B2 (en) * | 2022-02-25 | 2024-01-23 | Siemens Healthcare Gmbh | Trajectory planning for a medical robot system |
Also Published As
Publication number | Publication date |
---|---|
US8165660B2 (en) | 2012-04-24 |
US20090281452A1 (en) | 2009-11-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8165660B2 (en) | System and method for selecting a guidance mode for performing a percutaneous procedure | |
US11989338B2 (en) | Using optical codes with augmented reality displays | |
US20100111389A1 (en) | System and method for planning and guiding percutaneous procedures | |
US11992349B2 (en) | System and method for local three dimensional volume reconstruction using a standard fluoroscope | |
US20220313190A1 (en) | System and method for pose estimation of an imaging device and for determining the location of a medical device with respect to a target | |
US9202387B2 (en) | Methods for planning and performing percutaneous needle procedures | |
CN110741414B (en) | Systems and methods for identifying, marking, and navigating to a target using real-time two-dimensional fluoroscopic data | |
EP2004071B1 (en) | Targeting device, computer readable medium and program element | |
US20100292565A1 (en) | Medical imaging medical device navigation from at least two 2d projections from different angles | |
JP6713563B2 (en) | System and method for performing local three-dimensional volume reconstruction using standard fluoroscopy equipment | |
EP3254627A1 (en) | Fluoroscopic guidance system with offset light source and method of use | |
EP1727471A1 (en) | System for guiding a medical instrument in a patient body | |
US20220409290A1 (en) | Method and system for reproducing an insertion point for a medical instrument | |
US8731643B2 (en) | Imaging system and methods for medical needle procedures | |
US11576557B2 (en) | Method for supporting a user, computer program product, data medium and imaging system | |
WO2019075074A1 (en) | System and method for identifying and marking a target in a fluoroscopic three-dimensional reconstruction | |
EP4163874A1 (en) | Method and device for an improved display in a manual or a robot assisted intervention in a region of interest of the patient | |
Chen et al. | Accuracy and efficiency of an infrared based positioning and tracking system for image-guided intervention |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC., PENNSYLVANIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PFISTER, MARCUS;STROBEL, NORBERT;REEL/FRAME:022938/0107 Effective date: 20090629 |
|
AS | Assignment |
Owner name: SIEMENS AKTIENGESELLSCHAFT, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PFISTER, MARCUS;STROBEL, NORBERT;REEL/FRAME:027432/0102 Effective date: 20111213 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: SIEMENS HEALTHCARE GMBH, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS AKTIENGESELLSCHAFT;REEL/FRAME:039011/0400 Effective date: 20160610 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |
|
AS | Assignment |
Owner name: SIEMENS HEALTHINEERS AG, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS HEALTHCARE GMBH;REEL/FRAME:066088/0256 Effective date: 20231219 |