THREE-DIMENSIONAL ULTRASOUND IMAGING SYSTEM Field of the Invention
This invention relates in general to a three- dimensional imaging system for scanning animate and inanimate objects and creating three dimensional (3-D) images thereof. In the case of animate objects, the system of the present invention may be used for three dimensional (3-D) ultrasound imaging of human or animal organs such as the eye, prostate, and other organs suitable for ultrasound imaging. Background of the Invention
Clinical ultrasound diagnostic equipment is well known in the medical arts for viewing internal human organs. For example, opthamologists and radiologists often require images of the eye for investigating eye maladies and for making volumetric measurements. In diagnosing prostate cancer, a diagnostician uses trans- rectal ultrasound (TRUS) to identify whether lesions are present, and determines the location, size and extent of the lesions. However, such prior art ultrasound devices produce only 2-D images whereas the anatomies under investigation are three-dimensional.
Also, the ultrasound image in prior art 2-D images represents a single plane approximately 1 mm thick and at an arbitrary angle in the patient's body. Thus, it is generally difficult to localize the image plane in the organ, and very difficult to reproduce a particular image location at some later time.
Prior art ultrasound imaging systems typically comprise a probe for transmitting ultrasound signals into the human body and receiving reflected ultrasound signals therefrom, and a clinical ultrasound machine for receiving and processing analog ultrasound signals from the probe for generating multiple images of the organ. A number of patents have issued relating to prior art probes with internal mechanical sensors. Examples of these systems are disclosed in the following United
States Patents: 5,159,931 (Pini) ; 5,152,294 (Mochizuki et al) ; 4,819,650 (Goldstein); 4,841,979 (Dow et al) and 4,934,370 (Campbell).
Prior art systems are also known to use encoders for determining the position of the sensors and transmitting that information to a controlling computer. Examples of such systems are disclosed in the following United States Patents: 5,159,931 (Pini); 5,152,294 (Mochizuki et al) ; 4,932,414 (Coleman et al) ; 4,271,706 (Ledley) ; 4,341,120 (Anderson); 5,078,145 (Furuhata) ; 5,036,855 (Fry et al) ; 4,858,613 (Fry et al) and 4,955,365 (Fry et al) .
Other patents have issued which provide general background information on the subject of clinical ultrasound imaging systems. United States Patent 5,081,993 (Kitney et al) , discloses an intravascular probe for insertion into blood vessels. It incorporates an array of crystals surrounding a tube and generates a cross-sectional view. United States Patent 4,747,411 (Ledley) discloses a 3-D imaging system which requires the use of stereo eyeglasses. United States Patent
4,899,318 (Schlumberger et al) and United States Patent 4,028,934 (Sollish) relate to specific methods of stereoscopic 3-D visualization of objects. United States Patent 3,555,888 discloses a probe having a single crystal and a mechanical means for moving the single crystal. United States Patent 4,564,018 (Hutchison et al) discloses an ultrasonic diagnostic scanner for producing peak signals and count signals responsive generation of the peak signals for identifying a perceptible eye parameter. United States Patents 4,594,662; 4,562,540 and 4,598,366 relate to 3-D holography. United States Patent 4,866,614 (Tarn) teaches the use of a plurality of stationary ultrasound beams which are generated on the basis of multiple transducers. PCT application number PCT/EP92/00410 (Technomed
International) discloses a treatment probe inserted into the urethra and means for rotating and moving the probe
up and down on a stand.
United States Patent 4,932,414 (Coleman) is of interest for disclosing an extension of kidney stone acoustic shattering techniques, applied to the eye. In order to see what a surgeon is doing, an ultrasound imaging probe is provided for sweeping out a volume of the eye resulting in a 3-D impression when the volume is swept quickly enough. However, the Colemen Patent does not disclose any means for reconstruction of the 3-D image from the multiplicity of generated 2-D images.
Furthermore, the system of Coleman et al teaches the use of an encoder for determining the position of the probe. Summary of the Invention
According to the present invention, a 3-D ultrasound imaging system is provided which comprises a universal assembly onto which an ultrasound probe may be mounted. The assembly incorporates a motor and drive for either rotating or scanning the probe relative to the organ under investigation. Ultrasound signals from the probe are processed via a clinical ultrasound machine for generating multiple images of the organ. A computer is provided for executing software for controlling movement of the assembly to rotate or scan the probe, and for collecting the 2-D ultrasound images from the clinical ultrasound machine and re-constructing these images to form a 3-D display.
In essence, the present invention may be characterized by a novel combination of ultrasound probe, clinical ultrasound machine, probe assembly and computer for controlling movement of the probe assembly and reconstructing three dimensional images from a plurality of two dimensional images obtained from the clinical ultrasound machine. A unique advantage of the present invention is that it is adaptable to any prior art and manufacture of probe, thereby alleviating the prior art requirement for sophisticated probes with internal moving sensors, etc.
Brief Description of the Drawings
A detailed description of the preferred and alternative embodiments is provided hereinbelow with reference to the following drawings, in which: Figure l illustrates the configuration of a 3-D ultrasound imaging system according to the present invention;
Figure 2 is a perspective view of an assembly for mounting a conventional ultrasound probe for scanning the human eye or other organs, according to a first embodiment of the invention;
Figure 3 is a side view of the assembly shown in Figure 2;
Figure 4 is a detailed perspective view of the assembly shown in Figures 2 and 3;
Figure 5 is a perspective view (partly broken) of an assembly for rotating an end-firing ultrasound probe using axial rotation, according to a second embodiment of the present invention; Figure 6 is a side cross-sectional view of the assembly shown in Figure 5;
Figure 7 is a perspective view (partly broken) of an assembly for moving side-firing ultrasound probe using sector rotation, according to a third embodiment of the present invention;
Figure 8A is a perspective view (partly broken) of a further assembly for moving an ultrasound probe via lateral translation, according to a fourth embodiment of the invention; Figure 8B is a side cross-sectional view of the assembly shown in Figure 8A;
Figure 9 is a further embodiment of ultrasound probe assembly for rotating the probe via axial rotation;
Figures 10A-10D illustrate various geometries of probe sweeping movements for imaging the eye, according the various embodiments of ultrasound probe assemblies;
Figure 11 is block diagram showing functional
interconnection of various software modules of the present invention;
Figure 12 is a flow chart showing operation of the 3-D ultrasound imaging system of the present invention; Figure 13 is a flow chart showing details of the reconstruction and viewing steps of operating the 3-D imaging system of the present invention;
Figure 14 is a flow chart showing details of 3-D image reconstruction according to the preferred embodiment; Figure 15 is a flow chart showing details of the pixel mapping steps of the reconstruction process illustrated in Figure 14;
Figures 16A and 16B are reconstruction diagrams showing transformation of an input image captured by axial rotation and sector rotation, respectively;
Figure 17 is a reconstruction diagram schematically representing a temporary raster;
Figures 18A and 18B are reconstruction diagrams showing the computation of the contents of the temporary raster for an image captured by axial rotation;
Figures 19A and 19B are reconstruction diagrams showing computation of the contents of the temporary raster for sector rotation;
Figures 20A and 20B are reconstruction diagrams showing the creation of a list of contributing pixels in the temporary raster for an image captured by axial rotation;
Figures 21A and 2IB are reconstruction diagrams showing the creation of a list of contributing pixels in the temporary raster for an image captured by sector rotation;
Figure 22 is a reconstruction diagram showing extraction of a slice of a resultant image; and
Figure 23 is a reconstruction diagram showing the storage of the contents of the temporary raster in an array of output pixels forming a reconstructed 3-D image.
Detailed Description of the Preferred Embodiment
With reference to Figure 1, the overall configuration is illustrated of the 3-D ultrasound imaging system of the present invention. A conventional ultrasound probe 1 is rotated by means of a probe holder 3 which forms part of an assembly 5. The assembly 5 also includes a motor and output shaft for moving the holder 3 and associated probe 1 through a predetermined angular sweep for generating a plurality of images of an organ (not shown) under examination. It is contemplated that the system of the present invention may be used to provide 3-D ultrasound images of the eye, prostate, female breast, heart, arteries and veins, kidney, liver, and other organs suitable for ultrasound imaging. The signals from probe 1 are transmitted via a communication line 7 for processing by a conventional clinical ultrasound machine 9. The processed multiple images are transmitted via a communication line 10 from ultrasound machine 9 to a computer 11 which typically incorporates a video-based digitizer (not shown) . The series of 2-D images from ultrasound machine 9 are then reconstructed in computer 11 into a single 3-D image for interactive manipulation and display, or into a sequence of 3-D images. The computer 11 also generates control signals to a motor driver 13 via line 12, which in response generates further control signals along line 14 for controlling operation of the assembly 5 to sweep the probe l.
As discussed above, according to one embodiment of the invention, images of the human eye are scanned and reconstructed in the computer 11. With reference to Figure 2, an assembly 5 is shown comprising a head support platform 15 for supporting the head of a patient P. A base plate 17 is connected to the supporting platform 15. The horizontal and angular position of the plate 17 may be adjusted by means of the adjustment knob 19 which is threaded to a bolt (not shown) passing
through an adjustment slot 21 in the base plate 17. A vertical adjustment member 23 houses a threaded rod 25 to which is connected an adjustment knob 27 at a top end of the vertical adjustment member 23 (see Figures 3 and 4) . A bracket 29 is connected to vertical adjustment member 23 and has an internal threaded hole through which the adjustment bolt 25 passes, such that the height of the bracket 29 may be adjusted by rotating the knob 27.
A motor 31 is mounted in a motor box 33 which, in turn, is connected to the bracket 29. The motor 31 typically comprises a stepper motor along with gear reduction wheels 35 and 37, and an output shaft 39. An offset arm 41 is connected to the output shaft 39 at one end, while the opposite end of the offset arm 41 is connected to probe holder 3.
In operation, the pivoting offset arm 41 is positioned so that the probe 1 is placed firmly via a coupling gel (see Figure 10) adjacent the eye of the patient P. Once a desired scan region has been chosen, the motor 31 is actuated thereby causing the arm 41 to sweep across a scanning region while the probe 1 scans the eye to obtain a sequence of images. As discussed in greater detail below, the sequence of images is digitized rapidly and stored in computer 11 (Figure 1) . Preferably, approximately 100 images are collected in a single sweep of approximately 10 second duration, covering a probe scanning angle of approximately 30° (see Figure 3). The collected images are then reconstructed to form a 3-D volumetric representation of the eye which can then be viewed and manipulated in "real time", either on the monitor of computer 11 or a remote monitor, for diagnostic purposes. As discussed in greater detail below with reference Figures 10A and 10B, the embodiment of Figures 2-4 sweeps out a 3-D volume of the eye which is in the shape of a sector of a cylinder.
Turning to Figures 5 and 6, an alternative embodiment of assembly is shown for axially rotating an
end-firing ultrasound probe 51. Thus, the image volume swept out by the probe 51 in the embodiment of Figure 5 is generally conical, rather than being in the shape of a sector as in the case of the embodiment in Figures 2-4. The assembly 53 comprises an external housing which contains a motor 57 with an output shaft 59 extending from one end. The probe 51 is held securely by a barrel 60 supported by a pair of ball bearing mounts 61 and 63. Gear reduction wheels 65 and 67 are connected to the output shaft 59 and barrel 60, respectively, and a belt 69 translates rotary motion of the output shaft 59 to a similar (but gear reduced) rotation of the barrel 60 and consequently also of the probe 51 in a reciprocating circular motion in either direction as illustrated by the arrow A.
The axially rotating embodiment of Figures 5 and 6 may be useful for scanning internal organs such as the prostate.
With reference to the embodiment of Figure 7, an assembly is shown for rotating a side-firing ultrasound probe 70. Thus, the image volume is swept out by the probe 70 in the embodiment of Figure 7 is generally in the shape of a sector, rather than being conical as in the embodiment of Figures 5 and 6. The assembly 71 comprises a probe rotator 72 and motor 73 having an output shaft 74 extending from one end. The probe 70 is held securely by a barrel 75 supported by a pair of ball bearing mounts 76 and 77. Gear reduction wheels 78 and 79 are connected to the output shaft 74 and barrel 75, respectively, and a belt 80 translate rotary motion of the output shaft 74 to a similar (but gear reduced) rotation of the barrel 75 and consequently also of the probe 70 in a circular motion both clockwise and counterclockwise. The embodiment of Figure 7 is also useful for scanning internal organs such as the prostate.
Turning now to the alternative embodiment of Figures
8A and 8B, an assembly 80 is shown comprising a motor 81 within a housing 82. The motor 81 includes gear reduction wheels, belts, etc., which cause a threaded output shaft 83 thereof, to rotate. An I-block 84 is mounted to the output shaft 83 via an internal threaded hole so that the I-block 84 moves according to a linear motion in either direction, as represented by the arrow in Figure 8A. The holder 3 for probe 1 is connected to I-block 84 via screws 85 and 86. The angle of inclination of the probe 1 relative to the direction of movement, is adjustable via the screw 86.
As shown in Figure 8B, the assembly of Figures 8A and 8B may be used to collect images of an internal organ or lesions such as breast tumors within the trunk of the patient P, there being shown a layer of coupling gel 87 intermediate the probe 1 and the patient P.
Turning to yet another alternative embodiment of assembly illustrated in Figure 9, an ultrasound probe 91 is shown for trans-esophageal imaging of the heart. The probe 91 is mounted within an annular gear-toothed wheel 93 which, in turn, engages a further gear reduction wheel 95 connected to a motor 97 via an output shaft 99 thereof. Upon rotating the output shaft under control of the motor 97, gear wheels 95 and 93 rotate, which causes the ultrasound probe 91 to rotate in consequence.
Although five embodiments of assembly have been illustrated, for diagnostic investigation of the eye, prostate, heart, breast and other internal organs, it is contemplated that additional assemblies may be designed for clinical evaluation of female breasts, the heart, etc. All such embodiments are believed to be within the sphere and scope of the present invention.
Turning to Figures 10A-10D, various scanning geometries are illustrated. In Figure 10A, the sweeping motion of probe 1 results in a plurality of 2-D images of the eye E conforming to a fan of planes. In Figure 10B, the probe 1 is separated from the eye E by an acoustic
window of gel G such that a fan of two dimensional images is collected wherein the axis of rotation is behind the eye E. The probe rotation mode represented by Figures 10A and 10B is referred to herein as "sector" rotation. In Figures IOC and 10D, the axis of rotation is longitudinal so as to provide 180° rotation of the probe 1. The probe rotation mode represented by Figures IOC and 10D is referred to herein as "axial" rotation.
The method of reconstructing 3-D images for the geometries of Figures 10A-10D is discussed in greater detail below with reference to Figures 14-23. However, reference is made firstly to the block diagram of Figure 11 which shows the overall system functionality of the present invention. A user interface 1101 is provided in the form of a mouse, keyboard or other input to the computer 11 (Figure 1) and software. The software operating within the computer 11 is divided into various functional modules, as follows: ultrasound scanning module 1103, volume reconstruction module 1105 and volume viewing module 1107. In the case of ultrasound scanning 1103, commands 1109 may be sent to control the probe assembly, (e.g. assembly 5 in Figures 2-4, assembly 53 in Figures 5 and 6, or assembly 71 in Figures 7 and 8 via the motor driver 13 (Figure 1)) . A succession of 2-D images from the ultrasound probe 1 are transmitted from clinical ultrasound machine 9 via line 10 as an input 1111 to the ultrasound scanning module 1103. These analog images are digitized via the ultrasound scanning module 1103 and stored in a computer memory 1113. Computer memory 1113 can then store the digitized 2-D images via image data transfer in a file storage device 1115, or, upon receipt of an appropriate user command from interface 1101, the 2-D images may be reconstructed via volume reconstruction software module 1105, as discussed in greater detail below. Once the 3-D image has been reconstructed in module 1105 and stored in computer memory 1113, upon receipt of appropriate user
commands from user interface 1101, the volume viewer module 1107 manipulates the view perspective of the reconstructed 3-D image for output to a monitor (output to monitor module 1117) . According to the preferred embodiment, motor 31
(Figure 4), 57 (Figure 5), 73 (Figure 7), 81 (Figures 8A and 8B) and 97 (Figure 9) are precision motors which rotate their respective shafts with constant angular increment or velocity. Thus, when the 2-D images are acquired, they are equally spaced (i.e. either with constant angular intervals when sector or axial geometries are implemented, or with constant distance in the case of a linear geometry) . In some cases the selected motor may include an internal position encoder to ensure shaft rotation at a constant angular velocity. One example of such a preferred motor is the Model C57-51 stepper motor manufactured by Compumotor (division of Parker Corporation) .
Turning now to Figure 12, upon starting a typical 3-D ultrasound imaging procedure (start step 1201) , the probe 1 is inserted into the scanner assembly (step 1203) , the anatomy to be scanned is located (step 1205) , and an image of the anatomy is placed in an ROI frame (step 1207) . The ROI frame is a frame drawn on the output monitor by a frame grabber program executed by computer 11, where ROI represent an area of interest. Next, the user confirms the anatomy coverage by manual scan control (step 1209) . If the anatomy is outside of the ROI frame, then operation returns to step 1207, otherwise, (step 1211) the operation proceeds to step 1213 wherein the user fixes or holds the probe scanner assembly in place. Next, in step 1215, a volume of image is acquired and the user may review the acquired frames in step 1217. If the data has been acquired incorrectly (step 1219), the process control returns to step 1205.
Otherwise, process control proceeds to step 1221 in which the acquired volume is saved and stored to computer
memory as a stack of 2-D images I (x,y, z) , (see 1113 of Figure 11) . Next, the steps of reconstruction and viewing of the 3-D image are effected (step 1223 is discussed in greater detail below with reference to Figures 14-23) . If the user wishes to perform another 3- D scan (step 1225) , the process control returns to step 1205. Otherwise, the procedure is exited (step 1227) .
The steps for utilizing the reconstruction and viewing algorithm discussed in greater detail below with reference to Figures 14-23, are shown in Figure 13, beginning with a start step 1301. If the user does not wish to reconstruct a 3-D image (step 1303), then the user merely exits the system (step 1313) . Otherwise, the system reconstructs the 3-D image (step 1305) , and saves the reconstructed image to computer memory (step 1307) . The user is then given the opportunity to view the 3-D image (step 1309) . If the user does not elect to view the 3-D image, then process control returns to set 1303. However, if the user wishes to view the captured and reconstructed 3-D image, then the image is viewed and if required image slices may be extracted (step 1311) . After viewing the 3-D image, the procedure is exited (step 1313) .
Turning now to Figures 14-23, the 3-D image reconstruction algorithm of module 1105 (Figure 11) is discussed in greater detail. After initializing the module (step 1401), the stack of 2-D images I (x,y, z) is retrieved from computer memory 1113 (step 1403) . The stack of images I (x,y, z) is transformed into a resultant image R (x,y, z) . For axial rotation reconstruction (i.e. the embodiment of Figures 10C and 10D) , the transformation function is f: (x,y, z)→(x, z,y) . For sector reconstruction, the transformation is effected by f : (x,y, z)→(z,y, c-x) , where C is the x-dimension I(x,y, z) -1.
Considering the transformation of I (x,y, z) to R (x,y, z) for sector reconstruction, in order to access a
vertical line of the image stored in computer memory 1113, the transformation is effected as shown in Figure 16B to provide easy access to each vertical line of each image slice. The most efficient procedure for accessing each vertical line of the image is to do so on the zy- plane. Hence, as shown in Figure 16B, the image is rotated so that 2-D images on the yz-plane are transformed to the xy-plane.
With respect to axial rotation (Figure 16A) , the same geometry of 2-D images is originally stored in computer memory 1113, except that instead of accessing vertical lines, for 3-D reconstruction the system captures horizontal lines. The horizontal 2-D image "slices" are accessed from top-to-bottom of the 3D volume rather than from front-to-back, to provide greater speed of image reconstruction with the fewest number of calculations.
The transformations of step 1405 are conducted on a pixel-by-pixel basis, utilizing known methodologies. In step 1407, a temporary raster T(x,y) is created which must be large enough to store a z-slice of the reconstructed image. Figure 17 shows a temporary raster T(χ f Y) comprising a plurality of grid points or pixels 1701. The temporary raster T(x,y) is used to hold a single z-slice of the reconstructed image, which is represented by a 3-D array V(x,y, z) .
Upon creating the temporary raster T(x,y) all computer calculations are performed on the x,y-plane (i.e. on individual z-slices of the image) . The temporary raster T(x,y) must be created to sufficiently large dimensions to accommodate the creation of bigger images in axial rotation geometries. For each pixel in the temporary raster T(x,y) the module must find the exact coordinate in the original slice of the resultant image R (x,y, z) (i.e. the exact color or grey level must be computed for each pixel in T from R) .
In step 1409, for each pixel p(x,y) in T(x,y) a list
L (x,y) of contributing pixels is computed in the source image from which each such pixel receives a gray level, discussed in greater detail below with reference to Figure 15. Figures 18A and 18B shown computation of the contents of temporary raster T(x,y) from a z-slice of the resultant image R (x,y, z, ) , wherein the x-dimension of A(χ /Y) is Λ;. For axial rotation, the geometry is such that there is a line which rotates according to a circular path to compute the original location for a pixel T, therefor the use of polar coordinates is adopted. Starting from the origin, the radial distance "r" from the pixel and the angle "a" are computed. After "r" and "a" are computed, indexing is used for each image slice in R since the location of the first horizontal line of the image is known. By using the angle "a" the module computes where the pixel should be located in R. Once the location of the pixel has been computed, one of either a "nearest neighbor algorithm" or an averaging algorithm or weighting function are used to approximate the appropriate gray-level for the pixel. The arrow A in Figure 18B shows the direction with which the original images in I(x,y, z) were captured.
Figures 19A and 19B show the computation of contents of the temporary raster T(x,y) from A(x,y) for sector reconstruction.
For sector rotation, the geometry is similar to that of axial rotation except that the axis of rotation is outside of the image. As discussed above with reference to axial reconstruction, the angular displacement and distance of each pixel is computed from the first line of the image. However, a smaller value of "r" is used in sector reconstruction than the actual distance from the pixel to the axis of rotation since there is no need to extend the radius all of the way back to the axis of rotation. Using well known linear algebra techniques, the vector established by the calculated polar coordinates can be transformed to have the same origin
(i.e. it is possible to ignore the axis of rotation outside of the image volume) .
As shown in Figures 20A, 2OB, 21A and 2IB, the list L of p(x0,y0) may contain some or all of the pixels neighbors in A (x,y) , depending on the scheme adopted (i.e. nearest neighbor, averaging or interpolation of neighbors) . In step 1411, the module queries whether all z-slices of the final 3-D image V(x,y, z) have been constructed. If the answer to the query is no, then a new z-slice A(x,y) is extracted from the resultant image R (x,y, z) and the grey level or color of each pixel T(x,y) from A(x,y) is computed according to the L list of that pixel by means of nearest neighbor, or averaging or interpolation of neighbor pixels (step 1417) . Once the list of neighbors has been computed for each pixel T, then the entire 3-D volume can be reconstructed. In particular, each slice in R is accessed and reconstruction is executed, one slice at a time according to the list of neighbors L. After finding the exact location of the pixel J?, the list L of contributing pixels is computed. In Figures 20A and 21A, the location of the pixels is shown for axial rotation and sector rotation, respectively. Indexing is then performed to revert to the individual z-slice of R (referred to as A) , which changes the geometry back to a rectangularly shaped image (Figures 20B and 2IB) . At this stage, all four neighbors of the pixel are accessed to determine the appropriate grey-level or color for the selected pixel of T(x,y) . In Figure 22, the z-slice A(xf y) of R (x,y, z) is extracted and the computation of step 1417 is effected to compute all values of the plane that are contributed by the resultant image R(x,y, z) for storage into the final volume. After finishing all slices of R(x,y, z) , the reconstruction is complete and the contents of T(x,y) are stored in the final 3-D image V(x,y, z) .
Steps 1411-1419 are re-executed until all z-slices
of the final 3-D image have been reconstructed, after which the module is exited (step 1413) .
Turning briefly to Figure 15, the pixel mapping step
1409 is shown in greater detail as a sub-module. The module is initiated at step 1501, following which the procedure queries whether all pixels of the temporary raster T(x,y) has been processed (step 1503) . If not, a pixel p(x,y) of the temporary raster
T(χ /Y) is retrieved in step 1505. A mapping from p(x,y) to polar coordinates p' (r,a) is computed in step 1507 using the techniques discussed above with reference to
Figures 18A, 18B, 19A and 19B.
Next, a list L (x,y) is created of the contributing pixels in the source image according to p' (r, a) (step 1509) . The creation of this list is shown with reference to Figures 20A and 20B for axial rotation and 21A and 2IB for sector reconstruction.
The list L(x,y) is then stored at step 1511 and process control is returned to the query of step 1503. Once all pixels of the temporary raster T(x,y) have been processed, the sub-module is exited at step 1513.
As discussed above, once the reconstruction has been completed and the reconstructed image array V(x,y, z) has been stored in computer memory 1113, the image may be viewed and manipulated in a well known manner utilizing volume viewer module 1107 (Figure 11) (e.g. Analyze™) . Although the volume reconstruction technique has been described in an ultrasound 3-D imaging system having an ultrasound probe which is moveable to scan the target, it should be apparent to those of skill in the art that the volume reconstruction technique may be used in conjunction with 3-D imaging systems using a different type of ultrasound imaging probe and/or sweep geometry to acquire the 2-D images. Moreover, it should also be realized that the volume reconstruction technique is not limited to ultrasound imaging but may also be used in different environments where a 3-D image of a target
needs to be generated from a succession of 2-D images.
In addition, although the system has been described as including a clinical ultrasound machine and a computer, it is contemplated that a single machine performing the functions of both of these components be used.
Other embodiments and modifications of the invention are possible within the sphere and scope defined by the claims appended hereto.