System and method for creating a panoramic view of a volumetric image

Information

  • Patent Grant
  • 7609910
  • Patent Number
    7,609,910
  • Date Filed
    Wednesday, April 6, 2005
    19 years ago
  • Date Issued
    Tuesday, October 27, 2009
    15 years ago
Abstract
A system and method for creating a panoramic projection of a volumetric image of an object is disclosed. A virtual camera is placed in a center of a cubic space within the object. Images obtained from the camera are projected onto a single plane. A front image is mapped in a window at a center of the plane. A left, right, top and bottom image of the cubic space are mapped around the rectangular window such that the left, right, top and bottom images form a continuous image surrounding the rectangular window.
Description
FIELD OF THE INVENTION

The present invention is directed to a system and method for creating a panoramic view of a volumetric image, and more particularly, to a system and method for creating a three dimensional panoramic display of a volumetric image of an object such as a colon.


BACKGROUND OF THE INVENTION

Virtual endoscopy is a non-invasive diagnostic procedure aimed at exploring the inner surface of anatomical structures inside the human body. Using advanced image-processing techniques, three dimensional (3D) models are reconstructed from a series of high-resolution two dimensional (2D) images. A physician can then navigate inside the 3D virtual model either manually or using an automatic navigation algorithm.


One of the most promising uses for virtual endoscopy is the screening of patients for colorectal cancer. Virtual endoscopy mimics conventional colonoscopy with the added advantage that it is less invasive and potentially more attractive as a screening method. An additional benefit of virtual colonoscopy over conventional colonoscopy is the ability to fly-through the colon in both an antegrade and retrograde direction. Although this dual directional navigation significantly increases the amount of surface displayed, substantial sections of the colon remain obstructed behind haustral folds.


A number of techniques have been proposed to create image projections that overcome some of the limitations of the standard 3D visualization technique. These techniques seek to display more of the surface of the colon and thereby increase the polyp detection sensitivity. Some of the views are generated by placing the virtual camera in the center of a viewing space, and projecting images onto the corresponding walls.


One known technique uses a cubic viewing space with the cameras located in the center of the cube and projecting on each face an image with a 90 degree viewing angle. The cube is then unfolded into a single plane presenting a 360 degree field of view of the colon surface. An example of this projection is shown in FIG. 1. To reduce the disturbance due to the discontinuities that arise with this representation, small images are added as flaps adjacent to each face. Off-line animated image sequences are generated from a number of point samples selected along the central path through the colon. FIG. 1 shows a single frame of an animated sequence using this cubing mapping. Each square shown in FIG. 1 represents a different side of the cube. As shown, the sides are labeled as left 102, bottom 104, front 106, top 108, right 110 and back 112. The unfolded cube forms a cross-like structure. While this technique shows a projection of multiple views in the same frame, the construction of the layout is difficult to follow and has black areas which disrupt the viewing area. As a result, a polyp that appears near an edge of the cube can be split into two or more windows when the cube is unfolded.


Another known technique uses map projections that visualize the entire surface of a viewing sphere. Using Mercator and stereographic projections, the surface of the sphere is transformed onto a flat surface. The major drawback of this technique is that, as any projection of the surface of a sphere on a plane, it introduces some degree of deformation of the image. In particular, the Mercator projection maps the poles of the globe infinitely far away while displaying the objects near the equator with minimum distortion. There is a need for a method for creating a panoramic endoscopic view from a volumetric image that efficiently displays the views and is minimally distorted.


SUMMARY OF THE INVENTION

The present invention is directed to a system and method for creating a panoramic projection of a volumetric image of an object. A virtual camera is placed in a center of a cubic space within the object. Images obtained from the camera are projected onto a single plane. A front image is mapped in a window at a center of the plane. A left, right, top and bottom image of the cubic space are mapped around the rectangular window such that the left, right, top and bottom images form a continuous image surrounding the rectangular window.





BRIEF DESCRIPTION OF THE DRAWINGS

Preferred embodiments of the present invention will be described below in more detail, wherein like reference numerals indicate like elements, with reference to the accompanying drawings:



FIG. 1 illustrates a prior art technique for creating a projection of an image of a colon;



FIG. 2 is a system block diagram of a system for creating a panoramic view from a volumetric image of a colon in accordance with the present invention;



FIG. 3 illustrates a panoramic projection of an image of a colon in accordance with the present invention;



FIG. 4 illustrates a panoramic projection of an image of a colon˜in accordance with another embodiment of the present invention;



FIG. 5 illustrates a method for creating a disc-shaped panoramic view from a volumetric image in accordance with the present invention;



FIG. 6 illustrates a square-shaped panoramic view in accordance with the present invention;



FIG. 7 illustrates a method for finding direction vectors from unfolded cube coordinates in accordance with the present invention;



FIG. 8 illustrates an image of a colon and unseen areas after a retrograde fly through using standard camera projection; and



FIG. 9 illustrates an image of a colon and unseen areas after a retrograde fly through using the panoramic projection of the present invention.





DETAILED DESCRIPTION

The present invention is directed to a system and method for creating a panoramic endoscopic view from a volumetric image of an object. The present invention will be described in the context of performing a medical examination of a patient using a medical imaging system. More particularly, the present invention will be described in the context of using a Computed Tomography (CT) system to perform a virtual colonoscopy.


It is to be understood by those skilled in the art that the present invention may be used with any type of two or three dimensional imaging system for the purpose of viewing the internal structure of a patient. Such structures may include organs, blood vessels, tissue areas or other defined structures. The patient may be a human or an animal. Furthermore, the approach may be applied to images of any dimension since the filter is defined for higher dimensions as well as 2D.


The present invention may also be used in non-medical applications that involve two or three dimensional imaging. One such example could be for security screening to determine the contents or packages or otherwise contained objects. Such a system could be used in a postal delivery environment or for screening belongings in an airport or other high security venue. The present invention could also have industrial applications and be used to inspect or count items being manufactured or assembled.


The present invention utilizes a panoramic projection of a volumetric image of an object so that an observer, such as a physician or medical technician can view the maximal amount of surface without obstruction and in a way that it appears as if the observer is truly inside the object being viewed. The panoramic projection is designed to increase surface visibilities during virtual endoscopies. The present invention creates a projection that renders five faces of a cubic viewing space into a plane in a continuous fashion. Using this type of visualization technique as a screening method for colon cancer could significantly reduce the evaluation time as it avoids having to fly-through the colon in both directions and prevents the occlusion of potential polyps behind haustral folds.



FIG. 2 schematically depicts an exemplary CT system in accordance with the present invention. The CT system in conjunction with a Computed Aided Detection (CAD) system can be used to perform a non-invasive procedure on a patient to screen for various medical conditions. Examples of such procedures are virtual colonoscopies or chest screenings for the presence of lesions and possibly cancerous tissue. It is to be understood by those skilled in the art that other imaging modalities can be used without departing from the scope and spirit of the present invention. For example, magnetic resonance imaging (MRI), fluoroscopy or ultrasound imaging could also be used to obtain the imagines. For purposes of explanation, the present invention will be described in the context of use of a CT system.


The CT system is equipped with an X-ray source 1, emitting a pyramidal X-ray beam 2, whose marginal rays are represented by the dot-dashed lines in FIG. 2 that penetrates an object 3 to be examined, for example, a patient, and that is incident on a radiation detector 4. The X-ray source 1 and the radiation detector 4 are, in the exemplary embodiment, mounted opposite to one another on an annular gantry 7.


The X-ray source 1 and the radiation detector 4 thus form a metrological system that can be rotated around the system axis 8 so that the patient 3 can be X-rayed at various projection angles relative to said system axis 8 and at various positions along the system axis 8. The resultant output signals of the individual detector elements are read out by a data acquisition system 10. The signals are sent to a signal processor 12 that computes an image of the patient 3 that, in turn, can be displayed on a monitor 13.


The images scanned by the CT system and computed by the signal processor 12 are transmitted to a CAD system 20 for further processing. The CAD system 20 applies a filter to the image to obtain a response image. This is then used to generate candidates. Features are then collected from the candidates (including features derived from the filter values) and used to classify them as true or false positives. Only the candidates classified as true positives are presented to the physician.


In accordance with the present invention, a cubic viewing space in used in which the back projection is not displayed. A virtual camera is placed in the center of a cubic space and images are projected into a single plane using a 90 degree viewing angle. In accordance with the present invention, the front image is mapped into a rectangular window at the center of the plane while the other four images are projected into trapezoidal shaped windows as illustrated in FIG. 3. This approach resolves the problem of discontinuity found in the prior art (see FIG. 1) and at the same time makes better use of the screen space. As can be seen in FIG. 3, the center of the image 310 is the central view of the camera. Surrounding this view is the left view 302, top view 304, right view 306 and bottom view 308. The arrangement of the views provides a more realistic perspective of being inside the colon.


A fast raycast algorithm is used to render each frame in real time. Image 312 is a single frame during a fly-through sequence. The only distortion introduced by the projection arises from the mapping from the cubic walls into the trapezoidal projections. In order to limit the distortions that appear at the edges of the unfolded cube, in accordance with another embodiment of the present invention, another projection can be used. FIG. 4 illustrates a panoramic view in which the front face 410 is mapped into a square and the other four faces 402-408 are mapped around it into a disk. The projection guarantees a smooth transition around the five faces. The size of the frontal view can be adjusted.


A method for creating a panoramic projection from a volumetric image will now be described in more detail with reference to FIGS. 5-7. As shown, the axes of the image coordinate system are labeled x, y and z. The x axis extends from the right to the left of the image. The y axis extends from the bottom to the top of the image. The z axis extends from the back to the front of the image. The panoramic endoscopic view is obtained by casting rays from the center of a virtual endoscope positioned inside the volume. The intensity of a point in the endoscopic view is an arbitrary function of the volume intensity values along the corresponding ray.


From a given endoscopic position, each ray is uniquely characterized by a direction vector. As such, a panoramic view is modeled as a vector field (i.e., a collection of direction vectors). The vector field is computed for a fixed endoscopic position. It is assumed that the endoscope is placed at the origin of the volume coordinate system and points in the z direction. Arbitrary location and orientation of the endoscope can be taken into account by translation and rotation of the vector field.


The two panoramic views described above (i.e., the disc and square view) are improvements of the prior art unfolded cubic view. These panoramic views are designed to obtain a continuous projection, a large field of view and limited distortion.


In accordance with the present invention, the projection is a two step process. First, for each point in the panoramic view, the corresponding point in one of the faces of the unfolded cube is found. Second the corresponding direction vector is found.



FIG. 5 illustrates a method for the continuous panoramic projection of an unfolded cube onto a disc in accordance with the present invention. The endoscopic view is modeled as a disc inscribed in a square of side length L (i.e., the disc has a radius equal to L/2). Any point inside this square is characterized by a position vector {right arrow over (p)} with respect to its center. The inner square of side length l corresponds to the front view of the unfolded cube. The remaining sections of the disc contain transformed versions of the top, left, right and bottom views of the unfolded cube.


Four vectors are introduced: {right arrow over (v)}1, {right arrow over (v)}2, {right arrow over (v)}3, {right arrow over (v)}4 as shown in FIG. 5. The term v(j) is denoted as the j-th component of a vector {right arrow over (v)} and ∥{right arrow over (v)}∥ its norm. The expression |x| denotes the absolute value of a scalar x. The operator · corresponds to the scalar product of two vectors.


A coordinate system is attached to each face of the unfolded cube (where (0,0) is the upper-left corner of a face and the side length is 1). On a given face, points are located by a set of coordinates (r,c) as shown in FIG. 5. The mapping between the disc and the cube's faces is as follows:














If ∥{right arrow over (p)}∥ > radius then set point intensity value to zero (the point is outside


the disk)


Else

















If






max


(




p


(
1
)




,



p


(
2
)





)



<


l
2






then


















(The point is in the inner square − “front” view)



(What are the corresponding coordinates?)















r
=

0.5
+


p


(
1
)


l








c
=

0.5
+


p


(
2
)


l
























Else









(The point is in a lateral view)



(Which view?)
















n


=


p





p












If







n


·


v


2



>



2

2


then

















(The point is in the “top” view)







cos_angle
=


n


·


v


2








(

Cosine





of





the





angle





between






p







and







v


2


)





















norm_p

_inside

_front

=

l

2
·
cos_angle











(Length of the section of {right arrow over (p)} located inside the inner square)













r=radius-pradius-norm_p_inside_front(=d1d2seeFIG.1)c=0.5+n(2)2·cos_angle(d3l=n(2)2·cos_angleseeFIG.1)













Else





If







n


·


v


3



>



2

2


then












 (The point is in the “right” view)



 [. . .] (Apply same principle to the new direction)












Else





If







n


·


v


4



>



2

2


then












 (The point is in the “left” view)



 [. . .] (Apply same principle to the new direction)












Else





If







n


·


v


1



>



2

2


then












 (The point is in the “bottom” view)



 [. . .] (Apply same principle to the new direction)



End









End







End










FIG. 6 illustrates a method for the continuous panoramic projection of an unfolded cube onto a square in accordance with the present invention. The square projection view is a variant of the disc projection view, where the unfolded cube is mapped onto a square. There is no regions left unused, but there is some distortion along the diagonals of the endoscopic view. The mapping between the square and the cube's faces is as follows:



















If






max


(




p


(
1
)




,



p


(
2
)





)



<


l
2






then

















(The point is in the inner square − “front” view)



(What are the corresponding coordinates?)















r
=

0.5
+


p


(
1
)


l








c
=

0.5
+


p


(
2
)


l






















Else









(The point is in a lateral view)



(Which view?)
















n


=


p





p












If







n


·


v


2



>



2

2


then























(The point is in the “top” view)







cos_angle
=


n


·


v


2








(

Cosine





of





the





angle





between






p







and







v


2


)

















norm_p

_inside

_front

=

l

2
·
cos_angle













 (Length of the section of {right arrow over (p)} located inside the inner square)















r
=




radius
/
cos_angle

-



p







radius
/
cos_angle

-

norm_p

_inside

_front





(

=



d
1


d
2







see






FIG
.




2



)








c
=

0.5
+



n


(
2
)



2
·
cos_angle








(



d
3

l

=





n


(
2
)



2
·
cos_angle









see






FIG
.




2



)






























Else





If







n


·


v


3



>



2

2


then












 (The point is in the ”right” view)



 [. . .] (Apply same principle to the new direction)












Else





If







n


·


v


4



>



2

2


then












 (The point is in the “left” view)



 [. . .] (Apply same principle to the new direction)












Else





If







n


·


v


1



>



2

2


then












 (The point is in the “bottom” view)



 [. . .] (Apply same principle to the new direction)



End







End









The following explanations refer to FIG. 7 which shows a method for finding direction vectors from the unfolded cube coordinates. To each face of the unfolded cube corresponds a frustum (see FIG. 5 and FIG. 7). A frustum is characterized by a front plane (a square of edge length lfront placed at distance dfront of the endoscope location) and a back plane (a square of edge length lback placed at distance dback of the endoscope location).


The following describes how to find the direction vector {right arrow over (v)} of the ray corresponding to the location (r,c) on one of the cube faces. The “top” view is used as an example; the adaptation to the other views is immediate.







Let






p
1


=


[






l
front

2

-

c
·

l
front








d
front







-


l
front

2


+

r
·

l
front






]







(

See






FIG
.




3


)













And






p
2


=


[






l
back

2

-

c
·

l
back








d
back







-


l
back

2


+

r
·

l
back






]







(

See






FIG
.




3


)










Then






v
->


=


p
2

-

p
1















It is to be understood that the methods presented above can be combined to obtain a more compact and efficient implementation. The symmetry of the vector field can also be exploited to reduce the computational cost.


To demonstrate the usefulness of the present invention, the percentage of colon surface visualized using the standard camera projection is compared to the colon surface visualized using the panoramic projection of the present invention. A single flight through the virtual colon is performed for each projection. FIG. 8 illustrates a global view of the colon after a fly through in the retrograde direction using the conventional virtual camera projection. The dark gray voxels represent the areas not visible.



FIG. 9 illustrates the results of the same flight using the panoramic projection of the present invention. Since these projections show simultaneously a forward, left, right, up and down view, there is nearly 100% surfactant coverage with one single pass.


To achieve a good coverage, a conventional virtual navigation typically executes the additional fly through in the other direction. The added fly through not only increments the reading time as some parts of the surface are examined twice, but even then some areas remain hidden behind haustral folds. The panoramic projection therefore could be used to speed up the 3D reading since the number of fly throughs can be cut in half. It is to be understood by those skilled in the art that the panoramic projection of the present invention is not limited to the examination of the colon and can be used to display the inner surface of any virtual organ.


Having described embodiments for a method for creating a panoramic projection of a volumetric image, it is noted that modifications and variations can be made by persons skilled in the art in light of the above teachings. It is therefore to be understood that changes may be made in the particular embodiments of the invention disclosed which are within the scope and spirit of the invention as defined by the appended claims. Having thus described the invention with the details and particularity required by the patent laws, what is claimed and desired protected by Letters Patent is set forth in the appended claims.

Claims
  • 1. A method for performing a virtual fly-through inside of an anatomical lumen, represented by medical image data, using a medical image processing system, comprising: placing a virtual camera within the lumen;obtaining five images of the interior of the lumen from the point of view of the virtual camera, at 90° viewing angles, the five images comprising a forward image, a left image, a right image, a top image and a bottom image;displaying a first virtual fly-through frame by mapping the forward image to a rectangular window at the center of the frame while the remaining four images are projected to windows to the left, right, top, and bottom of the rectangular window, respectively; andrepeating steps of advancing the position of the virtual camera down the lumen and displaying a subsequent virtual fly-through frame in the manner in which the first virtual fly-through frame was displayed until the virtual fly-through is completed.
  • 2. The method of claim 1, wherein he remaining four images are projected to trapezoidal shaped windows such that all five display windows form a continuous panoramic view in the shape of a rectangle.
  • 3. The method of claim 1, wherein the remaining four images are projected to windows such that all five display windows form a continuous panoramic view in the shape of a disc.
  • 4. The method of claim 1, wherein a fast raycasting technique is used to obtain the five images of the interior of the lumen from the point of view of the virtual camera.
  • 5. The method of claim 1, wherein a size of the forward image can be adjusted.
  • 6. The method of claim 1, wherein the anatomical lumen is a human organ.
  • 7. The method of claim 6, wherein the human organ is a colon.
  • 8. The method of claim 1, wherein the medical image data is obtained using a Computed Tomography system.
  • 9. A system for performing a virtual fly-through inside of an anatomical lumen, represented by medical image data, comprising: an image acquisition apparatus for acquiring the medical image data; andan image processing system for: receiving the acquired medical image data;placing a virtual camera within the lumen;obtaining five images of the interior of the lumen from the point of view of the virtual camera, at 90° viewing angles, the five images comprising a forward image, a left image, a right image, a top image and a bottom image;displaying a first virtual fly-through frame by mapping the forward image to a rectangular window at the center of the frame while the remaining four images are projected to windows to the left, right, top, and bottom of the rectangular window, respectively; andrepeating steps of advancing the position of the virtual camera down the lumen and displaying a subsequent virtual fly-through frame in the manner in which the first virtual fly-through frame was displayed until the virtual fly-through is completed.
  • 10. The system of claim 9, wherein the remaining four images are projected to trapezoidal shaped windows such that all five display windows form a continuous panoramic view in the shape of a rectangle.
  • 11. The system of claim 9, wherein the remaining four images are projected to windows such that all five display windows form a continuous panoramic view in the shape of a disc.
  • 12. The system of claim 9, wherein a fast raycasting technique is used to obtain the five images of the interior of the lumen from the point of view of the virtual camera.
  • 13. The system of claim 9, wherein a size of the forward image can be adjusted.
  • 14. The system of claim 9, wherein the anatomical lumen is a human organ.
  • 15. The system of claim 14, wherein the human organ is a colon.
  • 16. The system of claim 9, wherein the medical image data is obtained using a Computed Tomography system.
CROSS REFERENCE TO RELATED APPLICATION

This application claims the benefit of U.S. Provisional Patent Application Ser. No. 60/561,183, filed on Apr. 9, 2004, which is incorporated by reference in its entirety.

US Referenced Citations (16)
Number Name Date Kind
5831623 Negishi et al. Nov 1998 A
5971767 Kaufman et al. Oct 1999 A
5999187 Dehmlow et al. Dec 1999 A
6514082 Kaufman et al. Feb 2003 B2
6947039 Gerritsen et al. Sep 2005 B2
7102634 Kim et al. Sep 2006 B2
7123777 Rondinelli et al. Oct 2006 B2
7232409 Hale et al. Jun 2007 B2
7245761 Swaminathan et al. Jul 2007 B2
7269241 Siltanen et al. Sep 2007 B2
7324104 Bitter et al. Jan 2008 B1
20020190980 Gerritsen et al. Dec 2002 A1
20030152897 Geiger Aug 2003 A1
20050151730 Lobregt Jul 2005 A1
20050168616 Rastegar et al. Aug 2005 A1
20070052724 Graham et al. Mar 2007 A1
Foreign Referenced Citations (1)
Number Date Country
WO 0055812 Sep 2000 WO
Related Publications (1)
Number Date Country
20050226483 A1 Oct 2005 US
Provisional Applications (1)
Number Date Country
60561183 Apr 2004 US