In vivo imaging device with a small cross sectional area

Information

  • Patent Grant
  • 7753842
  • Patent Number
    7,753,842
  • Date Filed
    Monday, August 27, 2007
    17 years ago
  • Date Issued
    Tuesday, July 13, 2010
    14 years ago
Abstract
An in vivo imaging device may include a capsule shaped housing and an image sensor. The capsule shaped housing may have a longitudinal axis and a window. The image sensor (e.g., a CMOS sensor) may include a pixel array portion and a circuitry portion. The circuitry portion may be segregated, for example longitudinally, from the pixel array portion. The pixel array portion may be disposed within said housing substantially parallel to said longitudinal axis.
Description
BACKGROUND OF THE INVENTION

Devices and methods for performing in-vivo imaging of passages or cavities within a body are known in the art. Such devices may include, inter alia, various endoscopic imaging systems and devices for performing imaging in various internal body cavities.


Reference is now made to FIG. 1 which is a schematic diagram illustrating an embodiment of an autonomous in-vivo imaging device. The device 10A typically includes a capsule-like housing 18 having a wall 18A. The device 10A has an optical window 21 and an imaging system for obtaining images from inside a body cavity or lumen, such as the GI tract. The imaging system may include an illumination unit 23. The illumination unit 23 may include one or more light sources 23A. The one or more right sources 23A may be a white light emitting diode (LED), or any other suitable light source, known in the art, The imaging system of the device 10A may include an imager 24, such as a CMOS or CCD, which acquires the images and an optical system 22 which focuses the images onto the imager 24. Typically, the imager 24 is arranged so that it's surface 27 is perpendicular to the longitudinal axis 19 of the device 10A. The illumination unit 23 illuminates the inner portions of the body lumen through an optical window 21. Device 10A further includes a transmitter 26 and an antenna 27 for transmitting the image signal of the imager 24, and one or more power sources 25. The power source(s) 26 may be any suitable power sources such as but not limited to silver oxide batteries, lithium batteries, or other electrochemical cells having a high energy density, or the like. The power source(s) 25 may provide power to the electrical elements of the device 10A.


Typically, in the gastrointestinal application, as the device 10A is transported through the gastrointestinal (GI) tract, the imager, such as but not limited to a multi-pixel CMOS imager acquires images (frames) which are processed and transmitted to an external receiver/recorder (not shown) worn by the patient for recording and storage. The recorded data may then be downloaded from the receiver/recorder to a computer or workstation (not shown) for display and analysis. During the movement of the device 10A through the GI tract, the imager may acquire frames at a fixed or at a variable frame acquisition rate. For example, in one embodiment the imager (such as, but not limited to a CMOS imager) may acquire images at a fixed rate of two frames per second (2 Hz). However, other different frame rates may also be used, depending, inter alia, on the type and characteristics of the specific imager or camera or sensor array implementation which is used, and on the available transmission bandwidth of the transmitter 26. The downloaded images may be displayed by the workstation by replaying them at a desired frame rate. This way, the expert or physician examining the data is provided with a movie-like video playback which may enable the physician to review the passage of the device through the GI tact.


It may generally be desirable to decrease the size and particularly the cross sectional area of in vivo imaging devices, such as the device 10A of FIG. 1, or of imaging devices that are to be inserted into working channels of endoscope-like devices, or integrated into catheter-like devices which may be used in conjunction with guide wires, or the like. Smaller catheter like devices with reduced area may be inserted into narrower body cavities or lumens, such as for example, the coronary arteries, the urethra, the common bile duct, or the like and may also be easier to insert into working channels of other devices such as endoscopes, laparoscopes, gastroscopes, or the like.


Decreasing the cross-sectional area of such devices may be limited by the cross-sectional area of the imaging sensor, such as for example the imager 24 of FIG. 1


SUMMARY OF THE INVENTION

In one embodiment, an image sensor (e.g., a CMOS sensor) includes a pixel array portion and a circuitry portion, where the circuitry portion may be segregated, for example longitudinally, from the pixel array portion.





BRIEF DESCRIPTION OF THE DRAWINGS

The invention is herein described, by way of example only, with reference to the accompanying drawings, in which like components are designated by like reference numerals, wherein:



FIG. 1 is a schematic diagram illustrating an embodiment of a prior art autonomous in-vivo imaging device;



FIG. 2 is a schematic front view of the surface layout of a typical prior art CMOS imager.



FIG. 3 is a schematic front view of the imager 24 of FIG. 1;



FIG. 4 is a schematic cross sectional view illustrating a CMOS imager having a square cross-sectional area disposed within a housing having a circular cross-section;



FIG. 5 is a schematic top view showing the layout of a rectangular CMOS imager having segregated pixel array area and support circuitry areas, in accordance with a preferred embodiment of the present invention;



FIG. 6 is a cross sectional view illustrating an in-vivo imaging device having a reduced cross-sectional area and including the CMOS imager of FIG. 5, in accordance with a preferred embodiment of the present invention;



FIG. 7 is a cross-sectional view illustrating an in-vivo imaging device having a reduced cross-sectional area and including the CMOS imager of FIG. 5, in accordance with another preferred embodiment of the present invention; and



FIG. 8 is a schematic cross sectional view of part of an insertable device for in vivo imaging, in accordance with a preferred embodiment of the present invention.





DETAILED DESCRIPTION OF THE INVENTION

In the following description, various aspects of the present invention will be described. For purposes of explanation, specific configurations and details are set s forth in order to provide a thorough understanding of the present invention. However, it will also be apparent to one skilled in be art that the present invention may be practiced without the specific details presented herein. Furthermore, well known features may be omitted or simplified in order not to obscure the present invention.


The present invention is based on providing an imager having a small cross-sectional area for implementing in vivo imaging devices such as, but not limited to, swallowable autonomous in-vivo imaging devices (capsule-like or shaped otherwise), and wired or wireless imaging units which are integrated into endoscope-like devices, catheter-like devices, or any other type of in-vivo imaging device that can be introduced into a body cavity or a body lumen.


It is noted that while the embodiments of the invention shown hereinbelow are adapted for imaging of the gastrointestinal (GI) tract, the devices and methods disclosed may be adapted for imaging other body cavities or spaces.



FIG. 2 is a schematic front view of the surface layout of a typical prior art CMOS imager. The CMOS imager 30 is an integrated circuit, which is typically implemented on a silicon wafer. The CMOS imager 30 includes an imaging sensor part 30A which is usually located at one of the corners of the frontal surface of the CMOS imager 30. The imaging sensor part 30A may include the two dimensional array of light sensitive diodes (not shown in detail) comprising the sensor pixels (not shown in detail), and may also include integrated amplification circuitry (not shown) and switching circuitry (not shown) for controlling the pixel sampling or readout, and may also include electrical con ducting paths for connecting the pixels to the functional units that perform the readout of the pixels.


The remaining surface portion 30B of the CMOS imager 30, may include, inter alia, integrated circuitry (not shown in detail) for performing various control and timing functions, analog to digital (A/D) conversion circuitry (not shown) for converting the analog signal sampled from the individual pixels and various input/output (I/O) circuitry (not shown in detail for sending the image digitized data and control signals as output signals to devices or circuitry (not shown) such as but not limited to a transmitter like the transmitter 26 of FIG. 1, or any other signal processing circuit or element or unit which may be connected to the CMOS imager 30, as is known in the art.


Reference is now made to FIG. 3 which is a schematic front view of the imager 24 of FIG. 1. The imager 24 of FIG. 1 may be a CMOS imager and may be configured to have a square shape having a side length S (as seen in the front view of FIG. 3).


The CMOS imager 24 may include an imaging sensor part 24A which is may be typically located at the center of the frontal surface of the CMOS imager 24. The imaging sensor part 24A may include the two dimensional array of light sensitive diodes (not shown in detail) comprising the sensor pixels (not shown in detail), and may also include integrated amplification circuitry (not shown) and switching circuitry (not shown) for controlling the pixel sampling or readout, and may also include electrical conducting paths for connecting the pixels to the functional units that perform the readout of the pixels.


The remaining surface portion 24B of the of the CMOS imager 24, which surrounds the imaging sensor part 24A, may include, inter alia, integrated circuitry (not shown in detail for performing various control and timing functions, analog to digital (A/D) conversion circuitry (not shown) for converting the analog signal sampled from the individual pixels and various input/output (I/O) circuitry (not shown in detail) for sending the image digitized data and control signals as output signals to devices or circuitry (not shown) such as but not limited to a transmitter like the transmitter 26 of FIG. 1, or any other signal processing circuit or element or unit which may be connected to the CMOS imager 24.


The advantage of the imager configuration shown in FIG. 3, is that it makes possible to position the optical system 22 (best seen in FIG. 1) in the center of the optical window 21 (See FIG. 1) without unduly increasing the cross-sectional area of the entire device 10A of FIG. 1). However, with the configuration (shown in FIG. 3) of the CMOS imager 24, if it is desired to minimize the cross-sectional area of the device 10A, the diminishing of the cross-sectional area of the entire device IA may be limited by tee diagonal of the entire CMOS imager 24.


Reference is now made to FIG. 4 which is a schematic cross sectional view illustrating a CMOS imager having a square cross-sectional area disposed within a housing having a circular cross-section.


The CMOS imager 24 is shown disposed within a housing 28 having a wall 28A. The housing 28 has a circular cross-section. The wall 28A may be part of a capsule like housing, such as the housing 18 shown in the capsule-like device 10A of FIG. 1). The wall 28A may also be the wall of an imaging unit which comprises part of an elongated device such as a catheter-like device (not shown), or an endoscope-like device (not shown) or the like.


Similar to the CMOS imager 24 of FIG. 3, the square CMOS imager 34 of FIG. 4 may have a square imaging sensor part 34A (having a side length of Ai) including the imaging pixels (the pixels are not shown in detail) and a remaining surface part 34B which surrounds the imaging sensor part 34A and which may include all the supporting circuitry (not shown) as disclosed in detail hereinabove for the surface part 24B of FIG. 3. The entire CMOS imager 34 has a side S, and diagonal D. It can be seen from FIG. 4 that the internal diameter of the wall 28A of the housing 28 may not be smaller than the length of the diagonal D of the square CMOS imager 34. Thus, using the configuration of the CMOS imager 34, it is not possible to decrease the diameter of the housing 28 to a length smaller than the diagonal D of the entire CMOS imager 34.


Reference is now made to FIG. 5 which is a schematic top view showing the layout of a rectangular CMOS imager having segregated pixel array area and support circuitry areas, in accordance with a preferred embodiment of the present invention. The CMOS imager 54 may include a square imaging sensor part 54A (having a side length of Ai) including the imaging pixels (the pixels are not shown in detail) and a remaining surface part 54B which is longitudinally segregated from the imaging sensor part 54A and which may include all the necessary supporting circuitry (not shown) as disclosed in detail hereinabove.


The length Li of the CMOS imager 54 is adapted such that all the necessary supporting circuitry may be accommodated in the part 54B. Thus, when the CMOS imager 54 is compared to the CMOS imager 34 of FIG. 4, the dimensions of the imaging sensor parts 34A and 54A are the same (both imaging sensor parts may be a square having a side length of Ai), the CMOS imager 34 has a square shape having a side Si while the CMOS imager 54 has an elongated shape having a length Li.


Reference is now made to FIG. 6 which is a cross-sectional view illustrating an in-vivo imaging device having a reduced cross-sectional area and including the CMOS imager of FIG. 5, in accordance with a preferred embodiment of the present invention.


The device 50 includes a capsule-like housing 58 which has an optical window 21A. The device 50 may include an optical system 22A and a mirror 55. The device 50 may include the CMOS imager 54 of FIG. 5. The mirror 55 may be substituted by any suitable light deflecting element such as a suitably configured prism, or the like, for deflecting the light rays collected by the 22A optical system towards the part 54A of the CMOS imager 54. The optical system 22A may be a lens, a group of lenses, a zoom lens, a composite lens, a wide angle lens or any other suitable image forming optical element known in the art.


The device 50 further includes the power sources 25A which may be similar to the power sources 25 of FIG. 1 but may be smaller in size due to the reduced cross sectional area of the device 50. The device 50 also includes an illuminating unit 53 comprising the light sources 53A. The construction and operation of the illuminating unit 53 and the light sources 53A may be as disclosed in detail for the illuminating unit 23 and the light sources 23A of FIG. 1, and possibly, as disclosed in WO 01/65995. The device 50 may also include a wireless transmitter 26A and an antenna 27A for transmitting the data of the acquired images.


The CMOS imager 54 is disposed longitudinally within the housing 58 such that the light rays 47 generated by the light sources 53A are reflected from the intestinal wall and pass through the optical window 21A. The reflected light rays are collected by the optical system 22A and are deflected towards the part 54A of the CMOS imager 54 to create an image to be sensed by the light sensing pixels (not shown) included in the part 54A. Preferably, but not necessarily, the CMOS imager 54 is disposed such that its longitudinal axis (not shown) is aligned parallel to the longitudinal axis 59 of the device 50. The angle α between the surface of the mirror 55 and the surface of the CMOS imager may be 45°, but may also be smaller than 45°. If the angle α is smaller than 45°, the image projected upon the pixels of the part 54A of the CMOS imager 54 may be distorted. Thus, the optical system 22A may be configured to suitably change the collected image in order to compensate for the distortion before the image reaches the part 54A of the CMOS imager 54. Alternatively, the distortion in the acquired image may be corrected after the acquisition by suitably processing the image data at a stage later than image acquisition. For example a distorted image may be processed in a workstation (not shown) after the image has been transmitted by the transmitter 26A. Such a distortion may be compensated by suitable computational algorithms, as is known in the art.


It is noted that in comparison to the configuration of the CMOS imaging unit 24 of FIG. 1 and the CMOS imaging unit 34 of FIG. 4 which are oriented perpendicular the longitudinal axis of the in vivo imaging device 10A, the configuration of the CMOS imager 54 being longitudinal and parallel to the axis 59 of the device 50 may enable a substantial reduction in the cross sectional area of the device 50 since the combination of the reduced area of the tight sensitive part 54A and the use of the mirror 55 allow such a reduction in cross-sectional area. It is noted that the cross sectional area (Ai)2 of the part 54A of FIG. 6 is substantially smaller than the cross sectional area (D)2 of the entire CMOS imager 34 of FIG. 4. The cross-sectional area of a cross section taken in a direction perpendicular to the longitudinal axis the device 50 may therefore be substantially reduced in comparison to the cross sectional area of a cross section taken in a direction perpendicular to the longitudinal axis 19 of the device 10A.


It is noted that additional configurations of the optical components of the in vivo imaging device may be possible.


Reference is now made to FIG. 7 which is a cross-sectional view illustrating an in-vivo imaging device having a reduced cross-sectional area and including the CMOS imager of FIG. 5, in accordance with another preferred embodiment of the present invention.


The device 60 includes a capsule-like housing 58 which has an optical window 21A. The device 60 may include an aperture 51, an optical system 65A and a mirror 55. The device 60 may include the CMOS imager 54 of FIG. 6. The mirror 55 may be substituted by any suitable light deflecting element such as a suitably configured prism (not shown), or the like, for deflecting the light rays 47C passing through by the aperture 51, towards the optical system 65A which projects an image on the part 54A of the CMOS imager 54. The optical system 65A may be a lens, a group of lenses, a zoom lens, a composite lens, a wide angle lens or any other suitable image forming optical element known in the art.


The device 60 further includes the power sources 25A which may be similar to the power sources 25A of FIG. 6.


The CMOS imager 54 is disposed longitudinally within the housing 68 such that the light rays 47A and 47B generated by the light sources 53A are reflected from the intestinal wall 45 and pass through the optical window 21A as tight rays 47C and 47D, respectively. The reflected light rays 47C and 47D are deflected by the mirror 55 towards the optical system 22A. The optical system 65A thus focuses an image on the part 54A of the CMOS imager 54 the image may be sensed by the light sensing pixels (not shown) included in the part 54A. Preferably, but not necessarily, the CMOS imager 54 is disposed such that its longitudinal axis (not shown) is aligned parallel to the longitudinal axis 69 of the device 60. The angle α between the surface of the mirror 55 and the surface of the CMOS imager may be 45°, but may also be smaller than 45°.


If the angle α is smaller than 45°, the image projected upon the pixels of the part 54A of the CMOS imager 54 may be distorted. Thus, the optical system 65A may be configured to suitably change the collected image in order to compensate for the distortion before the image reaches the part 54A of the CMOS imager 54. Alternatively, the distortion in the acquired image may be corrected after the acquisition by suitably processing the Image data at a stage later than image acquisition. For example, a distorted image may be processed in a workstation (not shown) after the image has been transmitted by the transmitter 26A. Such a distortion may be compensated by suitable computational algorithms, as is known in the art.


The configuration of the imager 54 and the mirror 55 and the optical system 65A of the device 60 may also enable the device 60 to have a reduced cross sectional area for the reasons disclosed in detail hereinabove.


Thus, the combination of the CMOS imager having segregated imaging and support circuitry parts disclosed hereinabove, and the longitudinal arrangement of such a segregated CMOS imager within the device allow the construction of autonomous or non-autonomous in vivo imaging devices with a small cross sectional area. The non-autonomous devices may include but are not limited to imaging heads or imaging units or imaging assemblies which are constructed as an integral part of, or are included within ,or are attached to catheter like devices, endoscope-like devices, trocars, or any other type of device which may be used for in vivo surgical and/or diagnostic purposes requiring imaging capabilities and may benefit from the reduced cross-sectional area of such imaging heads or imaging units or imaging assemblies.


Reference is now made to FIG. 8 which is a schematic cross sectional view of part of an insertable device for in vivo imaging, in accordance with a preferred embodiment of the present invention.


The insertable device 80 may include an elongated (preferably flexible) housing 78. The housing 78 may have an optical window 21A sealingly attached at the end 80A thereof. The device 80 may further include the CMOS imager 54 which is arranged longitudinally within the housing 78. The CMOS imager 54 has an imaging light sensitive part 54A segregated from other support circuitry part 54B as disclosed in detail hereinabove with respect to FIGS. 5-7.


The device 80 may include a mirror 55 which is inclined at an angle α to the imaging part 54A of the CMOS imager 54, as disclosed hereinabove. The angle α may be equal to 45° or may be different than 45°. Making the angle α smaller than 45°, may enable further reduction of the diameter or the cross sectional area of the device 80 as disclosed hereinabove for the devices 50 and 60. An optical system 22C may be suitably aligned along the longitudinal axis 89 of the device 80. The optical system 22C may include a single lens, multiple lenses, or other suitable optical elements like filters, as disclosed in detail hereinabove for me optical systems 22, 22A and 65A of FIGS. 1,6 and 7, respectively.


The device 80 may include an illumination unit 23H which may include light sources 23B. The light sources 23B may be the while LED light sources disclosed in detail hereinabove, and possibly, as disclosed in WO 01/15995., but may also be any other suitable miniature light sources known in the art. The CMOS imager 54 may be connected to a suitable power source 85 by suitable electrically conducting wires 92 connected to the power source 85. The CMOS imager 54 may be connected to suitable electrically conducting wires 82 for transmitting the image data to an external device (not shown) for further processing and for displaying of the acquired images.


It is noted that in accordance with another preferred embodiment of the present invention, the image data may be transmitted wirelessly to a receiver or a receiver/recorder, as is disclosed in detail for the autonomous in vivo imaging device 10A, and possibly, as disclosed in WO 01/65995. In such a case the device 80 may include an internal power source, a wireless transmitter (such as but not limited to the transmitters 26 or 26A of FIGS. 1, 6 and 7) and an antenna (such as but not limited to the antenna 27 of the device 10A).


The light sources 23B may be connected to the power source 85 by suitable electrically conducting wires 90 disposed within the housing 78. Alternatively, the light sources 23B may receive power and may be controlled by being coupled to the CMOS imager 54 (connections are not shown).


The power source 85 may be any suitable electrical power source, including but not limited to a mains operated power source, a battery, or the like. Alternatively, the power source 85 may be disposed within the housing 78 (not shown).


Other alternative designs of the optical arrangement in the device 80 may also be used. For example, the device 80 may have an optical aperture (such as, for example, the optical aperture 51 of FIG. 7) and an optical system (such as, for example, the optical system 65A of FIG. 7) disposed between the part 54A of the CMOS imager 54 and the mirror 55.


If the angle α is smaller than 45°, and a correction is needed for the distortion in the image, the distortion may be corrected by the optical system 22C (or by the optical system disposed between the part 54A of the CMOS imager 54 and the mirror 55, if the alternative optical arrangement is being used). Alternatively, image distortion may be corrected computationally by suitably processing the image data in a post-acquisition step, as is known in the art.


The segregated circuitry design of the CMOS imager 54 and the use of the mirror 55, enable the reduction of the transverse cross sectional area of the device 80 as explained in detail hereinabove for the devices 50 and 60 of FIGS. 6 and 7 respectively.


It is noted that, for the sake of simplicity of illustration, the insertable device 80 is illustrated as having imaging capabilities only. The device 80 may for example be inserted into a working channel of an endoscope (not shown) and may be used for insertion into and imaging of narrow cavities into which the endoscope cannot be inserted. However, The device 80 itself may include one or more working channels (not shown in FIG. 8 for the sake of clarity of illustration). The working channels may be used for performing insufflation and/or irrigation as is known in the art. Additionally, one or more instruments may be inserted into such working channels for performing a variety of surgical operations. Such instruments may includes but are not limited to scissors, other surgical blades, wire snares, banding devices, surgical laser devices, cauterizers, or any other surgical or diagnostic instrument known in the art and insertable through such a working channel. Thus, if the device 80 includes one or more working channels, the device 80 may be operated as an endoscopy device as is known in the art.


While the invention has been described with respect to a limited number of embodiments, it will be appreciated that many variations, modifications and other applications of the invention may be made which are within the scope and spirit of the invention.

Claims
  • 1. An in vivo imaging device, comprising: a capsule shaped housing having a longitudinal axis and an optical window at an end of the housing;a CMOS image sensor comprising a pixel array portion and a circuitry portion, wherein said pixel array portion is disposed within said housing substantially parallel to said longitudinal axis;a light deflecting element for deflecting received light to said pixel array portion, wherein the angle between said light deflecting element and said pixel array portion is smaller than 45° thereby distorting an image projected onto said pixel array portion; anda mechanism to compensate for said distortion.
  • 2. The device of claim 1 wherein the pixel array portion is continuous with the circuitry portion.
  • 3. The device of claim 1 wherein the circuitry portion comprises circuitry selected from the group consisting of timing circuitry, analog to digital conversion circuitry, input/output (I/O) circuitry, transmitting circuitry or a combination thereof.
  • 4. The device of claim 1, wherein the light deflecting element is a mirror.
  • 5. The device of claim 1, wherein said mechanism to compensate for said distortion includes an optical system configured to change an image to compensate for distortion of an image by the light deflecting element before the image reaches said pixel array portion.
  • 6. The device of claim 5, wherein said optical system comprises a lens within said housing, configured to collect light rays coming through said optical window.
  • 7. The device of claim 1 further comprising an aperture for reflected light to progress through.
  • 8. The device of claim 1,further comprising a plurality of light emitting devices.
  • 9. The device of claim 1, wherein said mechanism to compensate for said distortion comprises a workstation for processing distorted image data after acquisition.
  • 10. The device of claim 1, comprising an optical system, wherein the optical system is substantially parallel to said longitudinal axis.
  • 11. The device of claim 1, comprising an optical system, wherein the optical system is disposed between the image sensor and the light deflecting element.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 10/482,218, now abandoned, which was filed Dec. 29, 2003, as a National Phase Application of PCT International Application No. PCT/IL02/00526, International Filing Date Jun. 27, 2002, which claims priority of U.S. Provisional Patent Application No. 60/301,141, filed Jun. 28, 2001, all of which are incorporated herein by reference in their entireties.

US Referenced Citations (119)
Number Name Date Kind
3683389 Hollis Aug 1972 A
3888237 Mori Jun 1975 A
3971362 Pope et al. Jul 1976 A
4217045 Ziskind Aug 1980 A
4253447 Moore et al. Mar 1981 A
4278077 Mizumoto Jul 1981 A
4425117 Hugemann et al. Jan 1984 A
4467361 Ohno et al. Aug 1984 A
4622954 Arakawa et al. Nov 1986 A
4689621 Kleinberg Aug 1987 A
4692608 Cooper et al. Sep 1987 A
4710807 Chikama et al. Dec 1987 A
4741327 Yabe May 1988 A
4809680 Yabe et al. Mar 1989 A
4832003 Yabe May 1989 A
4844076 Lesho et al. Jul 1989 A
4857724 Snoeren et al. Aug 1989 A
4858002 Zobel et al. Aug 1989 A
4882619 Hasegawa et al. Nov 1989 A
4890159 Ogiu Dec 1989 A
4915113 Holman Apr 1990 A
4918521 Yabe et al. Apr 1990 A
4986642 Yokota et al. Jan 1991 A
5108407 Geremia et al. Apr 1992 A
5184223 Mihara et al. Feb 1993 A
5191879 Krauter Mar 1993 A
5216512 Bruijns et al. Jun 1993 A
5279607 Schentag et al. Jan 1994 A
5381784 Adair Jan 1995 A
5421337 Richards-Kortum et al. Jun 1995 A
5444574 Ono et al. Aug 1995 A
5450243 Nishioka et al. Sep 1995 A
5494483 Adair et al. Feb 1996 A
5547455 McKenna et al. Aug 1996 A
5598205 Nishioka et al. Jan 1997 A
5603687 Hori et al. Feb 1997 A
5604531 Iddan et al. Feb 1997 A
5723844 Dow et al. Mar 1998 A
5734418 Danna Mar 1998 A
5776050 Chen et al. Jul 1998 A
5782771 Hussman Jul 1998 A
5797837 Minami Aug 1998 A
5819736 Avny et al. Oct 1998 A
5833603 Kovacs et al. Nov 1998 A
5853005 Scanlon Dec 1998 A
5904647 Ouchi May 1999 A
5905597 Mizouchi et al. May 1999 A
5929901 Adair et al. Jul 1999 A
5938585 Donofrio Aug 1999 A
5940126 Kimura et al. Aug 1999 A
5980453 Forkey et al. Nov 1999 A
5986693 Adair et al. Nov 1999 A
5993378 Lemelson Nov 1999 A
5999844 Gombrich et al. Dec 1999 A
6010453 Fiddian-Green Jan 2000 A
6019721 Holmes et al. Feb 2000 A
6034823 Togino Mar 2000 A
6043839 Adair et al. Mar 2000 A
6074349 Crowley Jun 2000 A
6100920 Miller et al. Aug 2000 A
6142930 Ito et al. Nov 2000 A
6165128 Cespedes et al. Dec 2000 A
6204524 Rhodes Mar 2001 B1
6206825 Tsuyuki et al. Mar 2001 B1
6222620 Jung et al. Apr 2001 B1
6240312 Alfano et al. May 2001 B1
6266550 Selmon et al. Jul 2001 B1
6281506 Fujita et al. Aug 2001 B1
6285897 Kilcoyne et al. Sep 2001 B1
6458074 Matsui et al. Oct 2002 B1
6475145 Baylor Nov 2002 B1
6490490 Uchikubo et al. Dec 2002 B1
6527753 Sekine et al. Mar 2003 B2
6549796 Sohrab Apr 2003 B2
6579311 Makower Jun 2003 B1
6612982 Ouchi Sep 2003 B1
6632175 Marshall Oct 2003 B1
6648814 Kim et al. Nov 2003 B2
6659940 Adler et al. Dec 2003 B2
6692432 Yarush et al. Feb 2004 B1
6709387 Glukhovsky et al. Mar 2004 B1
6719684 Kim et al. Apr 2004 B2
6951536 Yokoi et al. Oct 2005 B2
7009634 Iddan et al. Mar 2006 B2
7044908 Montalbo et al. May 2006 B1
7122001 Uchiyama et al. Oct 2006 B2
7160258 Imran et al. Jan 2007 B2
7241262 Adler et al. Jul 2007 B2
7261728 Long et al. Aug 2007 B2
20010035902 Iddan et al. Nov 2001 A1
20010040211 Nagaoka et al. Nov 2001 A1
20010049497 Kalloo et al. Dec 2001 A1
20010051766 Gazdzinski Dec 2001 A1
20020042562 Meron et al. Apr 2002 A1
20020103417 Gazdzinski Aug 2002 A1
20020109774 Meron et al. Aug 2002 A1
20020138009 Brockway et al. Sep 2002 A1
20020156347 Kim et al. Oct 2002 A1
20020165589 Imran et al. Nov 2002 A1
20030013370 Glukhovsky Jan 2003 A1
20030020810 Takizawa et al. Jan 2003 A1
20030023150 Yokai et al. Jan 2003 A1
20030092964 Kim et al. May 2003 A1
20030117491 Avni et al. Jun 2003 A1
20030130562 Barbato et al. Jul 2003 A1
20030167000 Mullick et al. Sep 2003 A1
20030174409 Nagaoka et al. Sep 2003 A1
20030214726 Mihara Nov 2003 A1
20040073087 Glukhovsky et al. Apr 2004 A1
20040092825 Madar et al. May 2004 A1
20040097791 Tokuda et al. May 2004 A1
20040133076 Kobayashi et al. Jul 2004 A1
20040153008 Sharf et al. Aug 2004 A1
20040176664 Iddan Sep 2004 A1
20040225189 Kimoto et al. Nov 2004 A1
20050143624 Iddan Jun 2005 A1
20060004255 Iddan et al. Jan 2006 A1
20060167339 Gilad et al. Jul 2006 A1
20070270651 Gilad et al. Nov 2007 A1
Foreign Referenced Citations (33)
Number Date Country
34 40 177 May 1986 DE
0967656 Dec 1999 EP
57-45833 Mar 1982 JP
59129819 Jul 1984 JP
61059308 Mar 1986 JP
63-070820 Mar 1988 JP
63-226615 Sep 1988 JP
2036835 Feb 1990 JP
3-289779 Dec 1991 JP
4109927 Apr 1992 JP
1992-144533 May 1992 JP
4-180736 Jun 1992 JP
5015515 Jan 1993 JP
06-114036 Apr 1994 JP
6142081 May 1994 JP
7289504 Nov 1995 JP
09327447 Dec 1997 JP
10-65131 Mar 1998 JP
11326786 Nov 1999 JP
2001-112740 Apr 2001 JP
2001-170002 Jun 2001 JP
2001224553 Aug 2001 JP
WO 9221307 Dec 1992 WO
WO 9811816 Mar 1998 WO
WO 0108548 Feb 2001 WO
WO 0150941 Jul 2001 WO
WO 0153792 Jul 2001 WO
WO 0165995 Sep 2001 WO
WO 0187377 Nov 2001 WO
WO 0226103 Apr 2002 WO
WO 03090618 Nov 2003 WO
WO 2004028336 Apr 2004 WO
WO 2004058041 Jul 2004 WO
Related Publications (1)
Number Date Country
20080200757 A1 Aug 2008 US
Provisional Applications (1)
Number Date Country
60301141 Jun 2001 US
Continuations (1)
Number Date Country
Parent 10482218 US
Child 11892815 US