Scope sensing in a light controlled environment

Information

  • Patent Grant
  • 11185213
  • Patent Number
    11,185,213
  • Date Filed
    Monday, January 7, 2019
    5 years ago
  • Date Issued
    Tuesday, November 30, 2021
    2 years ago
Abstract
The disclosure extends to methods, systems, and computer program products for detecting whether an illumination source of an endoscope is in use (inside the body of a patient) versus not in use (outside the body of a patient). The disclosure relies on the fact that the working environment is lit solely by the endoscope and its components. Thus, communication between the illumination or light source controller and the imaging device, such as a surgical camera, is required. When the illumination or light source is turned off and the endoscope is outside the body, a sensor will detect ambient light alerting the illumination source controller that it is outside the body, which then keeps the illumination source off or at a low intensity level. Conversely, when the illumination source is turned off and the endoscope is inside the body, the sensor will not detect any light (or will detect only a very low level of light). Based on this logic, if the imaging device, such as a camera, knows that the light is off during a specific period of time the frame(s) from that time period can be analyzed and the level of light gathered in the frame(s) will show the scope location either inside or outside of the body.
Description
STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT

Not Applicable.


BACKGROUND

Current minimally invasive surgical procedures rely on endoscopes for visualization of the surgical site. In the arthroscopy, laparoscopy, urology, gynecology, and ENT (ear, nose, and throat) specialties, rigid endoscopes are primarily used. A rigid endoscope is constructed of an inner lumen containing multiple glass lens elements for visualization and an outer lumen containing a bundle of fiber optic strands for carrying light from a light source to the surgical site.


Conventional surgical light systems are very inefficient. From the light engine, which is typically a metal halide bulb, halogen bulb, xenon bulb, or LED(s) (light emitting diode), to the surgical site over ninety-five percent of the light is lost. These losses occur at multiple locations, the first being at the optic placed in front of the light engine to gather the light from a wide dispersion angle and focus it into a collimated beam with a diameter small enough to transmit to a fiber optic light cable. The second loss point is the junction of the focusing optic and the aforementioned fiber optic light cable. The fiber optic light cable is a bundle, typically with a diameter of five millimeters, of small fiber optic strands and measures one to three meters in length. The third loss point is over the length of the fiber bundle due to the attenuation rate of the bulk fiber strands. The fiber optic light cable transmits light from the light source to the endoscope in sterile field. The fourth loss point is the junction between the light cable and the proximal end of the endoscope.


Due to the losses in the light transmission path, the light source must generate a significant amount of light. This results in a significant amount of heat generated, particularly at each of the junction points and at the distal tip of the scope. The heat generated, specifically at the distal scope tip and at the junction between the light cable and scope, can present a safety risk to the surgical patient. The heat is such that if the scope is inadvertently rested on the patient for a period of time, a burn can occur. This is an issue with all conventional light sources and every year a few such incidents occur and are reported to the FDA (Food and Drug Administration).





BRIEF DESCRIPTION OF THE DRAWINGS

Non-limiting and non-exhaustive implementations of the present disclosure are described with reference to the following figures, wherein like reference numerals refer to like parts throughout the various views unless otherwise specified. Advantages of the present disclosure will become better understood with regard to the following description and accompanying drawings where:



FIG. 1 illustrates a flow chart of an example method for controlling power to an endoscopic light source in a light controlled environment according to one implementation;



FIG. 2 illustrates an imaging device located outside of a light deficient environment, such as a patient's body, wherein the light source is turned off due to the amount of ambient light present in accordance with the system and method described herein;



FIG. 3 illustrates an imaging device located in a light deficient environment, such as a patient's body, wherein the light source is turned on due to the lack of ambient light present in accordance with the system and method described herein;



FIG. 4 illustrates a light threshold value and comparison chart of an example method according to one implementation;



FIG. 5 illustrates an interval in a constant light system and a pulsed light system wherein a frame is captured for analysis in accordance with the system and method described herein;



FIG. 6 illustrates a system for controlling power to an endoscopic light source in a light controlled environment according to one implementation;



FIGS. 7A and 7B illustrate a perspective view and a side view, respectively, of an implementation of a monolithic sensor having a plurality of pixel arrays for producing a three dimensional image in accordance with the teachings and principles of the disclosure;



FIGS. 8A and 8B illustrate a perspective view and a side view, respectively, of an implementation of an imaging sensor built on a plurality of substrates, wherein a plurality of pixel columns forming the pixel array are located on the first substrate and a plurality of circuit columns are located on a second substrate and showing an electrical connection and communication between one column of pixels to its associated or corresponding column of circuitry; and



FIGS. 9A and 9B illustrate a perspective view and a side view, respectively, of an implementation of an imaging sensor having a plurality of pixel arrays for producing a three dimensional image, wherein the plurality of pixel arrays and the image sensor are built on a plurality of substrates.





DETAILED DESCRIPTION

The disclosure extends to methods, systems, and computer program products for detecting whether an endoscopic illumination or light source is in use (inside the body of a patient) versus not in use (outside the body of a patient). The methods, systems and computer program products rely on the fact that the working environment is lit solely by the endoscope and its components. Thus, communication between the illumination or light source controller and the imaging device, such as a surgical camera, is required. In the following description of the present disclosure, reference is made to the accompanying drawings, which form a part hereof, and in which is shown by way of illustration specific implementations in which the disclosure may be practiced. It is understood that other implementations may be utilized and structural changes may be made without departing from the scope of the present disclosure.


It must be noted that, as used in this specification and the appended claims, the singular forms “a,” “an,” and “the” include plural referents unless the context clearly dictates otherwise.


As used herein, the terms “comprising,” “including,” “containing,” “characterized by,” and grammatical equivalents thereof are inclusive or open-ended terms that do not exclude additional, unrecited elements or method steps.


Further, where appropriate, functions described herein can be performed in one or more of: hardware, software, firmware, digital components, or analog components. For example, one or more application specific integrated circuits (ASICs) can be programmed to carry out one or more of the systems and procedures described herein. Certain terms are used throughout the following description and Claims to refer to particular system components. As one skilled in the art will appreciate, components may be referred to by different names. This document does not intend to distinguish between components that differ in name, but not function.


For safety and power consumption reasons, what is needed are methods and systems for detecting when an illumination or light source is in use and when it is not in use. As will be seen, the disclosure provides methods and systems that can do this in an efficient and elegant manner.


Referring now to the figures, it will be appreciated that the disclosure relates to a detection mechanism for operating the illumination source when an endoscope is in use (inside the body of a patient) versus not in use (outside the body of a patient). The disclosure relies on the fact that the working environment is lit solely by the endoscope and its components. Thus, communication between the illumination or light source controller and the imaging device, such as a surgical camera, is required.


For safety reasons it is preferable to have the light source off while the endoscope is not in use. This removes the risk of burning a patient if, for example, the user inadvertently leaves the endo scope resting on the patient while performing other tasks. Every year there are reported cases of patient burns resulting from such misuse of conventional endoscopic video systems.


When the light is turned off and the endoscope is outside the body, the sensor will detect ambient light. Conversely, when the light is turned off and the endoscope is inside the body, the sensor will not detect any light (or will detect only a very low level of light). Based on this logic, if the camera knows that the light is off during a specific period of time the frame(s) from that time period can be analyzed and the level of light gathered in the frame(s) will show the scope location.


Knowing the location of the scope (inside or outside the body) allows the system to keep the light source off while outside the body and only turn the light source on when the endoscope is put into the body for use.


Alternately, the light source output intensity can be reduced to a low, safe level while the scope is outside the body and then increased to a high level when inside the body and in use. This implementation may be preferred for usability reasons. Users who are not familiar with the system described herein may suspect a functional problem with the system if the light source is completely off while the scope is not in use.


Referring now to FIG. 1, there is illustrated a method for controlling power to an endoscopic light source in a light controlled environment. As illustrated in the figure, at 110 the method may comprise sending an electric communication signal from an imaging device controller to a light source controller at a specified interval. It will be appreciated that the signal may be an instruction to turn off the light source for a predetermined sample period during which a single sample frame or a plurality of sample frames may be collected from an image sensor. At 120, the imaging device controller may receive a communication or data from the image sensor. Data may be collected from the image sensor for a single frame or a plurality of frames obtained during the predetermined sample period. The data may be related to an amount of ambient light received by the image sensor. At 130, the imaging device controller may analyze the sample frame or plurality of frames against a predetermined or predefined light threshold value. Analyzing the single frame or plurality of frames obtained during the predetermined sample period against a specified, predetermined threshold value is also illustrated in FIG. 4. The operation of the light source may be controlled based on the data received from the image sensor.


At 140, a determination is made by the image device controller. If the ambient light is above or below the predetermined light threshold value, then one of two processes may be followed. Specifically, if the measured light from the image sensor is determined to be below the predetermined light threshold value, then at 150 it is determined that the image sensor is in a light deficient environment. When it is determined that the light source is in a light deficient environment, that determination signifies that the imaging device is in-use. At 152, the light source remains in an operable state, thereby providing light to the light deficient environment. At 154, the light source may be turned off for a predetermined sample period and the process starts over again.


At 140, if the measured light from the image sensor is determined to be above the predetermined light threshold value, then at 160 it is determined that the image sensor is not in-use because it is outside of a light deficient environment. In such a circumstance, at 162, the light source is turned off, thereby providing a safety mechanism for controlling power to the light source. It will be appreciated that in one implementation, at 164, the turned off state may be a complete power down of the light source. In another implementation, at 166, the turned off state may be a reduction in power to the light source, such that the light source is only emitting a small amount of light energy. As noted previously, the method may include sampling at a plurality intervals, such as a second interval, to determine whether data received from the image sensor regarding a single frame is above or below the predetermined light threshold value.


Referring now to FIGS. 2 and 3, it will be appreciated that in one implementation, the default mode when the endoscopic light source system starts-up is for the light source to be turned off or in an off state. At this time, the imaging device controller is in communication with the light source controller and knows the light source is turned off or in an off state. At specified intervals, such as for example every 30th frame, the imaging device captures a single frame and analyzes the light level against the predetermined light threshold value. It is to be understood that the specified intervals may be at any frequency that produces the desired functional result. Further, it is to be understood that the light threshold value may be defined as an amount of total light received by the image sensor; or the threshold value may be defined as an average of an amount of light received per pixel on the image sensor.


If the light source is determined to be not in-use as illustrated in FIG. 2, then the imaging device controller communicates information, instructions or data to the light source controller to remain off or in an off state. Conversely, if the light source is determined to be in-use as illustrated in FIG. 3, then the imaging device controller communicates information, instructions or data to the light source controller to turn on. Once the light source is turned on, a new pattern begins. Thus, at predetermined, specified intervals the light source is turned off for a predetermined sample period during which time the imaging device captures a single frame and analyzes the light level against the predetermined threshold value. It is understood that the sample period may be any length that is long enough for the imaging device to capture one frame, but short enough that is does not negatively affect video quality or user experience. As illustrated best in FIG. 4, if the data received from the image sensor is below the predetermined threshold value, then the imaging device recognizes the light source as being in-use in a light deficient environment, and the imaging device controller communicates with the light source controller to turn on. Whereas, if the data received from the image sensor is above the predetermined threshold value, then the image device recognizes the light source as being not-in-use and is outside the light deficient environment, and the imaging device controller communicates with the light source controller to remain off.


Referring now to FIG. 5, the light source may be a pulsed light system. In an implementation, the light in the pulsed light system may be obtained from laser light. In an implementation, the light in the pulsed light system may be obtained from one or more light emitting diodes. In another implementation, the light source may be a constant light system.


It will be appreciated that the sampling interval may be every 30th frame as described above, or it may be any other frequency that provides the desired results. It is within the scope of the disclosure for the interval frequency may be different during the “in-use” condition and the “not-in-use” condition.


In an implementation, the imaging device, such as a camera, may provide constant control over the light source. In an implementation, the light source may have a default state that is changed by the imaging device as required.


The method and system of the disclosure may require communication between the light source controller and the imaging device controller. The disclosure also contemplates use of a light source with a response time that is fast enough that the “off” pulse during the sample period, during the “in-use” condition, does not adversely affect the video quality. LED and laser light sources may be used, while a metal halide bulb, halogen bulb, or xenon bulb may not be used in this implementation.


During use, the light source can be kept on constantly with a periodic “off” pulse or the light source can be pulsed “on” during normal use, illustrated best in FIG. 5, with an “on” pulse skipped for the black frame analysis.


In an implementation, the light intensity level can be reduced to a predetermined safe level while in the “not-in-use” state. In this implementation the default mode on startup could be a low light intensity level that poses no risk of burning. Then, as previously described, at predetermined intervals the light is turned off for the sample period and this sample frame is analyzed. If the result is “not-in-use”, the light is turned back on at the previous safe level and the pattern repeats. If the result is “in-use”, the light is turned on at the higher functional level.


In an implementation, the light could be pulsed light of a particular colors (including, but not limited to, RBG or YCbCr) rather than white light. In this implementation it may be desirable to change from pulsed colored light while “in-use” to pulsed or constant white light while “not-in-use” using the same techniques previously described. The default mode on startup could be a low level of pulsed or constant white light. Then, as previously described, at predetermined intervals the light is turned off for the sample period and this sample frame is analyzed. If the result is “not-in-use”, the white light is turned back on at the previous safe level and the pattern repeats. If the result is “in-use”, the pulsed color pattern is initiated.


In an implementation, the system may be comprised of a light source that is kept in a constant on-state with a mechanical shutter providing the periodic black frame. This shutter may be controlled by the imaging device, such that there would be no imaging device control of the light source needed. This shutter could be placed at any interface in the light path from the source to the distal tip of the endoscope. In this implementation there is no restriction on light source technology because there is no requirement for the light source to have a fast response time. Instead, the mechanical shutter requires a response time that is fast enough that the “off” pulse during the sample period, during the “in-use” condition, does not adversely affect the video quality.


In any implementation, a visual or audible signal could be given to inform the user of whether the system is in the “in-use” or “not-in-use” state. Alternately, the signal could inform the user when the state changes from “in-use” to “not-in-use” or from “not-in-use” to “in-use” or both.


A black frame would disrupt the video output. During image processing, the black frame can be removed and the previous frame can be displayed in its place. Conversely, multiple frames before and/or after the black frame can be used to construct a substitute frame.


Referring now to FIG. 6, a system for controlling power to an endoscopic light source in a light controlled environment is illustrated. The system may comprise an imaging device 200 comprising an imaging device controller 220, a light source comprising a light source controller 230; and an image sensor 240. It will be appreciated that the imaging device controller may cause the system to perform the following processes: send an electric communication signal to a light source controller at a specified interval; turn off the light source for a predetermined sample period based on the electric communication signal; collect data from the image sensor for a single frame obtained during the predetermined sample period, wherein the data relates to an amount of ambient light received by the image sensor; analyze the single frame obtained during the predetermined sample period against a specified, predetermined threshold value; and control the operation of the light source based on the data received from the image sensor.


Referring now to FIGS. 7A and 7B, the figures illustrate a perspective view and a side view, respectively, of an implementation of a monolithic sensor 700 having a plurality of pixel arrays for producing a three dimensional image in accordance with the teachings and principles of the disclosure. Such an implementation may be desirable for three dimensional image capture, wherein the two pixel arrays 702 and 704 may be offset during use. In another implementation, a first pixel array 702 and a second pixel array 704 may be dedicated to receiving a predetermined range of wave lengths of electromagnetic radiation, wherein the first pixel array 702 is dedicated to a different range of wave length electromagnetic radiation than the second pixel array 704.



FIGS. 8A and 8B illustrate a perspective view and a side view, respectively, of an implementation of an imaging sensor 800 built on a plurality of substrates. As illustrated, a plurality of pixel columns 804 forming the pixel array are located on the first substrate 802 and a plurality of circuit columns 808 are located on a second substrate 806. Also illustrated in the figure are the electrical connection and communication between one column of pixels to its associated or corresponding column of circuitry. In one implementation, an image sensor, which might otherwise be manufactured with its pixel array and supporting circuitry on a single, monolithic substrate/chip, may have the pixel array separated from all or a majority of the supporting circuitry. The disclosure may use at least two substrates/chips, which will be stacked together using three-dimensional stacking technology. The first 802 of the two substrates/chips may be processed using an image CMOS process. The first substrate/chip 802 may be comprised either of a pixel array exclusively or a pixel array surrounded by limited circuitry. The second or subsequent substrate/chip 806 may be processed using any process, and does not have to be from an image CMOS process. The second substrate/chip 806 may be, but is not limited to, a highly dense digital process in order to integrate a variety and number of functions in a very limited space or area on the substrate/chip, or a mixed-mode or analog process in order to integrate for example precise analog functions, or a RF process in order to implement wireless capability, or MEMS (Micro-Electro-Mechanical Systems) in order to integrate MEMS devices. The image CMOS substrate/chip 802 may be stacked with the second or subsequent substrate/chip 806 using any three-dimensional technique. The second substrate/chip 806 may support most, or a majority, of the circuitry that would have otherwise been implemented in the first image CMOS chip 802 (if implemented on a monolithic substrate/chip) as peripheral circuits and therefore have increased the overall system area while keeping the pixel array size constant and optimized to the fullest extent possible. The electrical connection between the two substrates/chips may be done through interconnects 803 and 805, which may be wirebonds, bump and/or TSV (Through Silicon Via).



FIGS. 9A and 9B illustrate a perspective view and a side view, respectively, of an implementation of an imaging sensor 900 having a plurality of pixel arrays for producing a three dimensional image. The three dimensional image sensor may be built on a plurality of substrates and may comprise the plurality of pixel arrays and other associated circuitry, wherein a plurality of pixel columns 904a forming the first pixel array and a plurality of pixel columns 904b forming a second pixel array are located on respective substrates 902a and 902b, respectively, and a plurality of circuit columns 908a and 908b are located on a separate substrate 906. Also illustrated are the electrical connections and communications between columns of pixels to associated or corresponding column of circuitry.


It will be appreciated that the teachings and principles of the disclosure may be used in a reusable device platform, a limited use device platform, a re-posable use device platform, or a single-use/disposable device platform without departing from the scope of the disclosure. It will be appreciated that in a re-usable device platform an end-user is responsible for cleaning and sterilization of the device. In a limited use device platform the device can be used for some specified amount of times before becoming inoperable. Typical new device is delivered sterile with additional uses requiring the end-user to clean and sterilize before additional uses. In a re-posable use device platform a third-party may reprocess the device (e.g., cleans, packages and sterilizes) a single-use device for additional uses at a lower cost than a new unit. In a single-use/disposable device platform a device is provided sterile to the operating room and used only once before being disposed of.


Additionally, the teachings and principles of the disclosure may include any and all wavelengths of electromagnetic energy, including the visible and non-visible spectrums, such as infrared (IR), ultraviolet (UV), and X-ray.


The foregoing description has been presented for the purposes of illustration and description. It is not intended to be exhaustive or to limit the disclosure to the precise form disclosed. Many modifications and variations are possible in light of the above teaching. Further, it should be noted that any or all of the aforementioned alternate implementations may be used in any combination desired to form additional hybrid implementations of the disclosure.


Further, although specific implementations of the disclosure have been described and illustrated, the disclosure is not to be limited to the specific forms or arrangements of parts so described and illustrated. The scope of the disclosure is to be defined by the claims appended hereto, any future claims submitted here and in different applications, and their equivalents.

Claims
  • 1. A system for controlling power to an endoscopic light source in a light controlled environment comprising: an imaging device comprising an imaging device controller;a light source comprising a light source controller;an image sensor;wherein the imaging device controller causes the system to perform the following processes:send an electric communication signal to a light source controller at a specified interval;turn off the light source for a predetermined sample period based on the electric communication signal, the predetermined sample period having a length sufficient for the image sensor to capture at least a single frame;collect data from the image sensor for the single frame obtained during the predetermined sample period, wherein the data relates to an amount of ambient light received by the image sensor;analyze the single frame obtained during the predetermined sample period against a specified, predetermined threshold value; andcontrol the operation of the light source based on the data received from the image sensor.
  • 2. The system of claim 1, wherein if the data received from the image sensor is below the predetermined threshold value, then the light source is recognized as being in-use by the imaging device controller and is in a light deficient environment.
  • 3. The system of claim 2, wherein the light source remains in an operable state to thereby provide light to the light deficient environment when the data received from the image sensor is below the predetermined threshold value.
  • 4. The system of claim 1, wherein if the data received from the image sensor is above the predetermined threshold value, then the light source is recognized as being not in-use by the imaging device controller and is outside of a light deficient environment.
  • 5. The system of claim 4, wherein the light source is turned off, thereby providing a safety mechanism for controlling power to the light source.
  • 6. The system of claim 1, wherein a default mode when the endoscopic light source system starts up is for the light source to be turned on.
  • 7. The system of claim 6, wherein at predetermined, specified intervals the imaging device controller communicates with the light source controller to turn off the light source for a predetermined sample period, the imaging device captures a single frame and analyzes the light level against the predetermined threshold value.
  • 8. The system of claim 7, wherein if the data received from the image sensor is below the predetermined threshold value, then the imaging device recognizes the light source as being in-use in a light deficient environment, and the imaging device controller communicates with the light source controller to return to the default mode.
  • 9. The system of claim 7, wherein if the data received from the image sensor is above the predetermined threshold value, then the imaging device recognizes the light source as being not-in-use and is outside the light deficient environment, and the imaging device controller communicates with the light source controller to remain off.
  • 10. The system of claim 1, wherein a default mode when the endoscopic light source system starts up is for the light source to be turned off.
  • 11. The system of claim 10, wherein the imaging device controller is in communication with the light source controller.
  • 12. The system of claim 11, wherein at specified intervals the imaging device captures the single frame and analyzes the light level against the predetermined threshold value.
  • 13. The system of claim 12, wherein the predetermined threshold value is an amount of total light received by the image sensor.
  • 14. The system of claim 12, wherein the predetermined threshold value is an average of an amount of light received per pixel.
  • 15. The system of claim 10, wherein if the light source is determined to be not in-use then the imaging device controller communicates to the light source controller to remain off.
  • 16. The system of claim 10, wherein if the light source is determined to be in-use then the imaging device controller communicates to the light source controller to turn on.
  • 17. The system of claim 16, wherein once the light source is turned on, a light pulsing pattern begins, such that at predetermined, specified intervals the light source is turned off for a predetermined sample period, the imaging device captures a single frame and analyzes the light level against the predetermined threshold value.
  • 18. The system of claim 17, wherein if the data received from the image sensor is below the predetermined threshold value, then the imaging device recognizes the light source as being in-use in a light deficient environment, and the imaging device controller communicates with the light source controller to turn on.
  • 19. The system of claim 17, wherein if the data received from the image sensor is above the predetermined threshold value, then the imaging device recognizes the light source as being not-in-use and is outside the light deficient environment, and the imaging device controller communicates with the light source controller to remain off.
  • 20. The system of claim 1, wherein the imaging device controller communicating with the system provides a user with information regarding the current state of the system.
  • 21. The system of claim 20, wherein the information is provided visually.
  • 22. The system of claim 20, wherein the information is provided audibly.
  • 23. The system of claim 1, wherein the imaging device controller communicating with the system provides a user with information regarding a change in the state of the system.
  • 24. The system of claim 23, wherein the information is provided visually.
  • 25. The system of claim 23, wherein the information is provided audibly.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a division of U.S. application Ser. No. 14/214,334, filed Mar. 14, 2014, and claims the benefit of U.S. Provisional Application No. 61/791,685, filed Mar. 15, 2013, which are hereby incorporated herein by reference in their entireties, including but not limited to those portions that specifically appear hereinafter, the incorporation by reference being made with the following exception: In the event that any portion of the above-referenced applications is inconsistent with this application, this application supersedes said above-referenced applications.

US Referenced Citations (501)
Number Name Date Kind
3666885 Hemsley et al. May 1972 A
4011403 Epstein et al. Mar 1977 A
4356534 Hattori Oct 1982 A
4363963 Ando Dec 1982 A
4433675 Konoshima Feb 1984 A
4436095 Kruger Mar 1984 A
4473839 Noda Sep 1984 A
4644403 Sakai et al. Feb 1987 A
4651226 Motoori et al. Mar 1987 A
4692606 Sakai et al. Sep 1987 A
4740837 Yanagisawa et al. Apr 1988 A
4741327 Yabe May 1988 A
4742388 Cooper et al. May 1988 A
4745471 Takamura et al. May 1988 A
4773396 Okazaki Sep 1988 A
4782386 Ams et al. Nov 1988 A
4786965 Yabe Nov 1988 A
4832003 Yabe May 1989 A
4845555 Yabe et al. Jul 1989 A
4853772 Kikuchi Aug 1989 A
4853773 Hibino et al. Aug 1989 A
4865018 Kanno Sep 1989 A
4866526 Ams et al. Sep 1989 A
4884133 Kanno Nov 1989 A
4884134 Tsuji et al. Nov 1989 A
4908701 Udagawa Mar 1990 A
4918521 Yabe et al. Apr 1990 A
4924856 Noguchi May 1990 A
4938205 Nudelman Jul 1990 A
4942473 Zeevi et al. Jul 1990 A
4947246 Kikuchi Aug 1990 A
4953539 Nakamura et al. Sep 1990 A
4959710 Uehara et al. Sep 1990 A
4963960 Takami Oct 1990 A
5001556 Nakamura et al. Mar 1991 A
5016975 Sasaki et al. May 1991 A
5021888 Kondou et al. Jun 1991 A
5047846 Uchiyama et al. Sep 1991 A
RE33854 Adair Mar 1992 E
5103497 Hicks Apr 1992 A
5111804 Funakoshi May 1992 A
5133035 Hicks Jul 1992 A
5187572 Nakamura et al. Feb 1993 A
5188094 Adair Feb 1993 A
5196938 Blessinger Mar 1993 A
5200838 Nudelman et al. Apr 1993 A
5220198 Tsuji Jun 1993 A
5228430 Sakamoto Jul 1993 A
5233416 Inoue Aug 1993 A
5241170 Field, Jr. et al. Aug 1993 A
5255087 Nakamura et al. Oct 1993 A
5264925 Shipp et al. Nov 1993 A
5313306 Kuban et al. May 1994 A
5325847 Matsuno Jul 1994 A
5365268 Minami Nov 1994 A
5402768 Adair Apr 1995 A
5408268 Shipp Apr 1995 A
5411020 Ito May 1995 A
5427087 Ito et al. Jun 1995 A
5454366 Ito et al. Oct 1995 A
5494483 Adair Feb 1996 A
5523786 Parulski Jun 1996 A
5550595 Hannah Aug 1996 A
5558841 Nakagawa Sep 1996 A
5594497 Ahern et al. Jan 1997 A
5627584 Nishikori May 1997 A
5658238 Suzuki Aug 1997 A
5665959 Fossum et al. Sep 1997 A
5704836 Norton et al. Jan 1998 A
5730702 Tanaka et al. Mar 1998 A
5734418 Danna Mar 1998 A
5748234 Lippincott May 1998 A
5749830 Kaneko et al. May 1998 A
5754313 Pelchy et al. May 1998 A
5783909 Hochstein Jul 1998 A
5784099 Lippincott Jul 1998 A
5857963 Pelchy et al. Jan 1999 A
5887049 Fossum Mar 1999 A
5924978 Koeda Jul 1999 A
5929901 Adair et al. Jul 1999 A
5949483 Fossum et al. Sep 1999 A
5957834 Mochida Sep 1999 A
5986693 Adair et al. Nov 1999 A
6023315 Harrold et al. Feb 2000 A
6038067 George Mar 2000 A
6043839 Adair et al. Mar 2000 A
6139489 Wampler et al. Oct 2000 A
6141505 Miyata et al. Oct 2000 A
6142930 Ito et al. Nov 2000 A
6166768 Fossum et al. Dec 2000 A
6184922 Saito et al. Feb 2001 B1
6184940 Sano Feb 2001 B1
6215517 Takahashi et al. Mar 2001 B1
6222175 Krymski Apr 2001 B1
6239456 Berezin et al. May 2001 B1
6272269 Naum Aug 2001 B1
6275255 Adair et al. Aug 2001 B1
6292220 Ogawa et al. Sep 2001 B1
6294775 Seibel Sep 2001 B1
6310642 Adair et al. Oct 2001 B1
6320331 Iida et al. Nov 2001 B1
6331156 Haefele et al. Dec 2001 B1
6389205 Muckner May 2002 B1
6416463 Tsuzuki et al. Jul 2002 B1
6429953 Feng Aug 2002 B1
6444970 Barbato Sep 2002 B1
6445022 Barna et al. Sep 2002 B1
6445139 Marshall et al. Sep 2002 B1
6464633 Hosoda et al. Oct 2002 B1
6466618 Messing et al. Oct 2002 B1
6485414 Neuberger Nov 2002 B1
6512280 Chen et al. Jan 2003 B2
6567115 Miyashita et al. May 2003 B1
6627474 Barna et al. Sep 2003 B2
6631230 Campbell Oct 2003 B1
6659940 Adler Dec 2003 B2
6665013 Fossum et al. Dec 2003 B1
6677992 Matsumoto et al. Jan 2004 B1
6687534 Tsujita Feb 2004 B2
6690466 Miller et al. Feb 2004 B2
6692431 Kazakevich Feb 2004 B2
6707499 Kung et al. Mar 2004 B1
6772181 Fu et al. Aug 2004 B1
6773392 Kikuchi et al. Aug 2004 B2
6791739 Ramanujan et al. Sep 2004 B2
6796939 Hirata et al. Sep 2004 B1
6799065 Niemeyer Sep 2004 B1
6809358 Hsieh et al. Oct 2004 B2
6836288 Lewis Dec 2004 B1
6838653 Campbell et al. Jan 2005 B2
6841947 Berg-johansen Jan 2005 B2
6847399 Ang Jan 2005 B1
6856712 Fauver et al. Feb 2005 B2
6873363 Barna et al. Mar 2005 B1
6879340 Chevallier Apr 2005 B1
6899675 Cline et al. May 2005 B2
6900829 Orzawa et al. May 2005 B1
6906745 Fossum et al. Jun 2005 B1
6921920 Kazakevich Jul 2005 B2
6933974 Lee Aug 2005 B2
6947090 Komoro et al. Sep 2005 B2
6961461 MacKinnon et al. Nov 2005 B2
6970195 Bidermann et al. Nov 2005 B1
6977733 Denk et al. Dec 2005 B2
6982740 Adair et al. Jan 2006 B2
6998594 Gaines et al. Feb 2006 B2
6999118 Suzuki Feb 2006 B2
7009634 Iddan et al. Mar 2006 B2
7009648 Lauxtermann et al. Mar 2006 B2
7030904 Adair et al. Apr 2006 B2
7037259 Hakamata et al. May 2006 B2
7068878 Crossman-Bosworth et al. Jun 2006 B2
7071979 Ohtani et al. Jul 2006 B1
7079178 Hynecek Jul 2006 B2
7102682 Baer Sep 2006 B2
7105371 Fossum et al. Sep 2006 B2
7106377 Bean et al. Sep 2006 B2
7119839 Mansoorian Oct 2006 B1
7151568 Kawachi et al. Dec 2006 B2
7159782 Johnston et al. Jan 2007 B2
7184084 Glenn Feb 2007 B2
7189226 Auld et al. Mar 2007 B2
7189961 Johnston et al. Mar 2007 B2
7208983 Imaizumi et al. Apr 2007 B2
7252236 Johnston et al. Aug 2007 B2
7258663 Doguchi et al. Aug 2007 B2
7261687 Yang Aug 2007 B2
7280139 Pahr et al. Oct 2007 B2
7298938 Johnston Nov 2007 B2
7312879 Johnston Dec 2007 B2
7319478 Dolt et al. Jan 2008 B2
7355155 Wang Apr 2008 B2
7356198 Chauville et al. Apr 2008 B2
7365768 Ono et al. Apr 2008 B1
7369140 King et al. May 2008 B1
7369176 Sonnenschein et al. May 2008 B2
7385708 Ackerman Jun 2008 B2
7455638 Ogawa et al. Nov 2008 B2
7470229 Ogawa et al. Dec 2008 B2
7476197 Wiklof et al. Jan 2009 B2
7532760 Kaplinsky et al. May 2009 B2
7540645 Choi May 2009 B2
7544163 MacKinnon et al. Jun 2009 B2
7545434 Bean et al. Jun 2009 B2
7564935 Suzuki Jul 2009 B2
7567291 Bechtel et al. Jul 2009 B2
7573516 Krymski et al. Aug 2009 B2
7573519 Phan et al. Aug 2009 B2
7583872 Seibel et al. Sep 2009 B2
7608807 Hick Oct 2009 B2
7616238 Avni et al. Nov 2009 B2
7630008 Sarwari Dec 2009 B2
7744528 Wallace et al. Jun 2010 B2
7783133 Dunki-Jacobs et al. Aug 2010 B2
7784697 Johnston et al. Aug 2010 B2
7791009 Johnston et al. Sep 2010 B2
7792378 Liege et al. Sep 2010 B2
7794394 Frangioni Sep 2010 B2
7796870 Wang Sep 2010 B2
7813538 Carroll et al. Oct 2010 B2
7901974 Venezia et al. Mar 2011 B2
7914447 Kanai Mar 2011 B2
7916193 Fossum Mar 2011 B2
7935050 Luanava et al. May 2011 B2
7944566 Xie May 2011 B2
7952096 Rhodes May 2011 B2
7969097 Van De Ven Jun 2011 B2
7995123 Lee et al. Aug 2011 B2
8018589 MacKinnon Sep 2011 B2
8040394 Fossum et al. Oct 2011 B2
8054339 Fossum et al. Nov 2011 B2
8059174 Mann et al. Nov 2011 B2
8100826 MacKinnon et al. Jan 2012 B2
8159584 Iwabuchi et al. Apr 2012 B2
8193542 Machara Jun 2012 B2
8194061 Wang Jun 2012 B2
8212884 Seibel et al. Jul 2012 B2
8213698 Wang Jul 2012 B2
8231522 Endo et al. Jul 2012 B2
8300111 Iwane Oct 2012 B2
8372003 St. George Feb 2013 B2
8382662 Soper et al. Feb 2013 B2
8396535 Wang et al. Mar 2013 B2
8423110 Barbato et al. Apr 2013 B2
8471938 Altice, Jr. et al. Jun 2013 B2
8476575 Mokhuatyuk Jul 2013 B2
8482823 Cheng Jul 2013 B2
8493474 Richardson Jul 2013 B2
8493564 Brukilacchio et al. Jul 2013 B2
8523367 Ogura Sep 2013 B2
8537203 Seibel et al. Sep 2013 B2
8559743 Liege et al. Oct 2013 B2
8582011 Dosluoglu Nov 2013 B2
8602971 Farr Dec 2013 B2
8605177 Rossi et al. Dec 2013 B2
8610808 Prescher et al. Dec 2013 B2
8614754 Fossum Dec 2013 B2
8625016 Fossum et al. Jan 2014 B2
8638847 Wang Jan 2014 B2
8648287 Fossum Feb 2014 B1
8649848 Crane et al. Feb 2014 B2
8668339 Kabuki et al. Mar 2014 B2
8668639 Kagaya Mar 2014 B2
8675125 Cossairt et al. Mar 2014 B2
8698887 Makino et al. Apr 2014 B2
8836834 Hashimoto et al. Sep 2014 B2
8848063 Jo et al. Sep 2014 B2
8858425 Farr et al. Oct 2014 B2
8885034 Adair et al. Nov 2014 B2
8941308 Briggs Jan 2015 B2
9182337 Kamee et al. Nov 2015 B2
9349764 Lee et al. May 2016 B1
9509917 Blanquart et al. Nov 2016 B2
9516239 Blanquart et al. Dec 2016 B2
9634878 Bench et al. Apr 2017 B1
9762879 Blanquart et al. Sep 2017 B2
9777913 Talbert et al. Oct 2017 B2
10084944 Henley et al. Sep 2018 B2
10251530 Henley et al. Apr 2019 B2
10277875 Blanquart et al. Apr 2019 B2
20010016064 Tsuruoka et al. Aug 2001 A1
20010017649 Yaron Aug 2001 A1
20010030744 Chang Oct 2001 A1
20010055462 Seibel Dec 2001 A1
20020045801 Niida Apr 2002 A1
20020054219 Jaspers May 2002 A1
20020064341 Fauver et al. May 2002 A1
20020080248 Adair et al. Jun 2002 A1
20020080359 Denk et al. Jun 2002 A1
20020140844 Kurokawa et al. Oct 2002 A1
20020158976 Vni Oct 2002 A1
20020158986 Baer Oct 2002 A1
20030007087 Hakamata et al. Jan 2003 A1
20030007686 Roever Jan 2003 A1
20030107664 Suzuki Jun 2003 A1
20030189663 Dolt et al. Oct 2003 A1
20030189705 Pardo Oct 2003 A1
20040082833 Adler et al. Apr 2004 A1
20040170712 Sadek El Mogy Sep 2004 A1
20040215059 Homan Oct 2004 A1
20050009982 Inagaki et al. Jan 2005 A1
20050027164 Barbato et al. Feb 2005 A1
20050038322 Banik Feb 2005 A1
20050041571 Ichihara et al. Feb 2005 A1
20050052680 Okamura Mar 2005 A1
20050113641 Bala May 2005 A1
20050122530 Denk et al. Jun 2005 A1
20050151866 Ando et al. Jul 2005 A1
20050200291 Naugler, Jr. et al. Sep 2005 A1
20050234302 MacKinnon et al. Oct 2005 A1
20050237384 Jess et al. Oct 2005 A1
20050261552 Mori et al. Nov 2005 A1
20050267329 Konstorum et al. Dec 2005 A1
20050277808 Sonnenschein et al. Dec 2005 A1
20050288546 Sonnenschein et al. Dec 2005 A1
20060038823 Arcas Feb 2006 A1
20060069314 Farr Mar 2006 A1
20060087841 Chern et al. Apr 2006 A1
20060106284 Shouji May 2006 A1
20060197664 Zhang et al. Sep 2006 A1
20060202036 Wang et al. Sep 2006 A1
20060221250 Rossbach et al. Oct 2006 A1
20060226231 Johnston et al. Oct 2006 A1
20060264734 Kimoto et al. Nov 2006 A1
20060274335 Wittenstein Dec 2006 A1
20070010712 Negishi Jan 2007 A1
20070010713 Negishi Jan 2007 A1
20070029629 Yazdi Feb 2007 A1
20070041448 Miller et al. Feb 2007 A1
20070066868 Shikii Mar 2007 A1
20070083085 Birnkrant Apr 2007 A1
20070092283 Sugihara Apr 2007 A1
20070129601 Johnston et al. Jun 2007 A1
20070147033 Ogawa et al. Jun 2007 A1
20070182723 Imai et al. Aug 2007 A1
20070182842 Sonnenschein et al. Aug 2007 A1
20070225560 Avni et al. Sep 2007 A1
20070244364 Luanava et al. Oct 2007 A1
20070244365 Wiklof Oct 2007 A1
20070274649 Takahashi Nov 2007 A1
20070276187 Wiklof et al. Nov 2007 A1
20070279486 Bayer et al. Dec 2007 A1
20070285526 Mann et al. Dec 2007 A1
20070293720 Bayer Dec 2007 A1
20080045800 Farr Feb 2008 A2
20080049132 Suzuki Feb 2008 A1
20080088719 Jacob et al. Apr 2008 A1
20080107333 Mazinani et al. May 2008 A1
20080136953 Barnea et al. Jun 2008 A1
20080158348 Karpen et al. Jul 2008 A1
20080164550 Chen et al. Jul 2008 A1
20080165360 Johnston Jul 2008 A1
20080167523 Uchiyama Jul 2008 A1
20080192131 Kim et al. Aug 2008 A1
20080208077 Iddan Aug 2008 A1
20080218598 Harada et al. Sep 2008 A1
20080218615 Huang et al. Sep 2008 A1
20080218824 Johnston et al. Sep 2008 A1
20080249369 Seibel et al. Oct 2008 A1
20080287742 St. George Nov 2008 A1
20090012361 MacKinnon et al. Jan 2009 A1
20090012368 Banik Jan 2009 A1
20090021588 Border et al. Jan 2009 A1
20090024000 Chen Jan 2009 A1
20090028465 Pan Jan 2009 A1
20090074265 Huang et al. Mar 2009 A1
20090091645 Trimeche et al. Apr 2009 A1
20090137893 Seibel et al. May 2009 A1
20090147077 Tani et al. Jun 2009 A1
20090154886 Lewis et al. Jun 2009 A1
20090160976 Chen et al. Jun 2009 A1
20090189530 Ashdown et al. Jul 2009 A1
20090208143 Yoon et al. Aug 2009 A1
20090227847 Tepper et al. Sep 2009 A1
20090232213 Jia Sep 2009 A1
20090259102 Koninckx et al. Oct 2009 A1
20090268063 Ellis-Monaghan et al. Oct 2009 A1
20090274380 Wedi Nov 2009 A1
20090292168 Farr Nov 2009 A1
20090309500 Reisch Dec 2009 A1
20090316116 Melville et al. Dec 2009 A1
20090322912 Blanquart Dec 2009 A1
20100004513 MacKinnon Jan 2010 A1
20100026722 Kondo Feb 2010 A1
20100049180 Wells et al. Feb 2010 A1
20100069713 Endo et al. Mar 2010 A1
20100102199 Negley et al. Apr 2010 A1
20100121142 OuYang et al. May 2010 A1
20100121143 Sugimoto et al. May 2010 A1
20100123775 Shibasaki May 2010 A1
20100134608 Shibasaki Jun 2010 A1
20100134662 Bub Jun 2010 A1
20100135398 Wittmann et al. Jun 2010 A1
20100137684 Shibasaki et al. Jun 2010 A1
20100149421 Lin et al. Jun 2010 A1
20100157037 Iketani et al. Jun 2010 A1
20100157039 Sugai Jun 2010 A1
20100165087 Corso et al. Jul 2010 A1
20100171429 Garcia Jul 2010 A1
20100182446 Matsubayashi Jul 2010 A1
20100198009 Farr et al. Aug 2010 A1
20100198134 Eckhouse et al. Aug 2010 A1
20100201797 Shizukuishi et al. Aug 2010 A1
20100208056 Olsson et al. Aug 2010 A1
20100228089 Hoffman et al. Sep 2010 A1
20100261961 Scott et al. Oct 2010 A1
20100274082 Iguchi et al. Oct 2010 A1
20100274090 Ozaki et al. Oct 2010 A1
20100305406 Braun et al. Dec 2010 A1
20100309333 Smith et al. Dec 2010 A1
20110028790 Farr et al. Feb 2011 A1
20110034769 Adair et al. Feb 2011 A1
20110051390 Lin et al. Mar 2011 A1
20110063483 Rossi et al. Mar 2011 A1
20110122301 Yamura et al. May 2011 A1
20110149358 Cheng Jun 2011 A1
20110181709 Wright et al. Jul 2011 A1
20110181840 Cobb Jul 2011 A1
20110184239 Wright Jul 2011 A1
20110184243 Wright et al. Jul 2011 A1
20110208004 Feingold Aug 2011 A1
20110212649 Stokoe et al. Sep 2011 A1
20110237882 Saito Sep 2011 A1
20110237884 Saito Sep 2011 A1
20110245605 Jacobsen et al. Oct 2011 A1
20110245616 Kobayashi Oct 2011 A1
20110255844 Wu et al. Oct 2011 A1
20110274175 Sumitomo Nov 2011 A1
20110279679 Samuel et al. Nov 2011 A1
20110288374 Hadani et al. Nov 2011 A1
20110291564 Huang Dec 2011 A1
20110292258 Adler et al. Dec 2011 A1
20110295061 Haramaty et al. Dec 2011 A1
20120004508 McDowall et al. Jan 2012 A1
20120014563 Bendall Jan 2012 A1
20120029279 Kucklick Feb 2012 A1
20120033118 Lee et al. Feb 2012 A1
20120041267 Benning et al. Feb 2012 A1
20120041534 Clerc et al. Feb 2012 A1
20120050592 Oguma Mar 2012 A1
20120078052 Cheng Mar 2012 A1
20120098933 Robinson et al. Apr 2012 A1
20120104230 Eismann et al. May 2012 A1
20120113506 Gmitro et al. May 2012 A1
20120120282 Goris May 2012 A1
20120140302 Xie et al. Jun 2012 A1
20120155761 Matsuoka Jun 2012 A1
20120157774 Kaku Jun 2012 A1
20120172665 Alyn et al. Jul 2012 A1
20120194686 Liu et al. Aug 2012 A1
20120197080 Murayama Aug 2012 A1
20120200685 Kawasaki et al. Aug 2012 A1
20120209071 Bayer et al. Aug 2012 A1
20120242975 Min et al. Sep 2012 A1
20120262621 Sato et al. Oct 2012 A1
20120281111 Jo et al. Nov 2012 A1
20120296238 Chernov Nov 2012 A1
20120319586 Riesebosch Dec 2012 A1
20120327270 Shirakawa et al. Dec 2012 A1
20130016200 Ovod Jan 2013 A1
20130018256 Kislev et al. Jan 2013 A1
20130035545 Ono Feb 2013 A1
20130053642 Mizuyoshi et al. Feb 2013 A1
20130070071 Peltie et al. Mar 2013 A1
20130126708 Blanquart May 2013 A1
20130127934 Chiang May 2013 A1
20130135589 Curtis et al. May 2013 A1
20130144120 Yamazaki Jun 2013 A1
20130155215 Shimada et al. Jun 2013 A1
20130155305 Shintani Jun 2013 A1
20130158346 Soper et al. Jun 2013 A1
20130184524 Shimada et al. Jul 2013 A1
20130211217 Yamaguchi et al. Aug 2013 A1
20130242069 Kobayashi Sep 2013 A1
20130244453 Sakamoto Sep 2013 A1
20130274597 Byrne et al. Oct 2013 A1
20130289347 Ito Oct 2013 A1
20130292571 Mukherjee et al. Nov 2013 A1
20130296651 Ito Nov 2013 A1
20130296652 Farr Nov 2013 A1
20130300837 DiCarlo et al. Nov 2013 A1
20130342690 Williams et al. Dec 2013 A1
20140012078 Coussa Jan 2014 A1
20140022365 Yoshino Jan 2014 A1
20140031623 Kagaya Jan 2014 A1
20140005532 Choi et al. Feb 2014 A1
20140052004 D'Alfonso et al. Feb 2014 A1
20140066711 Farin et al. Mar 2014 A1
20140073852 Banik et al. Mar 2014 A1
20140073853 Swisher et al. Mar 2014 A1
20140078278 Lei Mar 2014 A1
20140088363 Sakai et al. Mar 2014 A1
20140094649 Ito Apr 2014 A1
20140104466 Fossum Apr 2014 A1
20140110485 Toa et al. Apr 2014 A1
20140142383 Blumenzweig et al. May 2014 A1
20140160318 Blanquart et al. Jun 2014 A1
20140163319 Blanquart et al. Jun 2014 A1
20140198249 Tanaka et al. Jul 2014 A1
20140203084 Wang Jul 2014 A1
20140225215 Chien et al. Aug 2014 A1
20140267655 Richardson et al. Sep 2014 A1
20140267851 Rhoads Sep 2014 A1
20140267890 Lelescu et al. Sep 2014 A1
20140268860 Talbert et al. Sep 2014 A1
20140275764 Shen et al. Sep 2014 A1
20140288365 Henley et al. Sep 2014 A1
20140300698 Wany Oct 2014 A1
20140316197 St. George et al. Oct 2014 A1
20140316199 Kucklick Oct 2014 A1
20140354788 Yano Dec 2014 A1
20140364689 Adair et al. Dec 2014 A1
20150237245 Renard et al. Aug 2015 A1
20150271370 Henley et al. Sep 2015 A1
20160183775 Blanquart et al. Jun 2016 A1
20170085853 Blanquart et al. Mar 2017 A1
20170230574 Richardson et al. Aug 2017 A1
20190028621 Henley et al. Jan 2019 A1
20190133416 Henley et al. May 2019 A1
20190174058 Richardson et al. Jun 2019 A1
20190253685 Blanquart et al. Aug 2019 A1
Foreign Referenced Citations (62)
Number Date Country
1520696 Aug 2004 CN
101079966 Nov 2007 CN
201239130 May 2009 CN
101449575 Jun 2009 CN
101634749 Jan 2010 CN
101755448 Jun 2010 CN
102469932 May 2012 CN
103185960 Jul 2013 CN
0660616 Jun 1995 EP
0904725 Mar 1999 EP
1079255 Feb 2001 EP
1116473 Jul 2001 EP
1637062 Mar 2006 EP
1712177 Oct 2006 EP
1819151 Aug 2007 EP
2359739 Aug 2011 EP
2371268 Aug 2011 EP
3459431 Mar 2014 EP
63-234941 Sep 1988 JP
H04-039789 Apr 1992 JP
H07-240931 Sep 1995 JP
1995-240931 Mar 1997 JP
2000-051150 Feb 2000 JP
2000-199863 Jul 2000 JP
2000270230 Sep 2000 JP
2001-308531 Nov 2001 JP
2002-020816 Jan 2002 JP
2002-028125 Jan 2002 JP
2002-045329 Feb 2002 JP
2002-112961 Apr 2002 JP
2005-204741 Aug 2005 JP
2007-029746 Feb 2007 JP
2007143963 Jun 2007 JP
2007-240931 Sep 2007 JP
2008514304 May 2008 JP
2008-153313 Jul 2008 JP
2008264539 Nov 2008 JP
2008295929 Dec 2008 JP
2009-537283 Oct 2009 JP
2010-17377 Jan 2010 JP
2010-068992 Apr 2010 JP
2010-125284 Jun 2010 JP
2010-158415 Jul 2010 JP
2011-055327 Mar 2011 JP
2011514605 May 2011 JP
2012-000160 Jan 2012 JP
2012024450 Feb 2012 JP
2013-27432 Feb 2013 JP
2011-267098 Jun 2013 JP
2014514782 Jun 2014 JP
5682812 Jan 2015 JP
2015525642 Sep 2015 JP
A61-62440 Jun 2017 JP
2015001195 Jan 2016 MX
346174 Mar 2017 MX
1996005693 Feb 1996 WO
2006037034 Apr 2006 WO
2009018613 Feb 2009 WO
2009115885 Sep 2009 WO
2009120228 Oct 2009 WO
2012043771 Apr 2012 WO
2012137845 Oct 2012 WO
Non-Patent Literature Citations (2)
Entry
Jack, Keith “Video Demystified: A Handbook for the Digital Engineer,” 2007 Fifth Edition. ISBN: 978-0-7506-8395-1, p. 21.
Blumenfeld, et al. Three-dimensional image registration of MR proximal femur images for the analysis of trabecular bone parameters. Oct. 2008. [retrieved on Jul. 30, 2014] Retrieved from the internet: <URL: http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2673590/>.
Related Publications (1)
Number Date Country
20190133416 A1 May 2019 US
Provisional Applications (1)
Number Date Country
61791685 Mar 2013 US
Divisions (1)
Number Date Country
Parent 14214334 Mar 2014 US
Child 16241118 US