Bypassing pixel clock generation and CRTC circuits in a graphics controller chip

Information

  • Patent Grant
  • 7800623
  • Patent Number
    7,800,623
  • Date Filed
    Thursday, July 29, 2004
    20 years ago
  • Date Issued
    Tuesday, September 21, 2010
    14 years ago
Abstract
In a video processor unit, a method of providing a video data stream at a clock rate that is independent of a pixel clock rate. Receiving native video data from a video source at a native clock rate, storing the video data in a memory unit, reading selected portions of the video data at a memory clock rate, rasterizing the selected video data, packetizing the rasterized video data, sending the packetized video data to a display unit by way of a link at a link rate, wherein the link rate is directly related to the memory clock rate.
Description
BACKGROUND
I. Field of the Invention

The invention relates to display devices. More specifically, the invention describes a method and apparatus for using driving LCD panel drive electronics.


Overview


Until most recently, almost all TVs in use today rely on a device known as the cathode ray tube, or CRT, to display their images. In order to display an image on the entire screen, electronic circuits inside the TV use magnetic deflection coils (a horizontal deflection coil to move the beam from left to right and a vertical deflection coil to move the beam up and down) to move the electron beam in a “raster scan” pattern across and down the screen. FIG. 1 illustrates a raster 100 on a conventional cathode ray tube where an electron beam paints one line across the screen from left to right 102a and then quickly moves back to the left side 104, and paints another horizontal line 102b, while continuously moving down slightly. When the beam reaches the right side of the bottom line 107, it is moved back to the upper left corner of the screen, as represented by line 106. When the beam is “painting” lines 102 it is on, and when it is “retracing” lines 104, it is off so that it does not leave a trail on the screen. The term horizontal retrace is used to refer to the beam moving back to the left at the end of each line, while the term vertical retrace refers to its movement from the bottom to the top of the raster 106.


In order to display an image on a CRT display screen, each image is transmitted as a sequence of frames each of which includes a number of horizontal scan lines. Typically, time reference signals are provided in order to divide the video signal into horizontal scan lines and frames. These reference signals include a VSYNC signal that indicates the beginning of a frame and an HSYNC signal that indicates the beginning of a next source scan line. In this way, the image is divided into a number of points where each point is displayed as a pixel having a corresponding pixel clock that specifies the rate at which pixels are generated (in pixels per second). Therefore, in order to display video data from a video source, such as a CPU, DVD player, etc., the video data is processed by a conventional video processor unit (VPU) that converts the incoming video data (at a native format) to a video data at a video display format at a clock rate determined by a memory unit into which the data is temporarily stored.


A cathode ray tube (CRT) controller unit then converts the display formatted video data to CRT compatible display timing by adding timing and control signals (such Hsync and Vsync) regardless of the whether or not the display unit is a cathode ray tube based display or a fixed pixel display such as an LCD. In this way, the CRT controller unit converts the video data that is read from the video source to a pixel clock rate Φpixel based solely upon the assumption that the display requires CRT type signals. It should be noted therefore, that strictly based upon CRT legacy considerations, the incoming video signal has been converted from a native video clock Φnative to a pixel clock Φpixel and then sent to the display unit at a link rate LR that must be faster than the pixel clock Φpixel. At the display, the video signal must then be processed again by a pixel clock regenerator circuit in order for the video data to be properly displayed.


Therefore, even in those situations where a pixel clock rate is not meaningful (such as with fixed pixel displays, such as LCDs), conventional video processors require that a CRT controller unit force a video signal that will not be displayed on a CRT type display to conform to legacy CRT requirements.


Therefore, in those cases where a display is not CRT based, a video processor that does not enforce CRT legacy requirements is desirable.


SUMMARY OF THE INVENTION

What is provided is a video processor architecture embodied as a method, apparatus, and system suitable for implementation with digital displays, such as liquid crystal displays (LCDs), that provides a video signal that is directly related to a memory clock rate and independent of a pixel rate.


In a video processor unit, a method of providing a video data stream at a clock rate that is independent of a pixel clock rate. Receiving native video data from a video source at a native clock rate, storing the video data in a memory unit, reading selected portions of the video data at a memory clock rate, rasterizing the selected video data, packetizing the rasterized video data, sending the packetized video data to a display unit by way of a link at a link rate, wherein the link rate is directly related to the memory clock rate.


In another embodiment, a video processor for displaying video data on a non-CRT type display unit coupled thereto by way of a link is disclosed that includes a video memory unit having an associated memory clock rate and a video packet transmitter unit for transmitting the video data to the non-CRT type display in the form of video data packets at a link rate that is directly related to the memory clock rate Φmemory such that the video processor does not require a CRT controller unit.


In still another embodiment, a configurable video processor for providing displayable video data to a display unit coupled thereto is disclosed. The video processor includes a CRTC circuit, a selectable CRT controller circuit coupled to the CRTC circuit and a raster engine coupled to the selectable CRT controller circuit. The processor also includes a display interrogator unit coupled to the display unit and the selectable CRT controller circuit for determining a display type corresponding to the display unit, wherein when the display type is a non-CRT type display, the display interrogator unit sends a first signal to the selectable CRT controller circuit that disables (if enabled) the CRTC circuit such that the displayable video data is sent directly to the display unit from the raster engine and wherein when the display interrogator determines that the display is a CRT type display, then the display interrogator sends a second signal to the selectable CRT controller circuit that enables the CRTC circuit (if disabled) such that the displayable video data is processed by the CRTC circuit prior to being sent to the display unit by way of the link.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 illustrates a raster on a conventional cathode ray tube.



FIG. 2 shows an exemplary video processor unit (VPU) in accordance with an embodiment of the invention.



FIG. 3 shows an exemplary LCD display suitable for use with the VPU shown in FIG. 3.



FIG. 4 shows a high-level diagram of a data stream for transmission over the link in accordance with an embodiment of the invention.



FIG. 5 illustrates a system that can be used to implement the invention.





DETAILED DESCRIPTION OF SELECTED EMBODIMENTS

Reference will now be made in detail to a particular embodiment of the invention an example of which is illustrated in the accompanying drawings. While the invention will be described in conjunction with the particular embodiment, it will be understood that it is not intended to limit the invention to the described embodiment. To the contrary, it is intended to cover alternatives, modifications, and equivalents as may be included within the spirit and scope of the invention as defined by the appended claims.


Until most recently, almost all TVs in use today rely on a device known as the cathode ray tube, or CRT, to display their images. In order to display an image on the entire screen, electronic circuits inside the TV use magnetic deflection coils (a horizontal deflection coil to move the beam from left to right and a vertical deflection coil to move the beam up and down) to move the electron beam in a “raster scan” pattern across and down the screen. In order to display video data from a video source, such as a CPU, DVD player, etc., the video data provided by a video source at a native clock rate Φnative is processed by a conventional video processor unit (VPU) by storing the video data in a video memory unit until such time as it is read from the memory at a memory clock rate Φmemory and processed by a raster engine that converts the video data native format to a video signal having a displayable format based upon the characteristics of the display unit for which the image will be displayed.


A cathode ray tube (CRT) controller unit then converts the video signal using display timing by adding timing and control signals (such Hsync and Vsync) regardless of the whether or not the display unit is a cathode ray tube based display or a fixed pixel display such as an LCD. In this way, the CRT controller unit converts the video data that is read from the memory unit at the memory clock rate Φmemory to a pixel clock rate Φpixel based solely upon the assumption that the display requires CRT type signals. Once the CRT controller unit has added the CRT based control and timing signals, the video signal is then passed by way of a link to the display unit at a link rate LR. Therefore, even in those situations where a pixel clock rate is not meaningful (such as with fixed pixel displays, such as LCDs), conventional video processors require that a CRT controller unit force a video signal that will not be displayed on a CRT type display to conform to legacy CRT requirements.


Accordingly, the invention describes a method, system, and apparatus that does not enforce CRT legacy requirements for those situations where a video is to be displayed on a non-CRT type display unit, such as an LCD. More specifically, the inventive video processor does not force the video signal to conform to a pixel clock but preserves the memory clock rate by transmitting the video data in video data packets at a link rate that is directly related to the memory clock rate. In this way, the inventive video processor does not require a CRT controller unit for those situations where a display unit so connected is not a CRT type display. In some embodiments, the inventive processor is dedicated to a fixed pixel display and therefore does away with the CRT controller and associated circuitry entirely whereas in other embodiments, in order to maintain compatibility with all possible display environments, a selectable CRT controller circuit is included. The selection (or de-selection) of the CRT controller and associated circuitry is based upon a query of the display attributes during an initialization process whereby the video processor queries the display whether or not it is a CRT type display.


The invention will now be described in terms of a representative LCD panel. However, it should be noted that any digital fixed pixel display, be it LCD, plasma, DLP based, is also suitable and therefore the use of an LCD panel in the following description should not be considered to limit either the scope or the intent of the invention. It should be noted that the invention is also well suited to be used in conjunction with any packet based video display interface such as described in copending U.S. patent application Ser. No. 10/726,794 entitled “PACKET BASED VIDEO DISPLAY INTERFACE AND METHODS OF USE THEREOF” by Kobayashi filed Dec. 3, 2003 and U.S. patent application Ser. No. 10/909,103 entitled “USING PACKET TRANSFER FOR DRIVING LCD PANEL DRIVER ELECTRONICS” by Kobayashi filed concurrently with this application each of which are incorporated herein by reference for all purposes.


Accordingly, FIG. 2 shows an exemplary video processor unit (VPU) 200 in accordance with an embodiment of the invention. The VPU 200 includes an interface 202 coupled to a video source (not shown) such as a CPU, DVD player, etc. capable of providing a video signal V1 at a native clock rate Φnative and a native video format. It should be noted that the video source can include either or both a digital image (i.e. still or digital video) source and/or an analog image (i.e., still or analog video) source. Accordingly, the video source provides various video signals that can have any number and type of well-known formats, such as composite video, serial digital, parallel digital, RGB, or consumer digital video. The video signal can be an analog video signal provided the source includes some form of an analog video source such as for example, an analog television, still camera, analog VCR, DVD player, camcorder, laser disk player, TV tuner, set top box (with satellite DSS or cable signal) and the like. The source can also include a digital image source such as for example a digital television (DTV), digital still camera, and the like. The digital video signal can be any number and type of well known digital formats such as, SMPTE 274M-1995 (1920×1080 resolution, progressive or interlaced scan), SMPTE 296M-1997 (1280×720 resolution, progressive scan), as well as standard 480 progressive scan video.


Once received, the video data V1 is then stored (by way of a write/modify operation) in a video memory unit 204 coupled to the interface 202 at a memory space location corresponding to the video window being displayed. In most cases, the memory unit 204 is of a size sufficient to store at least a full frame of video data In order to display the video data stored in the memory unit 204, selected portions of the video data V1 are read from the memory 204 at a memory clock rate Φmemory and processed by a raster engine 206 that converts the video data native format to a video signal V2 having a displayable format based upon the characteristics of a display unit 208. For example, if the native format is 8 bit video data and the display unit 208 requires 24 bit video data, then the raster engine 206 has the option of utilizing a color look up table (CLUT) 210 in the conversion process.


Once the video signal V2 has been properly formatted for display on the display unit 208, a packetizer 212 packetizes the video signal V2 into a number of data packets 214 in the form of a data stream 215. The data stream 215 is transmitted to the display 208 by way of a link 216 at a transmission rate referred to as a link rate LR that is independent of the native stream rate Φnative of the video data but is, however, directly related to the memory clock rate Φmemory (i.e., LR=α×Φmemory). In the described embodiment, the speed, or link rate, of the link 216 can be configured to include a number of logical data channels (not shown) that can be adjusted to compensate for link conditions. For example, at 2.5 Gbps per channel, the link 216 can support SXGA 60 Hz with a color depth of 18 bits per pixel over a single channel. It should be noted that a reduction in the number of channels reduces not only the cost of interconnect, but also reduces the power consumption which is an important consideration (and desirable) for power sensitive applications such as portable devices and the like. However, by increasing the number of channels to four, the link 216 can support WQSXGA (3200×2048 image resolution) with a color depth of 24-bits per pixel at 60 Hz. or QSXGA (2560×2048) with a color depth of 18-bits per pixel at 60 Hz, without data compression. Even at the lowest rate of 1.0 Gbps per channel, only two channels are required to support an uncompressed HDTV (i.e., 1080i or 720p) data stream.


It should be noted, however, that the bandwidth of the link 216 must be greater than the aggregate bandwidth of all data stream(s) being transmitted over the link 216. In this way, the data packets 214 are received by and appropriately processed by a properly configured display 208 an example of which is described below without the need to generate a pixel clock at either the VPU 200 or the display 208. In one embodiment, the VPU 200 can optionally include a switch coupled to a CRT controller unit that provides the necessary CRT control signals and timing only in those cases where the display is a legacy CRT type display. In this way, the VPU can be used for any type display.



FIG. 3 shows an exemplary LCD display 300 suitable for use with the VPU 200. Accordingly, the LCD display 300 includes an LCD panel 302 having a number of picture elements 304 (pixels) that are arranged in a matrix connected to a data driver 306 by way of a plurality of data lines 308 and a plurality of gate lines 310. In the described embodiment, these picture elements 304 take the form of a plurality of thin film transistors (TFTs) 312 that are connected between the data lines 308 and the gate lines 310. During operation, each of a number of data latches 314 into which video data is stored, outputs digital data signals to an associated digital to analog converter (DAC) 316 by way of the data lines 308. Concurrently, each of a number of logic circuits 318 included in a gate driver 320 outputs a predetermined scanning signal to the gate lines 310 in sequence at timings that are in sync with a horizontal synchronizing signal. In this way, the TFTs 312 are turned ON when the predetermined scanning signal is supplied to the gate lines 310 to transmit the analog data signals supplied by the DACs 316 by way of the data lines 408 that ultimately drive selected ones of the picture elements 304.


In order to transmit the video data, the VPU 200 forms the data stream 215 that includes a number of the data packets 214 which are then received and processed by a display interface 322. In the described embodiment, the data packets 214 are then forwarded to directly to the data latches 314 included in the column driver 306 in such a way that all the video data (in the form of pixel data) used for the display of a particular frame line n of the video frame is provided to the data latches 314 within a line period τ. Therefore, once each data latch 314 has appropriate pixel data stored therein, the data driver 306 drive appropriate ones of the TFTs 312 in the LCD array 302.



FIG. 4 shows a high-level diagram of a data stream 400 for transmission over the link 216 formed of a number of video data packets 402 and audio data packets 404 multiplexed into the single data stream 400. In this example the video data packets 402 are consistent with UXGA graphics 1280×720p video (Stream ID=1) having an associated audio in the form of the audio packets 504 (Stream ID=2). In this example, each frame line is formed of at least 1280 pixels (or 3840 sub-pixels) therefore requiring 3840 data latches be used to store a single frame line of video data within the line period τ. For example, in one embodiment, when the data stream 400 is received at the display interface 322, a group of 3840 data packets (as defined by corresponding packet headers 406) are stored in a memory that takes the form of a frame buffer or a line buffer. It should also be noted, however, that the memory can be bypassed or be absent altogether if a strictly pipelined architecture is desired.



FIG. 5 illustrates a system 500 that can be used to implement the invention. The system 500 is only an example of a graphics system in which the present invention can be implemented. System 500 includes central processing unit (CPU) 510, random access memory (RAM) 520, read only memory (ROM) 525, one or more peripherals 530, graphics controller 560, primary storage devices 540 and 550, and digital display unit 570. CPU 510 is also coupled to one or more input/output devices 590. Graphics controller 560 generates image data and corresponding reference signals, and provides both to digital display unit 570. The image data can be generated, for example, based on pixel data received from CPU 510 or from an external circuitry.


Although only a few embodiments of the present invention have been described, it should be understood that the present invention may be embodied in many other specific forms without departing from the spirit or the scope of the present invention. The present examples are to be considered as illustrative and not restrictive, and the invention is not to be limited to the details given herein, but may be modified within the scope of the appended claims along with their full scope of equivalents.


While this invention has been described in terms of a preferred embodiment, there are alterations, permutations, and equivalents that fall within the scope of this invention. It should also be noted that there are many alternative ways of implementing both the process and apparatus of the present invention. It is therefore intended that the invention be interpreted as including all such alterations, permutations, and equivalents as fall within the true spirit and scope of the present invention.

Claims
  • 1. A pipelined method of displaying video data at a display unit, the display unit comprising: a display unit interface, a single line buffer used for storing video data to be displayed, the single line buffer having a size in accordance with a scan line for display; and a plurality of display elements, the method comprising:(a) receiving video data packets associated with the scan line for display at the display unit interface;(b) directly populating the single line buffer using the video data packets received at the display unit interface within a line period τ, the line period τ corresponding to the scan line for display;(c) only when the single line buffer is full, passing the video data stored in the single line buffer directly to selected of the plurality of display elements, corresponding to the scan line for display; and(d) concurrently with passing the video data (c), directly populating the single line buffer using video data packets received at the display unit interface, the received video data packets associated with a next scan line for display.
  • 2. The method of claim 1, wherein the display unit does not generate a clock signal.
  • 3. The method of claim 1, wherein each of the video data packets includes at least a packet header and a packet payload, wherein the packet header includes a packet ID and wherein the packet payload includes video data suitable for driving a corresponding pixel.
  • 4. The method of claim 1, wherein the plurality of display elements comprise thin film transistors.
  • 5. The method of claim 1, wherein the video data packets are directly received by the display unit interface without using Hsync and Vsync timing and control signals and without using a pixel clock signal.
  • 6. A display apparatus comprising: a display unit interface to receive packetized video data associated with a scan line for display directly from a video processing unit (VPU);a plurality of display elements; anda single line buffer used for storing video data to be displayed, wherein the single line buffer is configured to have a size in accordance with a scan line, wherein the display unit interface directly forwards the packetized video data to the single line buffer, filling the single line buffer within a line period T, the line period T corresponding to the scan line for display, wherein only when the single line buffer is full, the packetized video data in the single line buffer is directly passed to selected of the plurality of display elements, wherein concurrently with passing the packetized video data in the single line buffer directly to selected of the plurality of display element, directly populating the single line buffer using packetized video data received at the display unit interface, the received packetized video data associated with a next scan line for display, and wherein the display apparatus does not generate a pixel clock.
  • 7. The display apparatus of claim 6, wherein each of the video data packets of the packetized video data includes at least a packet header and a packet payload, wherein the packet header includes a packet ID and wherein the packet payload includes video data suitable for driving a corresponding pixel.
  • 8. The display apparatus of claim 6, wherein the plurality of display elements comprise thin film transistors.
  • 9. A non-transitory computer-readable medium having stored thereon data representing instructions that, when performed by a processing element of a display unit having a single line buffer having a size in accordance with a scan line and a plurality of display elements, the processor performs operations comprising: (a) receiving video data packets associated with a scan line at a display unit interface;(b) directly populating the single line buffer using the video data packets received at the display unit interface within a line period τ, the line period τ corresponding to the scan line for display;(c) only when the single line buffer is full, passing the video data stored in the single line buffer directly to selected of the plurality of display elements; and(d) concurrently with passing the video data (c), directly populating the single line buffer using video data packets received at the display unit interface, the received video data packets associated with a next scan line for display.
  • 10. The computer-readable medium of claim 9, wherein the display unit does not generate a clock signal.
  • 11. The computer-readable medium of claim 9, wherein the plurality of display elements comprise thin film transistors.
  • 12. An integrated circuit configured to: receive video data packets associated with a scan line for display;directly populate a single line buffer using the received video data packets within a line period τ, the line period τ corresponding to the scan line for display, wherein the single line buffer has a size in accordance with a scan line;forward the video data packets stored in the single line buffer to selected display elements only when the single line buffer becomes full; andconcurrently with forwarding the data packets to selected display elements, directly populate the single line buffer using received video data packets associated with a next scan line for display.
CROSS REFERENCE TO RELATED APPLICATIONS

This patent application takes priority under 35 U.S.C. 119(e) to (i) U.S. Provisional Patent Application No. 60/504,060 filed on Sep. 18, 2003, entitled “DIGITAL/ANALOG VIDEO INTERCONNECT AND METHODS OF USE THEREOF” by Kobayashi, and (ii) U.S. Provisional Patent Application No. 60/562,737 filed on Apr. 15, 2004, entitled “BYPASSING PIXEL CLOCK GENEARATION AND CRTC CIRCUITS IN A GRAPHICS CONTROLLER CHIP” by Kobayashi each of which are hereby incorporated by reference herein in their entirety. This application is also related to the following co-pending U.S. Patent applications each of which are incorporated by reference, (i) U.S. patent application Ser. No. 10/726,802, entitled “METHOD OF ADAPTIVELY CONNECTING A VIDEO SOURCE AND A VIDEO DISPLAY” naming Kobayashi as inventor; (ii) U.S. patent application Ser. No. 10/726,438, entitled “METHOD AND APPARATUS FOR EFFICIENT TRANSMISSION OF MULTIMEDIA DATA PACKETS” naming Kobayashi as inventor; (iii) U.S. patent application Ser. No. 10/726,440, entitled “METHOD OF OPTIMIZING MULTIMEDIA PACKET TRANSMISSION RATE”, naming Kobayashi as inventor; (iv) U.S. patent application Ser. No. 10/727,131, entitled “USING AN AUXILARY CHANNEL FOR VIDEO MONITOR TRAINING”, naming Kobayashi as inventor; (v) U.S. patent application Ser. No. 10/726,350, entitled “TECHNIQUES FOR REDUCING MULTIMEDIA DATA PACKET OVERHEAD”, naming Kobayashi as inventor; (vi) U.S. patent application Ser. No. 10/726,362, entitled “PACKET BASED CLOSED LOOP VIDEO DISPLAY INTERFACE WITH PERIODIC STATUS CHECKS”, naming Kobayashi as inventor; (vii) U.S. patent application Ser. No. 10/726,895, entitled “MINIMIZING BUFFER REQUIREMENTS IN A DIGITAL VIDEO SYSTEM”, naming Kobayashi as inventor; and (viii) U.S. patent application Ser. No. 10/726,441, entitled “VIDEO INTERFACE ARRANGED TO PROVIDE PIXEL DATA INDEPENDENT OF A LINK CHARACTER CLOCK”, naming Kobayashi as inventor; (ix) U.S. patent application Ser. No. 10/726,934, entitled “ENUMERATION METHOD FOR THE LINK CLOCK RATE AND THE PIXEL/AUDIO CLOCK RATE”, naming Kobayashi as inventor, and (x) U.S. patent application Ser. No. 10/726,794, entitled “PACKET BASED VIDEO DISPLAY INTERFACE AND METHODS OF USE THEREOF” naming Kobayashi as inventor. This application is also related to the following co-pending applications: (x) U.S. patent application Ser. No. 10/909,103, entitled “USING PACKET TRANSFER FOR DRIVING LCD PANEL DRIVER ELECTRONICS” filed Jul. 29, 2004, naming Kobayashi as inventor; and (xi) U.S. patent application Ser. No. 10/909,085, entitled “PACKET BASED STREAM TRANSPORT SCHEDULER AND METHODS OF USE THEREOF” filed Jul. 29, 2004, naming Kobayashi as inventor.

US Referenced Citations (154)
Number Name Date Kind
4479142 Buschman et al. Oct 1984 A
4796203 Roberts Jan 1989 A
4868557 Perlman Sep 1989 A
5245612 Kachi et al. Sep 1993 A
5258983 Lane et al. Nov 1993 A
5369775 Yamasaki et al. Nov 1994 A
5425101 Woo et al. Jun 1995 A
5515296 Agarwal May 1996 A
5541919 Yong et al. Jul 1996 A
5608418 McNally Mar 1997 A
5615376 Ranganathan Mar 1997 A
5625379 Reinert et al. Apr 1997 A
5629715 Zenda May 1997 A
5670973 Bassetti et al. Sep 1997 A
5739803 Neugebauer Apr 1998 A
5745837 Fuhrmann Apr 1998 A
5786844 Rogers et al. Jul 1998 A
5790083 Bassetti Aug 1998 A
5801776 Tamura et al. Sep 1998 A
5805173 Glennon et al. Sep 1998 A
5838875 Cho et al. Nov 1998 A
5852630 Langberg et al. Dec 1998 A
5909465 Bottomley et al. Jun 1999 A
5918002 Klemets et al. Jun 1999 A
5926155 Arai et al. Jul 1999 A
5940137 Hulvey Aug 1999 A
5949437 Clark Sep 1999 A
6005861 Humpleman Dec 1999 A
6020901 Lavelle et al. Feb 2000 A
6038000 Hurst, Jr. Mar 2000 A
6049316 Nolan et al. Apr 2000 A
6049769 Holmes et al. Apr 2000 A
6069929 Yabe et al. May 2000 A
6151334 Kim et al. Nov 2000 A
6151632 Chadda et al. Nov 2000 A
6154225 Kou et al. Nov 2000 A
6175573 Togo et al. Jan 2001 B1
6177922 Schiefer et al. Jan 2001 B1
6219736 Klingman Apr 2001 B1
6223089 Page Apr 2001 B1
6249319 Post Jun 2001 B1
6326961 Lin et al. Dec 2001 B1
6330605 Christensen et al. Dec 2001 B1
6337964 Inami et al. Jan 2002 B2
6353594 Tooker et al. Mar 2002 B1
6356260 Montalbo Mar 2002 B1
6437768 Kubota et al. Aug 2002 B1
6441857 Wicker et al. Aug 2002 B1
6446130 Grapes Sep 2002 B1
6477252 Faber et al. Nov 2002 B1
6542967 Major Apr 2003 B1
6543053 Li et al. Apr 2003 B1
6545688 Loveridge et al. Apr 2003 B1
6577303 Kim Jun 2003 B2
6587480 Higgins et al. Jul 2003 B1
6598161 Kluttz et al. Jul 2003 B1
6600469 Nukiyama et al. Jul 2003 B1
6608828 Balachandran Aug 2003 B1
6614800 Genty et al. Sep 2003 B1
6661422 Valmiki et al. Dec 2003 B1
6697376 Son et al. Feb 2004 B1
6704310 Zimmermann et al. Mar 2004 B1
6765931 Rabenko et al. Jul 2004 B1
6778168 Mamiya et al. Aug 2004 B2
6862606 Major et al. Mar 2005 B1
6865188 Stirling et al. Mar 2005 B1
6873625 Yoo et al. Mar 2005 B1
6903716 Kawabe et al. Jun 2005 B2
6909442 Hiyama et al. Jun 2005 B2
6914637 Wolf et al. Jul 2005 B1
6963968 Kori Nov 2005 B2
6973069 Spear et al. Dec 2005 B1
6975645 Suzuki et al. Dec 2005 B1
7046631 Giroux et al. May 2006 B1
7075987 Kim et al. Jul 2006 B2
7099277 Sahinoglu et al. Aug 2006 B2
7136415 Yun et al. Nov 2006 B2
7177329 Kobayashi et al. Feb 2007 B2
7194554 Short et al. Mar 2007 B1
7248590 Liu Jul 2007 B1
7256790 Valmiki et al. Aug 2007 B2
7295578 Lyle et al. Nov 2007 B1
7453479 Le et al. Nov 2008 B2
7525975 Caspi et al. Apr 2009 B2
20010030649 Mamiya et al. Oct 2001 A1
20010036193 Kori Nov 2001 A1
20010038387 Tomooka et al. Nov 2001 A1
20020007452 Traw et al. Jan 2002 A1
20020011996 Inoue et al. Jan 2002 A1
20020060676 Kim May 2002 A1
20020061024 Malkemes et al. May 2002 A1
20020062394 Bunn et al. May 2002 A1
20020071055 Ooshima et al. Jun 2002 A1
20020071390 Reeves et al. Jun 2002 A1
20020075902 Abbas et al. Jun 2002 A1
20020080468 Crummey et al. Jun 2002 A1
20020085582 Kim Jul 2002 A1
20020089517 Ludtke et al. Jul 2002 A1
20020122515 Bodenschatz Sep 2002 A1
20020136219 Ding et al. Sep 2002 A1
20020149617 Becker Oct 2002 A1
20020163598 Pasqualino Nov 2002 A1
20020164022 Strasser et al. Nov 2002 A1
20020184327 Major et al. Dec 2002 A1
20020190974 Morita Dec 2002 A1
20020190978 Agarwal et al. Dec 2002 A1
20030035442 Eng Feb 2003 A1
20030048852 Hwang et al. Mar 2003 A1
20030063077 Koyama Apr 2003 A1
20030076282 Ikeda et al. Apr 2003 A1
20030080971 Hochmuth et al. May 2003 A1
20030112822 Hong et al. Jun 2003 A1
20030145258 Warner et al. Jul 2003 A1
20030149987 Pasqualino et al. Aug 2003 A1
20030152160 Bauch et al. Aug 2003 A1
20030174156 Katsuhara et al. Sep 2003 A1
20030174795 Bruhnke et al. Sep 2003 A1
20030177423 Komatsu et al. Sep 2003 A1
20030212811 Thornton Nov 2003 A1
20040049705 Liebenow Mar 2004 A1
20040080671 Siemens et al. Apr 2004 A1
20040081151 Greis et al. Apr 2004 A1
20040088469 Levy May 2004 A1
20040103333 Martwick et al. May 2004 A1
20040114607 Shay et al. Jun 2004 A1
20040198386 Dupray Oct 2004 A1
20040203383 Kelton et al. Oct 2004 A1
20040210805 Kimelman et al. Oct 2004 A1
20040218598 Kobayashi Nov 2004 A1
20040218599 Kobayashi Nov 2004 A1
20040218624 Kobayashi Nov 2004 A1
20040218625 Kobayashi Nov 2004 A1
20040218627 Kobayashi Nov 2004 A1
20040221056 Kobayashi Nov 2004 A1
20040221180 Enami et al. Nov 2004 A1
20040221312 Kobayashi Nov 2004 A1
20040221315 Kobayashi Nov 2004 A1
20040228365 Kobayashi Nov 2004 A1
20040233181 Kobayashi Nov 2004 A1
20040240454 Yamauchi et al. Dec 2004 A1
20040243905 Merritt Dec 2004 A1
20050062711 Kobayashi Mar 2005 A1
20050066085 Kobayashi Mar 2005 A1
20050103333 Bonutti May 2005 A1
20050225547 Choi Oct 2005 A1
20060036788 Galang et al. Feb 2006 A1
20060059092 Burshan et al. Mar 2006 A1
20060117371 Margulis Jun 2006 A1
20060209890 MacMullan et al. Sep 2006 A1
20070049086 Sakane Mar 2007 A1
20070097885 Traversat et al. May 2007 A1
20070140298 Eng Jun 2007 A1
20080175277 Yin et al. Jul 2008 A1
20080284761 Knee et al. Nov 2008 A1
Foreign Referenced Citations (34)
Number Date Country
1353513 Jun 2002 CN
0 354 480 Feb 1990 EP
0 385 449 Sep 1990 EP
0 674 440 Sep 1995 EP
0 674 441 Sep 1995 EP
0788048 Jun 1997 EP
1 041 823 Oct 2000 EP
1 069 721 Jan 2001 EP
1089503 Apr 2001 EP
1 154 354 Nov 2001 EP
1 229 690 Aug 2002 EP
1 251 664 Oct 2002 EP
1 432 203 Jun 2004 EP
1 473 700 Nov 2004 EP
1 517 292 Mar 2005 EP
1 519 349 Mar 2005 EP
1 519 581 Mar 2005 EP
2 329 741 Mar 1999 GB
10145309 May 1998 JP
11175045 Jul 1999 JP
2001218082 Aug 2001 JP
2001036900 Sep 2001 JP
2002304168 Oct 2002 JP
110144 Apr 2005 SG
WO 9500917 Jan 1995 WO
9513681 May 1995 WO
WO 9841008 Sep 1998 WO
WO 9963513 Dec 1999 WO
WO 0020974 Apr 2000 WO
0225822 Mar 2002 WO
WO0225885 Mar 2002 WO
WO02065746 Aug 2002 WO
WO03058376 Jul 2003 WO
WO 03058376 Jul 2003 WO
Related Publications (1)
Number Date Country
20050062699 A1 Mar 2005 US
Provisional Applications (2)
Number Date Country
60504060 Sep 2003 US
60562737 Apr 2004 US