Numerous 3-D viewing concepts and devices have been developed over many years. Many of those approaches require the viewer to wear special 3-D viewing glasses or goggles or to view images through fixed eyepieces. Autostereoscopic (i.e., not requiring any special eyewear) display approaches have been developed. Autostereoscopic display approaches include holography, volumetric imaging, and multiple-image techniques such as integral imaging (“fly's eye” type lens arrays with each lens displaying its own complete image), light barriers containing slits through which light can pass at specific angles for viewing different images from different directions, and lenticular lens arrays. Holography, while fundamentally capable of producing highly realistic 3-D images, has not been developed beyond a rudimentary level for general 3-D viewing.
While someday a 3-D display may be available providing an autostereoscopic electronic display to an arbitrary number of freely-moving viewers located at arbitrary viewing angles and distances with smooth parallax and head-tipping for such uses as general-purpose TV viewing, there presently appears to be a vast potential commercial market for single-viewer 3-D television and personal computer applications.
One autostereoscopic technology that has received considerable attention is lenticular 3-D. A lenticular 3-D display basically consists of a series of cylindrical lenses located in front of a surface containing the image to be viewed. The cylindrical lenses in the lenticular sheet are adjacent to each other, parallel to each other, and typically vertically-oriented. The display surface containing the image to be viewed lies in the focal plane of the lenses.
Another technology, integral imaging, is essentially a photographic technique, although in some cases an electronic sensor sheet is used instead of photographic film. Each lens in a two-dimensional lens array records and displays its own unique view of a common scene, and does not image only a small element of a scene such as one or two pixels.
Conventional autostereoscopic displays have one or more of the following significant disadvantages: they restrict the viewed image to be located inside a limited volume; they are large, non-flat systems; they are highly complex and costly; and some systems require eye-tracking (i.e., the system tracks the locations of the viewers' eyes in order to direct different scenes precisely to the two eyes—or to more than two eyes for multiple viewers); intensive computing; and large data storage capability.
Conventional autostereoscopic lenticular display systems suffer from a number of additional deficiencies. Among the problematic issues are moire effects and other related effects. Another problem with the lenticular system is the imbalance between the horizontal and vertical resolutions, which produces some viewing discomfort. Lenticular 3-D displays also suffer from optical problems such as image distortion resulting from misalignment of lenticular arrays of cylindrical lenses. Generally, practical lenticular display systems cannot provide satisfactory 3-D display quality. This quality degradation is mainly due to two major problems, one that is intrinsic and the other extrinsic. Intrinsically, rays from a viewer's eye to the lenticulae placed on a display panel are not parallel to each other. However, a practical system may assume parallel rays for ease of manufacturing, thereby causing undesirable display distortion. The extrinsic problem is that lenticular arrays of cylindrical lenses may not be precisely aligned on the display panel, especially when lenticulae are slanted to solve the imbalance problem between horizontal and vertical resolutions. This alignment error causes considerable distortions in the 3-D display.
A technique for presenting a 3-D display includes the steps of providing a micro-lens array having a plurality of elements directing the outputs from respective pixel groups in mutually different directions and wherein at least two elements in the micro-lens array have a differing fixed optical property as a function of the distance of an object plane including a pixel group dividing line from a corresponding element of the micro-lens array, and an image-plane distance from the corresponding element and providing an stereoscopic image in disposed proximate to the micro-lens array, wherein the image provides a first eye perspective view and a second eye perspective view.
There are many possible applications of this new personal 3-D display. Exemplary applications include computer screens, control systems, medical displays (both for diagnostics and for operating-room use by surgeons), cellular telephone displays, various handheld wireless Internet devices, portable DVD players, personal-size TVs and electronic game players. Potential users include: researchers, designers, scientists, surgeons, teachers, cartographers, and the military for reconnaissance and other applications—as well as the public for general viewing and game playing. For many currently popular applications, the present invention appears to be a practical, relatively inexpensive, technologically feasible and sensible approach for high-quality three-dimensional viewing. Aspects of the invention offer a simple, effective means for satisfying the demand both for specialized and mass-market applications.
The added micro-lens array represents a relatively minor item of additional hardware. The display can be monochromatic or multi-color. For a monochromatic display, each pixel will exhibit some shade of the display's single color, say a shade of grey for a black-and-white display. In presently available color displays, each pixel contains three separate small color elements that together form the intended perceived image color at that pixel location. Embodiments of the invention can utilize both types of display.
There are fundamental differences between the type of lens used in a lenticular 3-D system and the type used in embodiments of the present invention. The basic optical characteristics of a cylindrical lens in lenticular 3-D systems are defined only in a plane perpendicular to the cylinder axis. With an optical object located in the focal plane of a cylindrical lens (i.e., at the focal distance from the front surface of the vertically-oriented lens) the image of that object in the horizontal plane is located at infinity. That is, for each point in the optical object, the rays emanating from that point in a horizontal plane emerge from the lens in that same plane and are directed parallel to each other in a particular direction in that plane. For rays not lying in the horizontal plane, the optical characteristics are not clearly defined. There is no focusing action by the lens; refraction at the lenticular lens surface directs the rays somewhere between the horizontal and vertical. The rays in a lenticular 3-D system are spread vertically over a large angle. In contrast, lenses in embodiments of the present invention form images at finite distances and controlled locations.
As described by G. L. Rogoff in “Optical System for Spatial Discrimination of Radiation from Extended Bodies,” Applied Optics, vol. 8, pp. 723-724 (1969)), a lens with an object at a focal length distance from the lens, a telescopic system, is primarily a directional device; it primarily determines the direction of rays that originate from points in its focal plane. It is not a device for forming images at finite distances. For a cylindrical lens, a rotational alignment error will distort the image. Rotation of a lens in the present invention will not distort the image. The optical characteristics of the lens elements in the present invention are rotationally symmetric about the optical axis of the lens.
For a lens element in the present invention, the optical characteristics are well defined both horizontally and vertically. Since the optical object is not located in the focal plane of the lens, the image of that object is at a finite and controlled distance from the lens. The light from a pixel is focused in both the horizontal and vertical directions to form an image in the viewing region.
Embodiments of the inventive display provide a new three-dimensional display intended for use by a single viewer. Additional viewers could view at the same time, but they would view in two dimensions rather than three; three-dimensional viewing by multiple viewers requires multiple displays. However, for the viewer viewing in 3-D no glasses are required, and the viewer's head need not be held in a fixed position; the head can be moved horizontally both parallel to the display surface and perpendicular to it to some extent while retaining the 3-D view. Horizontally, parallel to the display surface, the head can be moved a distance equal to the distance between the eyes. Some head movement is allowed vertically as well. This display system offers a degree of 3-D viewing comfort not found with long-term use of special glasses, goggles, eyepieces or any other special eyewear.
In one embodiment an autostereoscopic display apparatus includes an image display for providing a display output comprising a plurality of pixels grouped in a row and column array, a micro-lens array disposed adjacent the image display, having a plurality of elements directing the outputs from respective pixel groups in different directions so as to enable a stereoscopic image to be perceived and at least two elements in the micro-lens array have a differing fixed optical property as a function of a distance from an object plane including a pixel group dividing line to a center point of the corresponding element of the micro-lens array along the optical axis, and an image-plane distance from the corresponding element. Such a display provides 3-D viewing for a single viewer, without any special eyewear and with some flexibility as to the location of the viewer.
Further advantages over conventional lenticular 3-D displays include, but are not limited to: better resolution by avoiding vertical spreading of emitted light as in lenticular displays; more comfortable viewing by avoiding problems of visually uncomfortable imbalance between the horizontal and vertical resolutions of lenticular systems; brighter display image because light from the display surface is concentrated horizontally and vertically into the viewing region; avoidance of a Moire effect associated with aligning sets of linear objects; and less sensitivity to misalignment because lens rotation does not distort the image.
In another embodiment, a 3-D display includes an array of pixel pairs, each pixel pair including at least one first pixel and at least one second pixel, an array of lenses disposed in front of the array of pixel pairs, each lens in the array of lenses aligned with a respective pixel pair and configured with a focal pattern such that a first image of a first pixel of one of the pixel pairs is projected into a first viewing area and second image of a second pixel of the same one of the pixel pairs is projected into a second viewing area. The first viewing area being a first location for viewing by a respective one of a viewer's left and right eye and the second viewing area being a second location for viewing by the viewer's other eye.
The foregoing will be apparent from the following description of particular embodiments disclosed herein, as illustrated in the accompanying drawings in which like reference characters refer to the same parts throughout the different views. The drawings are not necessarily to scale, emphasis instead being placed upon illustrating the principles disclosed herein.
In one embodiment, the autostereoscopic display apparatus disclosed herein provides a 3-D display comprising a plurality of pixel groups arranged in a row and column array. Each pixel group is a pixel pair including at least one first pixel and at least one second pixel. As an example, the first and second pixels of each pixel group can be disposed side by side one another. An array of micro-lens elements is provided and is disposed in front of the array of pixel pairs. In one configuration, each lens in the array of lenses is aligned with a respective pixel pair to allow focusing of imagery produced from that pixel pair. Specifically, each lens in the array of lenses is configured with a focal pattern that projects an image of the first pixel of a pixel pair in a first viewing area to enable the first pixel to be viewable, for example by one of a left and a right eye of a viewer. That same lens projects an image of the second pixel of that same pixel pair in a second viewing area to enable the second pixel to be viewable by the other of the right and left eye of the same viewer. In this manner, the first viewing area (i.e., the images from all first pixels of all pixel pairs on a display) is perceived by a respective one of a viewer's left and right eye and the second viewing area that includes or contains the projected images of the other pixels of the pixel pairs is perceived by the viewer's other eye. Thus each micro-lens element projects the images of the first and second pixels (there may be more than one first pixel and more than one second pixel, and hence the term pixel group is not limited to just two pixels) into two viewing areas each utilized by a respective eye of a viewer viewing the display. In this manner, image signals sent to each first pixel of each pixel pair collectively combine to form an image viewed by the right eye of a viewer, while images sent to each second pixel of each pixel pair collectively form an image viewed by the left eye of the viewer. In this manner, the two images (one being sent as a signal to all first pixels, and the other being sent as a signal to all second pixels) can be viewed as a 3-D image by the viewer.
Now referring to
In one embodiment the display surface does not lie in the focal plane of any of the micro-lens elements. Also note that each pixel-lens combination is an independent optical system with its exit pupil located where the eyes are expected to be located. All the independent exit pupils overlap, with all having in common at least the common image point along the perpendicular from the display surface center.
In one embodiment, the micro-lens array 104 is retrofitted to an available conventional display unit or it can be partially or fully integrated with the image display 102. The micro-lens array 104 may be added to a display in various ways. For example in one embodiment, the micro-lens array 104 is mounted with separate spacers between the micro-lens array 104 and the surface of the image display 102; or the micro-lens array 104 may be constructed with built-in spacers extending to the surface of image display 102 to contact the surface of image display 102 directly in selected regions; or the micro-lens array 104 may be constructed with its side that faces the image display 102 made entirely flat (or curved) to contact the surface of image display 102 directly everywhere. In the latter case the side of the micro-lens array 104 facing away from the surface of image display 102 would be suitably shaped to provide the optical imaging required for this invention. The micro-lens array 104 can be retrofitted to an existing display or can be fabricated as an integral part of a display. For example, image display 102 front outer layer and the micro-lens array 104 can be constructed as a single unit. To achieve a flat outer surface, the micro-lens array 104 may be shaped only on its surface facing toward the surface of image display 102.
For clarity, aspects of the invention will be described first as a monochromatic black-and-white version. Color displays with each pixel group including three distinct colors are discussed further below. Now referring to
By providing a display 102 such as a laptop computer display, television display, or other video display in conjunction with the micro-lens array 104 disposed in front of groups of pixel pairs 113 on the display 102, and configuring the elements of the micro-lens array 104 to project images of half the pixels of the display (i.e., all first pixels of the pixel pairs) to a viewing area 114a for a viewer's right eye, and to project images of the other pixels to a viewing area 114b for the viewer's left eye, 3-D imagery (i.e., two distinct images) can be provided. As a viewer views the display 102, the right eye viewing area 114a is a region in which imagery from all first pixels is projected by the set of lens elements, and the left eye viewing area 114b is a region in which imagery from all second pixels is projected. By carefully aligning each element 110 of the micro-lens array 104, when a viewer positions his or her head in front of the display in a top-to-bottom and left-to-right centered manner, so long as the viewer's right and left eyes are located within each of the two different viewing areas in the viewing region 114, the viewer can experience 3-D imagery. Slight movement of the viewer's head is allowed while still perceiving the 3-D effect, so long as the viewer's right and left eyes do not move so much as to cause the right or left eye to stray (from head movement) into the first or second viewing area intended to be utilized by the other eye. If this occurs, the viewer can still view the display but the 3-D effect will be lost until their head moves back to a position in which each eye is capturing imagery from its respective first or second viewing area. The viewing areas 114a and 114b are thus regions in space into which the lens elements 110 project pixel images and in which a viewer's left and right eyes must respectively be located to perceive the images in a 3-D effect.
In one embodiment, two abutting pixels side-by-side form a pixel group behind each lens element. To distinguish between the two pixels associated with a given lens element, the pixel on the left side of their abutting line (facing the display surface) is referred to as pixel “a” 132a and the pixel on the right side will be called pixel “b” 132b in the following explanation. The distance, position and orientation of each lens with respect to its corresponding pixel pair are configured such that the image of the line dividing its pixel pair intersects a perpendicular to the display surface at (or near) its center, with the images of all the dividing lines of all the pixel pairs intersecting that perpendicular at the common image point 116. It is understood that the lens elements need not be parallel to the corresponding pixel surface.
Then one of a viewer's eyes, located to one side of the perpendicular near that common image point, will see only the “a” pixels, and the other eye, located to the other side of the perpendicular near that common image point, will see only the “b” pixels. Since the images of the “a” and “b” pixels in the viewing region are reversed from their positions in the display, the “a” pixels on the left sides of the pixel dividing lines would be seen by the right eye, and the “b” pixels would be seen by the left eye. Thus one eye sees an image on the display formed by the “a” pixels while the other eye sees an image formed by the “b” pixels. Pixel pairs need not be aligned vertically; they can be staggered as shown in conjunction with
When a viewer is looking at the display and the viewer's eyes are positioned such that they are located on opposite sides of the common image point, the viewer will perceive a 3-D image. As indicated above, the viewer's head and eyes need not be restrained to a fixed position but can be moved horizontally a distance equal to the distance between the eyes with the 3-D view still seen by the viewer. If both of the viewer's eyes are moved to the viewing region on the same side of the surface perpendicular, the viewer will see a two-dimensional image; both eyes will see either the “a” display image only or the “b” display image only.
The set of “a” pixels and set of “b” pixels are separately addressable and are programmed so that the “a” pixels display one image of a stereo pair of images and the “b” pixels display the other image of a stereo pair of images, with the two sets of pixels interlaced on the screen. Then one eye will see only one of those images and the other eye will see only the other of those images. The signals communicated to the sets of “a” pixels and “b” pixels are produced using techniques known in the art. Since each lens element will reverse the relative locations of the pixel images in its corresponding pixel pair, the group of “a” pixels will display the right-eye image, and the group of “b” pixels will display the left-eye image, and the viewer will see a stereoscopic three-dimensional image.
Referring now to
A distance 119 of a given lens element from the center of the boundary between pixels 115 of the corresponding pixel group and a distance 120 from the lens element 110 to the common image point 116 are in the ratio required for the width of the image of a pixel in the viewing region to be at least the same as the distance between the viewer's eyes. The distance of the lens array from the screen will generally be much smaller than the distance of the viewer from the screen. Thus, if the viewer's right eye is located to the right of the common image point 116, that eye will see only the set of “a” pixels, and if the left eye is located to the left of the common image point 116, that eye will see only the set of “b” pixels.
Also, high-quality images of the boundaries between pixels 115 are not required. It is important, however, that the boundary images intersect at the common image point 116 reasonably close to the plane where the eyes are expected to be located. This is accomplished by appropriately positioning and orienting each lens.
One way to accomplish this is to keep all the lenses parallel to each other in a plane (or to locate them on a non-planar surface that accommodates or conforms to the shape of the display surface) and to simply shift the position of each lens appropriately. In one embodiment, elements 110 of the micro-lens array 104 are disposed on a non-planar surface that conforms to the shape of the front surface of the image display. The outer shape of each lens may conform to the shape of the pair of pixels behind it, e.g., a rectangular lens for a pair of rectangular pixels, but it is not required to do so. Depending on the parameters chosen for the different lenses, they can have different outer shapes to allow them to remain close to each other and minimize dead space between adjacent lenses or to satisfy other requirements (e.g., to facilitate manufacturing).
In one embodiment, each lens is slightly smaller than its corresponding pixel pair, and the overall lens array is slightly smaller than the overall pixel array.
In another embodiment, at least two elements 110 in the micro-lens array 104 have a differing fixed optical property as a function of the distance from an object plane including the boundary between pixels 115 to a corresponding element 110 of the micro-lens array 104, and an image-plane distance from the corresponding element 110. In this embodiment, the optical property is the focal length of each lens element 110a-110n. For example, as an approximation, the thin lens equation is used to configure the focal length of each lens element 110a-110n. That equation is
1/f=1/p+1/q, (1)
where f is the focal length of the lens, p is the distance of the object plane (including the pixel-pair dividing line) from the lens, and q is the image-plane distance from the lens.
A given pixel group 113 in the pixel array of the display 102 and a lens element 110 can be identified by coordinates i and j, row and column respectively. With this notation, the focal length of the lens for a given pixel pair is denoted by fij; the perpendicular distance of the object plane including the pixel dividing line center to the optical center of the lens element is denoted by pij; the perpendicular distance of the image plane to the optical center of the lens element is denoted by qij; and Eq. (1) becomes:
1/fij=1/pij+1/qij, (2)
where subscript i represents the lens element row location; and subscript j represents the lens element column location. For a given pixel-pair group, lens element and distance between the display and viewer, combinations of values that satisfy Equation (2) can be chosen.
Referring again to
In one embodiment, a computer display 102 has dimensions 30 centimeters wide and 24 centimeters high, and a viewing distance (the distance to the common image point 116 from the display center) of 50 centimeters. In smaller displays, for example, a cell phone, a handheld wireless Internet device or game player, the viewing distance is considerably less, and slightly less for devices such as a portable DVD player or personal TV. In this embodiment, the value of qij is nominally 50 cm for i and j in the center of the display. With this configuration, a viewer's head can move horizontally a distance equal to the distance between the eyes while still perceiving the 3-D view. Each eye should be able to see its corresponding set of pixels “a” or “b” over a horizontal distance at least equal to the eye separation, which is on average approximately 6.3 cm. If a pixel is assumed to be square with dimension of approximately 0.5 mm, or 0.05 cm, then the required magnification is approximately 126, giving a value of pij equal to about 0.4 cm. Equation (2) provides:
f
ij
=p
ij
q
ij/(pij+qij). (3)
The values pij=0.4 cm and qij=50 cm give fij=0.4 cm. This is an approximate value. The focal length fij must actually be a bit smaller than 0.4 cm, the value of pij. If pij were exactly equal to fij, the image of the pixel dividing line would be at qij=∞. The outer dimensions of each lens can be about the same or smaller than the overall pixel-pair dimensions.
For clarity,
Now referring to
In another embodiment, the elements 110′ of the micro-lens array 104″ are disposed in a staggered fashion in a vertical direction in groups. That is, referring to
In conventional lenticular systems, the cylindrical lenses cannot be staggered vertically. When a pixel-based display is used to provide two stereoscopic images instead of a single 2D image, the number of pixels available for each of the two stereo images is half of the total number of pixels. Therefore the overall resolution of each of the two images is half the 2D image. If the lenses are not staggered, then a vertical column of pixels seen by the viewer in each stereo image is continuous and unchanged from its 2D version. For a given eye, alternate columns are missing, but the vertical columns that are seen have the same resolution as in the 2D display. The horizontal rows of pixels, however, are of necessity staggered. This imbalance between horizontal and vertical resolutions can be visually uncomfortable. This imbalance is avoided in embodiments of the present invention by staggering the lens elements as described above so that the vertical stacks of pixels are not all lined up vertically. The staggering of the lens elements in these embodiments alternates the vertical pixels seen by a given eye to match the horizontal staggering.
Referring again to
Embodiments discussed herein can be applied to a portion of a display screen rather than to the entire display screen. When applied to only a portion of a display screen, it is understood that, for example, references to the perpendicular from at or near the center of the display screen means from at or near the center of the portion of the display screen covered by the micro-lens array.
To some extent, the viewer's head can move either closer to the display screen or farther from the screen than the location of the common image point 116 and retain 3-D viewing. The distances that the viewer's head can move depend on the angular spread of rays from the pixels located farthest from the display center. Since the lenses are small compared with their distances from the common image point 116, the angular spread of rays from a given dividing line and lens will be relatively small, but, as the head is moved perpendicular to the screen, the overall width of the region occupied by the group of all dividing line rays increases more rapidly than does the width from a single pixel. That overall angular spread determines the separation between the regions from which one eye sees only the “a” pixels and the other eye sees only the “b” pixels. Thus the range of allowed viewing distances varies with display screen size and viewing distance.
The display apparatus is not limited to monochromatic or black-and-white displays. It will work with a display in which each pixel is of a single color—any color. However, many displays contain within each pixel three distinct stripes, or bands (sometimes called “sub-pixels”), each of a different color (generally red, green, and blue). Those displays depend on the brain's ability to mix the colors; the relative intensities of the three colors in a given pixel are adjusted for the viewer to perceive the intended brain-averaged single color. If in the present invention such a color pixel were to be accurately imaged where the eyes are located, each eye would tend to see only one of those color bands, with the observed color depending on the head position. To overcome this effect, a diffuser sheet can be used to achieve the intended averaged color. In one embodiment, the diffuser sheet is disposed between the image display and the micro-lens array to mix the outputs of each set of RGB pixels. In this embodiment, the thickness of the diffuser sheet is small relative to the distance between the lens array and the display surface. The diffuser sheet is placed in front of and close to the display surface and is constructed with separated diffusing regions, one diffusing region for each pixel. Those diffusing regions would be separated by light barriers to prevent light from crossing from one diffusing region to any adjacent region. Thus only the colors within a single pixel would be mixed to effectively replace the color-striped pixel with one of the single color intended to be viewed at that location on the display screen. The thickness of the diffuser sheet, placement of light barriers between diffuser sheets, and the proximity of the diffuser sheet to the display surface can be varied to improve optical performance. In one alternative embodiment, each color segment of each pixel is treated as a separate pixel with its own separate lens element. It is understood that for a color display, reference to pixels in this description means the diffused pixels.
There are two ways to view a two-dimensional picture on the invention's three-dimensional display. One way is to locate both eyes on the same side of the common image point 116. The second way is to provide for switching electrically either the “a” set of pixels or the “b” set to display additional pixels corresponding to the picture being presented by the other set. Then both sets of pixels would display the same picture. Thus the display system can be switched easily and reversibly between 3-D and 2-D viewing.
Although certain embodiments are described as an add-on to an available display, they can be built into a display. That can be done in its entirety or in part. For example, only the diffusing sheet might be built in, which would not impede normal viewing of the display alone and would allow the lens array to be an add-on. It is understood that embodiments of the invention can provide 3-D viewing of both dynamic and static (e.g., printed) pictures that can be represented by a stereo pair of images.
In another embodiment, non-reflecting light barriers (not shown) extending from the display surface between adjacent pixel pairs to the boundaries of associated lenses are included to reduce light from the pixels associated with a given lens reaching another lens and being directed to locations outside the intended viewing area.
In another embodiment, the lenses in the lens array are all constructed the same size and with identical optical parameters, allowing for a uniform micro-lens array. In this case, each lens element would be slightly smaller than a pixel group and the overall dimensions of the micro-lens array would be slightly smaller than the dimensions of the overall display surface.
Another embodiment (not shown) places a lens in front of each pixel instead of in front of a pair of pixels or group of pixels. In this embodiment, each pixel is imaged independently to one side or the other of the common image point 116 so that that pixel can be seen by only one eye when the viewer's head is in its expected viewing location. Since in this embodiment the part of each object to be imaged at the common image point 116 would be an object edge rather than a point near its center, higher-quality optics are required. Also, twice as many smaller lenses would be required.
Yet another embodiment (not shown) provides a separate lens for each color region within each color pixel group, eliminating the need for a diffuser sheet for a color display. Such an approach requires a greater number of smaller lenses.
While this invention has been particularly shown and described with references to preferred embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the present application as defined by the appended claims. Such variations are intended to be covered by the scope of this present application. As such, the foregoing description of embodiments of the present application is not intended to be limiting. Rather, any limitations to the invention are presented in the following claims.
The present application claims the benefit of U.S. Provisional Patent Application No. 61/086,324 filed on Aug. 5, 2008, which is incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
61086324 | Aug 2008 | US |