Optical positioning device using telecentric imaging

Information

  • Patent Grant
  • 8345003
  • Patent Number
    8,345,003
  • Date Filed
    Monday, July 26, 2010
    14 years ago
  • Date Issued
    Tuesday, January 1, 2013
    11 years ago
Abstract
One embodiment relates to an optical displacement sensor for sensing movement of a data input device across a surface by determining displacement of optical features in a succession of frames. The sensor includes at least an illuminator, telecentric imaging optics on the object (scattering surface) side, and an array of photosensitive elements. The illuminator is configured to illuminate a portion of the surface. The telecentric imaging optics is configured to image the optical features emanating from the illuminated portion of the surface, and the array of photosensitive elements is configured to detect intensity data relating to the optical features imaged by the telecentric imaging optics. Other embodiments are also disclosed.
Description
TECHNICAL FIELD

The present invention relates generally to an Optical Positioning Device (OPD), and to methods of sensing movement using same.


BACKGROUND OF THE INVENTION

Pointing devices, such as computer mice or trackballs, are utilized for inputting data into and interfacing with personal computers and workstations. Such devices allow rapid relocation of a cursor on a monitor, and are useful in many text, database and graphical programs. A user controls the cursor, for example, by moving the mouse over a surface to move the cursor in a direction and over distance proportional to the movement of the mouse. Alternatively, movement of the hand over a stationary device may be used for the same purpose.


Computer mice come in both optical and mechanical versions. Mechanical mice typically use a rotating ball to detect motion, and a pair of shaft encoders in contact with the ball to produce a digital signal used by the computer to move the cursor. One problem with mechanical mice is that they are prone to inaccuracy and malfunction after sustained use due to dirt accumulation, and such. In addition, the movement and resultant wear of the mechanical elements, particularly the shaft encoders, necessarily limit the useful life of the device.


One solution to the above-discussed with mechanical mice problems has been the development of optical mice. Optical mice have become very popular because they are more robust and may provide a better pointing accuracy.


The dominant conventional technology used for optical mice relies on a light emitting diode (LED) illuminating a surface at grazing incidence, a two-dimensional CMOS (complementary metal-oxide-semiconductor) detector which captures the resultant images, and software that correlates successive images to determine the direction, distance and speed the mouse has been moved. This technology typically provides good accuracy but suffers from low optical efficiency and relatively high image processing requirements.


Another approach uses one-dimensional arrays of photo-sensors or detectors, such as photodiodes. Successive images of the surface are captured by imaging optics, translated onto the photodiodes, and compared to detect movement of the mouse. The photodiodes may be directly wired in groups to facilitate motion detection. This reduces the photodiode requirements, and enables rapid analog processing. An example of one such a mouse is disclosed in U.S. Pat. No. 5,907,152 to Dandliker et al.


The mouse disclosed in Dandliker et al. differs from the standard technology also in that it uses a coherent light source, such as a laser. Light from a coherent source scattered off of a rough surface generates a random intensity distribution of light known as speckle. The use of a speckle-based pattern has several advantages, including efficient laser-based light generation and high contrast images even under illumination at normal incidence. This allows for a more efficient system and conserves current consumption, which is advantageous in wireless applications so as to extend battery life.


Although a significant improvement over the conventional LED-based optical mice, these speckle-based devices have not been wholly satisfactory for a number of reasons. In particular, mice using laser speckle have not demonstrated the accuracy typically demanded in state-of-the-art mice today, which generally are desired to have a path error of less than 0.5% or thereabout.


The present disclosure discusses and provides solutions to certain problems with prior optical mice and other similar optical pointing devices.


SUMMARY OF THE INVENTION

One embodiment relates to an optical displacement sensor for sensing movement of a data input device across a surface by determining displacement of optical features in a succession of frames. The sensor includes at least an illuminator, telecentric imaging optics, and an array of photosensitive elements. The illuminator is configured to illuminate a portion of the surface. The telecentric imaging optics is configured to image the optical features emanating from the illuminated portion of the surface, and the array of photosensitive elements is configured to detect intensity data relating to the optical features imaged by the telecentric imaging optics.


Another embodiment relates to a method of sensing movement of a data input device across a surface. A portion of the surface is illuminated and images of the illuminated portion of the surface are focused to an array of photosensitive elements using telecentric imaging optics. Displacement of optical features emanating from the illuminated portion of the surface are determined in a succession of frames to sense movement of the data input device across the surface in at least one dimension.


Another embodiment relates to an optical positioning device which includes at least a laser configured to illuminate a portion of a surface, an array of photosensitive elements; telecentric imaging optics, and a signal processor. The telecentric imaging optics is configured to map a speckle pattern generated by the illuminated portion of the surface along with finite aperture imaging optics to the array of photosensitive elements, and the signal processor is configured to determine displacement of the speckle pattern in a succession of frames.


Other embodiments are also disclosed.





BRIEF DESCRIPTION OF THE DRAWINGS

These and various other features and advantages of the present disclosure are understood more fully from the detailed description that follows and from the accompanying drawings, which however, should not be taken to limit the appended claims to the specific embodiments shown, but are for explanation and understanding only, where:



FIGS. 1A and 1B illustrate, respectively, a diffraction pattern of light reflected from a smooth surface and speckle in an interference pattern of light reflected from a rough surface;



FIG. 2 is a functional block diagram of a speckle-based OPD in accordance with an embodiment of the invention;



FIG. 3 is a block diagram of an array having interlaced groups of photosensitive elements in accordance with an embodiment of the invention;



FIG. 4 is a ray diagram which shows an example non-telecentric imaging system;



FIG. 5A is a ray diagram showing an example telecentric imaging system with chief rays normal to a surface according to an embodiment of the present invention;



FIG. 5B is a ray diagram showing an example telecentric imaging system with chief rays at a non-normal angle to a surface according to an embodiment of the present invention;



FIG. 6 is an optical diagram of a non-telecentric imaging system in which the aperture is placed at the lens location;



FIG. 7 is an optical diagram of an imaging system which is telecentric on the object side with chief rays normal to a surface according to an embodiment of the present invention;



FIG. 8 is an optical diagram of an imaging system which is telecentric on the image side;



FIG. 9 is an optical diagram of an imaging system which is telecentric on both object and image sides according to an embodiment of the present invention; and



FIG. 10 is a ray diagram of a system including illumination optics and imaging optics for a sensor according to an embodiment of the present invention.





DETAILED DESCRIPTION
Speckle “Boiling” Problem

One problem with prior speckle-based OPDs relates to the changing of the speckle pattern, or speckle “boiling”. In general, the speckle pattern from a surface moves as the surface is moved, and in the same direction with the same velocity. However, in many optical systems there will be additional changes in the phase front coming off of the surface. For example, if the set of rays that participate in the speckle pattern formation at the detector changes as the system is moved relative to the surface, then the detected speckle pattern will change (or “boil”) in a somewhat random manner instead of simply shift. This degrades the signal used to detect the surface motion (from the shift), leading to decreases in the accuracy and sensitivity of the system.


As discussed in detail below, one aspect of the present invention provides an OPD with negligible or reduced speckle boiling.


OPD Embodiments Disclosed Herein


The present disclosure relates generally to a sensor for an Optical Positioning Device (OPD), and to methods for sensing relative movement between the sensor and a surface based on displacement of a random intensity distribution pattern of light, known as speckle, reflected from the surface. OPDs include, but are not limited to, optical mice or trackballs for inputting data to a personal computer.


Reference in the specification to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the invention. The appearances of the phrase “in one embodiment” in various places in the specification do not necessarily all refer to the same embodiment.


Generally, the sensor for an OPD includes an illuminator having a light source and illumination optics to illuminate a portion of the surface, a detector having a number of photosensitive elements and imaging optics, and signal processing or mixed-signal electronics for combining signals from each of the photosensitive elements to produce an output signal from the detector.


In one embodiment, the detector and mixed-signal electronics are fabricated using standard CMOS processes and equipment. Preferably, the sensor and method of the present invention provide an optically-efficient detection architecture by use of structured illumination that produces uniform phase-front and telecentric speckle-imaging as well as a simplified signal processing configuration using a combination of analog and digital electronics. This architecture reduces the amount of electrical power dedicated to signal processing and displacement-estimation in the sensor. It has been found that a sensor using the speckle-detection technique, and appropriately configured in accordance with the present invention can meet or exceed all performance criteria typically expected of OPDs, including maximum displacement speed, accuracy, and % path error rates.


Introduction to Speckle-Based Displacement Sensors


This section discusses operating principles of speckle-based displacement sensors as understood and believed by the applicants. While these operating principles are useful for purposes of understanding, it is not intended that embodiments of the present disclosure be unnecessarily limited by these principles.


Referring to FIG. 1A, laser light of a wavelength indicated is depicted as a first incident wave 102 and second incident wave 104 to a surface, each making an angle of incidence θ with respect to the surface normal. A diffraction pattern 106 results which has a periodicity of λ/2 sin θ.


In contrast, referring to FIG. 1B, any general surface with morphological irregularities of dimensions greater than the wavelength of light (i.e. roughly >1 μm) will tend to scatter light 114 into a complete hemisphere in approximately a Lambertian fashion. If a coherent light source, such as a laser is used, the spatially coherent, scattered light will create a complex interference pattern 116 upon detection by a square-law detector with finite aperture. This complex interference pattern 116 of light and dark areas is termed speckle. The exact nature and contrast of the speckle pattern 116 depends on the surface roughness, the wavelength of light and its degree of spatial-coherence, and the light-gathering or imaging optics. Although often highly complex, a speckle pattern 116 is distinctly characteristic of a section of any rough surface that is imaged by the optics and, as such, may be utilized to identify a location on the surface as it is displaced transversely to the laser and optics-detector assembly.


Speckle is expected to come in all sizes up to the spatial frequency set by the effective aperture of the optics, conventionally defined in term of its numerical aperture NA=sin θ as shown FIG. 1B. Following Goodman [J. W. Goodman, “Statistical Properties of Laser Speckle Patterns” in “Laser Speckle and Related Phenomena” edited by J. C. Dainty, Topics in Applied Physics volume 9, Springer-Verlag (1984)—in particular, see page 39-40.], the size statistical distribution is expressed in terms of the speckle intensity auto-correlation. The “average” speckle diameter may be defined as









a
=


λ

sin





θ


=

λ
NA






(

Equation





1

)







It is interesting to note that the spatial frequency spectral density of the speckle intensity, which by Wiener-Khintchine theorem, is simply the Fourier transform of the intensity auto-correlation. The finest possible speckle, amin=λ/2NA, is set by the unlikely case where the main contribution comes from the extreme rays 118 of FIG. 1B (i.e. rays at ±θ), and contributions from most “interior” rays interfere destructively. The cut-off spatial frequency is therefore fco=1/(λ/2NA) or 2NA/λ.


Note that the numerical aperture may be different for spatial frequencies in the image along one dimension (say “x”) than along the orthogonal dimension (“y”). This may be caused, for instance, by an optical aperture which is longer in one dimension than another (for example, an ellipse instead of a circle), or by anamorphic lenses. In these cases, the speckle pattern 116 will also be anisotropic, and the average speckle size will be different in the two dimensions.


One advantage of a laser speckle-based displacement sensor is that it can operate with illumination light that arrives at near-normal incidence angles. Sensors that employ imaging optics and incoherent light arriving at grazing incident angles to a rough surface also can be employed for transverse displacement sensing. However, since the grazing incidence angle of the illumination is used to create appropriately large bright-dark shadows of the surface terrain in the image, the system is inherently optically inefficient, as a significant fraction of the light is reflected off in a specular manner away from the detector and thus contributes nothing to the image formed. In contrast, a speckle-based displacement sensor can make efficient use of a larger fraction of the illumination light from the laser source, thereby allowing the development of an optically efficient displacement sensor.


Disclosed Design for Speckle-Based Displacement Sensor


The detailed description below describes an architecture for one such laser-speckle-based displacement sensor using CMOS photodiodes with analog signal combining circuitry, moderate amounts of digital signal processing circuitry, and a low-power light source, such as, for example, a 850 nm Vertical Cavity Surface Emitting Laser (VCSEL). While certain implementational details are discussed in the detailed description below, it will be appreciated by those skilled in the art that different light sources, detector or photosensitive elements, and/or different circuitry for combining signals may be utilized without departing from the spirit and scope of the present invention.


A speckle-based mouse according to an embodiment of the present invention will now be described with reference to FIGS. 2 and 3.



FIG. 2 is functional diagram of a speckle-based system 200 according to an embodiment of the invention. The system 200 includes a laser source 202, illumination optics 204, imaging optics 208, at least two sets of multiple CMOS photodiode arrays 210, front-end electronics 212, signal processing circuitry 214, and interface circuitry 216. The photodiode arrays 210 may be configured to provide displacement measurements along two orthogonal axes, x and y. Groups of the photodiodes in each array may be combined using passive electronic components in the front-end electronics 212 to produce group signals. The group signals may be subsequently algebraically combined by the signal processing circuitry 214 to produce an (x, y) signal providing information on the magnitude and direction of displacement of the OPD in x and y directions. The (x,y) signal may be converted by the interface circuitry 218 to x,y data 220 which may be output by the OPD. Sensors using this detection technique may have arrays of interlaced groups of linear photodiodes known as “differential comb arrays.”



FIG. 3 shows a general configuration (along one axis) of such a photodiode array 302, wherein the surface 304 is illuminated by a coherent light source, such as a Vertical Cavity Surface Emitting Laser (VCSEL) 306 and illumination optics 308, and wherein the combination of interlaced groups in the array 302 serves as a periodic filter on spatial frequencies of light-dark signals produced by the speckle images.


Speckle generated by the rough surface 304 is mapped to the detector plane with imaging optics 310. Preferably, the imaging optics 310 are telecentric for optimum performance.


In one embodiment, the comb array detection is performed in two independent, orthogonal arrays to obtain estimations of displacements in x and y. A small version of one such array 302 is depicted in FIG. 3.


Each array in the detector consists of a number, N, of photodiode sets, each set having a number, M, of photodiodes (PD) arranged to form an MN linear array. In the embodiment shown in FIG. 3, each set consists of four photodiodes (4 PD) referred to as 1, 2, 3, 4. The PD1s from every set are electrically connected (wired sum) to form a group, likewise PD2s, PD3s, and PD4s, giving four signal lines coming out from the array. Their corresponding currents or signals are I1, I2, I3, and I4. These signals (I1, I2, I3, and I4) may be called group signals. Background suppression (and signal accentuation) is accomplished by using differential analog circuitry 312 to generate an in-phase differential current signal 314 (I13)=I1−I3 and differential analog circuitry 316 to generate a quadrature differential current signal 318 (I24)=I2−I4. These in-phase and quadrature signals may be called line signals. Comparing the phase of I13 and I24 permits detection of the direction of motion.


Preferably, to suppress the introduction of phase errors, which can translate directly into displacement error, the sensor of the present invention uses multiple comb arrays. Moreover, although the embodiments described herein use “4N” scheme for the individual arrays, the system design rationale is applicable (with appropriate modifications) for other array configurations or schemes, such as 3N, 5N, 6N, 7N, 8N, and so on. The terminology “4N” refers to a detector array in which every fourth detector is wired together, and the resulting four photocurrent signals are subtracted from each other as described in Dandliker, et al. (U.S. Pat. No. 5,907,152). However, many other groupings are possible with appropriate schemes for combining the signals.


Telecentric Imaging to Reduce Speckle Boiling Problem


The speckle pattern, such as the example shown in FIG. 1, may be use as the basis for displacement measurement where a surface undergoes a displacement relative to the measuring device, such as a speckle mouse. The displacement may be deduced from the displacement of the speckle pattern on the detector (after taking into account the effective optical magnification factor). Therefore, one factor for a successful measurement is that the speckle pattern undergo a nearly pure translation for small displacements [i.e. small compared to the detector field-of-view (FOV)], with only relatively minor changes in the pattern itself. In such situations, correlation or other signal-processing algorithms may be used to determine the physical displacement.


The change in speckle pattern comes from at least two sources. A first source of change is due to part of the surface leaving the detector FOV, while a new part is entering the detector FOV, as the detector system is displaced with respect to the surface. This source of speckle pattern change is unavoidable, but its effect be minimized by estimating the displacement in a time which is much faster than it takes the detector to move across its FOV.


A second source of speckle pattern change is due to changes of viewing angle with respect to points on the rough surface as the surface is displaced. As discussed in the present patent application, applicants note that this second source of change occurs if the imaging optics is non-telecentric on the object side (the scattering surface), meaning that the angular extent of ray depends on the field point—see FIG. 4, for example.


The illustration in FIG. 4 shows an angular extent of several (five) illustrative contributing rays being emitted from each of several points on a surface 402 under the lens 404 of an imaging system which is non-telecentric on the object side. Consider that the rays illustrated pass through a system aperture (not illustrated) of the non-telecentric image system, and further consider that the middle ray of each set of contributing rays is a chief ray in that it passes through the center of the system aperture. As seen from FIG. 4, the angle of the chief ray with respect to the surface at each image point depends on the location of that point in the field of view. In other words, the angular extent of the contributing rays at each image point is field-point dependent.


As the detector of an object-side non-telecentric imaging system is displaced, the set of rays from a given field point that contributes to the image changes. At an undesirable extreme, a completely uncorrelated speckle would occur when the angular change is greater than or equal to twice the lens numerical aperture. In order to maintain relatively translation-invariant speckle patterns, the angular change should be less than the lens numerical aperture (NA).


Imaging systems which are telecentric on the object side are illustrated in FIGS. 5A and 5B. In object-side telecentric imaging, the angular-extent of the contributing rays at each image point is substantially field-point independent, and the speckle pattern is substantially translation invariant for displacements small compared to the imaging system FOV. In other words, the chief rays from each image point are parallel.


The first example in FIG. 5A shows a lens 502 for an object-side telecentric imaging system configured such that the chief rays are normal to the surface 402.


The second example in FIG. 5B shows a lens 504 of an object-side telecentric imaging system configured such that the chief rays are at a non-normal angle to the surface 402. This embodiment permits the use of the same lens 504 for illumination and imaging functions. In such a configuration, in addition to the imaging function, the lens may be used to direct illuminating rays 506 at the field of view from a different angle.


Chief rays are rays that go through the center of the system aperture. As used in the present disclosure, an imaging system is telecentric on the object side if all or a substantial majority of the chief rays coming from the object (or surface) are parallel or practically parallel. Similarly, an imaging system is telecentric on the image side if all or a substantial majority of the chief rays going toward the image (or detector) are parallel or substantially parallel.


In an object-side telecentric system, moving the object axially (along the z-axis such that the object surface moves closer to or farther from the imaging system) should result in no or little change in the magnification of the image. This is because of the chief rays being parallel, as discussed above. Moving the object axially may defocus the image, but such defocus should not be problematic when tracking movement based on movement of the speckle pattern.


In accordance with an embodiment of the invention, an optical position sensing system may be adapted to have substantially reduced speckle boiling if it is telecentric on the object side and if its illumination beam has a substantially planar phase front before it strikes the surface. In contrast, substantial speckle boiling occurs without object-side telecentric optics and/or without a uniform phase front of the illumination beam. A non-uniform phase front may be, for example, divergent.


The imaging system that is telecentric on the object side and that has a substantially planar (i.e. has a substantially uniform) phase front before it impinges the surface results in reduced speckle boiling during translation. On the other hand, the imaging system that is telecentric on the imaging side, but not on the object side, does not reduce speckle boiling during translation.



FIGS. 6 and 7 show that telecentricity may be determined by the choice of aperture location in an imaging system.



FIG. 6 shows an imaging system configured such that the system aperture 604 is located immediately behind the lens 602. This is a non-telecentric system, as the chief ray (the thicker ray) goes through the lens 602 at an angle, which is different for every image point on the surface 402.



FIG. 7 shows an imaging system which is telecentric on the object side of the lens 702. In the configuration of FIG. 7, the system aperture 704 is located at the back focal plane, and all or substantially all of the chief rays are parallel (normal in this case) at the object surface 402. The back focal plane is located an optical distance away from the lens 702 equal to the focal length f of the lens. The optical distance equals the actual distance between the lens 702 and the back focal plane in a case where the medium is air or vacuum, as depicted in FIG. 7, but that if a material with an index of refraction different from one is used, then the actual distance between the lens 702 and the system aperture 704 is to be adjusted accordingly.


The object-side telecentric system illustrated in FIG. 7 has an aperture located at the back focal plane. Hence, a collimated beam from the object side should be focused and passed by the aperture efficiently. In other words, a substantial majority of the collimated beam should pass through the back focal plane aperture of the object-side telecentric system, but that is not necessarily true for a non-telecentric system.


While FIG. 7 shows an embodiment of an object-side telecentric system configured such that the chief rays are normal to the surface 402, alternate embodiments may be configured such that chief rays are at angles non-normal to the surface 402, for example, as discussed above in relation to FIG. 5B.


While FIG. 7 shows an imaging system which is telecentric on the object side, FIG. 8 shows an imaging system which is telecentric on the image side. In FIG. 8, the aperture 804 is located at the front focal plane of the lens 802. The system in FIG. 8 does not reduce speckle boiling during translation.


A system which is telecentric on both the object and image sides is shown in FIG. 9. Here f1 and f2 are the focal lengths of the left-hand (object-side) lens 902 and right-hand lens 904, respectively. The focal length of the front lens 902 is f1, and the focal length of the back lens 904 is f2. In this configuration, the aperture 906 is located at the back focal point of the front lens 902, which coincides with the front focal point of the back lens 904. The focal lengths may be chosen such that the magnification=f2/f1 is as desired, and to meet any size constraint on the total length f1+f2.


In accordance with an embodiment of the invention, in order to maintain a speckle image which translates without changing or with very little change, an image point at the detector may be formed from the same bundle of rays as the object (surface) is translated. In other words, the angular extent of the contributing rays should be invariant or near invariant as the surface translation occurs. Such invariance is present if the optical system is telecentric on the object side, for example, as in FIG. 7 or 9. Because the aperture and lenses move with the detector, telecentricity on the image side may have some advantages but is not required.


In the above discussion, the plane of the system aperture is located where the numerical aperture of the image is defined. The aperture may comprise a purposely introduced opaque structure with transparent sections, for example, as depicted in FIGS. 7 and 9. Alternatively, the aperture may comprise a lens in the system or may comprise other means for defining such an aperture.


Optics Layout


As discussed above, telecentric imaging optics (for example, as shown in FIG. 7 or 9) may be used to produce robust speckle signals that translate substantially without changing. Such imaging optics may be combined with illumination optics to form a complete system.


One embodiment of this combination of illumination optics 1002 and imaging optics 1004 is shown in FIG. 10. In this embodiment the illumination optics 1002 is configured such that the illumination 1006 comes at a non-zero incident angle θ with respect to the normal vector for the surface 402.


The foregoing description of specific embodiments and examples of the invention have been presented for the purpose of illustration and description, and although the invention has been described and illustrated by certain of the preceding examples, it is not to be construed as being limited thereby. They are not intended to be exhaustive or to limit the invention to the precise forms disclosed, and many modifications, improvements and variations within the scope of the invention are possible in light of the above teaching. It is intended that the scope of the invention encompass the generic area as herein disclosed, and by the claims appended hereto and their equivalents.

Claims
  • 1. An apparatus, comprising: an illuminator configured to illuminate a portion of a surface using a coherent light source;object-side telecentric imaging optics configured to reduce speckle boiling of optical features of a speckle pattern generated by the coherent light source and emanating from the illuminated portion of the surface; andan array of photosensitive elements comprising a comb array, wherein the array of photosensitive elements is configured to detect intensity data relating to the optical features imaged by the object-side telecentric imaging optics.
  • 2. The optical displacement sensor according to claim 1, wherein the optical features imaged by the object-side telecentric imaging optics are substantially invariant during translation of the surface with respect to the data input device.
  • 3. The optical displacement sensor according to claim 1, wherein the coherent light source comprises a Vertical Cavity Surface Emitting Laser (VCSEL), and wherein the photosensitive elements comprise photodiodes.
  • 4. The optical displacement sensor according to claim 1, wherein the object-side telecentric imaging optics is configured such that chief rays from points on the surface in a field of view are substantially parallel when the chief rays leave the surface.
  • 5. The optical displacement sensor according to claim 4, wherein the chief rays are normal to the surface when the chief rays leave the surface.
  • 6. The optical displacement sensor according to claim 4, wherein the chief rays leave the surface at an angle which is not normal to the surface.
  • 7. The optical displacement sensor according to claim 1, wherein the object-side telecentric imaging optics comprises: an aperture; anda first lens located between the aperture and the illuminated portion of the surface.
  • 8. The optical displacement sensor according to claim 7, wherein the first lens has a focal length substantially equal to an optical distance between the first lens and the aperture.
  • 9. The optical displacement sensor according to claim 8, further comprising image-side telecentric imaging optics which comprises: a second lens located between the aperture and the photosensitive elements, wherein the second lens has a focal length substantially equal to the optical distance between the second lens and the aperture.
  • 10. The optical displacement sensor according to claim 7, wherein the aperture comprises an optical element which defines a numerical aperture of light which is detected by the photosensitive elements.
  • 11. The optical displacement sensor according to claim 10, wherein the aperture further comprises an optical element which has both a substantially transparent section and a substantially opaque section.
  • 12. The optical displacement sensor according to claim 11, wherein the aperture has a substantially transparent section which is circular in shape.
  • 13. The optical displacement sensor according to claim 1, wherein the optical features comprise speckle, a complex interference pattern created by light reflected from the surface, and wherein the optical displacement sensor comprises a speckle-based displacement sensor adapted to sense movement of the data input device across the surface based on displacement of the speckle pattern.
  • 14. A method comprising: illuminating a portion of a surface using coherent light;reducing speckle boiling of a speckle pattern generated by the coherent light using object-side telecentric imaging optics;mapping images of the speckle pattern to an array of photosensitive elements comprising a comb array using the telecentric imaging optics; anddetermining displacement of optical features of the speckle pattern emanating from the illuminated portion of the surface in a succession of frames to sense movement of the data input device across the surface in at least one dimension.
  • 15. The method according to claim 14, wherein the illuminating is performed by using a laser.
  • 16. The method according to claim 14, further comprising detecting intensity data of the optical features.
  • 17. The method according to claim 14, further comprising sensing movement of an input device relative to a surface based on the displacement of the optical features.
  • 18. An optical positioning device comprising: a laser configured to illuminate a portion of the surface;an array of photosensitive elements;telecentric imaging optics on an object side configured to reduce speckle boiling of a speckle pattern from the illuminated portion of the surface and to focus the speckle pattern to the array of photosensitive elements, wherein the array of photosensitive elements comprises a comb array; anda signal processor configured to determine displacement of the speckle pattern in a succession of frames.
  • 19. The optical positioning device of claim 18, wherein the object-side telecentric imaging optics comprises: an aperture; anda lens located between the aperture and the illuminated portion of the surface.
  • 20. The optical positioning device of claim 18, wherein the illumination has a substantially uniform wave front before impinging upon the surface.
CROSS-REFERENCE TO RELATED APPLICATIONS

The present application is a continuation of U.S. nonprovisional application Ser. No. 11/125,858, now U.S. Pat. No. 7,773,070, filed May 9, 2005, which claims the benefit of U.S. provisional application No. 60/573,316, filed May 21, 2004, and U.S. provisional application No. 60/573,075, filed May 21, 2004, all of which are hereby incorporated by reference herein in their entirety.

US Referenced Citations (147)
Number Name Date Kind
3922093 Dandliker et al. Nov 1975 A
4546347 Kirsch Oct 1985 A
4560881 Briggs Dec 1985 A
4751380 Victor et al. Jun 1988 A
4799055 Nestler et al. Jan 1989 A
4814553 Joyce Mar 1989 A
4920260 Victor et al. Apr 1990 A
4936683 Purcell Jun 1990 A
4967093 Takemori Oct 1990 A
5006711 Hamashima et al. Apr 1991 A
5061860 Takemori Oct 1991 A
5086197 Liou Feb 1992 A
5288993 Bidiville et al. Feb 1994 A
5345257 Deguchi et al. Sep 1994 A
5345527 Lebby et al. Sep 1994 A
5391868 Vampola et al. Feb 1995 A
5473344 Bacon et al. Dec 1995 A
5578813 Allen et al. Nov 1996 A
5578817 Bidiville et al. Nov 1996 A
5606174 Yoshimura et al. Feb 1997 A
5627363 Paxman et al. May 1997 A
5644139 Allen et al. Jul 1997 A
D382550 Kaneko et al. Aug 1997 S
D385542 Kaneko et al. Oct 1997 S
5703356 Bidiville et al. Dec 1997 A
5729008 Blalock et al. Mar 1998 A
5729009 Daendliker et al. Mar 1998 A
5781229 Zediker et al. Jul 1998 A
5786804 Gordon Jul 1998 A
5825044 Allen et al. Oct 1998 A
5854482 Bidiville et al. Dec 1998 A
5907152 Daendliker et al. May 1999 A
5963197 Bacon et al. Oct 1999 A
5994710 Knee et al. Nov 1999 A
6031218 Piot et al. Feb 2000 A
6034760 Rees Mar 2000 A
6037643 Knee Mar 2000 A
6057540 Gordon et al. May 2000 A
6097371 Siddiqui et al. Aug 2000 A
6151015 Badyal et al. Nov 2000 A
6172354 Adan et al. Jan 2001 B1
6176143 Mo et al. Jan 2001 B1
6195475 Beausoleil, Jr. et al. Feb 2001 B1
6218659 Bidiville et al. Apr 2001 B1
6222174 Tullis et al. Apr 2001 B1
6225617 Daendliker et al. May 2001 B1
6226092 De Lega May 2001 B1
6233044 Brueck et al. May 2001 B1
6233368 Badyal et al. May 2001 B1
6246050 Tullis et al. Jun 2001 B1
6255643 Sayag Jul 2001 B1
6256016 Piot et al. Jul 2001 B1
6281881 Siddiqui et al. Aug 2001 B1
6281882 Gordon et al. Aug 2001 B1
6304330 Millerd et al. Oct 2001 B1
6326950 Liu Dec 2001 B1
6330057 Lederer et al. Dec 2001 B1
6351257 Liu Feb 2002 B1
6396479 Gordon May 2002 B2
6421045 Venkat et al. Jul 2002 B1
6424407 Kinrot et al. Jul 2002 B1
6433780 Gordon et al. Aug 2002 B1
6452683 Kinrot et al. Sep 2002 B1
6455840 Oliver et al. Sep 2002 B1
D464352 Kerestegian Oct 2002 S
6462330 Venkat et al. Oct 2002 B1
6476970 Smith Nov 2002 B1
6529184 Julienne Mar 2003 B1
6585158 Norskog Jul 2003 B2
6603111 Dietz et al. Aug 2003 B2
6618038 Bohn Sep 2003 B1
6621483 Wallace et al. Sep 2003 B2
6642506 Nahum et al. Nov 2003 B1
6657184 Anderson et al. Dec 2003 B2
6664948 Crane et al. Dec 2003 B2
6674475 Anderson Jan 2004 B1
6677929 Gordon et al. Jan 2004 B2
6703599 Casebolt et al. Mar 2004 B1
6707027 Liess et al. Mar 2004 B2
6710855 Shiraishi Mar 2004 B2
6737636 Dietz et al. May 2004 B2
6774351 Black Aug 2004 B2
6774915 Rensberger Aug 2004 B2
6795056 Norskog et al. Sep 2004 B2
6809723 Davis Oct 2004 B2
6819314 Black Nov 2004 B2
6823077 Dietz et al. Nov 2004 B2
6825998 Yoshida Nov 2004 B2
6869185 Kaminsky et al. Mar 2005 B2
6947459 Kurtz et al. Sep 2005 B2
6951540 Ebbini et al. Oct 2005 B2
7042575 Carlisle et al. May 2006 B2
7045763 Ballard May 2006 B2
7098894 Yang et al. Aug 2006 B2
7116427 Baney et al. Oct 2006 B2
7138620 Trisnadi et al. Nov 2006 B2
7161582 Bathiche et al. Jan 2007 B2
7189985 Xie et al. Mar 2007 B2
7205521 Gruhlke et al. Apr 2007 B2
7221356 Oliver et al. May 2007 B2
7227531 Lutian Jun 2007 B2
7248345 Todoroff et al. Jul 2007 B2
7268341 Lehoty et al. Sep 2007 B2
7285766 Carlisle et al. Oct 2007 B2
7292232 Ranta et al. Nov 2007 B2
7295324 Jones et al. Nov 2007 B2
7423279 Heinz et al. Sep 2008 B2
7435942 Lang Oct 2008 B2
7505033 Guo et al. Mar 2009 B2
7515280 Emtman et al. Apr 2009 B2
7746477 Huber et al. Jun 2010 B1
7773070 Trisnadi et al. Aug 2010 B2
20020113790 Hayashi Aug 2002 A1
20020130835 Brosnan Sep 2002 A1
20020145588 McCahon et al. Oct 2002 A1
20020158300 Gee Oct 2002 A1
20020190953 Gordon et al. Dec 2002 A1
20030034959 Davis et al. Feb 2003 A1
20030058506 Green et al. Mar 2003 A1
20030142288 Kinrot et al. Jul 2003 A1
20040000634 Ballard Jan 2004 A1
20040084610 Leong et al. May 2004 A1
20040189593 Koay Sep 2004 A1
20050024336 Xie et al. Feb 2005 A1
20050024623 Xie et al. Feb 2005 A1
20050024624 Gruhlke et al. Feb 2005 A1
20050035947 Lutian Feb 2005 A1
20050083303 Schroeder et al. Apr 2005 A1
20050094154 Baney et al. May 2005 A1
20050109961 Bittner May 2005 A1
20050117137 Hase Jun 2005 A1
20050156915 Fisher Jul 2005 A1
20050228838 Stetson et al. Oct 2005 A1
20050258346 Lehoty et al. Nov 2005 A1
20050259078 Roxlo et al. Nov 2005 A1
20050259097 Lehoty et al. Nov 2005 A1
20050259267 Carlisle et al. Nov 2005 A1
20060091301 Trisnadi et al. May 2006 A1
20060106319 Todoroff et al. May 2006 A1
20060118743 Lang Jun 2006 A1
20060187209 Lai et al. Aug 2006 A1
20060279545 Lan et al. Dec 2006 A1
20070057157 Trisnadi et al. Mar 2007 A1
20070126700 Wright Jun 2007 A1
20070165207 Xu et al. Jul 2007 A1
20080007526 Xu et al. Jan 2008 A1
20090027038 Garmire et al. Jan 2009 A1
Foreign Referenced Citations (2)
Number Date Country
9946603 Sep 1999 WO
0248853 Jun 2002 WO
Provisional Applications (2)
Number Date Country
60573316 May 2004 US
60573075 May 2004 US
Continuations (1)
Number Date Country
Parent 11124858 May 2005 US
Child 12843864 US