The present invention relates to a three-dimensional imaging system for robot vision comprising at least one micromirror array lens.
Robots have been widely used in the industry to replace human who had to perform dangerous and repetitive tasks. While earlier robots are usually required to manipulate objects in a limited range with less or no flexibility like robot arms in assembly lines, recent robots have become more intelligent and perceptive. Such robots are capable of performing more complex and difficult tasks including navigation, inspection, self-learning, and self-calibration thanks to advanced technologies for computation resources and sensory systems with lower cost. Some applications of these advanced robots include, but not limited to, housekeeping, underwater/space exploration, surgery with precision, and mine finding and mining.
Many aspects of sensory systems being used in robotics are adopted from human biological systems. Human senses including sight, hearing, smell, touch, and taste are very acute and efficient considering small sizes and fast processing times of their sensory systems. In the early stages of adopting these human sensing procedures, it was very difficult to create corresponding artificial sensory systems because of complexity, limited resources, and a lack of knowledge. Since then, a lot of efforts have been put into researches for these areas, and huge progresses have been made. Especially, human vision systems are relatively well understood and adopted in most advanced robot systems as a primary sensory system.
Some important features of human vision include three-dimensional perception, continuous tracking of a moving object, rapid object identification, and the like. Among them, three-dimensional perception is a fundamental element since it allows other features available. Many advanced robot systems that are required to perform navigation and/or manipulation in known or unknown environments have adopted three-dimensional vision systems, which collect and process environmental information surrounding a robot, and let the robot properly respond to stimuli without interruption from outside sources.
Typically, three-dimensional vision for robots can be accomplished by using stereo vision or optical flow methods, in which two images are compared in order to determine the three-dimensional location of an object. The former uses images taken by two parallel cameras that are disposed to view the object from different angles at the same time as disclosed in U.S. Pat. No. 5,432,712 to Chan, while the latter uses images taken at two different times by a single camera as disclosed in U.S. Pat. No. 5,109,425 to Lawton. Both methods require to find corresponding points of two different images using certain criteria such as color, shape, contrast, or other representative features. However, these processes can be very erroneous and time-consuming.
Research suggests that the human vision system is more efficient and effective in that it is capable of a rapid eye movement to the point of interest and contrasting high central visual resolution with low peripheral visual resolution in a wide field of view as disclosed in U.S. Pat. No. 5,103,306 to Weiman et al. These aspects demand fast changes of the optical axis and field of view of a lens system. However, it is difficult to accomplish such efficiency and effectiveness in a conventional robot vision system since those changes are usually performed by a complicated macroscopic servo mechanism.
To overcome the drawbacks of existing technologies, a desirable robot vision system requires a high-speed, accurate, miniaturized, and inexpensive three-dimensional imaging system.
The present invention provides a three-dimensional imaging system for robot vision, which is capable of three-dimensional positioning of an object, object identification, searching and tracking the object, and compensation for the aberration of the system.
One aspect of the invention is to provide a three-dimensional imaging system for robot vision that generates an all-in-focus image and three-dimensional position information of an object.
The three-dimensional imaging system for robot vision comprises at least one camera system having a lens system including at least one variable focal length micromirror array lens (MMAL), an imaging unit, and an image processing unit.
The variable focal length MMAL comprises a plurality of micromirrors. The following US patents and applications describe the MMAL: U.S. Pat. No. 6,934,072 to Kim, U.S. Pat. No. 6,934,073 to Kim, U.S. patent application Ser. No. 10/855,554 filed May 27, 2004, U.S. patent application Ser. No. 10/855,715 filed May 27, 2004, U.S. patent application Ser. No. 10/857,714 filed May 28, 2004, U.S. patent application Ser. No. 10/857,280 filed May 28, 2004, U.S. patent application Ser. No. 10/893,039 filed May 28, 2004, and U.S. patent application Ser. No. 10/983,353 filed Mar. 4, 2005, all of which are hereby incorporated by reference.
The variable focal length MMAL is suitable for the three-dimensional imaging system of the present invention since it has a fast focusing speed and a large range of focal length, and since it can be made to have either a small or a large aperture just adding more micromirrors to cover the aperture area.
The imaging unit includes one or more two-dimensional image sensors taking two-dimensional images at different focal planes. The detail for three-dimensional imaging using the variable focal length MMAL can be found in U.S. patent application Ser. No. 10/822,414 filed Apr. 12, 2004, U.S. patent application Ser. No. 10/979,624 filed Nov. 2, 2004, and U.S. patent application Ser. No. 11/208,115 filed Aug. 19, 2005.
The image sensor takes two-dimensional images of an object or scene with one or more focal planes that are shifted by changing the focal length of the variable focal length MMAL. The image processing unit extracts substantially in-focus pixels or areas from each two-dimensional image to generate a corresponding in-focus depthwise image. Each in-focus depthwise image represents a portion of the object or scene having the same image depth. Based on the known focal length of the two-dimensional image, the known distance from the lens to the image plane, and the magnification of the lens, three-dimensional position information of a portion of the object corresponding to each pixel of the in-focus depthwise image can be obtained. The focal length of the variable focal length MMAL can progressively increase or decrease, or vary in a selected order within a focal length variation range of the variable focal length MMAL such that any portion of the object or scene is imaged substantially in-focus at least once. A set of in-focus depthwise images taken at different focal lengths with a fast imaging rate represents the object or scene at a given moment. The object can remain still or be moving. For the case that the object is moving, the movement of the object can be ignored when the imaging rate is fast enough. The number of in-focus depthwise images representing the object at a given moment (number of depths) depends on the depth resolution requirement and the focusing speed of the variable focal length MMAL, and may increase for a better image quality. There are several methods for the image processing unit to generate an all-in-focus image of the object or scene from in-focus depthwise images thereof. Recent advances in both the image sensor and the image processing unit make them as fast as they are required to be. Three-dimensional position information of a portion of the object corresponding to each pixel of the all-in-focus image is obtained in the same way as in the in-focus depthwise image case. All the processes to obtain an all-in-focus image and three-dimensional position information of the object are achieved within a unit time which is at least persistent rate of the human eye.
Three-dimensional position information of an object or scene is a necessary element for robot's object manipulation and navigation. Also, the all-in-focus image and three-dimensional position information of the object enables the object identification to be more accurate and robust since three-dimensional object identification is less subject to the object orientation, illumination, and other environmental factors. The detail for the three-dimensional imaging system for pattern recognition using the variable focal length MMAL can be found in U.S. patent application Ser. No. (11/294,944) filed Dec. 6, 2005.
The next aspect of the invention is to provide an optical tracking unit using a three-dimensional imaging system for robot vision that performs searching and tracking an object of interest. Similar to the aforementioned human vision system, it is desirable that the object is searched in a wide field of view with low resolution images while being identified and tracked in a narrow field of view with higher resolution images. The variable focal length MMAL of the present invention has a large range of focal length variation, which can offer a variable field of view (a variable magnification); a wider field of view with a large magnification and a narrow field of view with a small magnification. The field of view is changed without macroscopic movements of the lens system because each micromirror of the variable focal length MMAL is adjusted for varying the focal length and actuated by the electrostatic force and/or electromagnetic force.
Tracking systems usually require that the object of interest be in the center of an image sensor. However, this entails a camera attitude control or a robot body attitude control for large-scale tracking. In the optical tracking unit of the present invention, the optical axis of the variable focal length MMAL can be adjusted in a limited range by controlling each micromirror of the variable focal length MMAL independently without using macroscopic servo mechanisms, which allows the robot vision system to be simplified and lightly weighted when compared to that having conventional tracking systems. The focusing speed of the variable focal length MMAL is so fast that a moving object can be identified and tracked. The principles of maintaining focus on a fast moving object are described in detail in U.S. patent Ser. No. 10/896,146 filed Jul. 21, 2005.
Another aspect of the invention is to provide a three-dimensional imaging system for robot vision that can compensate the aberration by controlling each micromirror independently. The present three-dimensional imaging system can produce a substantially sharp image through entire area without blurring or vignetting.
Still another aspect of the invention is to provide a small and compact three-dimensional imaging system for robot vision which can be used in a micro-robot vision system having limitation in space. Unlike conventional stereo vision systems that require at least two camera systems, the present invention can determine three-dimensional position information of an object or scene using only a single camera system, and this renders a simpler camera calibration and a more compact imaging device. Further, since the variable focal length MMAL can be made to have a small aperture, and the magnification and optical axis can be adjusted without macroscopic movements of the lens system, the three-dimensional imaging system for robot vision of the present invention can be made small and compact.
The present invention of the three-dimensional imaging system for robot vision using the variable focal length MMAL has the following advantages: (1) the system provides all-in-focus images; (2) the system provides three-dimensional position information of an object or scene using a single camera system; (3) The imaging processes to obtain an all-in-focus image and three-dimensional position information of an object or scene is achieved faster than the persistent rate of the human eye; (4) the system has a large variation of field of view since the system has a large range of focal depth; (5) the system uses a large field of view for searching and a small field of view for identifying or tracking an object of interest; (6) the system has a variable optical axis to locate an object image in the center of an image sensor; (7) the system can identify and track a moving object; (8) the system has a high depth resolution; (9) the production cost of the system is inexpensive because the variable focal length MMAL is inexpensive; (10) the system is very simple because there is no macroscopic mechanical displacement or deformation of the lens system; (11) the system is compact and suitable for the small as well as large robot vision system; (12) the system demands low power consumption when the variable focal length MMAL is actuated by electrostatic force.
Although the present invention is briefly summarized herein, the full understanding of the invention can be obtained by the following drawings, detailed description, and appended claims.
The lens system can comprise the second variable focal length MMAL 47 for a variable magnification. The first and second variable focal length MMAL 45, 47 are optically coupled and controlled to change the magnification of the system wherein the image of an object is optically magnified and to change the focal plane to form two-dimensional images in-focus at a given magnification. The objective lens 44 and the auxiliary lens 46 provide additional magnification. The field of view is adjusted without macroscopic movements of the lens system or time delay since each micromirror 48 of the variable focal length MMALs 45 and 47 is adjusted and actuated by electrostatic and/or electromagnetic force.
The image processing unit 43 generates an all-in-focus image and three-dimensional position information of an object 49 using two-dimensional images received from the imaging unit 42. The variable focal length MMAL 45 and 47 change their focal lengths so fast that the imaging processes to obtain the all-in-focus image and three-dimensional position formation of the object are achieved faster than the persistence rate of the human eye. Further, by controlling individual micromirrors of variable focal length MMALs, the optical axis of the lens system can be adjusted, as will be explained in
The
As shown in
In order to obtain a bright and sharp image, the variable focal length MMAL must meet the two conditions for forming a lens. One is that all the rays should be converged into the focus, and the other is that the phase of the converged rays must be the same. Even though the rays have different optical path lengths, the same phase condition can be satisfied by adjusting the optical path length difference to be integer multiples of the wavelength of the light. Each facet converges rays to one point, and rays refracted or reflected by different facets have an optical path length difference of integer multiples of the incident light.
To change the focal length of the MMAL, the translational motion and/or the rotational motion of each of the micromirrors are controlled to change the direction of light and to satisfy the phase condition of the light.
The variable focal length MMAL is also an adaptive optical component compensating the aberration of the imaging system by controlling the translational motion and/or the rotational motion of each micromirror.
The response speed of the micromirror 71 can exceed the persistent rate of the human eyes times the number of depths unless the depth resolution requirement is extremely high. It is possible to make the focal length change within hundreds of micro-seconds. The range of numerical aperture change of the MMAL is large since the range of focal length variation of the MMAL is large. So, the MMAL can have a greater range of image depths, which is an essential requirement for a three-dimensional imaging system.
This application is a continuation-in-part of, and claims priority to U.S. patent application Ser. No. 10/806,299 filed Mar. 22, 2004, U.S. patent application Ser. No. 10/822,414 filed Apr. 12, 2004, U.S. patent application Ser. No. 10/983,353 filed Nov. 8, 2004, U.S. patent application Ser. No. 10/872,241 filed Jun. 18, 2004, U.S. patent application Ser. No. 10/893,039 filed Jul. 16, 2004, U.S. patent application Ser. No. 10/979,619 filed Nov. 2, 2004, and U.S. patent application Ser. No. 10/979,624 filed Nov. 2, 2004, all of which are hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
2002376 | Mannheimer | May 1935 | A |
4407567 | Michelet | Oct 1983 | A |
4834512 | Austin | May 1989 | A |
5004319 | Smither | Apr 1991 | A |
5103306 | Weiman et al. | Apr 1992 | A |
5109425 | Lawton | Apr 1992 | A |
5212555 | Stoltz | May 1993 | A |
5369433 | Baldwin | Nov 1994 | A |
5402407 | Eguchi | Mar 1995 | A |
5432712 | Chan | Jul 1995 | A |
5467121 | Allcock | Nov 1995 | A |
5612736 | Vogeley | Mar 1997 | A |
5696619 | Knipe | Dec 1997 | A |
5881034 | Mano | Mar 1999 | A |
5897195 | Choate | Apr 1999 | A |
5986811 | Wohlstadter | Nov 1999 | A |
6025951 | Swart | Feb 2000 | A |
6028689 | Michaliek | Feb 2000 | A |
6064423 | Geng | May 2000 | A |
6084843 | Abe | Jul 2000 | A |
6104425 | Kanno | Aug 2000 | A |
6111900 | Suzudo | Aug 2000 | A |
6123985 | Robinson | Sep 2000 | A |
6272002 | Lee | Aug 2001 | B1 |
6282213 | Gutin et al. | Aug 2001 | B1 |
6315423 | Yu | Nov 2001 | B1 |
6329737 | Jerman | Dec 2001 | B1 |
6498673 | Frigo | Dec 2002 | B1 |
6507366 | Lee | Jan 2003 | B1 |
6549730 | Hamada | Apr 2003 | B1 |
6625342 | Staple | Sep 2003 | B2 |
6649852 | Chason | Nov 2003 | B2 |
6650461 | Atobe | Nov 2003 | B2 |
6658208 | Watanabe | Dec 2003 | B2 |
6711319 | Hoen | Mar 2004 | B2 |
6741384 | Martin | May 2004 | B1 |
6784771 | Fan | Aug 2004 | B1 |
6833938 | Nishioka | Dec 2004 | B2 |
6885819 | Shinohara | Apr 2005 | B2 |
6900901 | Harada | May 2005 | B2 |
6900922 | Aubuchon | May 2005 | B2 |
6906848 | Aubuchon | Jun 2005 | B2 |
6906849 | Mi | Jun 2005 | B1 |
6914712 | Kurosawa | Jul 2005 | B2 |
6919982 | Nimura | Jul 2005 | B2 |
6934072 | Kim | Aug 2005 | B1 |
6934073 | Kim | Aug 2005 | B1 |
6958777 | Pine | Oct 2005 | B1 |
6970284 | Kim | Nov 2005 | B1 |
6995909 | Hayashi | Feb 2006 | B1 |
6999226 | Kim | Feb 2006 | B2 |
7023466 | Favalora | Apr 2006 | B2 |
7031046 | Kim | Apr 2006 | B2 |
7046447 | Raber | May 2006 | B2 |
7068416 | Gim | Jun 2006 | B2 |
7077523 | Seo | Jul 2006 | B2 |
7161729 | Kim | Jan 2007 | B2 |
7212330 | Seo et al. | May 2007 | B2 |
7261417 | Cho et al. | Aug 2007 | B2 |
20020018407 | Komoto | Feb 2002 | A1 |
20020102102 | Watanabe | Aug 2002 | A1 |
20020135673 | Favalora | Sep 2002 | A1 |
20030058520 | Yu | Mar 2003 | A1 |
20030071125 | Yoo | Apr 2003 | A1 |
20030174234 | Kondo | Sep 2003 | A1 |
20030184843 | Moon | Oct 2003 | A1 |
20040009683 | Hiraoka | Jan 2004 | A1 |
20040012460 | Cho | Jan 2004 | A1 |
20040021802 | Yoshino | Feb 2004 | A1 |
20040052180 | Hong | Mar 2004 | A1 |
20040246362 | Konno | Dec 2004 | A1 |
20040252958 | Abu-Ageel | Dec 2004 | A1 |
20050024736 | Bakin | Feb 2005 | A1 |
20050057812 | Raber | Mar 2005 | A1 |
20050136663 | Gan | Jun 2005 | A1 |
20050174625 | Huiber | Aug 2005 | A1 |
20050180019 | Cho | Aug 2005 | A1 |
20050212856 | Temple | Sep 2005 | A1 |
20050224695 | Mushika | Oct 2005 | A1 |
20050225884 | Gim | Oct 2005 | A1 |
20050231792 | Alain | Oct 2005 | A1 |
20050264870 | Kim | Dec 2005 | A1 |
20060012766 | Klosner | Jan 2006 | A1 |
20060012852 | Cho | Jan 2006 | A1 |
20060028709 | Cho | Feb 2006 | A1 |
20060187524 | Sandstrom | Aug 2006 | A1 |
20060209439 | Cho et al. | Sep 2006 | A1 |
Number | Date | Country | |
---|---|---|---|
20070188883 A1 | Aug 2007 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 10806299 | Mar 2004 | US |
Child | 11319987 | US | |
Parent | 10822414 | Apr 2004 | US |
Child | 10806299 | US | |
Parent | 10872241 | Jun 2004 | US |
Child | 10822414 | US | |
Parent | 10893039 | Jul 2004 | US |
Child | 10872241 | US | |
Parent | 10979619 | Nov 2004 | US |
Child | 10893039 | US | |
Parent | 10979624 | Nov 2004 | US |
Child | 10979619 | US | |
Parent | 10983353 | Nov 2004 | US |
Child | 10979624 | US |