1. Field of Invention
The teaching presented herein relates to a method and apparatus for determining the alignment of vehicle wheels. More specifically, the teaching relates to a method and apparatus for determining the alignment of vehicle wheels using a three-dimensional target.
2. Discussion of Related Art
It is commonly known that if the wheels of a vehicle are out of alignment with each other, it can result in excessive or uneven wear of the tires and/or adversely affect the handling and stability of the vehicle. Therefore, the wheels of a vehicle need to be periodically checked to determine whether they are in alignment. Conventionally, to determine the alignment of wheels, a two-dimensional target is mounted onto the wheel to facilitate wheel alignment. A conventional two-dimensional target 100 is shown in
The two-dimensional target 100 can be used to facilitate wheel alignment, which is disclosed in U.S. Pat. Nos. 5,535,522 and 5,809,658. A wheel alignment system (as illustrated in FIG. 9 of U.S. Pat. No. 5,809,658) may be deployed in which a camera may be set up to capture a two-dimensional image of the two-dimensional target 100, in which the target elements 120 on the two-dimensional target 100 are visible. Certain features relating to the target elements may be computed by processing the captured two-dimensional image and such features can be used to determine the alignment of the wheel to which the two-dimensional target is attached using techniques well know in the wheel alignment art.
One problem associated with use of a two-dimensional target for wheel alignment is that a two-dimensional target of a large size is needed in order to achieve accurate wheel alignment determination.
The need to achieve accurate measurement such as wheel alignment determination is addressed by the present teaching. The present teaching provides an improved system using a 3D target.
One aspect of the present teaching relates to a method for determining the alignment of a motor vehicle wheel. A three-dimensional target is attached on the vehicle wheel, where the three-dimensional target has thereon a plurality of target elements that have certain known geometric characteristics and are configured in 3D space in accordance with certain known three-dimensional relationships with each other. A plurality of target element images corresponding to the plurality of target elements are detected from a 2D image of the three-dimensional target acquired by at least one camera. The alignment of the wheel is determined based on a spatial orientation of the three-dimensional target determined based on the target element images and the three-dimensional relationships among the target elements.
According to one embodiment, a three-dimensional target is attached on a vehicle, where the three-dimensional target has thereon a plurality of target elements, that have certain known geometric characteristics and are configured in 3D space in accordance with known three-dimensional relationships with each other. A 2D image of the three-dimensional target is acquired using at least one camera. The 2D image of the three-dimensional target is used to determine wheel alignment based on the three-dimensional target.
A different aspect of the present teaching relates to a system for determining the alignment of a motor vehicle wheel. A three-dimensional target is used for attaching to a vehicle wheel, where the three-dimensional target has thereon a plurality of target elements that have certain known geometric characteristics and are configured in 3D space in accordance with known three-dimensional relationships with each other. A 2D imaging system is deployed for acquiring a 2D image of the three dimensional target. A target element feature detecting system detects, from the 2D image, a plurality of target element images corresponding to the plurality of target elements. A wheel alignment determination system determines the alignment of the vehicle wheel based on a spatial orientation of the three-dimensional target determined in accordance with the detected target element images and the three-dimensional relationships among the target elements.
According to one embodiment of a system for determining the alignment of a motor vehicle wheel, a three-dimensional target is used that is attachable to a wheel to be aligned. The three-dimensional target has thereon a plurality of target elements that have certain known geometric characteristics and are configured in 3D space in accordance with certain known three-dimensional relationships with each other. An imaging system, having at least one camera, is configured capable of acquiring a 2D image of the three dimensional target. A wheel orientation determination system is configured for utilizing the 2D image of the three-dimensional target to determine wheel orientation based on the three-dimensional target.
Another aspect of the present teaching relates to a method for determining a measurement related to an object. In one embodiment, a three-dimensional target is associated with the object. The three-dimensional target has thereon a plurality of target elements that have certain known geometric characteristics and are configured in 3D space in accordance with certain known three-dimensional relationships with each other. A plurality of target element images corresponding to the plurality of target elements are detected from a 2D image of the three-dimensional target acquired by at least one camera. A measurement relating to the object is determined based on a spatial orientation of the three-dimensional target determined based on the target element images and the three-dimensional relationships among the target elements.
The inventions claimed and/or described herein are further described in terms of exemplary embodiments. These exemplary embodiments are described in detail with reference to the drawings. These embodiments are non-limiting exemplary embodiments, in which like reference numerals represent similar structures throughout the several views of the drawings, and wherein:
a-2e show exemplary constructions of three-dimensional targets, according to an embodiment of the present teaching;
The present teaching relates to method and system that utilize a three dimensional (3D) target associated with an object to make a measurement related to the object via image processing of a two dimensional (2D) image of the 3D target. In some embodiments, the object corresponds to a vehicle wheel. A 3D target can be mounted on the vehicle wheel that enables accurate wheel alignment. In some embodiments, the object corresponds to a handheld device. A 3D target can be attached to or associated with the device to enable ride height measurement. In some embodiments, the object corresponds to a camera. A 3D target attached or associated with the camera can be used to enable self-calibrating. Details relating to 3D target enabled measurement based on 2D image processing are provided below.
a-2e show exemplary constructions of three-dimensional targets, according to an embodiment of the present teaching. In
Target elements on each plane are made visually perceptible. This may be achieved by introducing contrast between target elements and the surface of the plane on which they reside. As shown in
In
b shows a different three-dimensional target 205, according to one embodiment of the present teaching. Three-dimensional target 205 has an overall shape substantially similar to a rigid cube with a plurality of facets, including top 206, front 207, left 208, bottom 209, back 210, and right 211. In a preferred embodiment, at least two of the facets have one or more target elements thereon. As seen in
c shows another exemplary construction of a three-dimensional target 212, according to an embodiment of the present teaching. As shown in
Within such a 3D construction, a plurality of two-dimensional target elements, 216, 217-a, 217-b, 217-c, 217-d, are spatially arranged on both surface 217 and surface 215 according to some pattern. In one preferred embodiment, the two-dimensional target elements are arranged so that all target elements are visible when viewed along a certain line of sight. Although all are visible, these target elements may or may not overlap. In one preferred embodiment, the line of sight is perpendicular to both surface 215 and 217.
d shows yet another exemplary construction of a three-dimensional target 220, according to one embodiment of the present teaching. The three-dimensional target 220 corresponds to a three-dimensional structure, which has at least two layers of planes in parallel in some hollow space. As shown in
In some embodiments, each of the planes has one or more target elements arranged thereon according to some pattern. In the illustrated embodiment as shown in
e shows a similar three-dimensional construct 231 as 220 (shown in
An example of an orientation determination system on which the present teaching may be implemented is illustrated in
Target devices 318, 320, 322, 324 are mounted on each of the wheels 326, 328, 330, 332 of the motor vehicle, with each target device 318, 320, 322, 324 including an attachment apparatus 338. The attachment apparatus 338 attaches the target device 318, 320, 322, 324 to wheel 326, 328, 330, 332. An example of an attachment apparatus is described in U.S. Pat. No. 5,024,001, entitled “Wheel Alignment Rim Clamp Claw” issued to Borner et al. on Jun. 18, 1991, incorporated herein by reference.
In operation, once the orientation determination system 300 has been calibrated, as described in U.S. Pat. Nos. 5,535,522 and 5,724,743, a vehicle can be driven onto the rack 340, and, if desired, the vehicle lifted to an appropriate repair elevation. The target devices 318, 320, 322, 324, once attached to the wheel rims, are then oriented so that the target devices face the respective camera 310, 312.
The location of the target devices 318, 320, 322, 324 relative to the rim of the wheels 326, 328, 330, 332 to which the target devices are attached are typically known. Once the target devices 318, 320, 322, 324 have been imaged in one position, the wheels 326, 328, 330, 332 are rolled to another position and a new image can be taken. Using the imaged location of the target devices 318, 320, 322, 324 in the two positions, the actual position and orientation of the wheels 326, 328, 330, 332 and wheel axis can be calculated by the vision imaging system 302. Although the distance between the two positions varies, the distance is often approximately 8 inches both forward and back.
During imaging, each point on the three-dimensional target 412, e.g., point Φ 416, denoted by Φ=(t0, t1, t2), where t0, t1, and t2 are the coordinates of the point Φ in the 3D target coordinate system, i.e. components of the unit vector axes U0, U1, and U2 of the 3D target coordinate system, is mathematically projected along vector r 418 and goes through the pinhole O 424 and arrives at point P on the 2D image plane 426 in the 2D image coordinate system. Such a 2D image point is denoted as P=(cX, cY), where cX and cY are the coordinates of this projected point in the 2D image coordinate system. The relationship between the 3D point Φ=(t0, t1, t2) on the three-dimensional target (expressed in terms of the 3D target coordinate system) and the 2D image point P=(CX, CY) is expressed as follows:
r=C+(t0*U0)+(t1*U1)+(t2*U2)
cX=F*(r·x)/(r·z),
cY=F*(r·y)/(r·z),
where r is the vector from the origin of the camera coordinate system to a point on the 3D target, C=(CX, CY, CZ) (not shown) is a vector from the origin of the camera coordinate system to the origin of the target coordinate system, U0, U1, and U2 are the orthogonal unit vector axes of the target coordinate system, defined relative to the camera coordinate system, and x, y, and z are the unit vectors of the camera coordinate system.
Substituting the expression of r, one can obtain the following:
r=C+(t0*U0)+(t1*U1)+(t2*U2)
cX=F*(CX+(t0*U0x)+(t1*U1x)+(t2*U2x))/cZ,
CY=F*(CY+(t0*U0y)+(t1*U1y)+(t2*U2Y))/cZ,
cZ=CZ+(t0*U0z)+(t1*U1z)+(t2*U2z)
Assume each target element is observed in the acquired 2D image as a blob. Each such blob may be characterized by a centroid, and all the target elements can be denoted by measured centroid coordinates (mxi, myi), where i is the index of a set of such centroids. Each such point (i) corresponds to a target element feature point Φ on the target.
To determine the orientation of a vehicle wheel relative to a camera in an imaging system as just described, from which misalignment of the vehicle wheel may be determined, the imaging system as depicted in
Assume that this measured set of centroids (mxi, myi) correspond to the projected set of points (cXi, cYi) from the set of target elements on the three-dimensional target, where i is the index of the set. To determine the orientation of the target relative to the camera, from which misalignment of a vehicle wheel mounted with a three-dimensional target as described herein may be determined, the following cost function can be minimized:
ρ=Σi((cXi−mxi)2+(cYi−myi)2)
where (mxi, myi) represents the measured centroid coordinate of the ith target element of the three-dimensional target mounted on a vehicle wheel, measured in the 2D image acquired during wheel alignment, and coordinate (cXi, cYi) represents a corresponding point projected from a target element on a hypothetical three-dimensional target.
In some embodiments, the hypothetical three-dimensional target is a 3D target model. This 3D target model has a known structure with a plurality of facets, each having a plurality of target elements. The centroid of each target element on the 3D target model may be mathematically projected or transformed onto a 2D image plane to yield a set of projected or model centroids. Each of such transformed model centroid has a coordinate or (cXi, cYi). In such a scenario, the model centroids can either be pre-stored or generated on the fly based on a plurality of stored parameters that are relevant to the transformation. Such parameters include camera parameters, the coordinate system for the 3D target model, the camera coordinate system, and the relationship between the camera coordinate system and the 3D target coordinate system.
The cost function p is a function of six independent parameters describing the 3D orientation of the target relative to the camera, because a coordinate (cXi, cYi) represents a point projected on the camera plane after a 3D point going through a 3D transformation with six degrees of freedom. For example, the six degrees of freedom can be realized via six independent parameters, e.g., CX, CY, CZ corresponding to translation in X-Y-Z directions, and yaw, tilt, and roll corresponding to rotations in the three dimensional space.
In minimizing the cost function p, the 3D coordinates of the hypothetical three-dimensional target are mathematically adjusted (via the 6 independent parameters) so that the difference between the two sets of 2D points, (cXi, cYi) and (mxi, myi), are minimized. The adjustment made to the six independent parameters with respect to a calibrated 3D position that yields a minimum p representing the orientation of the target being measured.
In operation, the imaging system 505 is set up according to the imaging geometry depicted in
The wheel alignment system 500 may be deployed to perform wheel alignment detection and correction thereof. When the three-dimensional target 502 is mounted on the vehicle wheel 501, e.g., in accordance with the system configuration as illustrated in
The detected 2D image features such as centroids (mxi, myi) are sent to the optimization system 525, which minimizes the cost function p by adjusting the 3D position of the hypothetical three-dimensional target or the 3D target model 503 with respect to six independent parameters as described herein. The adjustments made to the six independent parameters are then sent to the orientation determination system 545 where the orientation of the target 501 is determined based on the adjustment needed to minimize the cost function p. Then, the wheel alignment correction system 550 may compute the alignment parameters and any needed correction to the alignment of the wheel based on the measured orientation of the wheels relative to the each other and the vehicle wheel alignment specifications stored in the database.
The target element feature identification system 515 comprises an image component detection unit 620, a circle detection unit 630, and a centroid determination unit 640. Optionally, the target element feature identification system 515 may also include an image pre-processing unit 610. A 2D target image 510, acquired by the imaging system 505 may first be pre-processed by the image preprocessing system 610. Such pre-processing may include image filtering, enhancement, or edge detection.
The image component detection unit 620 analyzes a 2D image, either 510 or from the image pre-processing unit 610, to identify meaningful components in the 2D image. Such components may include 2D regions within the 2D image, representing 2D blobs. Each of such blobs may be obtained by, e.g., performing some image segmentation operations. For instance, when the imaged target elements have distinct contrast compared to the background, segmentation may be performed via a threshold operation with respect to the intensity of pixels to obtain individual regions for the target elements or an overlapped version thereof.
In some embodiments, based on the segmented image blobs, further image analysis may be performed to identify desired features. For example, if it is known that target elements are circles, the circle detection unit 630 may be invoked to detect boundaries of each image blob and compare such boundaries to the boundary shapes of such circles projected onto an image plane of an imaging system such as the one herein described. Additional analysis may be applied when there is overlap among image blobs. In some embodiments, algorithms known in the art may be employed to detect the circle boundaries of overlapped image blobs. Such detected circles may be used to derive a certain representation for each circle target element. For instance, the radius of a target element may be computed based on such a detected circle. The projected center of a detected circle may be used as an estimate of the centroid of the circle.
In some embodiments, centroids may be derived directly from image components detected by the image component detection unit 620. For example, for each image blob, algorithms known in the art may be applied to compute a centroid coordinate based on the coordinates of all pixels within the image blob. In some embodiments, the centroid of an image blob may also be derived based on the boundary points of the image blob such as a circle identified by the circle detection unit 630.
To perform wheel alignment, the constructed three-dimensional target is mounted, at 730, on a vehicle wheel according to certain geometric constraints as described herein. A calibrated camera in the system as shown in
Below, the process of optimizing p is described according to an embodiment of the present teaching. The cost function p is a non-linear function of six parameters. There is no analytical solution to p. Therefore, its optimization usually requires an iterative process, hence, it is computationally expensive. There is a wealth of literature related to such minimization procedures. For example, the well-known least squares approach can be employed to optimize p. To improve speed in wheel alignment, in some embodiments of the present teaching, a revised optimization process is employed.
In such a revised optimization process, the six independent parameters are separately adjusted. Thus at each step of the optimization, only one of the six parameters is considered a variable, and the other five parameters are treated as constants. In this case, the cost function p is still a non-linear function (a sum of ratios of polynomials) with no analytical solution. In some embodiments, the optimization with respect to one parameter may be carried out iteratively. In this case, each of the six parameters is adjusted, in a separate process, to minimize the cost function p until the changes in p caused by the adjustment is smaller than some threshold.
In some embodiments, the cost function with one parameter may be approximately solved. When the current parameter values are close to the values that minimize the cost function, the cost function p with one parameter is approximately a parabolic function with a differentiable, smoothly varying functional curve. Assume a parabolic or quadratic function in one parameter is expressed as: ρ(q)=a*q2+b*q+c, where q is a parameter (one of the six independent parameters). The first and second derivatives of this function correspond to: ρ′(q)=2a*q+b and ρ″(q)=2a. It is known that a minimum of ρ(q) occurs at q=q* when the first derivative of ρ(q) with respect to q is zero. That is, ρ′(q)=2a*q+b=0. Solving this equation, q*=−b/(2*a). Since ρ′(q=0)=b and ρ″(q=0)=2a, therefore, q*=−(ρ′(0)/ρ″(0)). In this way, the parameter value q* of parameter q minimizes the one parameter cost function ρ. Here, q* corresponds to the adjustment made to parameter q in order to minimize ρ. Applying this technique to each parameter in turn, the parameter value for each of the other five independent parameters that minimize the cost function ρ may be obtained.
The above discussed optimization process is applied to mathematical expressions corresponding to a perspective projection process. In some embodiments, a non-perspective solution may also be carried out. As discussed above, cZ=CZ+(t0*U0z)+(t1* U1z)+(t2*U2z). If Cz>>(t0*U0z)+(t1*U1z)+(t2*U2z), then cZ is approximately independent of U0z, U1z, and U2z. This permits an analytical computation of parameters C, U0, U1, and U2 instead of applying an iterative process such as a least-square fitting. Such a solution may be adequate as a final solution, or may be used as a starting point for the perspective calculation, giving parameter values close to the minimum, as required.
While the inventions have been described with reference to the certain illustrated embodiments, the words that have been used herein are words of description, rather than words of limitation. Changes may be made, within the purview of the appended claims, without departing from the scope and spirit of the invention in its aspects. Although the inventions have been described herein with reference to particular structures, acts, and materials, the invention is not to be limited to the particulars disclosed, but rather can be embodied in a wide variety of forms, some of which may be quite different from those of the disclosed embodiments, and extends to all equivalent structures, acts, and, materials, such as are within the scope of the appended claims.
This application is a Continuation of U.S. patent application Ser. No. 13/083,964, filed on Apr. 11, 2011 now abandoned, which is a Continuation of U.S. patent application Ser. No. 11/802,245, filed on May 21, 2007, now U.S. Pat. No. 7,953,247, the entire contents of each of which are hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
2552116 | Rodeghiero | May 1951 | A |
3152034 | Tompkins et al. | Oct 1964 | A |
3285803 | Baldwin et al. | Nov 1966 | A |
3409991 | Davis et al. | Nov 1968 | A |
3426991 | Rishovd | Feb 1969 | A |
3805399 | Price | Apr 1974 | A |
3861807 | Lescrenier | Jan 1975 | A |
3871360 | Van Horn et al. | Mar 1975 | A |
3952201 | Hounsfield | Apr 1976 | A |
4031884 | Henzel | Jun 1977 | A |
4108296 | Hayashi et al. | Aug 1978 | A |
4123892 | Asami | Nov 1978 | A |
4172326 | Henter | Oct 1979 | A |
4262306 | Renner | Apr 1981 | A |
4362176 | Watanabe | Dec 1982 | A |
4377038 | Ragan | Mar 1983 | A |
4394337 | Kummermehr | Jul 1983 | A |
4399175 | Kummermehr et al. | Aug 1983 | A |
4433489 | Boyce | Feb 1984 | A |
4460004 | Furuya | Jul 1984 | A |
4463425 | Hirano et al. | Jul 1984 | A |
4573275 | Bremer | Mar 1986 | A |
4853771 | Witriol et al. | Aug 1989 | A |
4994965 | Crawford et al. | Feb 1991 | A |
5048192 | Pascoal | Sep 1991 | A |
5080100 | Trotel | Jan 1992 | A |
5271055 | Hsieh et al. | Dec 1993 | A |
5279309 | Taylor et al. | Jan 1994 | A |
5295483 | Nowacki et al. | Mar 1994 | A |
5301435 | Buckley | Apr 1994 | A |
5315630 | Sturm et al. | May 1994 | A |
5389101 | Heilburn et al. | Feb 1995 | A |
5394875 | Lewis et al. | Mar 1995 | A |
5446548 | Gerig et al. | Aug 1995 | A |
5535522 | Jackson | Jul 1996 | A |
5538494 | Matsuda | Jul 1996 | A |
5582182 | Hillsman | Dec 1996 | A |
5603318 | Heilburn et al. | Feb 1997 | A |
5622187 | Carol | Apr 1997 | A |
5662111 | Cosman | Sep 1997 | A |
5724128 | January | Mar 1998 | A |
5724743 | Jackson | Mar 1998 | A |
5727554 | Kalend et al. | Mar 1998 | A |
5764723 | Weinberger et al. | Jun 1998 | A |
5771310 | Vannah | Jun 1998 | A |
5784431 | Kalend et al. | Jul 1998 | A |
5805289 | Corby et al. | Sep 1998 | A |
5809658 | Jackson et al. | Sep 1998 | A |
5820553 | Hughes | Oct 1998 | A |
5823192 | Kalend et al. | Oct 1998 | A |
5836954 | Heilbrun et al. | Nov 1998 | A |
5877100 | Smith et al. | Mar 1999 | A |
5886781 | Muller et al. | Mar 1999 | A |
5889550 | Reynolds | Mar 1999 | A |
5954647 | Bova et al. | Sep 1999 | A |
5987761 | Ohnesorge | Nov 1999 | A |
6064750 | January et al. | May 2000 | A |
6128585 | Greer | Oct 2000 | A |
6131293 | Maioli et al. | Oct 2000 | A |
6134792 | January | Oct 2000 | A |
6138302 | Sashin et al. | Oct 2000 | A |
6144875 | Schweikard et al. | Nov 2000 | A |
6146390 | Heilbrun et al. | Nov 2000 | A |
6148528 | Jackson | Nov 2000 | A |
6148585 | Baker | Nov 2000 | A |
6165181 | Heilbrun et al. | Dec 2000 | A |
6183855 | Buckley | Feb 2001 | B1 |
6185445 | Knuettel | Feb 2001 | B1 |
6185446 | Carlsen, Jr. | Feb 2001 | B1 |
6198959 | Wang | Mar 2001 | B1 |
6270836 | Holman | Aug 2001 | B1 |
6272368 | Alexanderscu | Aug 2001 | B1 |
6282799 | Warkotsch | Sep 2001 | B1 |
6296613 | Emmenegger et al. | Oct 2001 | B1 |
6300974 | Viala et al. | Oct 2001 | B1 |
6348058 | Melkent et al. | Feb 2002 | B1 |
6370455 | Larson et al. | Apr 2002 | B1 |
6384907 | Gooch | May 2002 | B1 |
6412183 | Uno | Jul 2002 | B1 |
6434507 | Clayton et al. | Aug 2002 | B1 |
6460004 | Greer et al. | Oct 2002 | B2 |
6478864 | Field | Nov 2002 | B1 |
6483577 | Stieff | Nov 2002 | B2 |
6501981 | Schweikard et al. | Dec 2002 | B1 |
6526665 | Jackson | Mar 2003 | B2 |
6600555 | McClenahan | Jul 2003 | B2 |
6611617 | Crampton | Aug 2003 | B1 |
6697761 | Akatsuka et al. | Feb 2004 | B2 |
6724930 | Kosaka et al. | Apr 2004 | B1 |
6770584 | Barney et al. | Aug 2004 | B2 |
6894771 | Dorrance et al. | May 2005 | B1 |
6901673 | Cobb et al. | Jun 2005 | B1 |
6973202 | Mostafavi | Dec 2005 | B2 |
6990215 | Brown et al. | Jan 2006 | B1 |
7144522 | Burchill et al. | Dec 2006 | B2 |
7373726 | Jackson et al. | May 2008 | B2 |
7444752 | Stieff et al. | Nov 2008 | B2 |
7703212 | Stieff et al. | Apr 2010 | B2 |
7810244 | Stieff et al. | Oct 2010 | B2 |
7869026 | Boyer et al. | Jan 2011 | B2 |
7876455 | Kawasaki et al. | Jan 2011 | B2 |
7877883 | Schommer et al. | Feb 2011 | B2 |
7930834 | Stieff et al. | Apr 2011 | B2 |
7953247 | Kassouf et al. | May 2011 | B2 |
8033028 | Stieff et al. | Oct 2011 | B2 |
20010034375 | Schwertfeger et al. | Oct 2001 | A1 |
20020092183 | Jackson | Jul 2002 | A1 |
20020164474 | Buckley | Nov 2002 | A1 |
20020164476 | Kahl et al. | Nov 2002 | A1 |
20030041981 | Cramer, III | Mar 2003 | A1 |
20030063292 | Mostafavi | Apr 2003 | A1 |
20030090682 | Gooch et al. | May 2003 | A1 |
20030187610 | Dorrance et al. | Oct 2003 | A1 |
20030210812 | Khamene et al. | Nov 2003 | A1 |
20030215640 | Ackerman et al. | Nov 2003 | A1 |
20040077738 | Field et al. | Apr 2004 | A1 |
20040150816 | Wakashiro et al. | Aug 2004 | A1 |
20050025952 | Field et al. | Feb 2005 | A1 |
20050047871 | Lee et al. | Mar 2005 | A1 |
20060090356 | Stieff | May 2006 | A1 |
20060096109 | Corghi | May 2006 | A1 |
20060152711 | Dale, Jr. et al. | Jul 2006 | A1 |
20060274302 | Shylanski et al. | Dec 2006 | A1 |
20070267498 | Marsh et al. | Nov 2007 | A1 |
20080209744 | Stieff et al. | Sep 2008 | A1 |
20080222903 | Abke | Sep 2008 | A1 |
20110001821 | Stieff et al. | Jan 2011 | A1 |
20110170089 | Stieff et al. | Jul 2011 | A1 |
20110316979 | Stieff et al. | Dec 2011 | A1 |
Number | Date | Country |
---|---|---|
0 806 629 | Nov 1997 | EP |
1 422 496 | May 2004 | EP |
1 434 169 | Jun 2004 | EP |
61-277010 | Dec 1986 | JP |
2002-90118 | Mar 2002 | JP |
2010-216969 | Sep 2010 | JP |
WO 0146909 | Jun 2001 | WO |
WO 0223121 | Mar 2002 | WO |
WO 02097362 | Dec 2002 | WO |
WO 2007124010 | Nov 2007 | WO |
Entry |
---|
U.S. Appl. No. 60/721,206, filed Sep. 28, 2005, Stieff et al. |
U.S. Appl. No. 60/938,947, filed May 18, 2007, Stieff et al. |
“Machine Vision Projects at VTT Automation, Machine Automation”, Machine Vision News. vol. 5. 2000. http:/www.automaatioseura.fi/jaostot/mvn/mvn5/vttautomation.html. |
International Preliminary Report on Patentability and Written Opinion of the International Searching Authority, issued in corresponding International Patent Application No. PCT/US2007/011978, mailed on Mar. 20, 2008. |
Chinese Office Action, and English translation thereof, issued in Chinese Application No. 2007-80053075.4 dated Mar. 28, 2012. |
Number | Date | Country | |
---|---|---|---|
20120170811 A1 | Jul 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13083964 | Apr 2011 | US |
Child | 13418091 | US | |
Parent | 11802245 | May 2007 | US |
Child | 13083964 | US |