The present invention relates to a system and method for measuring the surface appearance of a surface.
The application of paint to objects such as vehicle bodies is a sensitive process. The quality, durability, and color matching of the paint are important in producing a high quality product. In order to maintain quality, the surface appearance of the vehicle body is measured. Conventionally, the surface appearance is measured using a hand held imaging system that qualitatively measures surface appearance. These sensors are handled by an operator who must manually apply the sensor against the surface of the body of the vehicle in the position normal to the surface. Although these sensors are useful, there exists a wide horizon for improvement. Particularly, the use of these sensors are limited by an operator's inconsistencies in not accurately placing or aiming the sensor toward the vehicle body, which produces unreliable quantitative measurements for paint process control. Accordingly, the surface appearance measurements are typically inaccurate. Because a significant reliance is placed upon the operator, a significant amount of error is introduced in the quality control of the paint process.
The embodiments described herein were conceived in view of these and other disadvantages of conventional systems.
The embodiments described herein provide a system and method for measuring a surface appearance of a surface. The system includes an appearance sensor being positionable about a surface and being configured to measure the surface appearance. A surface relationship sensor is included that senses the distance between the appearance sensor and the surface. Accordingly, the surface relationship sensor provides a signal to the appearance sensor that corresponds to the distance. The appearance sensor is then positioned about the surface in response to the signal.
The method for measuring the surface appearance of a surface includes sensing a distance between an appearance sensor and the surface through the use of a surface relationship sensor. The method also includes generating a signal for the surface relationship sensor that corresponds to the distance. Additionally, the method includes positioning the appearance sensor about the surface in response to the signal. The method further includes measuring the surface appearance through the use of the appearance sensor.
The features of the present invention which are believed to be novel are set forth with particularity in the appended claims. The present invention, both as to its organization and manner of operation, together with further advantages thereof, may be best understood with reference to the following description, taken in connection with the accompanying drawings in which:
As required, detailed embodiments of the present invention are disclosed herein. However, it is to be understood that the disclosed embodiments are merely exemplary of the invention that may be embodied in various and alternative forms. The figures are not necessarily to scale, and some features may be exaggerated or minimized to show details of particular components. Therefore, specific functional details disclosed herein are not to be interpreted as limiting, but merely as a representative basis for the claims and/or as a representative basis for teaching one skilled in the art to variously employ in the present invention.
Referring to the drawings and in particular
As recognized by one of ordinary skill in the art, a surface appearance of a surface refers to the look of the surface. For example, the surface appearance may be described by terms such as glossy, dull, smooth, and the like. The embodiments described herein utilize an appearance sensor that is configured to generate signals, which are reflected by a surface and measured to quantify the surface appearance of that surface. The appearance sensor of the present invention is mounted on an operating end of a robotic arm. A hand-held embodiment of a surface appearance device is a wave-scan distinction of image (DOI) device available from BYK-Gardner, 9104 Guilford Road, Columbia, Md., 21046. Unlike the hand-held embodiment described above,
As will be described hereinafter, the sensor tool 36 also includes a surface relationship sensor for positioning the sensor tool 36 about a surface. The surface relationship sensor senses the relationship (i.e., distance) between the surface appearance sensor and the surface of a body. The surface relationship sensor may be a proximity sensor, an ultrasonic sensor, a camera and the like.
Now, referring to
The system 10 includes a conveyor station or measurement cell 28 located adjacent to the end of the oven zone 24 of the paint booth 14 for automatically measuring paint film thickness on the painted vehicle bodies 12. The system 10 includes a conveyor control system 30 (
As shown in
The sensor tool 36 includes at least one, preferably a plurality of ultrasonic pulse echo layer thickness (PELT) sensors (not shown), an appearance sensor 38, and a surface relationship sensor 40. It is recognized however, that in alternative embodiments, the PELT sensors, the appearance sensor 38, and the surface relationship sensor 40 or any combination thereof, may be installed on separate robots without departing from the scope of the present invention.
The sensor tool 36 on the robots 34 aligns the PELT sensors, the appearance sensor 38, and the surface relationship sensor 40 to specific locations on surfaces of the painted bodies 12.
The surface appearance is then measured without contacting the surface of the body 12. Additionally, unlike conventional systems, the present embodiments are not subject to errors associated with manual placement of appearance measurement devices. It is also recognized that in alternative embodiments, the appearance measurement may be taken first followed by the measuring of film thickness. In such a case, the film thickness will be measured at the same locations as the appearance measurements.
As shown in
The system 10 further includes an integrated paint quality control (IPQC) feedback control system 52 such as a programmable logic controller (PLC) 54, connected to the IPQC monitoring system 48, which receives the output information from the IPQC monitoring system 48. The IPQC feedback control system 44 outputs to and controls paint automation equipment such as the paint applicators, airflow control, etc. of the paint booth 14 based on the output from the IPQC monitoring system 48.
In general, a painted body 12 enters the cell 28 and the fixture 40 is placed on desired coordinates of the painted body 12. The computer system 42 of the AutoPelt system 32 communicates with the software of the sensor controls 44 until all designated areas are measured. The film thickness and surface appearance measurement information is then fed back to the computer system 50 of the IPQC monitoring system 48 to adjust the paint application parameters. The painted body 12 is then released back onto the moving conveyor of the paint booth 14. It should be appreciated that the number and location of the measurements will depend on the size of the painted body 12 and the paint application process.
While the best mode for carrying out the invention has been described in detail, those familiar with the art to which this invention relates will recognize various alternative designs and embodiments for practicing the invention as defined by the following claims.
Number | Name | Date | Kind |
---|---|---|---|
4702931 | Falcoff | Oct 1987 | A |
4715709 | Sekine et al. | Dec 1987 | A |
4951345 | Nappi, Sr. | Aug 1990 | A |
4977853 | Falcoff et al. | Dec 1990 | A |
4990261 | Ho | Feb 1991 | A |
5062298 | Falcoff et al. | Nov 1991 | A |
5110218 | Aizawa et al. | May 1992 | A |
5240594 | Ho | Aug 1993 | A |
5288166 | Allen et al. | Feb 1994 | A |
5328295 | Allen | Jul 1994 | A |
5518064 | Romanowski et al. | May 1996 | A |
5584336 | Romanowski et al. | Dec 1996 | A |
5639405 | Erikson | Jun 1997 | A |
5661250 | Katahira et al. | Aug 1997 | A |
5707659 | Erikson | Jan 1998 | A |
5781008 | Muller et al. | Jul 1998 | A |
5959211 | Wagner et al. | Sep 1999 | A |
6011947 | Acquaviva et al. | Jan 2000 | A |
6120833 | Bonnebat et al. | Sep 2000 | A |
6129079 | French et al. | Oct 2000 | A |
6388654 | Platzker et al. | May 2002 | B1 |
6478875 | Sampath et al. | Nov 2002 | B1 |
6484121 | Filev et al. | Nov 2002 | B1 |
6502059 | Filev et al. | Dec 2002 | B1 |
6528109 | Filev et al. | Mar 2003 | B1 |
6532066 | Filev et al. | Mar 2003 | B1 |
6543867 | Jones | Apr 2003 | B1 |
6561636 | Jones | May 2003 | B1 |
6565200 | Jones | May 2003 | B1 |
6565201 | Jones | May 2003 | B1 |
6572225 | Jones | Jun 2003 | B1 |
6573912 | Suzuki et al. | Jun 2003 | B1 |
6648435 | Jones | Nov 2003 | B1 |
6679591 | Jones | Jan 2004 | B2 |
6687015 | Waller et al. | Feb 2004 | B1 |
6701193 | Filev et al. | Mar 2004 | B1 |
6705710 | Jones et al. | Mar 2004 | B2 |
6709094 | Jones | Mar 2004 | B2 |
6719413 | Jones | Apr 2004 | B2 |
6722781 | Bazhenov et al. | Apr 2004 | B2 |
6858826 | Mueller et al. | Feb 2005 | B2 |
6862491 | Levin et al. | Mar 2005 | B2 |
6864903 | Suzuki | Mar 2005 | B2 |
6896033 | Yamamura et al. | May 2005 | B2 |
6936106 | Filev et al. | Aug 2005 | B2 |
6961133 | Caton et al. | Nov 2005 | B2 |
7054480 | Levin et al. | May 2006 | B2 |
7054674 | Cane et al. | May 2006 | B2 |
7098435 | Mueller et al. | Aug 2006 | B2 |
7104635 | Jones | Sep 2006 | B2 |
7146694 | Colson et al. | Dec 2006 | B2 |
7159641 | Yamamura et al. | Jan 2007 | B2 |
7200458 | Carman et al. | Apr 2007 | B2 |
7359564 | Keam et al. | Apr 2008 | B2 |
7406190 | Carman et al. | Jul 2008 | B2 |
7410737 | Levin et al. | Aug 2008 | B2 |
Number | Date | Country | |
---|---|---|---|
20080091360 A1 | Apr 2008 | US |