Laser scanner or laser tracker having a projector

Information

  • Patent Grant
  • 8638446
  • Patent Number
    8,638,446
  • Date Filed
    Thursday, September 8, 2011
    13 years ago
  • Date Issued
    Tuesday, January 28, 2014
    10 years ago
Abstract
A laser scanner or a laser tracker includes a light source that emits a light beam within an environment, and a data capture component that captures the light beam reflected back to the laser scanner or tracker from the environment. The laser scanner or tracker also includes a projector integrated within a body of the laser scanner or tracker or mounted to the body of the laser scanner or tracker at a predetermined location, the projector being operable to project visible information onto an object located within the environment, the projected visible information being indicative of images, data or information, the projected visible information being at least one of design intent information, information acquired by the laser scanner or tracker, or guidance to an operator.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims priority to U.S. Provisional Application No. 61/380,869, filed on Sep. 8, 2010; and to U.S. Non-Provisional application Ser. No. 13/006,507, filed on Jan. 14, 2011; which claims the benefit of priority to U.S. Provisional Application No. 61/296,555 filed on Jan. 20, 2010, to U.S. Provisional Application No. 61/351,347 filed on Jun. 4, 2010 and to U.S. Provisional Application No. 61/355,279 filed on Jun. 16, 2010. The present application also claims priority to U.S. Non-Provisional application Ser. No. 13/006,468, filed on Jan. 14, 2011; which claims the benefit of priority to U.S. Provisional Application No. 61/296,555 filed on Jan. 20, 2010, to U.S. Provisional Application No. 61/351,347 filed on Jun. 4, 2010 and to U.S. Provisional Application No. 61/355,279 filed on Jun. 16, 2010. The present application also claims priority to U.S. Non-Provisional application Ser. No. 13/006,524, filed on Jan. 14, 2011; which claims the benefit of priority to U.S. Provisional Application No. 61/296,555, filed on Jan. 20, 2012; the entire contents of each which are incorporated herein by reference.


FIELD OF THE INVENTION

The present invention relates to coordinate measurement devices, for example, laser scanners, laser trackers, and total stations, and more particularly to laser scanners and laser trackers having one or more relatively small projectors integrated therewith or added thereto by, e.g., mounting thereon, for projecting visual information in the form of images and/or data (e.g., CAD data or scanned point cloud data) onto various surfaces. The projected visual information may, for example, be of a type such as to provide guidance to an operator, such as written instructions, highlighted points to be measured, indicated areas where data are to be taken, and real time feedback on the quality of the data.


BACKGROUND

A laser scanner is one type of coordinate measurement device typically used for non-contact optical scanning of many different types of relatively large closed or open spaces or objects, for example, interior spaces of buildings, industrial installations and tunnels, or exterior shapes of planes, automobiles or boats. Laser scanners can be used for many different purposes, including industrial applications and accident reconstruction. A laser scanner optically scans and measures the environment around the laser scanner by emitting a rotating laser beam and detecting the laser beam as it is reflected back from the various objects in its path. Laser scanners typically collect a variety of data points with respect to the environment, including distance information for each object in its surrounding environment, a grey scale value (i.e., a measure of the intensity of light) for each distance measurement value, and coordinates (e.g., x, y, and z) for each distance measurement value. This scan data is collected, stored and sent to a processor that is typically remote from the laser scanner, where the data is processed to generate a three dimensional (3D) scanned image of the scanned environment with measurements. In order to generate the 3D scanned image, at least four values (x, y, z coordinates and grey scale value) are collected for each scanned data point.


Many contemporary laser scanners also include a camera mounted on the laser scanner for gathering digital images of the environment and for presenting the digital images to an operator of the laser scanner. The images can be oriented together with the scanned data to provide a more realistic image of the object being scanned. By viewing the images, the operator of the scanner can determine the field of view of the scanned data, and can adjust the settings on the laser scanner if the field of view needs adjusting. In addition, the digital images may be transmitted to the processor to add color to the 3D scanned image. In order to generate a 3D color scanned image, at least six values (x, y, z coordinates; and red value, green value, blue value or “RGB”) are collected for each data point. Examples of laser scanners are disclosed in U.S. Pat. No. 7,193,690 to Ossig et al.; U.S. Pat. No. 7,430,068 to Becker et al.; and U.S. Published Patent Application No. US2010/0134596 to Becker; each being incorporated by reference herein.


Another type of coordinate measurement device is a laser tracker, which measures the 3D coordinates of a certain point by sending a laser beam to the point, where the laser beam is typically intercepted by a retroreflector target. The laser tracker finds the coordinates of the point by measuring the distance and the two angles to the target. The distance is measured with a distance-measuring device such as an absolute distance meter (ADM) or an interferometer. The angles are measured with an angle-measuring device such as an angular encoder. A gimbaled beam-steering mechanism within the instrument directs the laser beam to the point of interest. The retroreflector may be moved manually by hand, or automatically, over the surface of the object. The laser tracker follows the movement of the retroreflector to measure the coordinates of the object. Exemplary laser trackers are disclosed in U.S. Pat. No. 4,790,651 to Brown et al., incorporated by reference herein; and U.S. Pat. No. 4,714,339 to Lau et al. The total station, which is most often used in surveying applications, may be used to measure the coordinates of diffusely scattering or retroreflective targets. The total station is closely related to both the laser tracker and the scanner.


A common type of retroreflector target is the spherically mounted retroreflector (SMR), which comprises a cube-corner retroreflector embedded within a metal sphere. The cube-corner retroreflector comprises three mutually perpendicular mirrors. The apex of the cube corner, which is the common point of intersection of the three mirrors, is located at the center of the sphere. It is common practice to place the spherical surface of the SMR in contact with an object under test and then move the SMR over the surface of the object being measured. Because of this placement of the cube corner within the sphere, the perpendicular distance from the apex of the cube corner to the surface of the object under test remains constant despite rotation of the SMR. Consequently, the 3D coordinates of the object's surface can be found by having a tracker follow the 3D coordinates of an SMR moved over the surface. It is possible to place a glass window on the top of the SMR to prevent dust or dirt from contaminating the glass surfaces. An example of such a glass surface is shown in U.S. Pat. No. 7,388,654 to Raab et al., incorporated by reference herein.


A gimbal mechanism within the laser tracker may be used to direct a laser beam from the tracker to the SMR. Part of the light retroreflected by the SMR enters the laser tracker and passes onto a position detector. The position of the light that hits the position detector is used by a tracker control system to adjust the rotation angles of the mechanical azimuth and zenith axes of the laser tracker to keep the laser beam centered on the SMR. In this way, the tracker is able to follow (track) the SMR as it is moved.


Angular encoders attached to the mechanical azimuth and zenith axes of the tracker may measure the azimuth and zenith angles of the laser beam (with respect to the tracker frame of reference). The one distance measurement and two angle measurements performed by the laser tracker are sufficient to completely specify the three-dimensional location of the SMR.


As mentioned, two types of distance meters may be found in laser trackers: interferometers and absolute distance meters (ADMs). In the laser tracker, an interferometer (if present) may determine the distance from a starting point to a finishing point by counting the number of increments of known length (usually the half-wavelength of the laser light) that pass as a retroreflector target is moved between the two points. If the beam is broken during the measurement, the number of counts cannot be accurately known, causing the distance information to be lost. By comparison, the ADM in a laser tracker determines the absolute distance to a retroreflector target without regard to beam breaks, which also allows switching between targets. Because of this, the ADM is said to be capable of “point-and-shoot” measurement. Initially, absolute distance meters were only able to measure stationary targets and for this reason were always used together with an interferometer. However, some modern absolute distance meters can make rapid measurements, thereby eliminating the need for an interferometer. Such an ADM is described in U.S. Pat. No. 7,352,446 to Bridges et al., incorporated by reference herein. The distances measured by interferometers and absolute distance meters are dependent on the speed of light through air. Since the speed of light varies with air temperature, barometric pressure, and air humidity, it is common practice to measure these quantities with sensors and to correct the speed of light in air to obtain more accurate distance readings. The distances measured by total stations and scanners also depend on the speed of light in air.


In its tracking mode, the laser tracker automatically follows movements of the SMR when the SMR is in the capture range of the tracker. If the laser beam is broken, tracking will stop. The beam may be broken by any of several means: (1) an obstruction between the instrument and SMR; (2) rapid movements of the SMR that are too fast for the instrument to follow; or (3) the direction of the SMR being turned beyond the acceptance angle of the SMR. By default, following the beam break, the beam may remain fixed at the point of the beam break, at the last commanded position, or may go to a reference (“home”) position. It may be necessary for an operator to visually search for the tracking beam and place the SMR in the beam in order to lock the instrument onto the SMR and continue tracking.


Some laser trackers include one or more cameras. A camera axis may be coaxial with the measurement beam or offset from the measurement beam by a fixed distance or angle. A camera may be used to provide a wide field of view to locate retroreflectors. A modulated light source placed near the camera optical axis may illuminate retroreflectors, thereby making them easier to identify. In this case, the retroreflectors flash in phase with the illumination, whereas background objects do not. One application for such a camera is to detect multiple retroreflectors in the field of view and measure each retroreflector in an automated sequence. Exemplary systems are described in U.S. Pat. No. 6,166,809 to Pettersen et al., and U.S. Pat. No. 7,800,758 to Bridges et al., incorporated by reference herein.


Some laser trackers have the ability to measure with six degrees of freedom (DOF), which may include three coordinates, such as x, y, and z, and three rotations, such as pitch, roll, and yaw. Several systems based on laser trackers are available or have been proposed for measuring six degrees of freedom. Exemplary systems are described in U.S. Pat. No. 7,800,758 to Bridges et al., U.S. Pat. No. 5,973,788 to Pettersen et al., and U.S. Pat. No. 7,230,689 to Lau.


It is desirable to provide a laser scanner or a laser tracker with one or more projectors, with each projector projecting visual information in the form of images and/or data (e.g., CAD data or scanned point cloud data) onto various surfaces. The projected visual information may, for example, be of a type such as to provide guidance to an operator, such as written instructions, highlighted points to be measured, indicated areas where data are to be taken, and real time feedback on the quality of the data.


SUMMARY OF THE INVENTION

According to an aspect of the present invention, a laser scanner includes a light source that emits a light beam within an environment, and a data capture component that captures the light beam reflected back to the laser scanner from the environment. The laser scanner also includes a projector integrated within a body of the laser scanner or mounted to the body of the laser scanner at a predetermined location, the projector being operable to project visible information onto an object located within the environment, the projected visible information being indicative of images, data or information, the projected visible information being at least one of design intent information, information acquired by the laser scanner, or guidance to an operator.


According to another aspect of the present invention, a laser tracker includes a light source that emits a light beam towards a target located within an environment, and a data capture component that captures the light beam reflected back to the laser scanner from the target located within the environment. The laser tracker also includes a projector integrated within a body of the laser tracker or mounted to the body of the laser tracker at a predetermined location, the projector being operable to project visible information onto an object located within the environment, the projected visible information being indicative of images, data or information, the projected visible information being at least one of design intent information, information acquired by the laser tracker, or guidance to an operator.





BRIEF DESCRIPTION OF THE DRAWINGS

Referring now to the drawings, exemplary embodiments are shown which should not be construed to be limiting regarding the entire scope of the disclosure, and wherein the elements are numbered alike in several FIGURES:



FIG. 1 is a front cross sectional view of a head portion of a laser scanner having a projector integrated therein in accordance with embodiments of the present invention;



FIG. 2 is an optical schematic view of the head portion of a laser scanner of FIG. 1 having a projector integrated therein in accordance with embodiments of the present invention;



FIG. 3 is a front cross sectional view of a head portion of a laser scanner having a projector externally mounted to the head portion in accordance with embodiments of the present invention;



FIG. 4 are two views showing visualization of movement over time of the Tower of Pisa using a projected image of the Tower of Pisa earlier in time utilizing the laser scanner having the projector according to the embodiments of FIGS. 1-3;



FIG. 5 shows a laser scanner with a projector according to embodiments of the present invention projecting “hidden features” onto a surface such as a wall;



FIG. 6 is a perspective view of a laser tracker having a projector integrated therein in accordance with embodiments of the present invention;



FIG. 7 is a perspective view of the laser tracker of FIG. 6 having computing and power supply elements attached thereto;



FIG. 8 is a perspective view of the laser tracker of FIG. 6 projecting a pattern onto a surface of an object or workpiece according to embodiments of the present invention;



FIG. 9 is a block diagram of various components including a projector within a portion of the laser tracker of FIG. 6 according to embodiments of the present invention;



FIG. 10 is a block diagram of various components including a projector within a portion of the laser tracker of FIG. 6 according to other embodiments of the present invention; and



FIG. 11 is a perspective view of alternative embodiments of the laser tracker of FIG. 6 with an external projector projecting a pattern onto a surface of an object or workpiece.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

Referring to FIG. 1, in accordance with embodiments of the present invention, there illustrated is a rotating scanning head portion 10 of a laser scanner 12 having a commercially available, relatively small or “miniature,” “ultraminiature,” or “pico” projector 14 integrated directly within the optical components (“optics”) located within the scanner head 10. The projector 14 may contain some amount of processing capability, as is known. The projector 14 may be connected with, or in communication with, a first computer or processor 15 of the laser scanner 12, where the computer or processor may be integral with the scanner 12 (e.g., located within the scanner head 10) or may be a second separate computer or processor 17 therefrom (e.g., a laptop computer). The scanner head 10 is typically mounted to a supporting tripod (not shown), which sits on the ground or other surface during laser scanner use. As described in more detail with respect to FIG. 2, the projector 14 sends various images, data or other information through the optics within the scanner head 10 and onto a rotating scanning mirror 16 that typically rotates relatively quickly through 360 degrees about a horizontal axis through the head 10, wherein the mirror 16 projects the images, data or other information towards a surface (not shown) for viewing. The scanner head 10 itself may rotate relatively more slowly through 360 degrees about a vertical axis through the head 10.


Various embodiments of the present invention include the integration or addition of such a relatively small image or data projector into equipment generally used for 3D measurement or metrology, including but not limited to, a laser scanner, laser tracker, white light scanner or similar type technological device or instrument. In embodiments of the present invention, the projector may be integrated within the laser scanner 12 or laser tracker, and the projected images, data or other information are controlled using data or information from the measurement equipment 12 itself, data or information previously captured by the measurement equipment 12, or data or information from some other source. As described in detail hereinafter, the projected images or data provide visualization of various types of information that is useful during a measurement session, or the projected images or data can assist in visualization of data previously taken by the equipment 12. The projected visual information may, for example, be of a type such as to provide guidance to an operator, such as written instructions, highlighted points to be measured, indicated areas where data are to be taken, and real time feedback on the quality of the data. This visual information provided to the operator may, for example, be in the form of visual cues, text or some other visual form of information.


Further, the projector may comprise one or more commercially available galvanometers or polygon scanners rather than one of the types of projectors mentioned herein above, for example, a miniature, ultraminiature, or pico-projector which might be based, for example, on microelectromechanical systems (MEMS) technology, liquid crystal display (LCD) or liquid crystal on silicon (LCOS) technology. For example, typically two galvanometers or two polygon scanners are used with associated mirrors to project the desired image, data, or information, in a desired pattern in two dimensions onto the surface of interest. In the case of a laser scanner 12, the galvanometer mirrors project the images or other information onto the rotating mirror where they are reflected off of towards the object of interest. The rotation of the main mirror in the laser scanner 12 and the rotation of the galvanometer mirror also in the laser scanner 12 creating the image or other projected information would be synchronized. As such, the laser scanner creates the images in the same way that images are generated for laser light shows. In the case of a laser tracker (discussed in more detail hereinafter), the galvanometer mirrors would project the images or other information directly onto the target of interest. The size of the pattern projected by a projector disposed on the tracker may be expanded by moving the tracker head to cover a relatively large area while, at the same time, dynamically changing the pattern from the projector to produce the desired image over a relatively large region of space. This way, the head of the laser tracker acts like a galvanometer mirror. When used, the galvanometers or polygon scanners may provide a relatively more powerful, brighter and more efficient laser beam for image or data projection purposes as compared to the light from a pico projector.


In many cases, it is advantageous to provide a focusing mechanism to make the projected image as sharp as possible on the surface of the object to which the two-dimensional pattern is projected. The focusing mechanism will generally include a mechanical actuator for moving one or more lenses.


With MEMS, LCD, LCOS, and other types of pico-projectors, it is common today to provide color projection patterns. Color may be used advantageously in providing information about an object.


Laser scanners and laser trackers frequently employ optics, sensors, mirrors, and/or laser sources mounted to motors and/or gimbals such that the scanner or tracker instrument or device 12 can automatically scan a large area or object or track a movable target (e.g., a retroreflector) within the working volume of the device 12 without the need to manually aim or move the sensor modules of the device 12.


Referring to FIG. 2, in some embodiments of a laser scanner 10, the laser light emitted from a laser light source 18 can be directed through use of a mirror 20. Techniques are known that allow the reflective surface of a mirror 20 to be coated in such a way (e.g., “dichroic” coating) as to reflect light at the wavelength of the source laser 18 while passing light of other wavelengths. Such embodiments allow a miniature projector 14 to be mounted behind an angled mirror 20 that reflects the laser beam emitted from the laser 18 to the rotating scanning mirror 16 (FIG. 1). In the embodiment shown in FIG. 2, the motors, encoders and drive circuitry used to point the laser beam also simultaneously direct the beam of the projector 14 via the scanning mirror 16 (FIG. 1).


Thus, in some embodiments, the computer or processor associated with the laser scanner 12 may be required to perform some mathematical calculations to correctly locate the image or data from the projector 14 onto the rotating scanning mirror 16. These calculations should be apparent to one of ordinary skill in the art. That is, the projected image or data is compensated for to account for the rotation of the mirror 16 so that the image or data is not distorted or smeared. For example, the image projected by the projector 14 onto the mirror 16 may be changed dynamically to provide an image that is stationary on a projection surface (e.g., a wall). The mirror 16 typically is made to rotate, in part, for laser safety reasons. In alternative embodiments, the laser that provides the laser beam to the rotating scanning mirror 16 for metrology purposes may be turned off, the mirror 16 may be held in a stationary position, and then the projector 14 may provide the relatively weaker light that comprises the image or data to the mirror 16. In these embodiments, no mathematical corrections for the now stationary mirror 16 are typically needed. In some cases, the size of the image projected onto the reflective surface of the mirror 16 is adjusted according to the distance from the scanner to the projection surface. This would be the case, for example, if the projector were to emit a diverging pattern of light and if the image on the projection surface were intended to have a fixed size. In this case, the distance measuring capability of the scanner may provide the information needed to enable the projector 14 to correctly size the projected image.


The image, data or other information projected from the projector 14 onto the surface of interest may have its timing controlled such that the image, data or other information may be mechanically or electronically strobed to coincide with certain angles of the rotating scanning mirror 16. Also, the laser beam used by the scanner 12 for metrology purposes may be provided in a mutually exclusive manner (e.g., multiplexed) with respect to the image, data or other information provided from the projector 14. That is, the laser beam and the projected light pattern or data may not be “on” (i.e., projected) at the same time, as that condition may not be necessary. Alternatively, the laser beam and the projected light pattern may be on at the same time. Typically, the projecting mode of the laser scanner 12, according to embodiments of the present invention, is not tied to or dependent upon the scanning mode of the scanner 12.


Referring to FIG. 3, in other embodiments of the present invention, the projector 14 can be mounted to multi-axis, motorized gimbals 22, for example, on top of the laser scanner head 10, as an alternative to being installed in-line with the measurement optics of the laser scanner 12 as in the embodiments of FIGS. 1 and 2. This allows the projection system 14 to be added to existing laser scanner equipment 12 that may not support the full integration embodiments of FIGS. 1 and 2. The embodiment of FIG. 3 may in some situations be simpler and less expensive to implement. In such an embodiment, the gimbals 22 upon which the projector 14 is mounted may be driven and aimed in synchronization with the optics of the 3D measurement device 12, thus assuring that the projector image is pointed to the same area that is of interest in taking measurements by the laser scanner 12. This has advantages when using the projector 12 for guidance or data presentation. Alternately, the independently mounted projector 14 may be commanded to project images in an area different than the primary device optics.


The position and orientation of a 3D metrology device such as a 3D laser scanner 12 or laser tracker relative to an object, part or structure to be measured may be established by identification of reference points or part or object features using known techniques. Once the coordinate system has been established, both the 3D metrology device (e.g., the laser scanner) 12 and the projector 14 can be synchronized and controlled with a relatively high degree of precision by an external computer connected to the device 12 as part of the device 12, or an on-board computer as part of the device 12 that can process the device position and control the orientation of the device optics. This allows the projected image to be shaped, scaled and controlled to match the surface onto which it projected, and it allows the image to update as the direction of the projector 14 changes such that it is always synchronized and locked to the environment.


Various implementations or usages of a projector 14 integrated within, or mounted to, a computer controlled 3D measurement device such as a laser scanner 12 according to embodiments of the present invention include, but are not limited to, projecting data, text, instructions, images or guidance in the form of, for example, visual cues or text or other forms of information on the surface of a part to be measured. They may also include providing a projected overlay of previously scanned/measured data or CAD data showing design intent for visualization of changes to a part or parameters. For 3D scanned data this may include: (1) comparison of a car body before and after an accident or before and after repair; (2) comparison of the CAD design of a planned equipment installation compared to the actual completed installation; (3) visualization of a proposed change to a part, equipment set-up, assembly line, wall, or building; (4) visualization of a part design compared to a drawing as a method of inspection; (5) visualization of hidden features (FIG. 5) such as studs, piping, electrical wiring and duct works behind a wall, ceiling or floor by projecting CAD data of the design, or scans taken during construction, onto the visible surface; (6) visualization of elements beneath the skin of the human or animal body by projecting an image of a 3D CAT scan, 3D X-ray or other 3D diagnostic data onto the body, which can provide visual assistance in locating organs, tumors, blood vessels, bones, or other physiological features as part of a surgical procedure; (7) visualization of crime scenes, before and after; (8) projection of contour maps onto a part, thereby indicating regions of the part that need to be removed, for example, by filing, sanding, or lapping, or filled in, for example, with epoxy filler; (9) projection of marks indicating areas in which subcomponents should be attached to the object, for example, with bolts, screws, or adhesive; (10) projection of lines, curves, marks, or fiducial features for aid in aligning of components; and (11) visualization of degradation or movement over time (via sequential scans) of archaeological sites, historic buildings, bridges, railways, roads, and other facilities that are subject to wear, settling, decomposition, weathering, or general deterioration over time, as illustrated, for example, in FIG. 4, which can be extended to the examination and visualization of wear and damage to large vehicles like ships, aircraft, spacecraft (e.g. space shuttle tiles). Specifically, the sole view 400 in FIG. 4A and the right hand leaning view 400 in FIG. 4 shows the Tower of Pisa tilted to the right side as viewed in FIGS. 4A and 4B, and also having a upright vertical image 410 of the Tower (shown in dotted lines in FIG. 4B) projected (in part) onto the leaning tower view 400 in FIG. 4B by the laser scanner 12 having the projector 14, according to various embodiments of the present invention. This illustrates the amount of movement or “tilt” to the right of the Tower of Pisa over time. Note that normally one would not see the left-hand (dotted line) portion of the upright vertical image 410 of the Tower since it would not be projected onto the tilted Tower nor would be projected onto any other surface. That is, this dotted line left-hand portion of the upright vertical image 410 of the Tower would be projected in free space. Instead, one would normally only see the shaded or filled in right-hand portion of the upright vertical image 410 projected onto the Tower 400. In FIG. 4B the entire upright vertical image 410 of the Tower is shown for exemplary purposes only.


To further extend the usefulness of the controlled projected image, multiple projectors 14 may be incorporated into a single device such as a laser scanner 12. This allows for a potential increase in the range of image coverage, ultimately to 360 degrees around the scanning device, with the possible exception of relatively small areas blocked by the laser scanner itself (e.g., at the location where the scanner head 10 attaches to the tripod).


In other embodiments of the present invention, synchronized images may be generated by multiple projectors 14 mounted independently of the laser scanner 12 or laser tracker, for example, on gimbaled computer controlled mounts or in fixed positions measured and known to the laser scanner 12 or tracker. In these embodiments, each projector 14 may be controlled by the laser scanner 12 or tracker, or by a computer attached to the laser scanner 12 or tracker and used to establish the coordinate system of the area. These embodiments may provide for relatively broader simultaneous coverage by the projected images while also supporting image projection in areas that might otherwise be blocked by equipment or features, including by the laser scanner 12 itself. Images projected by this array of projectors 14 can then be managed and controlled by the central computer or laser scanner 12 or other metrology device such that the projected images or data or other information is scaled, sized and properly aligned to the objects in the environment. For example, the projector 14 projects images, data or other information that is related to the direction and/or orientation that the laser scanner 14 is currently pointed to. As an example, the movement of a person or other object may be tracked by the laser scanner 14 and then images or data may be projected by the laser scanner 12 with the projector 14 according to the position and/or orientation of that person or object.


Embodiments of the present invention may be applied to any computer controlled aiming system that can establish a baseline coordinate system on a part or in an environment such that projected images can be aligned with the surface onto which they are projected. Other embodiments of the present invention may be utilized for entertainment purposes and may comprise, for example, projecting a movie on the surrounding walls of a room. For example, if a moving object (e.g., a person) can be tracked within a stationary environment, then the projected images within the environment can be automatically adjusted as a function of, for example, the person's movement, actions, or head orientation. This involves console gaming and virtual reality technologies. However, the 3D space reconstruction is different than the 2D gaming technologies. As an example, a system may have a laser scanner or tracker detect and follow a person walking around building, all the while projecting information on the wall that a person looks at. The projector cannot cover 360 degrees of space, but it can selectively project where someone is looking at, which gives the perception of projection over 3D space.


Referring to FIG. 5, there illustrated is a laser scanner 12 having a projector 14 according to embodiments of the present invention in which the projector 14 projects “hidden features” 24 onto a surface such as a wall 26. The hidden features may include objects such as, for example, studs, piping, electrical wiring and duct works that are located behind the wall 26, ceiling, floor or other visible surface. A worker may not know what is exactly positioned behind the wall surface 26 and/or does not know the exact positioning of these items behind the wall surface 26. It would be beneficial to provide the worker with an image of the items behind the wall surface 22 and the exact location of those items. Generally, this information about the hidden features is available as, e.g., CAD design data.


The projection of hidden features according to embodiments of the present invention may come about, for example, by first scanning a building such as a home using a laser scanner 12 during various construction phases (e.g., framing, wiring, plumbing, HVAC, etc.) to obtain scanned point cloud data of various structural details of the building. After completion of certain phases of the scanning to collect images and data, the laser scanner 12 with the projector 14 may then be used to project various “real” images and/or data obtained from the scanning process onto the walls, ceiling, floors, etc. Alternatively, CAD design “intent” data of the various surfaces of the building may be projected onto the surfaces. Regardless of whether real or intended images and/or data are projected, the projection of the hidden features onto these surfaces may assist someone in carrying out tasks, such as, for example, drilling a hole in a precise location of a stud behind a wall. These embodiments of the present invention allow a user of the laser scanner 12 with the projector 14 to identify the precise location of these objects or features such that no harm is caused to other objects or that no time is wasted trying to locate these hidden objects or features.


Similar to the embodiments illustrated in FIG. 5, the hidden features may comprise those within a human body that are covered by skin. For example, the projector 14 may project data onto the patient's skin to assist a doctor or surgeon in precisely locating internal human body parts to be accessed and/or surgically operated on. In an operating room, for example, a doctor may use a laser scanner 12 having a projector 14 to determine a precise location for making an incision or finding a tumor, correlating this location with 3D Computer Axial Tomography (CAT) data. In this case, the projector 14 may project an image on the patient, providing markers or actual replication of CAT scan imagery to guide the surgeon. Surgery performed remotely by manually operated robots may use such projection systems 14 in the same way as described above.


Besides displaying hidden components, for example, in a construction area or in an engineering device, the projector may display regions as they would appear following attachment. For example, before the wall surface 26 was in place in FIG. 5 and before the pipes and other construction elements were installed behind the wall surface 26, a scanner could display the desired appearance of the area, thereby providing guidance to the builder.


Referring now to FIGS. 6-11, there illustrated are embodiments of a laser tracker 30 having a projector integrated therein or mounted thereto, according to another aspect of the present invention. In FIG. 6, the laser tracker 30 includes a gimbaled beam-steering mechanism 32 that comprises a zenith carriage 34 mounted on an azimuth base 36 and rotated about an azimuth axis 38. A payload 40 is mounted on the zenith carriage 34 and is rotated about a zenith axis 42. The zenith mechanical rotation axis 42 and the azimuth mechanical rotation axis 38 intersect orthogonally, internally to the tracker 30, at a gimbal point 44, which is typically the origin for distance measurements. A laser beam 46 virtually passes through the gimbal point 44 and is pointed orthogonal to the zenith axis 42. In other words, the laser beam 46 is in the plane normal to the zenith axis 42. The laser beam 46 is pointed in the desired direction by motors located within the tracker 30 that rotate the payload 40 about the zenith axis 42 and the azimuth axis 38. Zenith and azimuth angular encoders (not shown) or transducers 41, 43, located internal to the tracker 30, are attached to the zenith mechanical axis 42 and to the azimuth mechanical axis 38, and indicate, to a relatively high degree of accuracy, the angles of rotation. The laser beam 46 travels to an external retroreflector 48 such as a spherically mounted retroreflector (SMR). By measuring the radial distance between the gimbal point 44 and the retroreflector 48 and the rotation angles about the zenith and azimuth axes 42, 38, the position of the retroreflector 48 is found within the spherical coordinate system of the tracker 30.


The laser beam 46 may comprise one or more laser wavelengths. For the sake of clarity and simplicity, a steering mechanism of the type shown in FIG. 6 is assumed in the following discussion. However, other types of steering mechanisms are possible. For example, it may be possible to reflect a laser beam off a mirror rotated about the azimuth and zenith axes 38, 42. An example of the use of a mirror in this way is disclosed in U.S. Pat. No. 4,714,339 to Lau et al. The techniques described here are applicable, regardless of the type of steering mechanism utilized.


In the laser tracker 30, one or more cameras 50 and light sources 52 are located on the payload 40. The light sources 52 illuminate the one or more retroreflector targets 48. The light sources 52 may be LEDs electrically driven to repetitively emit pulsed light. Each camera 50 may comprise a photosensitive array and a lens placed in front of the photosensitive array. The photosensitive array may be a CMOS or CCD array. The lens may have a relatively wide field of view, for example, thirty or forty degrees. The purpose of the lens is to form an image on the photosensitive array of objects within the field of view of the lens. Each light source 52 is placed near a camera 50 so that light from the light source 52 is reflected off each retroreflector target 48 onto the camera 50. In this way, retroreflector images are readily distinguished from the background on the photosensitive array as their image spots are brighter than background objects and are pulsed. In an embodiment, there are two cameras 50 and two light sources 52 placed symmetrically about the line of the laser beam 46. By using two cameras 50 in this way, the principle of triangulation can be used to find the three-dimensional coordinates of any SMR 48 within the field of view of the camera 50. In addition, the three-dimensional coordinates of the SMR 48 can be monitored as the SMR 48 is moved from point to point. A use of two cameras for this purpose is described in U.S. Published Patent Application No. US20100128259 to Bridges.


Other arrangements of one or more cameras 50 and light sources 52 are possible. For example, a light source 52 and a camera 50 can be coaxial or nearly coaxial with the laser beams 46 emitted by the tracker 30. In this case, it may be necessary to use optical filtering or similar methods to avoid saturating the photosensitive array of the camera 50 with the laser beam 46 from the tracker 30.


Another possible arrangement is to use a single camera 50 located on the payload or base 40 of the tracker 30. A single camera 50, if located off the optical axis of the laser tracker 30, provides information about the two angles that define the direction to the retroreflector 48 but not the distance to the retroreflector 48. In many cases, this information may be sufficient. If the 3D coordinates of the retroreflector 48 are needed when using a single camera 50, one possibility is to rotate the tracker 30 in the azimuth direction by 180 degrees and then to flip the zenith axis 42 to point back at the retroreflector 48. In this way, the target 48 can be viewed from two different directions and the 3D position of the retroreflector 48 can be found using triangulation.


Another possibility is to switch between measuring and imaging of the target 48. An example of such a method is described in international application WO 03/062744 to Bridges et al. Other camera arrangements are possible and can be used with the methods described herein.


As shown in FIG. 7, an auxiliary unit 60 is usually a part of the laser tracker 30. The purpose of the auxiliary unit 60 is to supply electrical power to the laser tracker body and in some cases to also supply computing and clocking capability to the system. It is possible to eliminate the auxiliary unit 60 altogether by moving the functionality of the auxiliary unit 60 into the tracker body. In most cases, the auxiliary unit 60 is attached to a general purpose computer 62. Application software loaded onto the general purpose computer 62 may provide application capabilities such as reverse engineering. It is also possible to eliminate the general purpose computer 62 by building its computing capability directly into the laser tracker 30. In this case, a user interface, preferably providing keyboard and mouse functionality is built into the laser tracker 30. The connection between the auxiliary unit 60 and the computer 62 may be wireless or through a cable of electrical wires. The computer 62 may be connected to a network, and the auxiliary unit 60 may also be connected to a network. Plural instruments, for example, multiple measurement instruments or actuators, may be connected together, either through the computer 62 or the auxiliary unit 60.


Referring to FIG. 8, there illustrated is a laser tracker 30 having an internal projector 94 (not shown) integrated within the tracker 30 (FIGS. 9-10) and projecting a pattern 70 onto a surface 72 of an object 74, such as a workpiece. Such a pattern 70 may be used, for example, to highlight the features 76 where measurements with the tracker 30 are to be taken through use of a circle 78, while also overlaying indicators 80 where the measurement device 30 would acquire the measurement points.


Referring to FIG. 9, there illustrated are various internal components within the laser tracker 30 of FIG. 6. The components include one or more distance meters 80, which may comprise an interferometer (IFM), an absolute distance meter (ADM), or both. Emitted from the distance meter 80 is one or more laser beams 82, which might be visible or infrared or both. The outgoing laser beam 82 passes through a first beam splitter 84. If the outgoing laser beam 82 is directed to a retroreflector 48 (FIG. 6), then on the return path this retroreflected laser beam 86 bounces off this first beam splitter 84 and travels to a position detector 88. The position of the light on the position detector 88 is used by the control system of the laser tracker 30 to keep the laser outgoing beam 82 centered on the retroreflector 48, thereby enabling the tracking function. If the outgoing laser beam 82 is directed to the workpiece 74 (FIG. 8) rather than a retroreflector 48, then the position of the returning laser beam 86 on the position detector 88 is not important. After passing through the first beam splitter 84, the outgoing laser beam 82 passes through a beam expander 90, which causes the diameter of the outgoing laser beam 82 to increase when the beam is traveling in the forward direction (out toward the retroreflector 48). The outgoing laser beam 82 then passes though a second beam splitter 92. Light from a projector 94 (similar to the projector 14 in the embodiments of FIGS. 1-5) sends a pattern of laser light 96 onto the second beam splitter 92. The reflected light 96 off of the second beam splitter 92 combines with the outgoing laser beam 82 from the distance meters 80, and the combined light 98 travels to either the retroreflector 48 or to the workpiece 74. In the case where the laser beam 98 is directed toward the workpiece 74, it may be possible to turn off the any visible light contained within the beam 82. This may allow the projected beam 98 to be more clearly seen.


Referring to FIG. 10, there illustrated is an embodiment of the various components of the laser tracker 30 similar to that of FIG. 9, except that the second beam splitter 92 and the projector 94 are both placed in front of the beam expander 90. The advantage of this approach is that the second beam splitter 92 can be made smaller than for the embodiment of FIG. 9. The disadvantage is that it may be more difficult to obtain proper alignment of the projector 94.


In the embodiments illustrated in FIGS. 6-10 and described herein, the laser tracker 30 has the projector 94 integrated within the internal components of the laser tracker 30. However, in other embodiments, it is possible for the projector 94 to be mounted on or otherwise attached to the laser tracker 30. For example, FIG. 11 illustrates such an embodiment in which the laser tracker 30 has the projector 14 mounted on top of the tracker body. The projector 14 may be mounted rigidly to the tracker body or the projector 14 may be mounted using a gimbal mechanism 22 similar to that of the embodiment of the laser scanner 12 shown in FIG. 3.


In another embodiment, the projector is offset from the optical axis that carries the laser beams 82 and 86. By moving the projector from the optical axis, the optical system that carries the light beam 82 may be made more compact and the distance from the projector to the region outside the tracker made smaller, thereby enabling creation of two-dimensional patterns having larger divergence angles. In this embodiment, it is not necessary to provide an independent axis for zenith (horizontal axis) rotation.


In the various embodiments of the present invention described hereinabove with respect to the laser tracker 30 of FIGS. 6-11, the projector 94, in a similar manner to the projector 14 of the laser scanner embodiments of the present invention described hereinabove with respect to FIGS. 1-5, may project images, data or other information. Such projected information provides visualization to an operator of various types of information that is useful during a measurement session using the laser tracker 30, or the projected images or data can assist in visualization of data previously taken by the equipment 12. The projected visual information may, for example, be of a type such as to provide guidance to an operator, such as written instructions, highlighted points to be measured, indicated areas where data are to be taken, and real time feedback on the quality of the data. This visual information provided to the operator may, for example, be in the form of visual cues, text or some other visual form of information. The uses to which the projected images may be put are generally the same as for a laser scanner. Of particular importance for laser trackers are (1) projections in which marks indicate where material is to be removed from or added to a structure and (2) projections in which marks indicate where components are to be added to a structure.


While preferred embodiments have been shown and described, various modifications and substitutions may be made thereto without departing from the spirit and scope of the invention. Accordingly, it is to be understood that the present invention has been described by way of illustrations and not limitation.


The presently disclosed embodiments are therefore to be considered in all respects as illustrative and not restrictive, the scope of the invention being indicated by the appended claims, rather than the foregoing description, and all changes which come within the meaning and range of equivalency of the claims are therefore intended to be embraced therein.

Claims
  • 1. A coordinate measuring device having a three-dimensional coordinate system, comprising: a light source configured to send a light beam to a first point;an optical detector configured to convert a portion of the light beam reflected back from the first point into a first electrical signal;a first processor configured to calculate a first distance from the coordinate measuring device to the first point, the calculation based at least in part on the first electrical signal and a speed of light in air;a projector integrated within a body of the coordinate measuring device or mounted to the body of the coordinate measuring device at a predetermined location, the projector being configured to project a two-dimensional pattern of visible light onto an object, the projected visible light being indicative of images, data, or information, the projector including an array of programmable pixels configured to produce the two-dimensional pattern in response to instructions received from a second processor;a beam steering mechanism configured to rotate the light beam and the two-dimensional pattern of light about a first axis by a first angle, the beam steering mechanism being further configured to rotate the light beam about a second axis by a second angle;a first angular transducer configured to measure the first angle; anda second angular transducer configured to measure the second angle.
  • 2. The coordinate measuring device of claim 1, wherein the first point is on a target, a portion of the target being in contact with the object.
  • 3. The coordinate measuring device of claim 2, wherein the target includes a retroreflector.
  • 4. The coordinate measuring device of claim 1, wherein the first point is on the object.
  • 5. The coordinate measuring device of claim 1, further comprising: a first optical system configured to project the light beam, the first optical system having a first optical axis; anda second optical system configured to project the two-dimensional pattern of light, the second optical system having a second optical axis.
  • 6. The coordinate measuring device of claim 5, wherein the first optical axis and the second optical axis are distinct.
  • 7. The coordinate measuring device of claim 5, wherein the first optical system and the second optical system share a common optical element.
  • 8. The coordinate measuring device of claim 1, wherein the beam steering mechanism is configured to rotate at least one of the first angle and the second angle by 180 degrees or more.
  • 9. The coordinate measuring device of claim 1, wherein the instructions from the second processor are based at least in part on the first distance.
  • 10. The coordinate measuring device of claim 1, wherein the instructions from the second processor are based at least in part on an orientation of the object with respect to the coordinate measuring device.
  • 11. The coordinate measuring device of claim 1, further comprising a beam splitter, wherein one of the light beam and the two-dimensional pattern of light is reflected off the beam splitter and the other of the light beam and the two-dimensional pattern of light is transmitted through the beam splitter.
  • 12. The coordinate measuring device of claim 11, wherein the beam splitter is a dichroic beam splitter.
  • 13. The coordinate measuring device of claim 1, wherein three-dimensional coordinates of the point are obtained, the three-dimensional coordinates based at least in part on the first distance, the first angle, and the second angle.
  • 14. The coordinate measuring device of claim 1, wherein the second processor is further configured to project hidden features onto a surface of the object, the hidden features representing elements behind the surface of the object.
  • 15. The coordinate measuring device of claim 1, wherein the second processor is further configured to project patterns to guide in machining of the object or attaching of a component to the object.
  • 16. The coordinate measuring device of claim 1, wherein the projector is configured to project colored light.
  • 17. The coordinate measuring device of claim 1, wherein the second processor is further configured to project a contour map, the contour map indicating regions of relative deviation from a model.
  • 18. The coordinate measuring device of claim 1, wherein the coordinate measuring device is one of a laser tracker, a laser scanner, and a total station.
  • 19. The coordinate measuring device of claim 1, wherein all pixels of the projector emit light synchronously.
  • 20. The coordinate measuring device of claim 1, wherein the projector is one of a microelectromechanical system (MEMS), a liquid crystal display (LCD), and a liquid crystal on silicon (LCOS) device.
US Referenced Citations (384)
Number Name Date Kind
1535312 Hosking Apr 1925 A
1918813 Kinzy Jul 1933 A
2316573 Egy Apr 1940 A
2333243 Glab Nov 1943 A
2702683 Green et al. Feb 1955 A
2748926 Leahy Jun 1956 A
2983367 Paramater et al. Jun 1958 A
2924495 Haines Sep 1958 A
2966257 Littlejohn Nov 1959 A
3066790 Armbruster Dec 1962 A
3458167 Cooley, Jr. Jul 1969 A
4138045 Baker Feb 1979 A
4340008 Mendelson Jul 1982 A
4379461 Nilsson et al. Apr 1983 A
4424899 Rosenberg Jan 1984 A
4430796 Nakagawa Feb 1984 A
4457625 Greenleaf et al. Jul 1984 A
4506448 Topping et al. Mar 1985 A
4537233 Vroonland et al. Aug 1985 A
4606696 Slocum Aug 1986 A
4659280 Akeel Apr 1987 A
4663852 Guarini May 1987 A
4664588 Newell et al. May 1987 A
4676002 Slocum Jun 1987 A
4714339 Lau et al. Dec 1987 A
4751950 Bock Jun 1988 A
4790651 Brown et al. Dec 1988 A
4816822 Vache et al. Mar 1989 A
4882806 Davis Nov 1989 A
4954952 Ubhayakar et al. Sep 1990 A
4982841 Goedecke Jan 1991 A
4996909 Vache et al. Mar 1991 A
5025966 Potter Jun 1991 A
5027951 Johnson Jul 1991 A
5189797 Granger Mar 1993 A
5205111 Johnson Apr 1993 A
5211476 Coudroy May 1993 A
5213240 Dietz et al. May 1993 A
5219423 Kamaya Jun 1993 A
5239855 Schleifer et al. Aug 1993 A
5289264 Steinbichler Feb 1994 A
5319445 Fitts Jun 1994 A
5332315 Baker et al. Jul 1994 A
5372250 Johnson Dec 1994 A
5373346 Hocker Dec 1994 A
5402582 Raab Apr 1995 A
5412880 Raab May 1995 A
5430384 Hocker Jul 1995 A
5455670 Payne et al. Oct 1995 A
5455993 Link et al. Oct 1995 A
5510977 Raab Apr 1996 A
5528505 Granger et al. Jun 1996 A
5535524 Carrier et al. Jul 1996 A
5611147 Raab Mar 1997 A
5623416 Hocker, III Apr 1997 A
5682508 Hocker, III Oct 1997 A
5724264 Rosenberg et al. Mar 1998 A
5752112 Paddock et al. May 1998 A
5754449 Hoshal et al. May 1998 A
5768792 Raab Jun 1998 A
5829148 Eaton Nov 1998 A
5832416 Anderson Nov 1998 A
5887122 Terawaki et al. Mar 1999 A
5926782 Raab Jul 1999 A
5956857 Raab Sep 1999 A
5973788 Pettersen et al. Oct 1999 A
5978748 Raab Nov 1999 A
5983936 Schwieterman et al. Nov 1999 A
5996790 Yamada et al. Dec 1999 A
5997779 Potter Dec 1999 A
D423534 Raab et al. Apr 2000 S
6050615 Weinhold Apr 2000 A
6060889 Hocker May 2000 A
6067116 Yamano et al. May 2000 A
6125337 Rosenberg et al. Sep 2000 A
6131299 Raab et al. Oct 2000 A
6151789 Raab et al. Nov 2000 A
6163294 Talbot Dec 2000 A
6166504 Iida et al. Dec 2000 A
6166809 Pettersen et al. Dec 2000 A
6166811 Long et al. Dec 2000 A
6219928 Raab et al. Apr 2001 B1
D441632 Raab et al. May 2001 S
6240651 Schroeder et al. Jun 2001 B1
6253458 Raab et al. Jul 2001 B1
6282195 Miller et al. Aug 2001 B1
6298569 Raab et al. Oct 2001 B1
6339410 Milner et al. Jan 2002 B1
6366831 Raab Apr 2002 B1
6408252 De Smet Jun 2002 B1
6418774 Brogårdh et al. Jul 2002 B1
6438856 Kaczynski Aug 2002 B1
6442419 Chu et al. Aug 2002 B1
6470584 Stoodley Oct 2002 B1
6477784 Schroeder et al. Nov 2002 B2
6519860 Bieg et al. Feb 2003 B1
D472824 Raab et al. Apr 2003 S
6547397 Kaufman et al. Apr 2003 B1
6598306 Eaton Jul 2003 B2
6611346 Granger Aug 2003 B2
6611617 Crampton Aug 2003 B1
6612044 Raab et al. Sep 2003 B2
6621065 Fukumoto et al. Sep 2003 B1
6626339 Gates et al. Sep 2003 B2
6633051 Holloway et al. Oct 2003 B1
6668466 Bieg et al. Dec 2003 B1
D491210 Raab et al. Jun 2004 S
6764185 Beardsley et al. Jul 2004 B1
6789327 Roth et al. Sep 2004 B2
6820346 Raab et al. Nov 2004 B2
6822749 Christoph Nov 2004 B1
6826664 Hocker, III et al. Nov 2004 B2
6868359 Raab Mar 2005 B2
6879933 Steffey et al. Apr 2005 B2
6892465 Raab et al. May 2005 B2
6895347 Dorny et al. May 2005 B2
6901673 Cobb et al. Jun 2005 B1
6904691 Raab et al. Jun 2005 B2
6920697 Raab et al. Jul 2005 B2
6925722 Raab et al. Aug 2005 B2
6931745 Granger Aug 2005 B2
6935748 Kaufman et al. Aug 2005 B2
6948255 Russell Sep 2005 B2
6965843 Raab et al. Nov 2005 B2
7003892 Eaton et al. Feb 2006 B2
7006084 Buss et al. Feb 2006 B1
7024032 Kidd et al. Apr 2006 B2
7032321 Raab et al. Apr 2006 B2
7040136 Forss et al. May 2006 B2
7051447 Kikuchi et al. May 2006 B2
7106421 Matsuura et al. Sep 2006 B2
7117107 Dorny et al. Oct 2006 B2
7120092 del Prado Pavon et al. Oct 2006 B2
7152456 Eaton Dec 2006 B2
7174651 Raab et al. Feb 2007 B2
7184047 Crampton Feb 2007 B1
7191541 Weekers et al. Mar 2007 B1
7193690 Ossig et al. Mar 2007 B2
7196509 Teng Mar 2007 B2
7199872 Van Cranenbroeck Apr 2007 B2
7230689 Lau Jun 2007 B2
7242590 Yeap et al. Jul 2007 B1
7249421 MacManus et al. Jul 2007 B2
7269910 Raab et al. Sep 2007 B2
7285793 Husted Oct 2007 B2
7296979 Raab et al. Nov 2007 B2
7306339 Kaufman et al. Dec 2007 B2
7312862 Zumbrunn et al. Dec 2007 B2
7313264 Crampton Dec 2007 B2
7319512 Ohtomo et al. Jan 2008 B2
7337344 Barman et al. Feb 2008 B2
7348822 Baer Mar 2008 B2
7352446 Bridges et al. Apr 2008 B2
7360648 Blaschke Apr 2008 B1
7372558 Kaufman et al. May 2008 B2
7372581 Raab et al. May 2008 B2
7383638 Granger Jun 2008 B2
7388654 Raab et al. Jun 2008 B2
7389870 Slappay Jun 2008 B2
7395606 Crampton Jul 2008 B2
7430068 Becker et al. Sep 2008 B2
7447931 Rischar et al. Nov 2008 B1
7449876 Pleasant et al. Nov 2008 B2
7454265 Marsh Nov 2008 B2
7463368 Morden et al. Dec 2008 B2
7508971 Vaccaro et al. Mar 2009 B2
7525276 Eaton Apr 2009 B2
7545517 Rueb et al. Jun 2009 B2
7546689 Ferrari et al. Jun 2009 B2
7552644 Haase et al. Jun 2009 B2
7561598 Stratton et al. Jul 2009 B2
7564250 Hocker Jul 2009 B2
7578069 Eaton Aug 2009 B2
D599226 Gerent et al. Sep 2009 S
7589595 Cutler Sep 2009 B2
7591077 Pettersson Sep 2009 B2
7591078 Crampton Sep 2009 B2
7602873 Eidson Oct 2009 B2
7604207 Hasloecher et al. Oct 2009 B2
7610175 Eidson Oct 2009 B2
7614157 Granger Nov 2009 B2
7624510 Ferrari Dec 2009 B2
D607350 Cooduvalli et al. Jan 2010 S
7656751 Rischar et al. Feb 2010 B2
7693325 Pulla et al. Apr 2010 B2
7701592 Saint Clair et al. Apr 2010 B2
7712224 Hicks May 2010 B2
7721396 Fleischman May 2010 B2
7735234 Briggs et al. Jun 2010 B2
7743524 Eaton et al. Jun 2010 B2
7752003 MacManus Jul 2010 B2
7765707 Tomelleri Aug 2010 B2
7769559 Reichert Aug 2010 B2
7774949 Ferrari Aug 2010 B2
7779548 Ferrari Aug 2010 B2
7779553 Jordil et al. Aug 2010 B2
7800758 Bridges et al. Sep 2010 B1
7804602 Raab Sep 2010 B2
7805851 Pettersson Oct 2010 B2
7805854 Eaton Oct 2010 B2
7809518 Zhu et al. Oct 2010 B2
RE42055 Raab et al. Jan 2011 E
RE42082 Raab et al. Feb 2011 E
7881896 Atwell et al. Feb 2011 B2
7903261 Saint Clair et al. Mar 2011 B2
7908757 Ferrari Mar 2011 B2
8052857 Townsend Nov 2011 B2
8065861 Caputo Nov 2011 B2
8082673 Desforges et al. Dec 2011 B2
8099877 Champ Jan 2012 B2
8123350 Cannell et al. Feb 2012 B2
20010004269 Shibata et al. Jun 2001 A1
20020032541 Raab et al. Mar 2002 A1
20020087233 Raab Jul 2002 A1
20020128790 Woodmansee Sep 2002 A1
20020170192 Steffey et al. Nov 2002 A1
20030033104 Gooche Feb 2003 A1
20030053037 Blaesing-Bangert et al. Mar 2003 A1
20030125901 Steffey et al. Jul 2003 A1
20030142631 Silvester Jul 2003 A1
20030167647 Raab et al. Sep 2003 A1
20030172536 Raab et al. Sep 2003 A1
20030172537 Raab et al. Sep 2003 A1
20030208919 Raab et al. Nov 2003 A1
20030221326 Raab et al. Dec 2003 A1
20040022416 Lemelson et al. Feb 2004 A1
20040040166 Raab et al. Mar 2004 A1
20040103547 Raab et al. Jun 2004 A1
20040111908 Raab et al. Jun 2004 A1
20040139265 Hocker, III et al. Jul 2004 A1
20040162700 Rosenberg et al. Aug 2004 A1
20040259533 Nixon et al. Dec 2004 A1
20050016008 Raab et al. Jan 2005 A1
20050028393 Raab et al. Feb 2005 A1
20050085940 Griggs et al. Apr 2005 A1
20050115092 Raab et al. Jun 2005 A1
20050144799 Raab et al. Jul 2005 A1
20050151963 Pulla et al. Jul 2005 A1
20050166413 Crampton Aug 2005 A1
20050188557 Raab et al. Sep 2005 A1
20050259271 Christoph Nov 2005 A1
20050276466 Vaccaro et al. Dec 2005 A1
20050283989 Pettersson Dec 2005 A1
20060016086 Raab et al. Jan 2006 A1
20060017720 Li Jan 2006 A1
20060026851 Raab et al. Feb 2006 A1
20060028203 Kawashima et al. Feb 2006 A1
20060053647 Raab et al. Mar 2006 A1
20060056459 Stratton et al. Mar 2006 A1
20060056559 Pleasant et al. Mar 2006 A1
20060059270 Pleasant et al. Mar 2006 A1
20060096108 Raab et al. May 2006 A1
20060123649 Muller Jun 2006 A1
20060129349 Raab et al. Jun 2006 A1
20060169050 Kobayashi et al. Aug 2006 A1
20060169608 Carnevali Aug 2006 A1
20060282574 Zotov et al. Dec 2006 A1
20060287769 Yanagita et al. Dec 2006 A1
20060291970 Granger Dec 2006 A1
20070030841 Lee et al. Feb 2007 A1
20070043526 De Jonge et al. Feb 2007 A1
20070050774 Eldson et al. Mar 2007 A1
20070055806 Stratton et al. Mar 2007 A1
20070058162 Granger Mar 2007 A1
20070097382 Granger May 2007 A1
20070105238 Mandl et al. May 2007 A1
20070142970 Burbank et al. Jun 2007 A1
20070147265 Eidson Jun 2007 A1
20070147435 Hamilton et al. Jun 2007 A1
20070147562 Eidson Jun 2007 A1
20070153297 Lau Jul 2007 A1
20070163134 Eaton Jul 2007 A1
20070176648 Baer Aug 2007 A1
20070177016 Wu Aug 2007 A1
20070183459 Eidson Aug 2007 A1
20070185682 Eidson Aug 2007 A1
20070217169 Yeap et al. Sep 2007 A1
20070217170 Yeap et al. Sep 2007 A1
20070221522 Yamada et al. Sep 2007 A1
20070223477 Eidson Sep 2007 A1
20070248122 Hamilton Oct 2007 A1
20070256311 Ferrari Nov 2007 A1
20070257660 Pleasant et al. Nov 2007 A1
20070258378 Hamilton Nov 2007 A1
20070282564 Sprague et al. Dec 2007 A1
20070294045 Atwell et al. Dec 2007 A1
20080046221 Stathis Feb 2008 A1
20080052936 Briggs et al. Mar 2008 A1
20080066583 Lott Mar 2008 A1
20080068103 Cutler Mar 2008 A1
20080080562 Burch et al. Apr 2008 A1
20080098272 Fairbanks et al. Apr 2008 A1
20080148585 Raab et al. Jun 2008 A1
20080154538 Stathis Jun 2008 A1
20080179206 Feinstein et al. Jul 2008 A1
20080183065 Goldbach Jul 2008 A1
20080196260 Pettersson Aug 2008 A1
20080204699 Benz et al. Aug 2008 A1
20080216552 Ibach et al. Sep 2008 A1
20080228331 McNerney et al. Sep 2008 A1
20080232269 Tatman et al. Sep 2008 A1
20080235969 Jordil et al. Oct 2008 A1
20080235970 Crampton Oct 2008 A1
20080240321 Narus et al. Oct 2008 A1
20080245452 Law et al. Oct 2008 A1
20080246943 Kaufman et al. Oct 2008 A1
20080252671 Cannell et al. Oct 2008 A1
20080256814 Pettersson Oct 2008 A1
20080257023 Jordil et al. Oct 2008 A1
20080263411 Baney et al. Oct 2008 A1
20080271332 Jordil et al. Nov 2008 A1
20080282564 Pettersson Nov 2008 A1
20080295349 Uhl et al. Dec 2008 A1
20080298254 Eidson Dec 2008 A1
20080309460 Jefferson et al. Dec 2008 A1
20090000136 Crampton Jan 2009 A1
20090016475 Rischar et al. Jan 2009 A1
20090031575 Tomelleri Feb 2009 A1
20090046140 Lashmet et al. Feb 2009 A1
20090046895 Pettersson et al. Feb 2009 A1
20090049704 Styles et al. Feb 2009 A1
20090083985 Ferrari Apr 2009 A1
20090089004 Vook et al. Apr 2009 A1
20090089078 Bursey Apr 2009 A1
20090089233 Gach et al. Apr 2009 A1
20090089623 Neering et al. Apr 2009 A1
20090109797 Eidson Apr 2009 A1
20090113183 Barford et al. Apr 2009 A1
20090113229 Cataldo et al. Apr 2009 A1
20090122805 Epps et al. May 2009 A1
20090125196 Velazquez et al. May 2009 A1
20090139105 Granger Jun 2009 A1
20090157419 Bursey Jun 2009 A1
20090165317 Little Jul 2009 A1
20090177435 Heininen Jul 2009 A1
20090177438 Raab Jul 2009 A1
20090187373 Atwell et al. Jul 2009 A1
20090241360 Tait et al. Oct 2009 A1
20090249634 Pettersson Oct 2009 A1
20090265946 Jordil et al. Oct 2009 A1
20100040742 Dijkhuis et al. Feb 2010 A1
20100057392 York Mar 2010 A1
20100078866 Pettersson Apr 2010 A1
20100095542 Ferrari Apr 2010 A1
20100122920 Butter et al. May 2010 A1
20100123892 Miller et al. May 2010 A1
20100128259 Bridges et al. May 2010 A1
20100134596 Becker Jun 2010 A1
20100148013 Bhotika et al. Jun 2010 A1
20100208062 Pettersson Aug 2010 A1
20100277747 Rueb et al. Nov 2010 A1
20100281705 Verdi et al. Nov 2010 A1
20100286941 Merlot Nov 2010 A1
20100312524 Siercks et al. Dec 2010 A1
20100318319 Maierhofer Dec 2010 A1
20100325907 Tait Dec 2010 A1
20110007305 Bridges et al. Jan 2011 A1
20110007326 Daxauer et al. Jan 2011 A1
20110013199 Siercks et al. Jan 2011 A1
20110019155 Daniel et al. Jan 2011 A1
20110023578 Grasser Feb 2011 A1
20110043515 Stathis Feb 2011 A1
20110094908 Trieu Apr 2011 A1
20110107611 Desforges et al. May 2011 A1
20110107612 Ferrari et al. May 2011 A1
20110107613 Tait May 2011 A1
20110107614 Champ May 2011 A1
20110111849 Sprague et al. May 2011 A1
20110112786 Desforges et al. May 2011 A1
20110164114 Kobayashi et al. Jul 2011 A1
20110173827 Bailey et al. Jul 2011 A1
20110173828 York Jul 2011 A1
20110178755 York Jul 2011 A1
20110178762 York Jul 2011 A1
20110178764 York Jul 2011 A1
20110178765 Atwell et al. Jul 2011 A1
20110192043 Ferrari Aug 2011 A1
20110273568 Lagassey Nov 2011 A1
20120181194 McEwan et al. Jul 2012 A1
20120210678 Alcouloumre et al. Aug 2012 A1
20120260611 Jones Oct 2012 A1
20130025143 Bailey et al. Jan 2013 A1
20130025144 Briggs et al. Jan 2013 A1
20130062243 Chang et al. Mar 2013 A1
Foreign Referenced Citations (92)
Number Date Country
2508896 Sep 2002 CN
2665668 Dec 2004 CN
3245060 Jul 1983 DE
4410775 Oct 1995 DE
29622033 Feb 1997 DE
19543763 May 1997 DE
19820307 Nov 1999 DE
10026357 Jan 2002 DE
202005000983 Apr 2005 DE
102004015668 Sep 2005 DE
19720049 Jan 2006 DE
10114126 Aug 2006 DE
102004010083 Nov 2006 DE
102005060967 Jun 2007 DE
102006023902 Nov 2007 DE
102006035292 Jan 2008 DE
102008039838 Mar 2010 DE
102005036929 Jun 2010 DE
102008062763 Jul 2010 DE
102009001894 Sep 2010 DE
0546784 Jun 1993 EP
0730210 Sep 1996 EP
0614517 Mar 1997 EP
1160539 Dec 2001 EP
1189124 Mar 2002 EP
0767357 May 2002 EP
1361414 Nov 2003 EP
1468791 Oct 2004 EP
1669713 Jun 2006 EP
1734425 Dec 2006 EP
1429109 Apr 2007 EP
1764579 Dec 2007 EP
1878543 Jan 2008 EP
1967930 Sep 2008 EP
2023077 Feb 2009 EP
2060530 May 2009 EP
2068067 Jun 2009 EP
2108917 Oct 2009 EP
2400261 Dec 2011 EP
2935043 Feb 2010 FR
894320 Apr 1962 GB
2255648 Nov 1992 GB
2341203 Mar 2000 GB
2420241 May 2006 GB
2452033 Feb 2009 GB
5581525 Jun 1955 JP
5827264 Feb 1983 JP
06313710 Nov 1994 JP
6313710 Nov 1994 JP
7210586 Aug 1995 JP
2003194526 Jul 2003 JP
2004257927 Sep 2004 JP
2005517908 Jun 2005 JP
2006241833 Sep 2006 JP
2006301991 Nov 2006 JP
2009524057 Jun 2009 JP
9208568 May 1992 WO
9808050 Feb 1998 WO
9910706 Mar 1999 WO
0014474 Mar 2000 WO
0033149 Jun 2000 WO
0034733 Jun 2000 WO
02101323 Dec 2002 WO
2004096502 Nov 2004 WO
2005072917 Aug 2005 WO
2005075875 Aug 2005 WO
2005100908 Oct 2005 WO
2006051264 May 2006 WO
2007002319 Jan 2007 WO
2007028941 Mar 2007 WO
2007125081 Nov 2007 WO
2007144906 Dec 2007 WO
2008027588 Mar 2008 WO
2008047171 Apr 2008 WO
2008064276 May 2008 WO
2008066896 Jun 2008 WO
2008075170 Jun 2008 WO
2008157061 Dec 2008 WO
2009001165 Dec 2008 WO
2009016185 Feb 2009 WO
2009083452 Jul 2009 WO
2009127526 Oct 2009 WO
2009130169 Oct 2009 WO
2009149740 Dec 2009 WO
2010040742 Apr 2010 WO
2010092131 Aug 2010 WO
2010108089 Sep 2010 WO
2010148525 Dec 2010 WO
2011000435 Jan 2011 WO
2011000955 Jan 2011 WO
2011057130 May 2011 WO
2012038446 Mar 2012 WO
Non-Patent Literature Citations (105)
Entry
German Office Action and English Language summary for DE 112011100292.0 filed Jul. 3, 2012, based on PCT Application US2011/021252 filed Jan. 14, 2011.
German Patent Application No. 11 2011 100 291.2 dated Dec. 20, 2012.
Japanese Office Action and English Language summary for JP2012-550042 filed Jul. 20, 2012; based on International Application No. PCT/US2011/021249 filed Jan. 14, 2011.
Japanese Office Action and English Language summary for JP2012-550044 filed Jul. 20, 2012; based on International Application No. PCT/US2011/021252 filed Jan. 14, 2011.
Japanese Office Action and English Language summary for JP2012-550043 filed Jul. 20, 2012; based on International Application No. PCT/US2011/021250 filed Jan. 14, 2011.
Examination Report for German Application No. 11 2011 100 193.2 Report dated Dec. 20, 2012; based on PCT/US2011/021249.
Faro Product Catalog; Faro Arm; 68 pages; Faro Technologies Inc. 2009; printed Aug. 3, 2009.
Romer Measuring Arms; Portable CMMs for the shop floor; 20 pages; Hexagon Metrology, Inc. (2009) http//us. ROMER.com.
Dylan, Craig R., High Precision Makes the Massive Bay Bridge Project Work. Suspended in MidAir—Cover Story—Point of Beginning, Jan. 1, 2010, [online] http://www.pobonline.com/Articles/Cover—Story/BNP—GUID—9-5-2006—A—10000000000 . . . [Retreived Jan. 25, 2010].
Franklin, Paul F., What IEEE 1588 Means for Your Next T&M System Design, Keithley Instruments, Inc., [on-line] Oct. 19, 2010, http://www.eetimes.com/General/DisplayPrintViewContent?contentItemId=4209746, [Retrieved Oct. 21, 2010].
HYDROpro Navigation, Hydropgraphic Survey Software, Trimble, www.trimble.com, Copyright 1997-2003.
Trimble—Trimble SPS630, SPS730 and SPS930 Universal Total Stations, [on-line] http://www.trimble.com/sps630—730—930.shtml (1 of 4), [Retreived Jan. 26, 2010 8:50:29AM].
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021253 mailed Mar. 22, 2012.
International Preliminary Report on Patentability for International Application No. PCT/US2011/021253; Date of Completion May 9, 2012.
International Preliminary Report on Patentability for International Application No. PCT/US2011/021274; Date of Completion Apr. 12, 2012.
International Search Report of the International Searching Authority for Application No. PCT/US2013/022186; Date of Mailing: May 29, 2013.
Written Opinion of the International Searching Authority for Application No. PCT/US2013/022186; Date of Mailing: May 29, 2013.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021246 International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021249 International filing date Jan. 14 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021250 International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021252 International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021247 International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021259. International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021262. International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021263. International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021264. International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021270. International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021272. International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021273. International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021276. International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021278. International filing date Jan. 14, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/021794. International filing date Jan. 20, 2011. Date of Issuance Jul. 24, 2012.
International Preliminary Report on Patentability for International Application Serial No. PCT/US2011/020625; International Filing Date of Jan. 10, 2011. Date of Issuance Jul. 17, 2012.
International Search Report for International Application No. PCT/US2011/021247 mailed Aug. 26, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021247 mailed Aug. 26, 2011.
Information on Electro-Optical Information Systems; EOIS 3D Mini-Moire C.M.M. Sensor for Non-Contact Measuring & Surface Mapping; Direct Dimensions, Jun. 1995.
ABB Flexible Automation AB: Product Manual IRB 6400R M99, On-Line Manual, Sep. 13, 2006, XP000002657684, Retrieved from the Internet: URL:http://pergatory.mit.edu/kinematiccouplings/case—studies/ABB—Robotics/general/6400R%20Product%20Manual.pdf [retrieved on Aug. 26, 2011].
Anonymous : So wird's gemacht: Mit T-DSL and Windows XP Home Edition gemeinsam ins Internet (Teil 3) Internet Citation, Jul. 2003, XP002364586, Retrieved from Internet: URL:http://support.microsfot.com/kb/814538/DE/ [retrieved on Jan. 26, 2006].
Cho, et al., Implementation of a Precision Time Protocol over Low Rate Wireless Personal Area Networks, IEEE, 2008.
Cooklev, et al., An Implementation of IEEE 1588 Over IEEE 802.11b for Syncrhonization of Wireless Local Area Network Nodes, IEEE Transactions on Instrumentation and Measurement, vol. 56, No. 5, Oct. 2007.
Hart, A., “Kinematic Coupling Interchangeability”, Precision Engineering, vol. 28, No. 1, Jan. 1, 2004, pp. 1-15, XP55005507, ISSN: 0141-6359, DOI: 10.1016/S0141-6359(03)00071-0.
International Search Report for International Application No. PCT/US2011/021246 mailed Apr. 12, 2011.
International Search Report for International Application No. PCT/US2011/021248 mailed Sep. 19, 2011.
International Search Report for International Application No. PCT/US2011/021249 mailed Apr. 21, 2011.
International Search Report for International Application No. PCT/US2011/021250 mailed Apr. 18, 2011.
International Search Report for International Application No. PCT/US2011/021252 mailed Apr. 27, 2011.
International Search Report for International Application No. PCT/US2011/021253 mailed Sep. 26, 2011.
International Search Report for International Application No. PCT/US2011/021259 mailed May 25, 2011.
International Search Report for International Application No. PCT/US2011/021262 mailed May 11, 2011.
International Search Report for International Application No. PCT/US2011/021263 mailed May 4, 2011.
International Search Report for International Application No. PCT/US2011/021264 mailed May 31, 2011.
International Search Report for International Application No. PCT/US2011/020270; Date of Mailing May 2, 2011.
International Search Report for International Application No. PCT/US2011/021272 mailed Apr. 7, 2011.
International Search Report for International Application No. PCT/US2011/020273; Date of Mailing Apr. 20, 2011.
International Search Report for International Application No. PCT/US2011/020274; Date of Mailing May 6, 2011.
International Search Report for International Application No. PCT/US2011/020276; Date of Mailing May 17, 2011.
International Search Report for International Application No. PCT/US2011/021278 mailed May 25, 2011.
International Search Report for International Application No. PCT/US2011/020625; Date of Mailing Feb. 25, 2011.
International Search Report for International Application No. PCT/US2011/021794 mailed Sep. 23, 2011.
Jasperneite, et al., Enhancements to the Time Synchronization Standard IEEE-1588 for a System of Cascaded Bridges, IEEE, 2004.
Romer “Romer Absolute Arm Product Brochure” (2010); Hexagon Metrology; www.hexagonmetrology.com; Hexagon AB 2010.
Romer “Romer Absolute Arm Maximum Performance Portable Measurement” (Printed Oct. 2010); Hexagon Metrology, Inc. http://us:ROMER.com; Hexagon Metrology, Inc 2010.
Sauter, et al., Towards New Hybrid Networks for Industrial Automation, IEEE, 2009.
Spada, et al., IEEE 1588 Lowers Integration Costs in Continuous Flow Automated Production Lines, XP-002498255, ARC Insights, Insight # 2003-33MD&H, Aug. 20, 2003.
Willoughby, P., “Elastically Averaged Precisoin Alignment”, In: “Doctoral Thesis”, Jun. 1, 2005, Massachusetts Institute of Technology, XP55005620, abstract 1.1 Motivation, Chapter 3, Chapter 6.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021246 mailed Apr. 12, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021248 mailed Sep. 19, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021249 mailed Apr. 21, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021250 mailed Apr. 18, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021252 mailed Apr. 27, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021253 mailed Sep. 26, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021259 mailed May 25, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021262 mailed May 11, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021263 mailed May 4, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021264 mailed May 31, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021270 mailed May 2, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021272 mailed Apr. 7, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021273 mailed Apr. 20, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021274 mailed May 6, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021276 mailed May 6, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021278 mailed May 25, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/020625: Filed Jan. 10, 2011.
Written Opinion of the International Searching Authority for International Application No. PCT/US2011/021794 mailed Sep. 23, 2011.
International Search Report for International Patent Application PCT/US2011/050787; mailing date Nov. 3, 2011.
Written Opinion of the International Searching Authority for International Patent Application PCT/US2011/050787; mailing date Nov. 3, 2011.
GoMeasure3D—Your source for all things measurement, Baces 3D 100 Series Portable CMM from GoMeasure3D, [online], [retrieved on Nov. 29, 2011], http://www.gomeasure3d.com/baces100.html.
It is Alive in the Lab, Autodesk University, Fun with the Immersion MicroScribe Laser Scanner, [online], [retrieved on Nov. 29, 2011], http://labs.blogs.com/its—alive—in—the—lab/2007/11/fun-with-the-im.html.
Ghost 3D Systems, Authorized MicroScribe Solutions, FAQs—MicroScribe 3D Laser, MicroScan Tools, & related info, [online], [retrieved on Nov. 29, 2011], http://microscribe.ghost3d.com/gt—microscan-3d—faqs.htm,.
Electro-Optical Information Systems, “The Handy Handheld Digitizer” [online], [retrieved on Nov. 29, 2011], http://vidibotics.com/htm/handy.htm.
Kreon Laser Scanners, Getting the Best in Cutting Edge 3D Digitizing Technology, B3-D MCAD Consulting/Sales [online], [retrieved on Nov. 29, 2011], http://www.b3-d.com/Kreon.html.
MicroScan 3D User Guide, RSI GmbH, 3D Systems & Software, Oberursel, Germany, email: info@rsi-gmbh.de, Copyright RSI Roland Seifert Imaging GmbH 2008.
Laser Reverse Engineering with Microscribe, [online], [retrieved on Nov. 29, 2011], http://www.youtube.com/watch?v=8VRz—2aEJ4E&feature=PlayList&p=F63ABF74F30DC81B&playnext=1&playnext—from=PL&index=1.
Examination Report under Section 18(3); Report dated Oct. 31, 2012; Application No. GB1210309.9.
Examination Report under Section 18(3); Report dated Nov. 1, 2012; Application No. GB1210311.5.
Examination Report under Section 18(3); Report dated Nov. 6, 2012; Application No. GB1210306.5.
Mg Lee; “Compact 3D LIDAR based on optically coupled horizontal and vertical Scanning mechanism for the autonomous navigation of robots” (13 pages) vol. 8037; downloaded from http://proceedings.spiedigitallibrary.org/ on Jul. 2, 2013.
Examination Report for German Application No. 11 2011 100 290.4 Report dated Jul. 16, 2013; based on PCT/US2011/021247.
P Ben-Tzvi, et al “Extraction of 3D Images Using Pitch-Actuated 2D Laser Range Finder for Robotic Vision” (6 pages) BNSDOCID <XP 31840390A—1—>.
International Search Report for International Application No. PCT/US2013/040309 mailed Jul. 15, 2013.
International Search Report for International Application No. PCT/US2013/040321 mailed Jul. 15, 2013.
International Search Report for International Application No. PCT/US/2013/041826 filed May 20, 2013; mailed Jul. 29, 2013.
Yk Cho, et al. “Light-weight 3D LADAR System for Construction Robotic Operations” (pp. 237-244); 26th International Symposium on Automation and Robotics in Construction (ISARC 2009).
Written Opinion for International Application No. PCT/US2013/040309 mailed Jul. 15, 2013.
Written Opinion for International Application No. PCT/US/2013/041826 filed May 20, 2013; mailed Jul. 29, 2013.
Written Opinion for International Application No. PCT/US2013/040321 mailed Jul. 15, 2013.
Related Publications (1)
Number Date Country
20120057174 A1 Mar 2012 US
Provisional Applications (8)
Number Date Country
61380869 Sep 2010 US
61296555 Jan 2010 US
61351347 Jun 2010 US
61355279 Jun 2010 US
61296555 Jan 2010 US
61351347 Jun 2010 US
61355279 Jun 2010 US
61296555 Jan 2010 US
Continuation in Parts (3)
Number Date Country
Parent 13006507 Jan 2011 US
Child 13227504 US
Parent 13006468 Jan 2011 US
Child 13006507 US
Parent 13006524 Jan 2011 US
Child 13006468 US