Weld training systems are used to provide training to welders who are unfamiliar with welding and/or with certain aspects of welding. Conventional weld training systems include suites of sensors and/or have very precise positioning requirements to ensure proper tracking of training.
Systems and methods are provided for weld training, substantially as shown in and/or described in connection with at least one of the figures, as set forth more completely in the claims.
These and other advantages, aspects and novel features of the present invention, as well as details of an illustrated embodiment thereof, will be more fully understood from the following description and drawings.
The figures are not necessarily to scale. Where appropriate, similar or identical reference numbers are used to refer to similar or identical components.
“Realistic” weld training systems that provide feedback to trainee welders have made great advancements in recent years. However, such realistic weld training systems can be very costly. Disclosed examples are capable of providing low cost or no cost weld training by using a reduced-complexity weld training system to teach fundamental concepts of welding for which a high degree of realism offered by conventional weld training systems is unnecessary.
Disclosed example weld training systems use commonly-available computing devices containing a display and one or more cameras to simulate the proper setup of welding equipment and simulate weld techniques while providing feedback by analyzing images captured using the one or more cameras. In some examples, a weld training system may be implemented using an application downloaded onto a computing device such as a tablet computer or a smartphone, a mounting device to hold the computing device in a desired orientation, and a real or model welding torch. In some examples, a real or model welding coupon may also be used as the workpiece for a training weld. In some other examples, the computing device may be used with actual welding equipment, where the computing device is positioned between the welder's eyes and the workpiece so as to obstruct the arc from the user's eyes.
Disclosed examples calculate and depict, in real-time and based on analyzing images captured through the camera, welding events such as spatter, burn back, burn-through, stubbing, and/or any other welding events based on measured and calculated performance of the weld. In some examples the welding events are also determined and based on the appropriateness of the selected weld parameters to the type of weld being performed.
Disclosed example weld training systems include a computing device having a display device on a first side and a camera on a second side. The computing device is configured to capture images with the camera and process the captured images to identify a first simulation device as a simulation weld torch and a second simulation device as a simulation workpiece. The computing device is further configured to display images of a simulated welding operation on the display device of the computing device, based on analyzing the captured images to detect indicia of weld performance. The images of the simulated welding operation reflect the indicia of weld performance.
Disclosed example non-transitory machine readable storage media store machine readable instructions may be executed by a processor of a computing device having a display device on a first side and a camera on a second side. The instructions cause the computing device to capture images using the camera and process the captured images to identify a first simulation device as a simulation weld torch and a second simulation device as a simulation workpiece. The instructions also cause the processor to display images of a simulated welding operation on the display device based on analyzing the captured images to detect indicia of weld performance. The images of the simulated welding operation reflect the indicia of weld performance.
Some examples further include a mounting device that holds the computing device to orient the camera of the computing device toward a simulation area. In some examples, the mounting device orients the display device of the computing device away from the simulation area. In some examples, the mounting device orients the display device such that a user of the first simulation device is facing the simulation area.
In some examples, the display device presents stereoscopic images. In some such examples, the mounting device includes one or more lenses to provide a stereoscopic view of the stereoscopic images. In some examples, the computing device recognizes when the computing device is connected to the mounting device. In some examples, the mounting device includes a protective housing to prevent damage to the computing device from an actual weld.
In some examples, the processing of the captured images includes calculating a distance between the first simulation device and the second simulation device. In some such examples, displaying the images of the simulated welding operation is based on the calculated distance as the indicia of weld performance. In some examples, the computing device enables selection of one or more weld variables. In some such examples, the computing device depicts welding events including at least one of spatter, burn back, burn-through, or wire stubbing, based on at least one of the indicia of weld performance or the one or more weld variables. In some examples, the computing device enables the selection of the one or more weld variables with at least one of a weld calculator view or a weld equipment view.
In some examples, the computing device processes the captured images based on input from a sensor of the computing device. In some such examples, the sensor includes at least one of an accelerometer, a magnetometer, a microphone, or an ambient light sensor.
Some examples include a plurality of cameras configured to capture images substantially simultaneously. In some examples, the computing device is a smartphone or a tablet computer. In some examples, the computing device processes the captured images without using additional sensors. In some examples, the camera generates stereoscopic images and the display device displays the stereoscopic images. In some examples, the indicia of weld performance comprise at least one of aim, travel speed, work angle, travel angle, or contact tip to work distance.
As used herein, the term “real-time” refers to performance of a process or other action relating to a system in which input data is processed substantially immediately (e.g., within milliseconds, as soon as possible, etc.) so that the result of processing is available virtually immediately as feedback. In this regard, “real-time” is used on contradistinction to post-processing.
The mounting device 108 is configured to hold the computing device 106 to orient a camera of the computing device 106 toward a simulation area 104 (e.g., toward the workpiece 112). The mounting device 108 also orients a display device of the computing device 106 away from the simulation area 104. The mounting device 108 may orient the display device such that a user of the welding torch 110 is facing the simulation area 104. The computing device 106 may be configured to recognize when the computing device 106 is connected to or mounted in the mounting device 108. For example, the mounting device 108 may trigger one or more inputs in the computing device 106 via magnets and/or capacitively charged elements that are recognized by corresponding sensors in the computing device 106.
For use with a real weld, the mounting device 108 may comprise a protective shield (e.g., glass, plastic, or air curtain) to protect the computing device from spatter, heat, etc.). Alternatively, the computing device 106 may be ruggedized (e.g., by a case which it goes in before being placed in the mount).
The position of the torch 110 and workpiece 112 in three-dimensional space is determined from images captured by a camera of the computing device 106, images from a camera of the mount (e.g., received by the computing device via USB, HDMI, or the like), and/or from output (e.g., conveyed wirelessly to the computing device 106 from one or more sensors mounted on the torch 110, workpiece 112, and/or simulation area 104. For both a simulated and real weld, this position information may be used to monitor the welder's technique (e.g., aim, speed, work angle, travel angle, contact tip to work distance, and/or other parameters). For a simulated weld operation, this position information may be used to generate a simulated arc and/or simulated bead.
As described in more detail below, the welder 102 may input parameters such as power source voltage, current, workpiece metal/thickness, and/or the like via a human machine interface (e.g., touchscreen, pressure-sensitive touchscreen, gestures captured by a forward facing camera of the computing device 106, and/or the like) of the computing device 106. These parameters may be used for monitoring the quality of the weld/assessing the technique of the welder 102. For a simulated weld operation, these parameters may be used for rendering a simulated arc and/or simulated bead. The welder 102 may select a profile for storing the welder's results of the training session to track progress over time.
For an actual weld operation, the system may deal with the extremely high contrast resulting from the presence of the weld arc in a variety of ways. For example, the computing device 106 may be operable to perform a variety of image processing techniques to provide an HDR mode such that, viewing the screen of the computing device while welding, the welder 102 can clearly see, simultaneously, the workpiece in close proximity to the arc (e.g., can see the weld puddle) and relatively far from the arc. This is in contrast to viewing the workpiece/arc directly with protective eyewear because when the arc is on, the eyewear is too dark to see well in areas that are not very brightly lit by the arc. The example computing device 106 may serve as an eye protection device in lieu of a welding helmet when placed between the arc of an actual weld and the user's eyes.
In some examples, the computing device 106 presents a three-dimensional or stereoscopic image. This may either be with the aid of 3D glasses or other lens (which may also be designed to meet requirements as protective eyewear for welding) or the display may be autostereoscopic. For example, the mounting device 108 may include one or more lenses to provide a stereoscopic view of stereoscopic images present on the computing device 106.
In some examples, the welding system 100 may switch between simulated welding mode and real welding mode via an input to a human machine interface. In this manner, the welder 102 can do a practice run and then very quickly switch to a real weld once s/he has a “feel” for the weld.
The mounting device 108 may be such that the computing device 106 is easily inserted and removed and/or repositioned within the mounting device 108. Although the mounting device 108 is shown attached to a workbench or table in the simulation area 104, it may be easily removable and re-mounted elsewhere (e.g., using clamps, magnets, that can be manipulated while wearing welding gloves and not requiring any tools). For example, the mounting device 108 may be adapted to permit mounting to different workstations (including different workstations of different sizes shapes, etc.), to welding equipment (e.g., power source, wire feeder, welding torch, welding robot, etc.), and/or to a workpiece itself.
In some examples, sensor information from the computing device 106 (e.g., images from its camera, outputs from its accelerometer, gyroscope, etc.) may be communicated to another computing device, such as a computing device integrated into the welder's helmet. For example, the welding helmet may comprise a near-to-eye display and data from the computing device 106 may be wirelessly communicated to the helmet and presented on the near-to-eye display of the helmet. Similarly, the display of the helmet may augment the display on the computing device 106. For example, the display in the helmet may display parameters captured by the accelerometer/gyroscope/etc. of the computing device 106 while the display of the computing device is fully allocated to presenting the images captured by its camera. As another example, a graphical user interface for interacting with the computing device may be presented on a display of the welder's helmet, a wristband, and/or the like. Similarly, a graphical user interface for interacting with/controlling, during a real or simulated weld operation, the welder's helmet, the welder's wristband, a welding power source, wire feeder, gas cylinder, welding robot, and/or the like may be presented on the display of the computing device 106.
The computing device 106 may generate a stereoscopic image such that, by changing the angle at which s/he looks at the display, the welder can see different angles of the torch/workpiece etc., just as if looking directly at the physical workpiece.
The computing device 106 may include a front facing camera that track the welder's head and/or eyes, and may analyze the images captured by the front facing camera to change the view of the simulated workpiece such that the 2 dimensional image moves along with movements of the welder's eyes and/or head to simulate the welder getting a different view of the weld operation in progress.
The mounting device 202 may include a headband 206 or other mounting system to hold the mounting device 202 and the computing device 204 on the head of the welder 102. In other examples, the mounting device 202 may integrate the computing device 204 into a welding helmet or other headwear. In other examples, the mounting device 202 may be attached to the welder's 102 clothing, helmet, and/or the like.
The example computing device 300 of
A bus 312 enables communications between the processor 302, the RAM 306, the ROM 308, the mass storage device 310, a network interface 314, and/or an input/output interface 316.
The example network interface 314 includes hardware, firmware, and/or software to connect the computing device 300 to a communications network 318 such as the Internet. For example, the network interface 314 may include IEEE 802.X-compliant wireless and/or wired communications hardware for transmitting and/or receiving communications.
The example I/O interface 316 of
The I/O device(s) 320 may also include a keyboard, a keypad, a mouse, a trackball, a pointing device, a microphone, an audio speaker, an optical media drive, a multi-touch touch screen, a gesture recognition interface, a magnetic media drive, and/or any other type of input and/or output device.
The example computing device 300 may access a non-transitory machine readable medium 322 via the I/O interface 316 and/or the I/O device(s) 320. Examples of the machine readable medium 322 of
Consistent with embedded systems, one or more of the processor 302, the random access memory 306, the read-only memory 308, the mass storage device 310, the bus 312, the network interface 314, and/or the I/O interface 316 may be implemented in a single package.
The weld calculator view 502 includes characteristics of the joint and/or the weld, such as values for input parameters including, but not limited to, a desired fillet size, a desired penetration depth, a penetration profile, a bead width, a bevel width, a gap width, a joint length, and/or a bevel angle. The weld calculator view 502 may additionally or alternatively include inputs for wire type, wire feed speed, shielding gas type, spin or weave pattern, and/or travel speed. In some examples, the weld calculator view 502 may permit the welder to request a recommendation for weld variables such as, but are not limited to, a weld process 96, a power source voltage setting, a power source current setting, a power source frequency, a polarity, and/or an operation mode (e.g., constant current CC, constant voltage CV, or pulse).
Example systems and methods that may be used to implement the weld calculator view 502 of
The weld equipment view 602 includes a process input 606 to select a welding process (e.g., flux cored, MIG, TIG, stick, etc.), an electrode input 608 to select a wire/rod/tungsten type and/or size, a thickness input 610 to select a material thickness, a voltage/current dial 612, a wire feed speed dial 614, and an auto-set toggle 616. The weld equipment view 602 also includes a display 618 to output information is the information could be shown on welding equipment. While example inputs and outputs are shown for the weld equipment view 602 of
The weld equipment view 602 includes a button 620 to change to the weld calculator view 502 described above with reference to
During the training weld, the computing device 106 captures images with the camera of the computing device 106 (e.g., the camera(s) 324, 404 of
As shown in
Processing of captured images may include calculating a distance between the simulation welding torch 110 and the simulation workpiece 112 and/or the displaying of the images of the simulated welding operation is based on the calculated distance as the indicia of weld performance.
The simulated welding view 702 may depict welding events such as spatter, burn back, burn-through, and/or wire stubbing, based on analyzing the images captured by the camera(s) to determine the user's welding performance and/or based on weld variables.
In some examples, one or more of the sensors of the computing device 106 may be used as part of the analysis. For example, one or more of an accelerometer, a magnetometer, a microphone, or an ambient light sensor of the computing device 106 may be used to determine information about from the training weld or the computing device 106. In other examples, the computing device 106 does not use any sensors other than the camera(s).
The example welding parameter is graphed in
The example view 802 includes a button 808 to enable the welder 102 to return to the weld calculator view 502 of
At block 902, an application (or “app) is opened on the computing device 106. For example, the welder 102 may select an app and/or the computing device 106 may recognize that the computing device 106 has been attached to the mounting device 108 and automatically open the app in response.
At block 904, the computing device 106 reads inputs relating to a weld training configuration. For example, the computing device 106 may read one or more sensors, such as an accelerometer, to determine whether the computing device 106 is oriented correctly for performing weld training. A correct orientation may be useful to ensure that a training weld is captured and displayed to the welder 102.
At block 906, the computing device 106 determines whether the computing device 106 is in a physical configuration (e.g., orientation) for weld training. A physical orientation for weld training may include being attached to the mounting device 108 and/or being oriented at a correct angle relative to gravity. If the computing device 106 is not in a physical configuration for weld training (block 906), at block 908 the computing device 106 determines whether the computing device 106 has been manually set for a weld training configuration. For example, the welder 102 may instruct the computing device 106 to enter a weld training mode even if the computing device 106 is not in a particular orientation. If the computing device 106 has not been manually set for a weld training configuration (block 908), control returns to block 904.
If the computing device 106 is in a physical configuration for weld training (block 906) or the computing device 106 has been manually set for a weld training configuration (block 908), at block 910 the computing device 106 guides the user through weld setup on the computing device 106. For example, the computing device 106 may present the weld calculator view 502 and/or the weld equipment view 602 of
At block 912, the computing device 106 performs weld training analysis and presentation. For example, while the welder 102 performs a training weld, the computing device 106 capture images with camera(s) of the computing device 106, processes the captured images to identify a first simulation device (e.g., the weld torch 110) as a simulation weld torch and a second simulation device (e.g., the workpiece 112) as a simulation workpiece, and displays images of a simulated welding operation on the display device (e.g., the view 702 of
At block 914, after the training weld is completed, the computing device 106 presents results of the training weld (e.g., in the view 802 of
At block 916, the computing device 106 determines whether the weld configuration is to be modified, such as in response to a selection of the buttons 808, 810 to return to the weld calculator view 502 and/or the weld equipment view 602. If the weld configuration is to be modified (block 916), control returns to block 910.
If the weld configuration is not to be modified (block 916), at block 918 the computing device 106 determines whether another training weld is to be performed with the same settings. If another training weld is to be performed (block 918), control returns to block 912. If no further training welds are to be performed (block 918), the example instructions 900 may end.
At block 1002, the computing device 106 captures images with one or more camera(s) of the computing device 106. For example, the computing device 106 may capture the images using the camera(s) 322, 404 of
At block 1004, the computing device 106 processes the captured images to identify a first simulation device as a simulation weld torch. In some examples, the computing device 106 may use image processing techniques to identify the welding torch 110 as a device held by the user's hand and/or as having distinct markings identifying the device as a welding torch.
At block 1006, the computing device 106 processes the captured images to identify a second simulation device as a simulation workpiece. For example, the computing device 106 may use image processing techniques to identify the workpiece 112 as having a particular shape, as an object distinct from a background or other surface on which the object is resting, and/or as having distinct markings identifying the device as a workpiece.
At block 1008, the computing device 106 determines whether sensor data is available that is relevant to the training weld. For example, relevant sensor data may include accelerometer and/or gyroscope data to determine an orientation and/or movement of the computing device 106 (e.g., if the computing device is mounted to the headwear of the welder 102). If relevant sensor data is not available (block 1008), at block 1010 the computing device 106 measures one or more indicia of welding performance based on the images captured by the camera(s). For example, the computing device 106 may calculate indicia such as aim, speed, work angle, travel angle, and/or contact tip to work distance using the images.
On the other hand, if relevant sensor data is available (block 1008), at block 1012 the computing device 106 measures one or more indicia of welding performance based on the images captured by the camera(s) and based on the sensor data.
After measuring the one or more indicia at block 1010 or block 1012, at block 1014 the computing device 106 calculates/simulates weld performance based on the measured one or more indicia. For example, the computing device 106 may use a model to calculate a weld result using the measured indicia, such as aim, travel speed, work angle, travel angle, and/or contact tip to work distance, and/or the selected weld parameters as inputs to the model.
At block 1016, the computing device 106 displays images of a simulated welding operation on the display device (e.g., the display devices 326, 402 of
At block 1018, the computing device 106 determines whether the training weld is complete. If the training weld is not complete (block 1018), control returns to block 1008 to continue monitoring the training weld. When the training weld is complete (block 1018), the example instructions 1000 end and control returns to block 914 of
As utilized herein the terms “circuits” and “circuitry” refer to physical electronic components (i.e. hardware) and any software and/or firmware (“code”) which may configure the hardware, be executed by the hardware, and or otherwise be associated with the hardware. As used herein, for example, a particular processor and memory may comprise a first “circuit” when executing a first one or more lines of code and may comprise a second “circuit” when executing a second one or more lines of code. As utilized herein, “and/or” means any one or more of the items in the list joined by “and/or”. As an example, “x and/or y” means any element of the three-element set {(x), (y), (x, y)}. In other words, “x and/or y” means “one or both of x and y”. As another example, “x, y, and/or z” means any element of the seven-element set {(x), (y), (z), (x, y), (x, z), (y, z), (x, y, z)}. In other words, “x, y and/or z” means “one or more of x, y and z”. As utilized herein, the term “exemplary” means serving as a non-limiting example, instance, or illustration. As utilized herein, the terms “e.g.,” and “for example” set off lists of one or more non-limiting examples, instances, or illustrations. As utilized herein, circuitry is “operable” to perform a function whenever the circuitry comprises the necessary hardware and code (if any is necessary) to perform the function, regardless of whether performance of the function is disabled or not enabled (e.g., by a user-configurable setting, factory trim, etc.).
The present method and/or system may be realized in hardware, software, or a combination of hardware and software. The present methods and/or systems may be realized in a centralized fashion in at least one computing system, or in a distributed fashion where different elements are spread across several interconnected computing systems. Any kind of computing system or other apparatus adapted for carrying out the methods described herein is suited. A typical combination of hardware and software may be a general-purpose computing system with a program or other code that, when being loaded and executed, controls the computing system such that it carries out the methods described herein. Another typical implementation may comprise an application specific integrated circuit or chip. Some implementations may comprise a non-transitory machine-readable (e.g., computer readable) medium (e.g., FLASH drive, optical disk, magnetic storage disk, or the like) having stored thereon one or more lines of code executable by a machine, thereby causing the machine to perform processes as described herein.
While the present method and/or system has been described with reference to certain implementations, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted without departing from the scope of the present method and/or system. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the present disclosure without departing from its scope. Therefore, it is intended that the present method and/or system not be limited to the particular implementations disclosed, but that the present method and/or system will include all implementations falling within the scope of the appended claims.
The application arises from a continuation application of U.S. patent application Ser. No. 15/400,548, filed Jan. 6, 2017, entitled “Systems and Methods to Provide Weld Training,” and claims priority to U.S. Provisional Patent Application Ser. No. 62/276,290, filed Jan. 8, 2016, entitled “Weld Training Systems and Methods.” The entirety of U.S. Provisional Patent Application Ser. No. 62/276,290 is incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
4453085 | Pryor | Jun 1984 | A |
4482960 | Pryor | Nov 1984 | A |
4602163 | Pryor | Jul 1986 | A |
4654949 | Pryor | Apr 1987 | A |
4753569 | Pryor | Jun 1988 | A |
4769700 | Pryor | Sep 1988 | A |
4788440 | Pryor | Nov 1988 | A |
5148591 | Pryor | Sep 1992 | A |
5506682 | Pryor | Apr 1996 | A |
5602967 | Pryor | Feb 1997 | A |
5608847 | Pryor | Mar 1997 | A |
5956417 | Pryor | Sep 1999 | A |
6044183 | Pryor | Mar 2000 | A |
6051805 | Vaidya | Apr 2000 | A |
6107601 | Shimagama | Aug 2000 | A |
6163946 | Pryor | Dec 2000 | A |
6167607 | Pryor | Jan 2001 | B1 |
6230327 | Briand | May 2001 | B1 |
6271500 | Hirayama | Aug 2001 | B1 |
6301763 | Pryor | Oct 2001 | B1 |
6314631 | Pryor | Nov 2001 | B1 |
6315186 | Friedl | Nov 2001 | B1 |
6317953 | Pryor | Nov 2001 | B1 |
6441342 | Hsu | Aug 2002 | B1 |
6476354 | Jank | Nov 2002 | B1 |
6479793 | Wittmann | Nov 2002 | B1 |
6750428 | Okamoto | Jun 2004 | B2 |
7358458 | Daniel | Apr 2008 | B2 |
7523069 | Friedl et al. | Apr 2009 | B1 |
8144193 | Melikian | Mar 2012 | B2 |
8274013 | Wallace | Sep 2012 | B2 |
8428926 | Choquet | Apr 2013 | B2 |
8512043 | Choquet | Aug 2013 | B2 |
8569646 | Daniel | Oct 2013 | B2 |
8648903 | Loipetsberger | Feb 2014 | B2 |
8657605 | Wallace | Feb 2014 | B2 |
8680432 | Uecker | Mar 2014 | B2 |
8680434 | Stoger et al. | Mar 2014 | B2 |
8747116 | Zboray et al. | Jun 2014 | B2 |
8777629 | Kreindl | Jul 2014 | B2 |
8834168 | Peters | Sep 2014 | B2 |
8851896 | Wallace | Oct 2014 | B2 |
8884177 | Daniel | Nov 2014 | B2 |
8911237 | Postlethwaite | Dec 2014 | B2 |
8915740 | Zboray | Dec 2014 | B2 |
8987628 | Daniel et al. | Mar 2015 | B2 |
8992226 | Leach | Mar 2015 | B1 |
9011154 | Kindig | Apr 2015 | B2 |
9012802 | Daniel | Apr 2015 | B2 |
9050678 | Daniel | Jun 2015 | B2 |
9050679 | Daniel | Jun 2015 | B2 |
9073138 | Wills | Jul 2015 | B2 |
9089921 | Daniel | Jul 2015 | B2 |
9101994 | Albrecht | Aug 2015 | B2 |
9104195 | Daniel | Aug 2015 | B2 |
9196169 | Wallace | Nov 2015 | B2 |
9218745 | Choquet | Dec 2015 | B2 |
9221117 | Conrardy | Dec 2015 | B2 |
9230449 | Conrardy | Jan 2016 | B2 |
9269279 | Penrod et al. | Feb 2016 | B2 |
9280913 | Peters | Mar 2016 | B2 |
9293056 | Zboray | Mar 2016 | B2 |
9293057 | Zboray | Mar 2016 | B2 |
9318026 | Peters | Apr 2016 | B2 |
9330575 | Peters | May 2016 | B2 |
9336686 | Peters | May 2016 | B2 |
9352411 | Batzler | May 2016 | B2 |
9368045 | Becker | Jun 2016 | B2 |
9468988 | Daniel | Oct 2016 | B2 |
9483959 | Wallace | Nov 2016 | B2 |
9583014 | Becker | Feb 2017 | B2 |
9583023 | Becker et al. | Feb 2017 | B2 |
9589481 | Becker et al. | Mar 2017 | B2 |
20060090135 | Fukuda | Apr 2006 | A1 |
20090231423 | Becker et al. | Sep 2009 | A1 |
20100048273 | Wallace | Feb 2010 | A1 |
20100062406 | Zboray | Mar 2010 | A1 |
20100079356 | Hoellwarth | Apr 2010 | A1 |
20110006047 | Penrod | Jan 2011 | A1 |
20110083241 | Cole | Apr 2011 | A1 |
20110117527 | Conrardy | May 2011 | A1 |
20120006800 | Ryan | Jan 2012 | A1 |
20120180180 | Steve | Jul 2012 | A1 |
20120189993 | Kindig | Jul 2012 | A1 |
20130163090 | Yu | Jun 2013 | A1 |
20130189658 | Peters | Jul 2013 | A1 |
20130200882 | Almalki | Aug 2013 | A1 |
20130206741 | Pfeifer et al. | Aug 2013 | A1 |
20130230832 | Peters | Sep 2013 | A1 |
20130252214 | Choquet | Sep 2013 | A1 |
20130288211 | Patterson | Oct 2013 | A1 |
20140014638 | Artelsmair | Jan 2014 | A1 |
20140017642 | Postlethwaite | Jan 2014 | A1 |
20140042135 | Daniel et al. | Feb 2014 | A1 |
20140042136 | Daniel et al. | Feb 2014 | A1 |
20140042137 | Daniel et al. | Feb 2014 | A1 |
20140065584 | Wallace | Mar 2014 | A1 |
20140205976 | Peters | Jul 2014 | A1 |
20140220522 | Peters | Aug 2014 | A1 |
20140234813 | Peters | Aug 2014 | A1 |
20140263224 | Becker | Sep 2014 | A1 |
20140263227 | Daniel et al. | Sep 2014 | A1 |
20140272835 | Becker | Sep 2014 | A1 |
20140272836 | Becker | Sep 2014 | A1 |
20140272837 | Becker | Sep 2014 | A1 |
20140272838 | Becker | Sep 2014 | A1 |
20140315167 | Kreindl | Oct 2014 | A1 |
20140322684 | Wallace | Oct 2014 | A1 |
20140346158 | Matthews | Nov 2014 | A1 |
20140349256 | Connor | Nov 2014 | A1 |
20150034618 | Langeder | Feb 2015 | A1 |
20150056584 | Boulware | Feb 2015 | A1 |
20150056585 | Boulware | Feb 2015 | A1 |
20150072323 | Postlethwaite | Mar 2015 | A1 |
20150125836 | Daniel | May 2015 | A1 |
20150154884 | Salsich | Jun 2015 | A1 |
20150170539 | Chica Barrera | Jun 2015 | A1 |
20150190875 | Becker | Jul 2015 | A1 |
20150190876 | Becker | Jul 2015 | A1 |
20150190887 | Becker | Jul 2015 | A1 |
20150190888 | Becker | Jul 2015 | A1 |
20150194072 | Becker | Jul 2015 | A1 |
20150194073 | Becker | Jul 2015 | A1 |
20150209887 | Delisio | Jul 2015 | A1 |
20150228203 | Kindig | Aug 2015 | A1 |
20150235565 | Postlethwaite | Aug 2015 | A1 |
20150248845 | Postlethwaite | Sep 2015 | A1 |
20150264992 | Happel | Sep 2015 | A1 |
20150268663 | Daniel et al. | Sep 2015 | A1 |
20150320601 | Gregg | Nov 2015 | A1 |
20150325153 | Albrecht | Nov 2015 | A1 |
20150348439 | Zboray | Dec 2015 | A1 |
20150348441 | Zboray | Dec 2015 | A1 |
20150356888 | Zboray | Dec 2015 | A1 |
20150375324 | Becker | Dec 2015 | A1 |
20150375327 | Becker | Dec 2015 | A1 |
20150379894 | Becker | Dec 2015 | A1 |
20160012750 | Wallace | Jan 2016 | A1 |
20160039034 | Becker | Feb 2016 | A1 |
20160039053 | Becker | Feb 2016 | A1 |
20160049085 | Beeson | Feb 2016 | A1 |
20160093233 | Boulware | Mar 2016 | A1 |
20160114418 | Jones | Apr 2016 | A1 |
20160125592 | Becker et al. | May 2016 | A1 |
20160125593 | Becker | May 2016 | A1 |
20160125594 | Becker | May 2016 | A1 |
20160125653 | Denis | May 2016 | A1 |
20160125761 | Becker | May 2016 | A1 |
20160125762 | Becker | May 2016 | A1 |
20160125763 | Becker | May 2016 | A1 |
20160125764 | Becker | May 2016 | A1 |
20160155358 | Zboray | Jun 2016 | A1 |
20160155359 | Zboray | Jun 2016 | A1 |
20160155360 | Zboray et al. | Jun 2016 | A1 |
20160155361 | Peters | Jun 2016 | A1 |
20160171906 | Matthews | Jun 2016 | A1 |
20160189559 | Peters | Jun 2016 | A1 |
20160203732 | Wallace | Jul 2016 | A1 |
20160203733 | Wallace | Jul 2016 | A1 |
20160203734 | Boulware | Jul 2016 | A1 |
20160203735 | Boulware | Jul 2016 | A1 |
20160236303 | Matthews | Aug 2016 | A1 |
20160267806 | Hsu | Sep 2016 | A1 |
20160288236 | Becker | Oct 2016 | A1 |
20160307460 | Peters | Oct 2016 | A1 |
20160321954 | Peters | Nov 2016 | A1 |
20160343268 | Postlethwaite | Nov 2016 | A1 |
20160358503 | Batzler | Dec 2016 | A1 |
20160361774 | Daniel et al. | Dec 2016 | A9 |
20160365004 | Matthews | Dec 2016 | A1 |
20170046974 | Becker | Feb 2017 | A1 |
20170046977 | Becker | Feb 2017 | A1 |
20170046982 | Wallace | Feb 2017 | A1 |
Number | Date | Country |
---|---|---|
1866317 | Nov 2006 | CN |
1919516 | Feb 2007 | CN |
101587659 | Nov 2009 | CN |
201364652 | Dec 2009 | CN |
202210364 | May 2012 | CN |
202741926 | Feb 2013 | CN |
203338594 | Dec 2013 | CN |
103942903 | Jul 2014 | CN |
103996322 | Aug 2014 | CN |
104471629 | Mar 2015 | CN |
104603860 | May 2015 | CN |
104708174 | Jun 2015 | CN |
105144266 | Dec 2015 | CN |
105160645 | Dec 2015 | CN |
2801966 | Nov 2014 | EP |
2005102230 | Nov 2005 | WO |
2009146359 | Dec 2009 | WO |
Entry |
---|
ARVIKA Forum Vorstellung Projeckt PAARA, BMW Group Virtual Reality Center, Nuernberg, 2003. |
Fite-Georgel, Pierre; “Is there a Reality in Industrial Augmented Reality?” 10th IEEE International Symposium on Mixed and Augmented Reality (ISMAR), 2011. |
Int'l Search Report and Written Opinion Application No. PCT/US2017/012558 dated Mar. 23, 2017 (12 pages). |
Klinker, Gudrun, Augmented Reality im prktischen Einsatz, Oct. 10, 2012 (40 pages). |
Klinker, Gudrun, Intelligent Welding Gun, 2002. |
Mrovlje, et al. “Distance measuring based on stereoscopic pictures”. 9th International PhD Workshop on Systems and Control: Young Generation Viewpoint 1.-3. Oct. 2008, Izola, Slovenia (Year: 2008). |
NAMeS Users Guide, N A Tech Neural Applications, Copyright 1997, 1998, 1999, 2000 Golden, CO (123 pages). |
Native American Technologies, “ArcSentry Weld Quality Monitoring System” web page, http://web.archive.org/web/20020608124903/http://www.natech-inc.com/arcsentry1/index.html, published Jun. 8, 2002. |
Native American Technologies, “P/NA.3 Process Modelling and Optimization” web pages, http://web.archive.org/web/20020608125619/http://www.natech-inc.com/pna3/index.html, published Jun. 8, 2002. |
Petrovai etal, “A stereovision based approach for detecting and tracking lane and forward obstacles on mobile devices”. 2015 IEEE Intelligent Vehicles Symposium (IV) Jun. 28-Jul. 1, 2015. COEX, Seoul, Korea (Year: 2015). |
Rehm Welding Technology, Invertig.Pro Digital, Sep. 16, 2013. |
Rehm Welding Technology, Product Range, Aug. 2013. |
Sandor, C., Klinker, G., A rapid prototyping software infrastructure for user interfaces in ubiquitous augmented reality, Pers Ubiquit Compu (2005) 9 169-185. |
Tig Welder How to Play, www.tradesgamer.com, Nov. 17, 2011. |
Wang et al. “Stereo vision-based depth of field rendering on a mobile device”. Journal of Electronic Imaging 23(2), 023009 (Mar.-Apr. 2014) (Year: 2014). |
Number | Date | Country | |
---|---|---|---|
20220254276 A1 | Aug 2022 | US |
Number | Date | Country | |
---|---|---|---|
62276290 | Jan 2016 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16887518 | May 2020 | US |
Child | 17677271 | US | |
Parent | 15400548 | Jan 2017 | US |
Child | 16887518 | US |