This disclosure relates to systems for inspection of structures.
Expansion joints are common features in bridges and other structures subject to thermal expansion and contraction. Expansion joints may also be referred to as structural bearings. Example types of expansion joints include pin-and-hanger assemblies, rocker bearings, and so on. Expansion joints allow structural members to expand or contract without damaging the structure.
In general, this disclosure relates to devices, systems, and methods for detection of expansion joint failures. As described herein, a camera captures images of expansion joint assemblies, such as pin-and-hanger assemblies, rocker bearing, and so on. Additionally, an instrument detects a temperature. A computing system determines, based on the temperature, an expected angle of the bearing relative to a base line. The computing system also determines an actual angle of the bearing relative to the base line. The computing system superimposes a first line on the image. The first line indicates the expected angle. Additionally, the computing system superimposes a second line on the image. The second line indicates the actual angle. In this way, a user may easily compare where the actual angle of the bearing to the angle that bearing should be at, given the temperature. The actual angle differing significantly from expected angle may be a strong indication that the bearing is locked up.
In one example, this disclosure describes a method for inspecting an expansion joint, the method comprising: receiving, by a computing system, an image captured by a camera, the image being of a structural bearing, wherein the structural bearing is a hanger bearing or a rocker bearing; receiving, by the computing system, a temperature measurement generated by an instrument; determining, by the computing system, based on the temperature, an expected angle of the structural bearing relative to a base line; determining, by the computing system, an actual angle of the structural bearing relative to the base line; superimposing, by the computing system, a first line on the image, the first line indicating the expected angle; and superimposing, by the computing system, a second line on the image, the second line indicating the actual angle.
In another example, this disclosure describes a computing system comprising: a transceiver configured to: receive an image captured by a camera, the image being of a structural bearing, wherein the structural bearing is a hanger bearing or a rocker bearing; and receive a temperature measurement generated by an instrument; and one or more processing circuits configured to: determine, based on the temperature, an expected angle of the structural bearing relative to a base line; determine an actual angle of the structural bearing relative to the base line; superimpose a first line on the image, the first line indicating the expected angle; and superimpose a second line on the image, the second line indicating the actual angle.
In another example, this disclosure describes a non-transitory computer-readable storage medium having instructions stored thereon that, when executed, configure a computing system to: receive an image captured by a camera, the image being of a structural bearing, wherein the structural bearing is a hanger bearing or a rocker bearing; receive a temperature measurement generated by an instrument; determine, based on the temperature, an expected angle of the structural bearing relative to a base line; determine an actual angle of the structural bearing relative to the base line; superimpose a first line on the image, the first line indicating the expected angle; and superimpose a second line on the image, the second line indicating the actual angle.
The details of one or more examples of the disclosure are set forth in the accompanying drawings and the description below. Other features, objects, and advantages will be apparent from the description, drawings, and claims.
Controller device 10 may, for example, be a general-purpose device such as a personal digital assistants (PDAs), a laptop or desktop computer, a tablet computer, a cellular or satellite radio telephone, a so-called “smart phones,” or any other such device. In examples where controller device 10 is a general-purpose device, controller device 10 may be loaded with and may be configured to execute software designed to control UAV 6. In other examples, controller device 10 may be a special-purpose device designed specifically for use with UAV 6.
Controller device 10 communicates with UAV 6 via communication link 18. Communication link 18 may, for example, be a direct link through a radio communication protocol, such as WiFi, Bluetooth, ZigBee, a proprietary protocol, or any other suitable protocol. In other examples, communication link 18 is a network-based link where controller device 10 communicates with UAV 6 through one or more intermediary devices such as gateways, routers, switches, repeaters, or other such network devices.
Computing system 14 may, for example, include a general-purpose device such as a PDA, a laptop or desktop computer, a tablet computer, a smart phone, or any other such device. Computing system 14 may be loaded with and configured to execute software designed to process data collected by UAV 6. In some examples, UAV 6 may be configured to stream data to computing system 14 in real-time or near real time via, for example, a wireless communication link. In other examples, UAV 6 may store data while in flight and transfer the data to computing system 14 at a later time, such as after the completion of a flight.
UAV 6 is shown in
UAV 6 may, for example, be configured to acquire any or all of audio data, still image data, or video data. In the example of
In accordance with a technique of this disclosure, camera 16 of UAV 6 may capture an image of a structural bearing, such as a hanger bearing or a rocker bearing. Additionally, an instrument mounted on UAV 6 detects a temperature. Various types of instruments mounted on UAV 6 may detect the temperature. For example, an infrared thermometer mounted on UAV 6 may detect a temperature of the bridge bearing or other components of the bridge. In another example, a thermometer may measure the temperature of air surrounding UAV 6. In some examples, instruments mounted on UAV 6 may detect both the air temperature and the temperature of bridge components.
However, structural bearings (e.g., expansion joints) are potential points of structural failure. For example, if a rocker bearing is locked up due to heavy rust, loading due to thermal expansion may be transferred to structural components that are unable to bear the load, potentially resulting in a structural failure. Accordingly, structural bearings are frequently the subject of inspection, especially in bridges and other structures subject to adverse environmental conditions. However, frequent inspection may be difficult and expensive.
UAV 6 may send image data, thermometer data, and other information to computing system 14 via a communication link 20. Communication link 20 may, for example, be a direct link through a radio communication protocol, such as WiFi, Bluetooth, ZigBee, a proprietary protocol, or any other suitable protocol. In other examples, communication link 20 may be a network-based link where controller device 10 communicates with UAV 6 through one or more intermediary devices such as gateways, routers, switches, repeaters, or other such network devices. In examples where UAV 6 stores data and transfers the data to computing system 14 after completion of a flight, communication link 20 may represent a wired connection, such as a USB connection, Lightning connection, or other such connection. In other examples, communication link 20 may represent the manual transfer of data from UAV 6 to computing system 14 by, for example, ejecting a computer readable medium device from UAV 6 and inserting the computer readable medium device into computing system 14.
Computing system 14 determines, based on the temperature, an expected angle of the bearing relative to a base line (e.g., a horizontal or vertical line). Computing system 14 may determine the expected angle based on the temperature in various ways. For example, computing system 14 may store historical images of the bearing along with corresponding temperature readings. In this example, if a temperature corresponding to a current image of the bearing matches a temperature corresponding to a historical image, computing system 14 may determine that the expected angle is the angle of the bearing as shown in the historical image. Furthermore, in this example, if a temperature corresponding to a current image of the bearing does not match a temperature corresponding to a historical image, computing system 14 may estimate the expected angle based on multiple historical images. For example, computing system 14 may store a first image of the bearing that was captured when the temperature is 10° C., a second image of the bearing that was captured when the temperature is 0° C., and so on. In this example, if the current temperature is 10° C., computing system 14 may determine that the expected angle of the bearing is the same as the angle of the bearing shown in the first image. In this example, if the current temperature 5° C., computing system 14 may determine that the expected angle of the bearing is halfway between the angles of the bearing shown in the first image and the second image, assuming linear expansion of the bridge members. In other example, computing system 14 may perform a similar calculation assuming non-linear expansion of bridge members. Thus, in such examples, computing system 14 may determine the expected angle of the structural bearing based on the current temperature, the historical images, and the historical temperature measurements.
In some examples, computing system 14 may determine the expected angle based on the temperature and engineering characteristics of the structure. For example, computing system 14 may have engineering specifications of the structure, such as data on lengths and materials of applicable structural members of the structure. In this example, computing system 14 may calculate the expected lengths of structural members of the structure given the temperatures of the structural members and determine the expected angle of the bearing accordingly. For instance, the bearing may get further from vertical as the temperature gets hotter or colder.
Additionally, computing system 14 may determine an actual angle of the bearing relative to the base line. For example, computing system 14 may determine that an expected angle of the bearing should by 95°, given the temperature. In this example, computing system 14 may determine that the actual angle of the bearing relative to the same base line is 85°. Hence, in this example, the 10° difference in angle may indicate that the bearing is locked up.
To help a user interpret the image, computing system 14 superimposes a first line on the image. The first line indicates the expected angle. Additionally, computing system 14 superimposes a second line on the image. The second line indicates the actual angle. Thus, a user reviewing the image can easily see differences between the temperature-appropriate angle and the actual angle. This may enable the user to determine whether the bearing is seized up. In some examples, computing system 14 also superimposes the baseline onto the image.
In some examples, a tilt detection instrument in UAV 6 may detect a physical tilt of UAV 6 at a time camera 16 captures the image. In some examples, computing system 14 may use a tilt measurement generated by the tilt detection instrument to determine the base line. In some examples, computing system 14 uses readings from the tilt detection instrument to rotate the image to compensate for tilt of UAV 6. Furthermore, in some examples, an orientation detection instrument in UAV 6, such as a compass or gyroscope may determine a yaw and/or attitude of camera 16 at a time camera 16 captures the image. Computing system 14 may apply skew effects to the image to compensate for yaw and attitude variations between images.
Processing circuits 52 are intended to represent all processing circuitry and all processing capabilities of UAV 6. Processing circuits 52 may, for example, include one or more digital signal processors (DSPs), general purpose microprocessors, integrated circuits (ICs) or a set of ICs (e.g., a chip set), application specific integrated circuits (ASICs), field programmable logic arrays (FPGAs), or other equivalent integrated or discrete logic circuitry. Accordingly, the term “processor,” as used herein may refer to any of the foregoing structures, or combination thereof, or any other structure suitable for implementation of the techniques described herein.
Memory 54 is intended to represent all of the various memory devices within UAV 6. Memory 54 constitutes a computer-readable storage medium and may take the form of either a volatile memory that does not maintain stored contents once UAV 6 is turned off or a non-volatile memory that stores contents for longer periods of time, including periods of time when UAV 6 is in an unpowered state. Examples of volatile memories include random access memories (RAM), dynamic random access memories (DRAM), static random access memories (SRAM), integrated random access memory (IRAM), thyristor random access memory (TRAM), zero-capacitor random access memory (ZRAM), or any other type of suitable volatile memory. Examples of non-volatile memory include optical disk drives, magnetic disk drives, flash memory, read only memory (ROM), forms of electrically programmable memories (EPROM) or electrically erasable and programmable memories (EEPROM), or any other such type of non-volatile memory.
The processing functionality of UAV 6 may be implemented by hardware, software, firmware, or combinations thereof. Memory 54 may store software and firmware that include sets of instructions. Processing circuits 52 and, other hardware components of UAV 6, may execute the instructions to perform certain parts of the techniques of this disclosure.
Processing circuits 52 may cause transceiver 56 to use antenna 58 to send data to controller device 10 (
Camera 62 is configured to capture images. In some examples, camera 62 captures images based on visible light. In some examples, camera 62 captures images based on infrared radiation.
Sensors 64 are intended to represent all the various sensors included in UAV 6. UAV 6 may, for example, include one or more sensors used for flight management, such as accelerometers, gyroscopes, magnetometers, barometers, Global Navigation Satellite Systems (GNSS) sensors, tilt sensors, inertial measurement sensors, speed sensors, and others. Particularly, in the example of
Communication channels 110 interconnect processing circuits 102, memory 104, transceiver 106, and display 108. Power supply 100 provides power to processing circuits 102, memory 104, transceiver 106 and display 108. Processing circuits 102, memory 104, transceiver 106, and display 108 may be implemented in a manner similar to processing circuits 52, memory 54, and transceiver 56 described above with respect to
In the example of
In the example of
Furthermore, in the example of
In some examples, to ease the determination of the expected angle and actual angle, angle unit 116 may rotate or skew an image such that the image appears to be taken from the same position as historical images of the same structural bearing. For instance, if the historical images are all taken with a tilt of 0° relative to the horizon, but a gust of wind occurring when UAV 6 captured a new image caused the new image to be taken with a tilt of 5° relative to the horizon, angle unit 116 may rotate the new image −5° to ensure that the new image is from an angle consistent with the historical images. Similarly, historical images of the structural bearing may be taken straight on at the structural bearing, but a camera of UAV 6 may be yawed 4° when taking a new image of the structural bearing. Accordingly, in this example, angle unit 116 may apply a skew of −4° to the new image to correct for the yaw.
Image modification unit 118 may superimpose a first line on an image of the structural bearing. The first line indicates the expected angle. Additionally, image modification unit 118 may superimpose a second line on the image. The second line indicates the actual angle.
In accordance with a technique of this disclosure, computing system 14 has superimposed an expected angle line 162, an actual angle line 164, and a baseline 166 on image 150. Expected angle line 162 indicates an angle that hanger 158 is expected to have given the temperature. Actual angle line 165 indicates an actual angle of hanger 158. Baseline 166 is a vertical line that a user may use for reference.
In some examples, structural member 152 is a suspended span and structural member 154 is an anchor span. Thus, an end of structural member 154 closest to the pin-and-hanger assembly is supported by a fixed point, such as a bridge pier. However, an end of structural member 154 is not supported by a fixed point, but rather is suspended from structural member 154. In some examples, as part of determining the expected angle of a structural bearing (e.g., the pin-and-hanger assembly of
In normal operation, an angle of rocker member 206 changes as structural member 202 expands and contracts. However, the rocker bearing may lock up if rocker member 206 is no longer able to rotate at pivot point 212, e.g., due to corrosion. If the rocker bearing is locked up, the actual angle of rocker member 206 might not correspond to an expected angle of rocker member 206. In some examples, the rocker bearing does not move correctly due to other conditions, such as an object being jammed into a gap between structural members.
In accordance with a technique of this disclosure, computing system 14 (
In the example of
The computing system determines, based on the temperature, an expected angle of the bearing relative to a base line (258). Additionally, the computing system determines an actual angle of the bearing relative to the base line (260). The computing system may determine the expected angle and the actual angle in accordance with any of the examples provided elsewhere in this disclosure.
Additionally, the computing system superimposes a first line on the image (262). The first line indicates the expected angle. The computing system also superimposes a second line on the image (264). The second line indicates the actual angle. The computing system may superimpose the first line and the second line on the image by changing the values of pixels in the image such that the first line and the second line are visible in the image. In some examples, the computing system may output the image, with the superimposed lines, for display (266). For example, the computing system may send signals representing the image to a monitor for display.
Furthermore, in the example of
In the example of
The computing system determines, based on the temperature, an expected angle of the bearing relative to a base line (308). Additionally, the computing system determines an actual angle of the bearing relative to the base line (310). The computing system may determine the expected angle and the actual angle in accordance with any of the examples provided elsewhere in this disclosure.
Furthermore, in the example of
Although the foregoing description has been described with respect to cameras and instruments for measuring temperature mounted on UAVs, the techniques of this disclosure may be implemented in other ways. For example, the camera and instrument may be mounted on a robot configured to crawl along a structure. In another example, the camera and/or instrument may be handheld or mounted on a support such as a tripod.
In one or more examples, the functions described may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software, the functions may be stored on or transmitted over, as one or more instructions or code, a computer-readable medium and executed by a hardware-based processing unit. Computer-readable media may include computer-readable storage media, which corresponds to a tangible medium such as data storage media, or communication media including any medium that facilitates transfer of a computer program from one place to another, e.g., according to a communication protocol. In this manner, computer-readable media generally may correspond to (1) tangible computer-readable storage media which is non-transitory or (2) a communication medium such as a signal or carrier wave. Data storage media may be any available media that can be accessed by one or more computers or one or more processors to retrieve instructions, code and/or data structures for implementation of the techniques described in this disclosure. A computer program product may include a computer-readable medium.
By way of example, and not limitation, such computer-readable storage media can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage, or other magnetic storage devices, flash memory, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer. Also, any connection is properly termed a computer-readable medium. For example, if instructions are transmitted from a website, server, or other remote source using a coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave are included in the definition of medium. It should be understood, however, that computer-readable storage media and data storage media do not include connections, carrier waves, signals, or other transient media, but are instead directed to non-transient, tangible storage media. Disk and disc, as used herein, includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk and Blu-ray disc, where disks usually reproduce data magnetically, while discs reproduce data optically with lasers. Combinations of the above should also be included within the scope of computer-readable media.
Instructions may be executed by one or more processors, such as one or more DSPs, general purpose microprocessors, ASICs, FPGAs, or other equivalent integrated or discrete logic circuitry. Accordingly, the term “processor,” as used herein may refer to any of the foregoing structure or any other structure suitable for implementation of the techniques described herein. In addition, in some aspects, the functionality described herein may be provided within dedicated hardware and/or software modules configured for encoding and decoding, or incorporated in a combined codec. Also, the techniques could be fully implemented in one or more circuits or logic elements.
The techniques of this disclosure may be implemented in a wide variety of devices or apparatuses, including a wireless handset, an integrated circuit (IC) or a set of ICs (e.g., a chip set). Various components, modules, or units are described in this disclosure to emphasize functional aspects of devices configured to perform the disclosed techniques, but do not necessarily require realization by different hardware units. Rather, as described above, various units may be combined in a codec hardware unit or provided by a collection of interoperative hardware units, including one or more processors as described above, in conjunction with suitable software and/or firmware.
Various examples have been described. These and other examples are within the scope of the following claims.
Number | Name | Date | Kind |
---|---|---|---|
3992707 | Schmidtlein et al. | Sep 1976 | A |
4550376 | Maciejczak | Oct 1985 | A |
4664340 | Jackson | May 1987 | A |
4789947 | Maciejczak | Dec 1988 | A |
4899296 | Khattak | Feb 1990 | A |
5150857 | Moffitt et al. | Sep 1992 | A |
5152478 | Cycon et al. | Oct 1992 | A |
5295643 | Ebbert et al. | Mar 1994 | A |
5559695 | Daily | Sep 1996 | A |
5575438 | McGonigle et al. | Nov 1996 | A |
5695153 | Britton et al. | Dec 1997 | A |
5904724 | Margolin | May 1999 | A |
6206324 | Smith | Mar 2001 | B1 |
6377875 | Schwaerzler | Apr 2002 | B1 |
6422508 | Barnes | Jul 2002 | B1 |
6450445 | Moller | Sep 2002 | B1 |
6502787 | Barrett | Jan 2003 | B1 |
6536553 | Scanlon | Mar 2003 | B1 |
6575402 | Scott | Jun 2003 | B1 |
6588701 | Yavnai | Jul 2003 | B2 |
6604706 | Bostan | Aug 2003 | B1 |
6622090 | Lin | Sep 2003 | B2 |
6665594 | Armstrong | Dec 2003 | B1 |
6691949 | Plump et al. | Feb 2004 | B2 |
6694228 | Rios | Feb 2004 | B2 |
6712312 | Kucik | Mar 2004 | B1 |
6721646 | Carroll | Apr 2004 | B2 |
6813559 | Bodin et al. | Nov 2004 | B1 |
6847865 | Carroll | Jan 2005 | B2 |
6873886 | Mullen et al. | Mar 2005 | B1 |
6925382 | Lahn | Aug 2005 | B2 |
6959895 | Cylinder | Nov 2005 | B2 |
6961441 | Hershey et al. | Nov 2005 | B1 |
7000883 | Mercadal et al. | Feb 2006 | B2 |
7032861 | Sanders, Jr. et al. | Apr 2006 | B2 |
7044074 | Harrie et al. | May 2006 | B2 |
7044422 | Bostan | May 2006 | B2 |
7107148 | Bodin et al. | Sep 2006 | B1 |
7130741 | Bodin et al. | Oct 2006 | B2 |
7149611 | Beck et al. | Dec 2006 | B2 |
7158877 | Carlsson et al. | Jan 2007 | B2 |
7204455 | Sinclair | Jan 2007 | B2 |
7228227 | Speer | Jun 2007 | B2 |
7231294 | Bodin et al. | Jun 2007 | B2 |
7249732 | Sanders, Jr. et al. | Jul 2007 | B2 |
7269513 | Herwitz | Sep 2007 | B2 |
7286913 | Bodin et al. | Oct 2007 | B2 |
7289906 | van der Merwe et al. | Oct 2007 | B2 |
7299130 | Mulligan et al. | Nov 2007 | B2 |
7302316 | Beard | Nov 2007 | B2 |
7343232 | Duggan et al. | Mar 2008 | B2 |
7603207 | Abraham et al. | Oct 2009 | B2 |
8060270 | Vian et al. | Nov 2011 | B2 |
9609288 | Richman | Mar 2017 | B1 |
20030201909 | Hilliard | Oct 2003 | A1 |
20040016065 | Steiger | Jan 2004 | A1 |
20040094662 | Sanders, Jr. et al. | May 2004 | A1 |
20040129828 | Bostan | Jul 2004 | A1 |
20050082421 | Perlo | Apr 2005 | A1 |
20050165517 | Reich | Jul 2005 | A1 |
20060008137 | Nagandaripour et al. | Jan 2006 | A1 |
20060049304 | Sanders, Jr. et al. | Mar 2006 | A1 |
20060097107 | Parks et al. | May 2006 | A1 |
20060102780 | Parks | May 2006 | A1 |
20060106506 | Nichols et al. | May 2006 | A1 |
20060121418 | DeMarco et al. | Jun 2006 | A1 |
20060192047 | Goossen | Aug 2006 | A1 |
20060197835 | Anderson et al. | Sep 2006 | A1 |
20060231675 | Bostan | Oct 2006 | A1 |
20060235584 | Fregene et al. | Oct 2006 | A1 |
20060253228 | Abraham et al. | Nov 2006 | A1 |
20060271248 | Cosgrove et al. | Nov 2006 | A1 |
20060287824 | Lin | Dec 2006 | A1 |
20070018052 | Eriksson | Jan 2007 | A1 |
20070022935 | Griffith et al. | Feb 2007 | A1 |
20070034734 | Yoeli | Feb 2007 | A1 |
20070034738 | Sanders, Jr. et al. | Feb 2007 | A1 |
20070034739 | Yoeli | Feb 2007 | A1 |
20070051848 | Mantych et al. | Mar 2007 | A1 |
20070069083 | Shams et al. | Mar 2007 | A1 |
20070093945 | Grzywna et al. | Apr 2007 | A1 |
20070129855 | Coulmeau | Jun 2007 | A1 |
20070131822 | Stallard | Jun 2007 | A1 |
20070193650 | Annati | Aug 2007 | A1 |
20070200027 | Johnson | Aug 2007 | A1 |
20070221790 | Goossen | Sep 2007 | A1 |
20070228214 | Horak | Oct 2007 | A1 |
20070244608 | Rath et al. | Oct 2007 | A1 |
20070246610 | Rath et al. | Oct 2007 | A1 |
20070262195 | Bulaga et al. | Nov 2007 | A1 |
20070271032 | Cheng et al. | Nov 2007 | A1 |
20070284474 | Olson et al. | Dec 2007 | A1 |
20070295298 | Mark | Dec 2007 | A1 |
20080023587 | Head et al. | Jan 2008 | A1 |
20080033604 | Margolin | Feb 2008 | A1 |
20080035786 | Bilyk et al. | Feb 2008 | A1 |
20080059068 | Strelow et al. | Mar 2008 | A1 |
20080071431 | Dockter et al. | Mar 2008 | A1 |
20080078865 | Bume | Apr 2008 | A1 |
20090265193 | Collins et al. | Oct 2009 | A1 |
20100215212 | Flakes, Jr. | Aug 2010 | A1 |
20110203212 | Matys | Aug 2011 | A1 |
20130235185 | Sobotka | Sep 2013 | A1 |
20170066459 | Singh | Mar 2017 | A1 |
20170227968 | Klinger | Aug 2017 | A1 |
20170236024 | Wang | Aug 2017 | A1 |
20180067003 | Michiwaki | Mar 2018 | A1 |
20180130196 | Loveland | May 2018 | A1 |
20190128771 | Santarone | May 2019 | A1 |
Number | Date | Country |
---|---|---|
105597308 | May 2016 | CN |
106054916 | Oct 2016 | CN |
106320173 | Jan 2017 | CN |
1767453 | Mar 2007 | EP |
1868008 | Dec 2007 | EP |
2394340 | Apr 2004 | GB |
0015497 | Mar 2000 | WO |
2004002821 | Jan 2004 | WO |
2007001369 | Jan 2007 | WO |
2007058643 | May 2007 | WO |
Entry |
---|
Gillins, et al., “Cost-Effective Bridge Safety Inspection Using Unmanned Aircraft Systems (UAS),” Feb. 2016, 10 pp. |
Adhikari, et al., “Image-Based Change Detection for Bridge Inspection,” retrieved from http://www.iaarc.org/publications/fulltext/isarc2013Paper158.pdf on Jun. 2017, 11 pp. |
“Unmanned Aerial Vehicle Bridge Inspection Demonstration Project,” Minnesota Department Transportation, Jul. 2015, 214 pp. [uploaded in 3 parts]. |
Number | Date | Country | |
---|---|---|---|
20190087658 A1 | Mar 2019 | US |