Municipal lighting audits entail the collection of many attributes of light poles, such as geo-location, the lamp and luminaire type, pole condition, and the height of the lighting fixture. Often, this is visually inferred based on past experience of the auditor. Moreover, the entire process is manual, and therefore slow and unscalable.
EP 1 953 568 (Bosch GMBH Robert 6 Aug. 2008 (2008-08-06) teaches an imager-semiconductor component that comprises two-dimensional integrated arrangement of imager-pixels for receiving an optical or infrared radiation and emission of picture signals. Measuring-cells are provided, which are integrated on the semiconductor component for determining a distance by measuring travelling time of the optical or infrared radiation.
The exemplary embodiments relate to a system and a method for automated luminaire detection.
According to an aspect of the present disclosure, the system detects a luminaire height. The system comprises an infrared recording component configured to capture a plurality of infrared images, a depth recording component configured to capture a plurality of depth images and a processing device. The processing device configured to identify a luminaire in at least one of the infrared images, correlate the identified luminaire to at least one of the depth images, and determine the height of the luminaire based on the at least one of the depth images.
In another aspect of the system, the infrared recording component and the depth recording component are one of synchronized to capture their respective images simultaneously or the respective images are captured at a rate independent of each other.
In another aspect, the system further comprises a memory that stores the plurality of the depth images and the plurality of the infrared images. A bounding box may be superimposed on a portion of an infrared image from the plurality of infrared images. Further, the correlation may be based on one of timestamps or GPS coordinates.
In another aspect, the system further comprises a further depth recording component configured to capture a further plurality of depth images. The processor may determine the height of the luminaire based on a measured-out ceiling grid that is generated based from at least one of the plurality of depth images or of the further plurality of depth images.
In a further aspect, a system for collecting luminaire information comprises a motor vehicle, a first data collection apparatus that collects first data related a plurality of luminaires, wherein the first data collection apparatus is mounted on or within the motor vehicle, a second data collection apparatus that collects second data related the plurality of luminaires, wherein the second data collection apparatus is mounted on or within the motor vehicle, and a processing device configured to analyze the first and second data to determine at least one characteristic for each of the plurality of luminaires.
In another aspect, the first data collection apparatus is an infrared recording component configured to capture a plurality of infrared images and the second data collection apparatus is a depth recording component configured to capture a plurality of depth images.
In another aspect, the first data collection apparatus is an image recording component configured to capture a plurality of photographic images and the second data collection apparatus is a spectrometer configured to measure a plurality of wavelength measurements.
In another aspect, the system further comprises a GPS component configured to record a GPS location of the first data collected and the second data collected. The processing device may a luminaire height based on the first data and the second data. In another aspect, the processing device identifies a luminaire cluster based on the first data and the second data.
In a further aspect, a system for detecting and identifying a luminaire cluster comprises an image recording component configured to capture a plurality of photographic images, a spectrometer configured to measure a plurality of wavelength measurements and a processing device. The processing device is configured to detect and group luminaire housing structures from the photographic images, identify and group luminaire types from the wavelength measurements and classify the grouped luminaire housing structures and the grouped luminaire types.
In another aspect, the system further comprises a GPS component configured to record a GPS location for each of the plurality of the photographic images and each of the plurality of wavelength measurements. Each of the plurality of the photographic images and each of the wavelength measurements may be matched by the GPS location. In yet another aspect, the image recording component records the plurality of photographic images at a different time than when the spectrometer records the wavelength measurements.
In another aspect, the processing device is further configured to organize the grouped luminaire housing structures into subgroups, wherein the subgroups comprise a first type of luminaire housing structures group and a first type of luminaire types.
In another aspect, the image recording component records the plurality of photographic images at rate, the rate correlating to a velocity of the system.
In another aspect, the grouping the luminaire housing structures comprises of comparing each of the luminaire housing structures from the plurality of the photographic images to a plurality of candidate images, assigning a score to each of the photographic images based on a number of matching points to each of the candidate images and grouping the luminaire housing structures based on the score.
The exemplary embodiments may be further understood with reference to the following description and the appended drawings wherein like elements are referred to with the same reference numerals. The exemplary embodiments relate to a system and method for automated detection of a luminaire.
The processing device 110 may be any type of computing device, such as a personal computer, tablet, smartphone, or other electronic device that can be used to record data. The infrared camera 120 may be any type of thermographic camera, such as those that form an image (e.g. an infrared image) using infrared radiation instead of visible light. The depth camera 130 may be any type of range-based imaging device used to resolve the distance between the depth camera 130 and an object(s). This may be done by measuring each pixel in a depth image, although other methods may be utilized.
The image camera 140 may be any type of photographic camera, such as those capable of capturing photographic images. The spectrometer 150 may be any type of device capable of measuring a spectrum, such as a graph that shows intensity as a function of wavelength. In particular, the spectrometer may be utilized to record measurement pertaining to the luminaire. The luminaire meter 160 may be any type of device capable of measuring a lux (or luminance eminence), a color, a temperature, a wavelength spectrum, a color renditioning index, and/or chromaticity coordinates of the luminaire. The GPS module 170 may be any type of device capable of accurately calculating a geographical location by utilizing information of GPS satellites. In particular, the GPS module may produce GPS coordinates of a measurement or an image recorded by any of the LIC components. For example, upon or after capture, the infrared image capture by the infrared camera 120 may be tagged with GPS coordinates. Other locations may also be used. In should be noted that any of the measurements or the images recorded by any of the LIC components may also be tagged with a timestamp, a date, etc.
The processor 210 may engage the LIC system 100 components to perform data collection. The memory 220 stores the data collected by the LIC system components as well as other data, such as but not limited to, the timestamp, the date, an elevation of the infrared camera 120 and/or the depth camera 130, etc.
The input device 240 may receive inputs from the user and includes a keyboard, a mouse, a touch screen and/or other input devices. Further, the input device 240 may receive inputs from the LIC system 100 components. The output device 250 may communicate data to the user via a monitor, a printer and/or other output devices. The receiver 260 and the transmitter 270 may be utilized for wired and/or wireless communications such as with a communications network. In an exemplary embodiment, a combined transceiver may be utilized to provide the functionalities of the receiver 260 and transmitter 270.
In step 310, the LIC system 100 records data collected by the infrared camera 120 and the depth camera 130. Specifically, the motor vehicle mounted with the LIC system 100 may drive along on the road. The infrared camera 120 may continuously capture the infrared images while the depth camera 130 continuously captures the depth images. The infrared camera 120 and the depth camera 130 may be synced to capture their respective images simultaneously. Alternatively, the infrared camera 120 and the depth camera 130 may capture their respective images at different rates.
Various methods for determining a rate of capture of the infrared images and depth images may be implemented. In one exemplary embodiment, the rate of capture may be set to a specified time interval. In a different exemplary embodiment, the rate of capture may be set to correlate with the velocity of the motor vehicle. For example, the rate of capture may be set to linearly or exponentially increase and decrease dependent on the velocity of the motor vehicle. In yet another exemplary embodiment, the rate of capture may be set to 0 if the motor vehicle is stationary, such as at a stop light or in a traffic jam.
In a further exemplary embodiment, the infrared camera 120 may be set to capture the infrared images when a heat signature is detected. As such, the depth camera 130 may be configured to take the depth images in response to actions of the infrared camera 120. For example, the depth camera 130 may be configured to capture the depth images at a simultaneous time as the infrared camera 120 captures the infrared images or on a specified time delay. Additionally, a ratio may be set regarding an amount of the infrared images captured to the amount of the depth images captured. For example, for every one of the infrared images captured, two depth images may be captured.
While preferable, in method 300, for the motor vehicle to drive during nighttime, the LIC system 100 can record data at any time of day as long as the luminaires are producing a heat signature. The infrared images and the depth images captured may be stored in the memory 220.
In step 320, the recorded data of step 310 (e.g. the infrared images and the depth images) is processed. While exemplary embodiments will make reference to this step and further steps of method 300 as being performed by the processor 210, is should be noted that the recorded data may be processed by a further processing unit. For example, the recorded data may be transferred from the memory 220 of the processing unit 110 to the further processing unit.
In an exemplary embodiment, the processing may comprise identifying the presence of the luminaire in the infrared image. As discussed above, this may be done by detecting infrared radiation in the infrared image. If no luminaire is detected, the processor 210 may proceed to the next infrared image. If the luminaire is detected, the processor 210 may further determine whether the luminaire is the same as a luminaire identified in a different infrared image. If the same luminaire is identified, the processor 210 may, again, proceed to the next infrared image. Infrared images with no detected luminaire may be discarded.
In exemplary embodiments, a bounding box may be utilized on a portion of the infrared image identified to be the luminaire. The bounding box may be a point set that fully encloses an area in the infrared image. For example, after detecting the heat signature in the infrared image, a bounding box may be superimposed to enclose the heat signature. As will be further explained below, a similar bounding box may then be superimposed in a similar location on a correlating depth image(s). By utilizing the bounding box, the amount of processing needed would be reduced since only the enclosed portion of the depth image(s) will need to be analyzed.
The processing may further comprise correlating the infrared image with the detected luminaire to the depth image(s). This may be done by using the timestamp or the GPS coordinates. Once the infrared image(s) and the depth image(s) are correlated, the bounding box may be also be imposed onto the depth image. In particular, if the infrared image and the depth image were captured at the same or nearly the same time, then the bounding box may be placed at the same position on the depth image as it was placed on the infrared image. In an alternative exemplary embodiment, the placement of the bounding box in the depth image may be appropriately shifted with respect to the placement of the bounding box on the infrared image to account for different placements of the infrared camera 120 and the depth camera 130, different fields of view by the infrared camera 120 and the depth camera 130, differences in time of capture and/or a speed of the vehicle at the time of capture, etc.
In a further exemplary embodiment, the accuracy of the identifying may be improved. In particular, the depth camera 130 may be mounted at a 45° angle, such that the depth camera 130 faces upwards as well as laterally. If the luminaire is detected in the infrared image, the processor 210 may verify whether a structural object is attached to the luminaire in the correlated depth image. In additional, the processor may verify whether the structural object extends to a street side view in the correlated depth image. It will be understood by those of skill in the art that a further depth camera may be used in this embodiment instead of the depth camera 130 being mounted at a 45° angle. It will also be understood by those skilled in the art that the angle of mounting for any embodiment may be altered for optimal measurements.
In step 330, the LIC system 100 determines the height of the luminaire. The height of the luminaire may be determined by combining the measured height between the road/ground and the depth camera 130 with a vertical height between the depth camera 130 and the luminaire. The vertical height may be measured or calculated.
In one exemplary embodiment, if the motor vehicle is able to drive either directly underneath or relatively close to the luminaire, the distance measured by the depth camera 130 is determined to be the vertical height. In order to increase the accuracy of the vertical height, a measure area, as seen in
In an alternate embodiment, as seen in
d12=(d3+c)2+h2
d22=c2+h2
With d1, d2 and d3 being known values, the processor 210 can calculate the values of c and h from the above formulas. The value of h would be the vertical height. The measured vertical height between the road/ground may be, either, but not limited to, the height of the first depth camera 130a, the height of the second depth camera 130b or an average of their respective heights. This method may be used when driving directly underneath or relatively close to the luminaire is not feasible.
In yet another embodiment, as seen in as seen in
By utilizing a measured-out ceiling grid on the depth image, an angle of each pixel can be determined. The ceiling grid may contain grid coordinates and be of a predetermined height. It should be noted that the resolution may be improved by narrowing a spacing between the grid coordinates. The angle offset of each of the grid coordinates may be determined through the following function:
θ=tan−1(z/r)
wherein:
r=a radius length along a horizontal plane from an origin to grid coordinates.
z=a vertical z-axis height of the grid coordinates.
θ=an angle formed by the horizontal plane and a line of sight to the grid coordinates.
The calculated θ is then inserted into formula z=sin(θ)*d. The variable d is the measurement taken by the depth camera 130. The resulting variable z produces the vertical height.
In step 340, the LIC system 100 determines if there is any remaining data available for processing. If there is, the method proceeds to step 320 in order to continue processing the data. If there are no images remaining, the method may end.
In step 710, the LIC system 100 records data collected by the image camera 140, the spectrometer 150 and the GPS module 170. Specifically, the motor vehicle mounted with the LIC system 100 may drive along on the road. The image camera 140 may continuously capture the photographic images while the spectrometer 150 records the wavelength measurements. Further, the GPS module 170 may record the GPS coordinates of each of the photographic images captured and/or each of the wavelength measurements recorded. The image camera 140 and the spectrometer 150 may be synced to capture/record their respective images and measurements simultaneously. Alternatively, the image camera 140 and the spectrometer 150 may capture/record their respective images and measurements at different rates.
In an exemplary embodiment, the capturing of the photographic images and the recording of the wavelength measurements may be conducted at separate times. Specifically, the photographic images may be captured during daytime. This would provide for improved detail of the photographic images. This would further negate a need for a flash device, as compared to capturing the photographic images during the nighttime. On the other hand, the recording of wavelength measurements may be conducted during nighttime. This would ensure that the luminaires are turned on and emitting their respective wavelengths. As disclosed above, the GPS module 170 may record the GPS coordinates of each of the photographic images captured and each of the wavelength measurement recorded.
Similar to step 310, various methods for determining a rate of capture of the photographic images and a rate of recordation of the wavelength measurements may be used. In one exemplary embodiment, the rate of capture of the photographic images may be set to a specified time interval. In a different exemplary embodiment, the rate of capture of the photographic images may be set to correlate with the velocity of the motor vehicle. For example, the rate of capture of the photographic images may be set to linearly or exponentially increase and decrease dependent on the velocity of the motor vehicle. In yet another exemplary embodiment, the rate of capture of the photographic images may be set to 0 if the motor vehicle is stationary, such as at a stop light or in traffic. It will be understood by those of skill in the art that similar methods may be applied for the rate of recordation of the wavelength measurements. In an alternate exemplary embodiment, the luminaire meter 160 may be used instead of the spectrometer 150.
In step 720, the recorded data of step 710 (e.g. the photographic images and the spectrometer measurements) is processed. While exemplary embodiments will make reference to this step and further steps in method 700 as being performed by the processor 210, is should be noted that the recorded data may be processed by the further processing unit.
In an exemplary embodiment, the processing may comprise detecting the luminaire housing structure from the photographic images. Specifically, a procedure may be utilized to detect a subset of pixels in each of the photographic images that correspond to the luminaire housing structure. Each of the photographic images that are determined to possess the luminaire housing structure may remain stored for further processing while each of the photographic images that are determined to not possess the luminaire housing structure may be discarded. The procedure utilized for detecting the luminaire housing structures may be a Cascade object detector, a Viola-Jones algorithm or may rely on Haar features.
The processing may further comprise grouping luminaire housing structure types. Specifically, a candidate image of a luminaire housing structure type may be selected from the photographic images. Remaining photographic images may be compared to the candidate image. For example, a score may be assigned to each of the remaining photographic images based on a number of matching points, as seen in
The processing may also comprise identifying the presence of the luminaire type from the spectrometer measurements. This may be done by comparing the intensity of the wavelengths from the spectrometer measurements to a spectral power distribution. For example, as seen in
Remaining spectrometer measurements may then be correlated to the respective photographic images, forming a compilation. In an exemplary embodiment, the correlating may be accomplished by matching the GPS coordinates of the spectrometer measurement to the photographic image. In another exemplary embodiment, the correlating may be accomplished by utilizing the timestamps.
It will be understood by those skilled in the art that the GPS coordinates may be used to eliminate identifying the presence of either the luminaire or the luminaire housing structure in their respective measurements/photographic images. In particular, if the photographic images are processed, the GPS coordinates may be used to eliminate spectrometer measurements in locations that were determined to not contain any luminaire housing structures, and vice-versa. This would reduce the amount of processing required.
In an exemplary embodiment, the grouped luminaire housing structure types may be further organized into subgroups. Specifically, the groups of luminaire housing structure types may be organized into the subgroups based on their correlating spectrometer measurements. For example, a group one, which contains photographic images of a luminaire housing structure type 1, may be split into subgroup 1 and subgroup 2, with subgroup 1 containing luminaire housing structure types having a first type of the luminaire and subgroup 2 containing luminaire housing structure types having a second type of the luminaire. Each of the groups and/or the subgroups may be identified as one of the plurality of clusters and assigned an identification number.
In step 730, the groups and/or subgroups may be classified. Specifically, a classification algorithm may be utilized to cycle thought the groups and/or subgroups and assign a classification indicative of the luminaire class and the luminaire housing structure type. For example, the algorithm may indicate that the luminaire class is one of a tungsten incandescent class, a mercury fluorescent class, a low-pressure sodium class, a high-pressure sodium class or a metal halide class. It will be understood by those of skill in the art the above luminaire class list is not exhaustive and that other luminaire classes may be included. In another exemplary embodiment, a person may be used to cycle thought the groups and/or subgroups and assign the classification indicative of the luminaire class and the lamp type. In particular, the person may be an expert capable of quickly identifying the group and/or subgroup.
Either the classification algorithm or the person may identify an improperly labeled compilation, group or subgroup. For example, a determination may be made that the compilation is not similar to a remainder of the group/subgroup. The non-similar compilation may then be removed from the group and properly identified.
In step 740, a feedback loop may be used to improve accuracy of the method 700. Specifically, the non-similar compilations may be analyzed to determine a reason as to why each of the non-similar compilations was improperly labeled. The process data step 720 may then be modified for improved performance.
Those skilled in the art will understand that the above-described exemplary embodiments may be implanted in any number of manners, including, as a separate software module, as a combination of hardware and software, etc. Further, those skilled in the art will understand that the above-described exemplary embodiments may be used separately or in combination. For example, as shown in
It is noted that the claims may include reference signs/numerals in accordance with PCT Rule 6.2(b). However, the present claims should not be considered to be limited to the exemplary embodiments corresponding to the reference signs/numerals.
A system for detecting and identifying a luminaire cluster, comprising:
an image recording component configured to capture a plurality of photographic images;
a spectrometer configured to measure a plurality of wavelength measurements; and
a processing device configured to:
The system of example 1, further comprising:
a GPS component configured to record a GPS location for each of the plurality of the photographic images and each of the plurality of wavelength measurements.
The system of example 2, wherein each of the plurality of the photographic images and each of the wavelength measurements are matched by the GPS location.
The system of example 3, wherein the image recording component records the plurality of photographic images at a different time than when the spectrometer records the wavelength measurements.
The system of example 1, wherein the processing device is further configured to:
organize the grouped luminaire housing structures into subgroups, wherein the subgroups comprise a first type of luminaire housing structures group and a first type of luminaire types.
The system of example 1, wherein the image recording component records the plurality of photographic images at rate, the rate correlating to a velocity of the system.
The system of example 1, wherein the grouping the luminaire housing structures comprises of:
comparing each of the luminaire housing structures from the plurality of the photographic images to a plurality of candidate images;
assigning a score to each of the photographic images based on a number of matching points to each of the candidate images; and
grouping the luminaire housing structures based on the score.
Number | Date | Country | Kind |
---|---|---|---|
16174028.7 | Jun 2016 | EP | regional |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/EP2017/060801 | 5/5/2017 | WO | 00 |
Number | Date | Country | |
---|---|---|---|
62332712 | May 2016 | US |