The present disclosure generally relates to a system and method for gas cloud detection and, in particular, to a system and method of detecting spectral signatures of hydrogen sulfide.
Spectral imaging systems and methods have applications in a variety of fields. Spectral imaging systems and methods obtain a spectral image of a scene in one or more regions of the electromagnetic spectrum to detect phenomena, identify material compositions or characterize processes. The spectral image of the scene can be represented as a three-dimensional data cube where two axes of the cube represent two spatial dimensions of the scene and a third axes of the data cube represents spectral information of the scene in different wavelength regions. The data cube can be processed using mathematical methods to obtain information about the scene. Some of the existing spectral imaging systems generate the data cube by scanning the scene in the spatial domain (e.g., by moving a slit across the horizontal dimensions of the scene) and/or spectral domain (e.g., by scanning a wavelength dispersive element to obtain images of the scene in different spectral regions). Such scanning approaches acquire only a portion of the full data cube at a time. These portions of the full data cube are stored and then later processed to generate a full data cube.
Furthermore, many conventional spectral imaging systems are unable to detect and identify hydrogen sulfide (H2S) gas, which is very dangerous to humans, even though such conventional systems may be able to detect other hydrocarbon gases.
Accordingly, there remains a continuing need for a spectral imaging system that can detect and identify hydrogen sulfide gas.
The systems, methods and devices of this disclosure each have several innovative aspects, no single one of which is solely responsible for the desirable attributes disclosed herein.
In one embodiment, an infrared (IR) imaging system for detecting a gas is disclosed. The imaging system can include an optical filter that selectively passes light having a wavelength in a range of 1500 nm to 1700 nm while attenuating light at wavelengths above 1700 nm and below 1500 nm. The imaging system can include an optical detector array sensitive to light having a wavelength of 1590 nm that is positioned rear of the optical filter
In one embodiment, an infrared (IR) imaging system for imaging a scene is disclosed. The imaging system can include an optical system comprising an optical focal plane array (FPA) unit and a plurality of spatially and spectrally different optical channels to transfer IR radiation from the scene towards the optical FPA unit, each optical channel positioned to transfer a portion of the IR radiation incident on the optical system from the scene towards the optical FPA unit. At least one of the plurality of optical channels can be in the short-wavelength infrared (SWIR) spectral range. The imaging system can be configured to acquire a first video image of the scene in the short-wavelength infrared spectral range.
In one embodiment, an infrared (IR) imaging system for detecting hydrogen sulfide (H2S) gas is disclosed. The imaging system can an optical detector array, and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector array. Within the band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of H2S gas can be greater than a convolution of the optical filter with an absorption spectrum of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O).
In one embodiment, an infrared (IR) imaging system for detecting hydrogen sulfide (H2S) gas is disclosed. The system can include an optical detector array and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector. The optical filter can comprise a passband that transmits within the passband a first signal representative of hydrogen sulfide (H2S) and a second signal representative of one of carbon dioxide (CO2), methane (CH4), water (H2O), and sulfur dioxide (SO2), the first signal comprising a first peak within the passband, the second signal comprising a second peak within the passband, wherein a ratio of the first peak to the second peak is in a range of 50 to 10,000.
In one embodiment, an optical filter is disclosed. The optical filter can include a filter element comprising a passband that selectively passes light within a band of infrared (IR) wavelengths. The passband can transmit a first signal representative of hydrogen sulfide (H2S) and a second signal representative of one of carbon dioxide (CO2), methane (CH4), water (H2O), and sulfur dioxide (SO2), the first signal comprising a first peak within the passband, the second signal comprising a second peak within the passband, wherein a ratio of the first peak to the second peak is in a range of 50 to 10,000.
In one embodiment, an optical filter is disclosed. The optical filter can include a filter element comprising a passband that selectively passes light within a band of infrared (IR) wavelengths. Within the band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of H2S gas can be greater than a convolution of the optical filter with an absorption spectrum of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O).
In one embodiment, an infrared (IR) imaging system for imaging a scene is disclosed. The imaging system can include an optical system comprising an optical focal plane array (FPA) unit and a plurality of spectrally different optical channels to transfer IR radiation from the scene towards the optical FPA unit. Each optical channel can be positioned to transfer a portion of the IR radiation incident on the optical system from the scene towards the optical FPA unit. The plurality spectrally different optical channels can be coupled to or integrally formed with the optical FPA unit. At least one of the plurality of optical channels can be in the short-wavelength infrared (SWIR) spectral range. The imaging system can be configured to acquire a first video image of the scene in the short-wavelength infrared spectral range.
In one embodiment, an infrared (IR) imaging system for detecting carbon dioxide (CO2) gas is disclosed. The imaging system can include an optical detector array and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector array. Within the band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of CO2 gas can be greater than a convolution of the optical filter with an absorption spectrum of hydrogen sulfide (H2S), or methane (CH4), or sulfur dioxide (SO2), or water (H2O).
In one embodiment, an infrared (IR) imaging system for detecting carbon dioxide (CO2) gas is disclosed. The imaging system can comprise an optical detector array and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector. The optical filter can comprise a passband that transmits within the passband a first signal representative of carbon dioxide (CO2) and a second signal representative of one of hydrogen sulfide (H2S), methane (CH4), water (H2O), and sulfur dioxide (SO2), the first signal comprising a first peak within the passband, the second signal comprising a second peak within the passband, wherein a ratio of the first peak to the second peak is in a range of 50 to 10,000.
In one embodiment, an infrared (IR) imaging system for detecting methane (CH4) gas is disclosed. The system can include an optical detector array and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector array. Within the band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of CH4 gas can be greater than a convolution of the optical filter with an absorption spectrum of hydrogen sulfide (H2S), or carbon dioxide (CO2), or sulfur dioxide (SO2), or water (H2O).
In one embodiment, an imaging system is disclosed. The imaging system can comprise imaging optics (such as one or more lenses) and one or more optical channels that convey infrared radiation to an optical detector array. The imaging system can include processing electronics configured to detect hydrogen sulfide gas from a scene in which multiple chemicals are present. For example, the one or more optical channels can selectively filter wavelengths at which hydrogen sulfide has high absorption characteristics as compared with other gases commonly found at the scene.
In one embodiment, an infrared (IR) imaging system for detecting hydrogen sulfide (H2S) gas is disclosed. The system can include an optical system comprising an optical detector array and one or more optical channels that transfer infrared radiation to the optical detector array. The system can include processing electronics configured to process image data received by the optical detector array, the processing electronics configured to detect H2S gas based on the captured image data.
Various embodiments of the systems described herein provide an infrared (IR) imaging system for determining a concentration of a target chemical species in an object (e.g., a gas plume). The imaging system includes (i) an optical system, having an optical focal plane array (FPA) unit configured to receive IR radiation from the object along at least two or more optical channels defined by components of the optical system, the at least two or more optical channels being spatially and spectrally different from one another; and (ii) a processor or processing electronics configured to acquire multispectral optical data representing said target chemical species from the received IR radiation in a single occurrence of data acquisition (or snapshot). The optical FPA unit includes an array of photo-sensitive devices that are disposed at the focus of one or more lenses. In various embodiments, the array of photo-sensitive devices can include a two-dimensional imaging sensor that is sensitive to radiation having wavelengths between 1 μm and 20 μm (for example, in mid infra-red wavelength range, long infra-red wavelength range, visible wavelength range, etc.). In various embodiments, the array of photo-sensitive devices can include CCD or CMOS sensors, bolometers or other detectors that are sensitive to infra-red radiation. The optical system may include an optical aperture (a boundary of which is defined to circumscribe or encompass the at least two or more spatially distinct optical channels) and one or more optical filters. In various implementations, the one or more optical filters can comprise at least two spectrally-multiplexed filters. Each of these optical filters can be associated with one of the at least two or more optical channels and configured to transmit a portion of the IR radiation received in the associated optical channel. In various embodiments, the one or more optical filters can be spectrally multiplexed and may include, for example, at least one of a longpass optical filter and a shortpass optical filter, or a band pass filter (with or without a combination with another filter such as a notch filter, for example). The optical system may further include at least two imaging lenses. The at least two imaging lenses, for example each of the imaging lens, may be disposed to transmit IR radiation (for example, between about 1 micron and about 20 microns), that has been transmitted through a corresponding optical filter towards the optical FPA unit. In one embodiment, the optical FPA unit is positioned to receive IR radiation from the object through the at least two imaging lenses to form respectively-corresponding two or more sets of imaging data representing the object. The processor or processing electronics is configured to acquire this optical data from the two or more sets of imaging data. In various embodiments of the imaging systems, the FPA unit may be devoid of cooling systems. In various embodiments, two or more of the array of photo-sensitive devices may be uncooled. In some embodiments, the system further comprises two or more temperature-controlled shutters removably positioned to block IR radiation incident onto the optical system from the object.
Also disclosed herein is an implementation of a method of operating an infrared (IR) imaging system. The method includes receiving IR radiation from an object through at least two optical channels defined by components of an optical system of the IR imaging system, which at least two optical channels are spatially and spectrally different from one another. The method further includes transmitting the received IR radiation towards an optical focal plane array (FPA) unit that is not being cooled in the course of normal operation. For example, in various embodiments of the imaging systems, the FPA unit may be devoid of cooling systems. In various embodiments, two or more of the array of photo-sensitive devices may be uncooled. Some embodiments further comprise removably positioning at least one of at least two temperature-controlled shutters in front of the optical system to block IR radiation incident onto the optical system from the object.
Various innovative aspects of the subject matter described in this disclosure can be implemented in the following embodiments:
An infrared (IR) imaging system, the imaging system comprising:
a plurality of cameras;
at least one thermal reference source having a known temperature placed in front of the plurality of cameras and configured to be imaged by the plurality of cameras; and
a data-processing unit comprising a processor, the imaging system configured to:
acquire with the plurality of cameras a plurality of frames having regions that correspond to the image of the reference source; and
apply a dynamic calibration correction to the plurality of cameras to allow every camera in the plurality of cameras to be calibrated to agree with every other camera in the plurality imaging the reference source.
The system of Embodiment 1, wherein the plurality of cameras comprises a FPA unit and a plurality of lenses.
The system of any of Embodiments 1-2, wherein the FPA unit comprises one FPA or a plurality of FPAs.
The system of any of Embodiments 1-3, wherein the at least one thermal reference source has a known spectrum.
The system of any of Embodiments 1-4, further comprising an additional thermal reference source imaged by the plurality of cameras.
The system of any of Embodiments 1-5, wherein the additional reference source has a temperature and a spectrum different from the known temperature and the known spectrum of the at least one reference source.
The system of any of Embodiments 1-6, wherein the temperature of the additional thermal reference source is less than the known temperature.
The system of any of Embodiments 1-7, wherein the temperature of the additional thermal reference source is greater than the known temperature.
The system of any of Embodiments 1-8, wherein the at least one reference source is displaced away from a conjugate image plane of the plurality of cameras such that the image of the at least one reference source captured by the plurality of cameras is blurred.
The system of any of Embodiments 1-9, wherein the at least one reference source is positioned at a conjugate image plane of the plurality of cameras.
The system of any of Embodiments 1-10, further comprising a plurality of mirrors configured to image the at least one reference source onto the plurality of cameras.
The system of any of Embodiments 1-11, wherein the plurality of mirrors are disposed outside a central field of view of the plurality of cameras.
The system of any of Embodiments 1-12, further comprising a first and a second temperature-controlled shutter removably positioned to block IR radiation incident on the system from reaching the plurality of cameras.
The system of any of Embodiments 1-13, wherein the system includes at least two spatially and spectrally different optical channels.
The system of any of Embodiments 1-14, wherein the system includes at least three optical channels.
The system of any of Embodiments 1-15, wherein the system includes at least four optical channels.
The system of any of Embodiments 1-16, wherein the system includes at least five optical channels.
The system of any of Embodiments 1-17, wherein the system includes at least six optical channels.
The system of any of Embodiments 1-18, wherein the system includes at least seven optical channels.
The system of any of Embodiments 1-19, wherein the system includes at least eight optical channels.
The system of any of Embodiments 1-20, wherein the system includes at least nine optical channels.
The system of any of Embodiments 1-21, wherein the system includes at least ten optical channels.
The system of any of Embodiments 1-22, wherein the system includes at least twelve optical channels.
The system of any of Embodiments 1-23, further comprising one or more sensors configured to measure a temperature of the at least one reference source.
The system of any of Embodiments 1-24, wherein the plurality of cameras is configured to image the same portion of the at least one reference source.
An infrared (IR) imaging system, the imaging system comprising:
a plurality of cameras;
a first temperature-controlled reference source imaged by the plurality of cameras;
a second temperature-controlled reference source imaged by the plurality of cameras; and
a data-processing unit comprising a processor, said data-processing unit configured to:
acquire with the plurality of cameras a plurality of frames having regions that correspond to the image of the reference source; and
dynamically calibrate the plurality of cameras so that various cameras imaging a scene are forced to agree on a temperature estimate of the first and second reference sources.
The imaging system of any of Embodiment 26, wherein the data-processing unit is configured to calculate a dynamic calibration correction and apply the correction to the plurality of cameras for each of the plurality of frames.
The system of any of Embodiments 26-27, wherein the first reference source is maintained at a first temperature.
The system of any of Embodiments 26-28, wherein the second reference source is maintained at a second temperature.
The system of any of Embodiments 26-29, wherein the first temperature is greater than the second temperature.
The system of any of Embodiments 26-30, wherein the first temperature is less than the second temperature.
The system of any of Embodiments 26-31, wherein the first and the second reference sources are displaced away from a conjugate image plane of the plurality of cameras such that the image of the first and the second reference sources captured by the plurality of cameras is blurred.
The system of any of Embodiments 26-32, wherein the first and the second reference sources are positioned at a conjugate image plane of the plurality of cameras.
The system of any of Embodiments 26-33, further comprising:
a first mirror configured to image the first reference onto the plurality of cameras; and
a second mirror configured to image the second reference source onto the plurality of cameras.
The system of any of Embodiments 26-34, further comprising a first and a second temperature-controlled shutter removably positioned to block IR radiation incident on the system from reaching the plurality of cameras.
The system of any of Embodiments 26-35, wherein the system includes at least two spatially and spectrally different optical channels.
The system of any of Embodiments 26-36, wherein the system includes at least four optical channels.
The system of any of Embodiments 26-37, wherein the system includes at least six optical channels.
The system of any of Embodiments 26-38, wherein the system includes at least eight optical channels.
The system of any of Embodiments 26-39, wherein the system includes at least ten optical channels.
The system of any of Embodiments 26-40, wherein the system includes at least twelve optical channels.
The system of any of Embodiments 26-41, further comprising one or more sensors configured to measure a temperature of the first or the second reference source.
The system of any of Embodiments 26-42, wherein the plurality of cameras is configured to image the same portion of the first reference source and wherein plurality of cameras is configured to image the same portion of the second reference source.
An infrared (IR) imaging system, the imaging system comprising:
a plurality of cameras;
a reference having an unknown temperature configured to be imaged by the plurality of cameras; and
a data-processing unit comprising a processor, the imaging system configured to:
acquire with the plurality of cameras a plurality of frames having regions that correspond to the image of the reference;
calculate a dynamic calibration correction using a temperature measured by one of the cameras in the plurality of cameras as a reference temperature; and
apply the calibration correction to the other cameras in the plurality of cameras to match the temperature estimate of the other cameras in the plurality of cameras with the reference temperature.
The system of any of Embodiment 44, wherein the reference is displaced away from a conjugate image plane of the plurality of cameras such that the image of the reference source captured by the plurality of cameras is blurred.
The system of any of Embodiments 44-45, wherein the reference is positioned at a conjugate image plane of the plurality of cameras.
The system of any of Embodiments 44-46, further comprising a plurality of mirrors configured to image the reference onto the plurality of cameras.
The system of any of Embodiments 44-47, further comprising a first and a second temperature-controlled shutter removably positioned to block IR radiation incident on the system from reaching the plurality of cameras.
The system of any of Embodiments 44-48, wherein the system includes at least two spatially and spectrally different optical channels.
The system of any of Embodiments 44-49, wherein the system includes at least three optical channels.
The system of any of Embodiments 44-50, wherein the system includes at least four optical channels.
The system of any of Embodiments 44-51, wherein the system includes at least five optical channels.
The system of any of Embodiments 44-52, wherein the system includes at least six optical channels.
The system of any of Embodiments 44-52, wherein the system includes at least seven optical channels.
The system of any of Embodiments 44-53, wherein the system includes at least eight optical channels.
The system of any of Embodiments 44-54, wherein the system includes at least nine optical channels.
The system of any of Embodiments 44-55, wherein the system includes at least ten optical channels.
The system of any of Embodiments 44-56, wherein the system includes at least twelve optical channels.
The system of any of Embodiments 44-57, wherein the plurality of cameras is configured to image the same portion of the reference.
An infrared (IR) imaging system, the imaging system comprising:
an optical system including an optical focal plane array (FPA) unit, the optical system includes components associated with at least two optical channels, said at least two optical channels being spatially and spectrally different from one another, each of the at least two optical channels positioned to transfer IR radiation incident on the optical system towards the optical FPA unit, the optical FPA unit comprising at least two detector arrays disposed at a distance from two corresponding focusing lenses;
at least one thermal reference having a known temperature, wherein radiation emitted from the at least one reference is directed towards the optical FPA unit and imaged by the at least two detector arrays; and
a data-processing unit, said data-processing unit configured to:
acquire a plurality of frames with the at least two detector arrays having regions in the plurality of image frames that correspond to the image of the reference; and
dynamically calibrate the at least two detector arrays to address a difference in the temperature estimate of the reference between the two detector arrays.
The system of any of Embodiment 60, wherein the at least one thermal reference has a known spectrum.
The system of any of Embodiments 60-61, further comprising an additional thermal reference, wherein radiation from the additional reference is directed towards the optical FPA unit and imaged by the at least two detector arrays.
The system of any of Embodiments 60-62, wherein the additional reference has a temperature and a spectrum different from the known temperature and the known spectrum of the at least one reference source.
The system of any of Embodiments 60-63, wherein the temperature of the additional thermal reference is less than the known temperature.
The system of any of Embodiments 60-64, wherein the temperature of the additional thermal reference is greater than the known temperature.
The system of any of Embodiments 60-65, wherein the at least one reference is displaced away from a conjugate image plane of the at least two detector arrays such that the image of the at least one reference captured by the at least two detector arrays is defocused.
The system of any of Embodiments 60-66, wherein the at least one reference is positioned at a conjugate image plane of the at least two detector arrays such that the image of the at least one reference captured by the at least two detector arrays is focused.
The system of any of Embodiments 60-67, further comprising at least two reflecting elements configured to direct radiation from the at least one reference source toward the at least two detector arrays.
The system of any of Embodiments 60-68, wherein the at least two reflecting elements are disposed outside a central field of view of the at least two detector arrays.
The system of any of Embodiments 60-69, further comprising a third detector array disposed between the at least two detector arrays.
The system of any of Embodiments 60-70, wherein the data-processing unit is configured to:
acquire a plurality of frames using the third detector array; and
dynamically calibrate the third detector array to match a temperature estimate of the third detector array with the temperature estimates of the at least two detector arrays.
The system of any of Embodiments 60-71, wherein radiation emitted from the at least one reference source is not imaged by the third detector array.
The system of any of Embodiments 60-72, wherein the third detector array has a field of view greater than a field of view of the at least two detector arrays.
The system of any of Embodiments 60-73, wherein the at least one reference is imaged by the third detector array.
The system of any of Embodiments 60-74, further comprising a third reflecting element disposed outside the field of view of the at least two detector arrays and configured to image the at least one reference onto the third detector array.
The system of any of Embodiments 60-75, further comprising a first and a second temperature-controlled element removably positioned to block IR radiation incident on the optical system from reaching the optical FPA unit.
The system of any of Embodiments 60-76, wherein the optical system includes components associated with three optical channels.
The system of any of Embodiments 60-77 wherein the optical system includes components associated with four optical channels.
The system of any of Embodiments 60-78, wherein the optical system includes components associated with six optical channels.
The system of any of Embodiments 60-79, wherein the optical system includes components associated with eight optical channels.
The system of any of Embodiments 60-80, wherein the optical system includes components associated with ten optical channels.
The system of any of Embodiments 60-81, wherein the optical system includes components associated with twelve optical channels.
The system of any of Embodiments 60-82, wherein the optical system includes components associated with sixteen optical channels.
The system of any of Embodiments 60-83, wherein the optical system includes components associated with twenty four optical channels.
The system of any of Embodiments 60-84, wherein each of the at least two detector arrays is configured to image the same portion of the at least one reference so as to consistently provide a common reference temperature.
The system of any of Embodiments 60-85, wherein the data-processing unit comprises processing electronics.
The system of any of Embodiments 60-86, wherein the data-processing unit comprises a processor.
The system of any of Embodiments 60-87, wherein the data-processing unit comprises one or more processors.
An infrared (IR) imaging system, the imaging system comprising:
an optical system including an optical focal plane array (FPA) unit, the optical system including at least two optical channels, said at least two optical channels being spatially and spectrally different from one another, each of the at least two optical channels positioned to transfer IR radiation incident on the optical system towards the optical FPA unit, the optical FPA unit comprising at least two detector arrays disposed at a distance from two corresponding focusing lenses;
a first temperature-controlled reference imaged by the at least two detector arrays;
a second temperature-controlled reference imaged by the at least two detector arrays; and
a data-processing unit configured to:
acquire a plurality of frames with the at least two detector arrays having regions in the plurality of image frames that correspond to the image of the first and second references; and
dynamically calibrate the at least two detector arrays to address a difference in a temperature estimate of the first and second references between the two detector arrays.
The system of Embodiment 89, wherein the first reference is maintained at a first temperature.
The system of any of Embodiments 89-90, wherein the second reference is maintained at a second temperature.
The system of any of Embodiments 89-91, wherein the first temperature is greater than the second temperature.
The system of any of Embodiments 89-92, wherein the first temperature is less than the second temperature.
The system of any of Embodiments 89-93, wherein the first and the second references are displaced away from a conjugate image plane of the at least two detector arrays such that the image of the first and the second references captured by the at least two detector arrays is defocused.
The system of any of Embodiments 89-94, wherein the first and the second references are positioned at a conjugate image plane of the at least two detector arrays such that the image of the first and the second references captured by the at least two detector arrays is focused.
The system of any of Embodiments 89-95, further comprising:
a first reflecting element configured to direct radiation from the first reference toward the at least two detector arrays; and
a second reflecting element configured to direct radiation from the second reference toward the at least two detector arrays.
The system of any of Embodiments 89-96, wherein the first reflecting element is disposed outside a field of view of the at least two detector arrays.
The system of any of Embodiments 89-97, wherein the second reflecting element is disposed outside a field of view of the at least two detector arrays.
The system of any of Embodiments 89-98, further comprising a third detector array disposed between the at least two detector arrays.
The system of any of Embodiments 89-99, wherein the data-processing unit is configured to:
acquire a plurality of frames using the third detector array; and
dynamically calibrate the third detector array to address a difference in the temperature estimate of the first and second references between the third detector array and the two detector arrays.
The system of any of Embodiments 89-100, wherein the first and second references are not imaged by the third detector array.
The system of any of Embodiments 89-101, wherein the third detector array has a field of view greater than a field of view of the at least two detector arrays.
The system of any of Embodiments 89-102, wherein r first and second references are imaged by the third detector array.
The system of any of Embodiments 89-103, further comprising a third reflecting element disposed outside the field of view of the at least two detector arrays and configured to image the first and second references onto the third detector array.
The system of any of Embodiments 89-104, further comprising a first and a second temperature-controlled element removably positioned to block IR radiation incident on the optical system from reaching the optical FPA unit.
The system of any of Embodiments 89-105, wherein the optical system includes components associated with three optical channels.
The system of any of Embodiments 89-106, wherein the optical system includes components associated with four optical channels.
The system of any of Embodiments 89-107, wherein the optical system includes components associated with five optical channels.
The system of any of Embodiments 89-108, wherein the optical system includes components associated with six optical channels.
The system of any of Embodiments 89-109, wherein the optical system includes components associated with seven optical channels.
The system of any of Embodiments 89-110, wherein the optical system includes components associated with eight optical channels.
The system of any of Embodiments 89-111, wherein the optical system includes components associated with ten optical channels.
The system of any of Embodiments 89-112, wherein the optical system includes components associated with twelve optical channels.
The system of any of Embodiments 89-113, wherein each of the at least two detector arrays is configured to image the same portion of the first reference source so as to consistently provide a common first reference temperature and wherein each of the at least two detector arrays is configured to image the same portion of the second reference source so as to consistently provide a common second reference temperature.
The system of any of Embodiments 89-114, further comprising a temperature controller configured to control the temperature of the first or second reference.
The system of any of Embodiments 89-115, further comprising one or more sensors configured to measure a temperature of the first or the second reference.
The system of any of Embodiments 89-116, wherein the one or more sensors are configured to communicate the measured temperature of the first or the second reference to a temperature controller.
The system of any of Embodiments 89-117, wherein the one or more sensors are configured to communicate the measured temperature of the first or the second reference to the data-processing unit.
The system of any of Embodiments 89-118, wherein the first or the second reference is associated with a heater.
The system of any of Embodiments 89-119, wherein the first or the second reference is associated with a cooler.
The system of any of Embodiments 89-120, wherein the data-processing unit comprises processing electronics.
The system of any of Embodiments 89-121, wherein the data-processing unit comprises a processor.
The system of any of Embodiments 89-122, wherein the data-processing unit comprises one or more processors.
An infrared (IR) imaging system, the imaging system comprising:
an optical system including components associated with at least two optical channels, said at least two optical channels being spectrally different from one another, each of the at least two optical channels positioned to transfer IR radiation incident on the optical system towards a plurality of cameras;
at least one calibration surface with unknown temperature and imaged by each of the plurality of cameras; and
a data-processing unit configured to:
acquire a plurality of image frames with the plurality of cameras including the imaged surface; and
adjust one or more parameters of the cameras in the plurality of cameras such that a temperature estimate of the calibration surface of the cameras in the plurality of cameras agree with each other.
The system of Embodiment 124, wherein the one or more parameters is associated with a gain of the cameras in the plurality of cameras.
The system of any of Embodiments 124-125, wherein the one or more parameters is associated with a gain offset of the cameras in the plurality of cameras.
The system of any of Embodiments 124-126, wherein the calibration surface is displaced away from a conjugate image plane of the plurality of cameras such that an image of the surface is defocused.
The system of any of Embodiments 124-127, wherein the calibration surface is positioned at a conjugate image plane of the plurality of cameras such that an image of the surface is focused.
The system of any of Embodiments 124-128, further comprising at least one reflecting element configured to image the surface onto the plurality of cameras.
The system of any of Embodiments 124-129, further comprising a first and a second temperature-controlled element removably positioned to block IR radiation incident on the optical system from reaching the plurality of cameras.
The system of any of Embodiments 124-130, wherein the optical system includes components associated with three optical channels.
The system of any of Embodiments 124-131, wherein the optical system includes components associated with four optical channels.
The system of any of Embodiments 124-132, wherein the optical system includes components associated with six optical channels.
The system of any of Embodiments 124-133, wherein the optical system includes components associated with eight optical channels.
The system of any of Embodiments 124-134, wherein the optical system includes components associated with ten optical channels.
The system of any of Embodiments 124-135, wherein the optical system includes components associated with twelve optical channels.
The system of any of Embodiments 124-136, wherein the data-processing unit comprises processing electronics.
The system of any of Embodiments 124-137, wherein the data-processing unit comprises a processor.
The system of any of Embodiments 124-138, wherein the data-processing unit comprises one or more processors.
The system of any of Embodiments 124-139, wherein the calibration surface comprises a sidewall of a housing of the system.
An infrared (IR) imaging system, the imaging system comprising:
at least four spatially and spectrally different optical channels configured to receive IR radiation from a common object, each of the at least four spatially and spectrally different optical channels comprising at least one imaging lens configured to image the object on a Focal Plane Array (FPA) unit; and
processing electronics in communication with the FPA unit,
wherein said infrared system is configured to:
acquire multispectral optical data from the at least four different optical channels; and
process the multispectral optical data to detect one or more target species present in the object.
The system of Embodiment 141, further comprising twelve optical channels.
The system of any of Embodiments 141-142, configured to simultaneously acquire multispectral optical data from the at least four different optical channels.
The system of any of Embodiments 141-143, further comprising a plurality of optical filters associated with the optical channels.
The system of any of Embodiments 141-144, wherein a number of optical filters is two.
The system of any of Embodiments 141-145, wherein a number of optical filters is three.
The system of any of Embodiments 141-146, wherein a number of optical filters is four.
The system of any of Embodiments 141-147, wherein the plurality of optical filters comprise at least one long pass (LP) filter.
The system of any of Embodiments 141-148, wherein the plurality of optical filters comprise multiple long pass (LP) filters.
The system of any of Embodiments 141-149, wherein the plurality of optical filters comprise at least one short pass (SP) filter.
The system of any of Embodiments 141-150, wherein the plurality of optical filters comprise multiple short pass (SP) filters.
The system of any of Embodiments 141-151, wherein the plurality of optical filters comprise at least one band pass (BP) filter.
The system of any of Embodiments 141-152, wherein the plurality of optical filters comprise at one short pass (SP) filter and one long pass (LP) filter.
The system of any of Embodiments 141-153, wherein the FPA unit comprises a plurality of FPAs.
The system of any of Embodiments 141-154, further comprising first and second temperature-controlled elements removably positioned to block IR radiation incident on the imaging system from reaching the FPA unit.
The system of any of Embodiments 141-155, further comprising a field reference configured for dynamically calibrating a plurality of the FPAs in the FPA unit.
The system of any of Embodiments 141-156, wherein the field reference is configured to obscure a peripheral region of an image generated by a plurality of the FPAs in the FPA unit.
The system of any of Embodiments 141-157, configured to compare spectral data in at least one of the four optical channels acquired at a first instant of time with spectral data in the at least one of the four optical channels acquired at a second instant of time to generate a temporal difference image.
The system of any of Embodiments 141-158, configured to use a difference between the multispectral optical data acquired by the two optical channels to correct parallax-induced imaging errors.
The system of any of Embodiments 141-159, configured to use a difference between the multispectral optical data acquired by the two optical channels to estimate a distance between the system and the object.
The system of any of Embodiments 141-160, configured to estimate a size of the object based on the estimated distance and an optical magnification factor of the two optical channels.
The system of any of Embodiments 141-161, configured to compare spectral data in one of the at least four optical channels with spectral data in another one of the at least four optical channels to generate a spectral difference image.
The system of any of Embodiments 141-162, further comprising a visible light imaging sensor.
The system of any of Embodiments 141-163, configured to use the visible light imaging sensor to compensate for motion-induced imaging errors.
The system of any of Embodiments 141-164, configured to process the multispectral optical data by cross-correlating multispectral optical data from at least one of the optical channels with a reference spectrum.
The system of any of Embodiments 141-165, configured to process the multispectral optical data using spectral unmixing.
The system of any of Embodiments 141-166, further comprising five optical channels.
The system of any of Embodiments 141-167, further comprising six optical channels.
The system of any of Embodiments 141-168, further comprising seven optical channels.
The system of any of Embodiments 141-169, further comprising eight optical channels.
The system of any of Embodiments 141-170, further comprising nine optical channels.
The system of any of Embodiments 141-171, further comprising ten optical channels.
The system of any of Embodiments 141-172, further comprising eleven optical channels.
The system of any of Embodiments 141-173, wherein a number of optical filters is five.
The system of any of Embodiments 141-174, wherein a number of optical filters is six.
The system of any of Embodiments 141-175, wherein a number of optical filters is seven.
The system of any of Embodiments 141-176, wherein a number of optical filters is eight.
The system of any of Embodiments 141-177, wherein a number of optical filters is nine.
The system of any of Embodiments 141-178, wherein a number of optical filters is ten.
The system of any of Embodiments 141-179, wherein a number of optical filters is eleven.
The system of any of Embodiments 141-180, wherein a number of optical filters is twelve.
The system of any of Embodiments 141-181, wherein the processing electronics comprises one or more processors.
The system of any of Embodiments 141-182, further comprising a thermal reference configured to be imaged onto the FPA unit such that a plurality of frames of the acquired multispectral optical data has an image of the thermal reference source.
The system of any of Embodiments 141-183, wherein the thermal reference has a known temperature.
The system of any of Embodiments 141-184, wherein the thermal reference is a temperature-controlled reference source.
The system of any of Embodiments 141-185, wherein the temperature-controlled reference source includes a heater.
The system of any of Embodiments 141-186, wherein the temperature-controlled reference source includes a cooler.
The system of any of Embodiments 141-187, further comprising a mirror configured to image the thermal reference onto the FPA unit.
The system of any of Embodiments 141-188, wherein a temperature of the thermal reference is unknown.
The system of any of Embodiments 141-189, wherein the thermal reference is a surface.
The system of any of Embodiments 141-190, wherein the surface comprises a wall of a housing of the system.
The system of any of Embodiments 141-191, wherein different optical channels receive IR radiation from the same portion of the thermal reference so as to consistently provide a common reference temperature.
The system of any of Embodiments 141-192, wherein a temperature of the same portion of the thermal reference is unknown.
The system of any of Embodiments 141-193, configured to acquire the multispectral optical data at a frame rate between about 5 Hz to about 200 Hz.
The system of any of Embodiments 141-194, wherein one or more of the at least four optical channels is configured to collect IR radiation to provide spectral data corresponding to a discrete spectral band located in the wavelength range between about 7.9 μm and about 8.4 μm.
The system of any of Embodiments 141-195, wherein each optical channel is configured to transfer spectrally distinct, two-dimensional image data of the common object to one or more imaging sensors.
The system of any of Embodiments 1-24, wherein the one or more sensors are configured to communicate the measured temperature of the at least one reference source to a temperature controller.
The system of any of Embodiments 1-24, wherein the one or more sensors are configured to communicate the measured temperature of the at least one reference source to the data-processing unit.
The system of any of Embodiments 26-42, wherein the one or more sensors are configured to communicate measured temperature of the first or the second reference to a temperature controller.
The system of any of Embodiments 26-42, wherein the one or more sensors are configured communicate the measured temperature of the first or the second reference to the data-processing unit.
An infrared (IR) imaging system for imaging a target species in an object, the imaging system comprising:
an optical system comprising an optical focal plane array (FPA) unit and defining a plurality of spatially and spectrally different optical channels to transfer IR radiation towards the optical FPA unit, each optical channel positioned to transfer a portion of the IR radiation incident on the optical system from the object towards the optical FPA unit; and
a programmable processor configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium, to acquire, in a single occurrence of data acquisition, multispectral optical data representing in spatial (x, y) and spectral (λ) coordinates said object and said target species from the IR radiation received at the optical FPA unit.
The system of Embodiment 200, wherein the multispectral data comprises a number of spectrally different images of the object obtained from IR image data transferred to the optical FPA unit by a corresponding optical channel.
The system of any of Embodiments 200-201, further comprising an optical filter corresponding to a particular optical channel and configured to transmit the portion of IR radiation towards the optical FPA unit.
The system of any of Embodiments 200-202, wherein the optical filter includes one of a longpass optical filter and a shortpass optical filter.
The system of any of Embodiments 200-203, further comprising one or more front objective lenses.
The system of any of Embodiments 200-204, wherein the optical system comprises a plurality of lenses, each lens corresponding to an optical channel.
The system of any of Embodiments 200-205, wherein each optical channel is defined at least in part by a corresponding filter and a corresponding lens.
The system of any of Embodiments 200-206, wherein the plurality of lenses comprises a lens array.
The system of any of Embodiments 200-207, further comprising a plurality of relay lenses configured to relay the IR radiation along the optical channels.
The system of any of Embodiments 200-208, further comprising a plurality of moveable temperature-controlled reference source removably positioned to block IR radiation incident onto the optical system from reaching the optical FPA unit.
The system of any of Embodiments 200-209, wherein the multispectral optical data from the plurality of optical channels is captured substantially simultaneously by the optical FPA unit.
The system of any of Embodiments 200-210, wherein the multispectral optical data from the plurality of optical channels is captured during one image frame acquisition by the optical FPA unit.
The system of any of Embodiments 200-211, further comprising first and second temperature-controlled moveable shutters removably positioned to block IR radiation incident onto the optical system from reaching the optical FPA unit.
The system of any of Embodiments 200-212, wherein the optical FPA unit is devoid of a cooling device.
The system of any of Embodiments 200-213, further comprising a filter array.
The system of any of Embodiments 200-214, wherein the processor is configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium to acquire said optical data from the two or more sets of imaging data.
The system of any of Embodiments 200-215, wherein the processor is configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium to process the acquired optical data to compensate for at least one of (i) parallax-induced differences between the two or more sets of imaging data and (ii) difference between the two or more sets of imaging data induced by changes in the object that are not associated with the target species.
The system of any of Embodiments 200-216, wherein the processor is configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium to process the acquired optical data to generate a temporal reference image.
The system of any of Embodiments 200-217, wherein the processor is configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium to use the temporal reference image to generate a temporal difference image.
The system of any of Embodiments 200-218, wherein the processor is configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium to process the acquired optical data to estimate a volume of a gas cloud.
The system of any of Embodiments 200-219, wherein IR radiation measured at a pixel comprises a spectrum comprising a sum of component spectra, and wherein the processor is configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium to unmix the spectrum.
The system of any of Embodiments 200-220, wherein the optical FPA unit includes a bolometer configured to operate without being cooled.
The system of any of Embodiments 200-221, further comprising a field reference for dynamically adjusting data output from the optical FPA unit.
The system of any of Embodiments 200-222, wherein the field reference comprises an array of field stops.
The system of any of Embodiments 200-223, wherein the field reference comprises a uniform temperature across its surface.
The system of any of Embodiments 200-224, wherein the field reference is adapted to obscure or block a peripheral portion of the IR radiation propagating from the object towards the optical FPA unit.
The system of any of Embodiments 200-225, further comprising a visible light imaging sensor.
The system of any of Embodiments 200-226, wherein the processor is configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium to process data received from an imaging sensor to compensate for motion-induced imaging errors.
The system of any of Embodiments 200-227, wherein the processor is configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium to process data received from the visible light imaging sensor to compensate for motion-induced imaging errors.
The system of any of Embodiments 200-228, wherein the processor is configured to execute instructions stored in a tangible, non-transitory computer-readable storage medium to construct, in the single occurrence of data acquisition, a multispectral data cube of the object, the multispectral data cube comprising a number of spectrally different images of the object, each spectrally different image comprising IR image data transferred to the optical FPA unit by a corresponding optical channel.
The system of any of Embodiments 200-229, wherein the portion of the IR radiation corresponds to a region of wavelengths of the spectrum of wavelengths, the region of wavelengths at least partially overlapping another region of wavelengths transferred by another optical channel.
An infrared (IR) imaging system, the imaging system comprising:
a plurality of spatially and spectrally different optical channels, at least some of the plurality of optical channels configured to receive IR radiation from a common object, each of the plurality of spatially and spectrally different optical channels comprising at least one imaging lens configured to image the object on a Focal Plane Array (FPA) unit; and
processing electronics in communication with the FPA unit,
wherein said infrared system is configured to:
acquire multispectral optical data from the plurality of different optical channels; and
process the multispectral optical data to detect one or more target species present in the object, and
wherein the system is further configured to compare spectral data in at least one of the plurality of optical channels acquired at a first instant of time with spectral data in at least one of the plurality of optical channels acquired at a second instant of time to generate a temporal difference image.
An infrared (IR) imaging system, the imaging system comprising:
a plurality of spatially and spectrally different optical channels, at least some of the plurality of optical channels configured to receive IR radiation from a common object, each of the plurality of spatially and spectrally different optical channels comprising at least one imaging lens configured to image the object on a Focal Plane Array (FPA) unit; and
processing electronics in communication with the FPA unit,
wherein said infrared system is configured to:
acquire multispectral optical data from the plurality of different optical channels; and
process the multispectral optical data to detect one or more target species present in the object, and
wherein the system is configured to use a difference between the multispectral optical data acquired by two optical channels to correct parallax-induced imaging errors.
The system of Embodiment 231, wherein spectral characteristics of the two optical channels are identical.
An infrared (IR) imaging system, the imaging system comprising:
a plurality of spatially and spectrally different optical channels, at least some of the plurality of optical channels configured to receive IR radiation from a common object, each of the plurality of spatially and spectrally different optical channels comprising at least one imaging lens configured to image the object on a Focal Plane Array (FPA) unit; and
processing electronics in communication with the FPA unit,
wherein said infrared system is configured to:
acquire multispectral optical data from the plurality of different optical channels; and
process the multispectral optical data to detect one or more target species present in the object, and
wherein the system is configured to use a difference between the multispectral optical data acquired by two optical channels to estimate a distance between the system and the object.
The system of Embodiment 233, wherein spectral characteristics of the two optical channels are identical.
An infrared (IR) imaging system, the imaging system comprising:
a plurality of spatially and spectrally different optical, at least some of the plurality of optical channels configured to receive IR radiation from a common object, each of the plurality of spatially and spectrally different optical channels comprising at least one imaging lens configured to image the object on a Focal Plane Array (FPA) unit;
a visible light imaging sensor; and
processing electronics in communication with the FPA unit,
wherein said infrared system is configured to:
acquire multispectral optical data from the plurality of different optical channels; and
process the multispectral optical data to detect one or more target species present in the object.
An infrared (IR) imaging system, the imaging system comprising:
a plurality of spatially and spectrally different optical, at least some of the plurality of optical channels configured to receive IR radiation from a common object, each of the plurality of spatially and spectrally different optical channels comprising at least one imaging lens configured to image the object on a Focal Plane Array (FPA) unit; and
processing electronics in communication with the FPA unit,
wherein said infrared system is configured to:
acquire multispectral optical data from the plurality of different optical channels; and
process the multispectral optical data to detect one or more target species present in the object, and
wherein the system is configured to compensate for motion-induced imaging errors.
An infrared (IR) imaging system, the imaging system comprising:
a plurality of spatially and spectrally different, at least some of the plurality of optical channels configured to receive IR radiation from a common object, each of the plurality of spatially and spectrally different optical channels comprising at least one imaging lens configured to image the object on a Focal Plane Array (FPA) unit; and
processing electronics in communication with the FPA unit,
wherein said infrared system is configured to:
acquire multispectral optical data from the plurality of different optical channels; and
process the multispectral optical data to detect one or more target species present in the object by cross-correlating multispectral optical data from at least one of the optical channels with a reference spectrum.
An infrared (IR) imaging system, the imaging system comprising:
a plurality of spatially and spectrally different optical, at least some of the plurality of optical channels configured to receive IR radiation from a common object, each of the plurality of spatially and spectrally different optical channels comprising at least one imaging lens configured to image the object on a Focal Plane Array (FPA) unit; and
processing electronics in communication with the FPA unit,
wherein said infrared system is configured to:
acquire multispectral optical data from the plurality of different optical channels; and
process the multispectral optical data to detect one or more target species present in the object by using spectral unmixing.
An infrared (IR) imaging system, the imaging system comprising:
an optical system including an optical focal plane array (FPA) unit, the optical system includes components associated with at least two optical channels, said at least two optical channels being spatially and spectrally different from one another, each of the at least two optical channels positioned to transfer IR radiation incident on the optical system towards the optical FPA unit, the optical FPA unit comprising at least two detector arrays disposed at a distance from two corresponding focusing lenses;
at least one thermal reference having a known temperature, wherein one of the at least two detector arrays is configured to image the at least one reference; and
a data-processing unit, said data-processing unit configured to:
acquire a plurality of frames with one of the at least two detector arrays having regions in the plurality of image frames that correspond to the image of the reference; and
dynamically calibrate another of the at least two detector array to match a temperature estimate of another of the at least two detector array with the temperature estimate of one of the at least two detector array.
The system of any of Embodiments 1-59, wherein the plurality of cameras are configured to acquire multispectral image data from an object continuously for a duration of time.
The system of any of Embodiments 1-59, comprising at least two spectrally and spatially distinct optical channels configured to transfer two-dimensional image data of an object to the plurality of cameras.
An infrared (IR) imaging system for imaging a scene, the imaging system comprising:
an optical system comprising an optical focal plane array (FPA) unit including a plurality of spatially and spectrally different optical channels to transfer IR radiation from the scene towards the optical FPA unit, each optical channel positioned to transfer a portion of the IR radiation incident on the optical system from the scene towards the optical FPA unit,
wherein at least one of the plurality of optical channels is in the mid-wavelength infrared spectral range and at least another one of the plurality of optical channels is in the long-wavelength infrared spectral range,
wherein the imaging system is configured to acquire a first video image of the scene in the mid-wavelength infrared spectral range and a second video image of the scene in the long-wavelength infrared spectral range.
The imaging system of Embodiment 242, wherein the at least one mid-wavelength optical channel comprises a cold stop filter.
The imaging system of any of Embodiments 242-243, wherein the at least one mid-wavelength optical channel comprises filter having a pass-band in the mid-wavelength infrared spectral range.
The imaging system of any of Embodiments 242-244, further comprising processing electronics configured to extract information from the first and second video image to detect presence of one or more chemical species in the scene.
The imaging system of Embodiment 245, wherein the processing electronics are configured to compare the extracted information to one or more spectral information stored in a database.
The imaging system of any of Embodiments 242-246, further comprising a feedback system configured to synchronize the first and the second video image.
The imaging system of any of Embodiments 242-247, wherein the first and/or second video image has a frame rate between about 5 frames per second and about 120 frames per second.
The imaging system of any of Embodiments 242-248, wherein the processing electronics are configured to extract information from the first and/or second video image on a frame by frame basis.
The imaging system of any of Embodiments 242-249, wherein the processing electronics are configured to extract information from the first video and second video image and detect presence of one or more chemical species in the scene within 5 seconds from a time when imaging of the scene started.
The imaging system of any of Embodiments 242-250, wherein the processing electronics are configured to extract information from the first video and second video image and detect presence of one or more chemical species in the scene within 1 second from a time when imaging of the scene started.
The imaging system of any of Embodiments 242-251, wherein the processing electronics are configured to extract information from the first video and second video image and detect presence of one or more chemical species in the scene in sufficiently real time from a time when imaging of the scene started.
The imaging system of any of Embodiments 242-252, wherein the one or more chemical species are selected from the group consisting of methane, cyclopropane, alkanes, alkenes, ammonia, freon, hydrogen cyanide, sulfur dioxide, and hydrogen sulfide.
The imaging system of any of Embodiments 242-253, wherein the scene includes a moving gas plume.
The imaging system of any of Embodiments 242-254, wherein the FPA unit includes at least one mid-wavelength infra-red FPA capable of detecting mid-wavelength infra-red radiation and at least one long-wavelength infra-red FPA capable of detecting long-wavelength infra-red radiation.
The imaging system of Embodiment 255, wherein the at least one mid-wavelength infra-red FPA is cooled to a temperature below 200 degree Kelvin.
The imaging system of Embodiment 256, wherein the at least one mid-wavelength infra-red FPA is cooled to a temperature between about 110 degree Kelvin and about 150 degree Kelvin.
The imaging system of Embodiment 257, wherein the at least one mid-wavelength infra-red FPA is cooled to a temperature of about 135 degree Kelvin.
The imaging system of any of Embodiments 255-258, wherein the at least one mid-wavelength infra-red FPA is coupled to a cooler.
The imaging system of any of Embodiments 255-259, wherein the at least one mid-wavelength infra-red FPA and the at least one long-wavelength infra-red FPA are configured to image one more reference sources.
The imaging system of Embodiment 260, wherein each of the one more reference sources has a known temperature.
The imaging system of any of Embodiments 242-261, wherein a number of the plurality of optical channels is at least 4.
The imaging system of any of Embodiments 242-262, wherein a number of the plurality of optical channels is at least 5.
The imaging system of any of Embodiments 242-263, wherein a number of the plurality of optical channels is at least 8.
The imaging system of any of Embodiments 242-264, wherein a number of the plurality of optical channels is at least 9.
The imaging system of any of Embodiments 242-265, wherein a number of the plurality of optical channels is at least 12.
The imaging system of any of Embodiments 242-266, wherein a number of the plurality of optical channels is at least 13.
The imaging system of any of Embodiments 242-267, wherein a number of the plurality of optical channels is at least 20.
The imaging system of any of Embodiments 242-268, wherein a number of the plurality of optical channels is between 4 and 50.
The imaging system of any of Embodiments 255-261, wherein at least one of the plurality of optical channels comprises a spectrally selective optical element configured to receive radiation from the scene and direct a portion of the radiation in the mid-wavelength infrared spectral range toward the mid-wavelength infra-red FPA and direct a portion of the radiation in the long-wavelength infrared spectral range toward the long-wavelength infra-red FPA.
A method of detecting one or more chemical species in a scene, the method comprising:
obtaining a first video image data of the scene in mid-wavelength infrared spectral range using a mid-wavelength infra-red FPA;
obtaining a second video image data of the scene in long-wavelength infrared spectral range using a long-wavelength infra-red FPA;
obtaining one or more spectra from the first and the second video image data; and
comparing the obtained one or more spectra with one or more reference spectra stored in a database to detect one or more chemical species in the scene.
The method of Embodiment 271, wherein the first and/or the second video image data has a frame rate between about 5 frames per second and about 60 frames per second.
The method of any of Embodiments 270-272, wherein the one or more spectra are obtained using hyperspectral video analytics.
The method of any of Embodiments 270-273, wherein the one or more spectra are obtained within 5 seconds from a time when imaging of the scene begins.
The method of any of Embodiments 270-274, wherein the one or more spectra are obtained within 1 second from a time when imaging of the scene begins.
The method of any of Embodiments 270-275, wherein the one or more spectra are obtained in sufficiently real time from a time when imaging of the scene begins.
The method of any of Embodiments 270-276, wherein the one or more chemical species detected are selected from the group consisting of methane, cyclopropane, alkanes, alkenes, ammonia, freon, hydrogen cyanide sulfur dioxide, and hydrogen sulfide.
The imaging system or method of any of Embodiments 242-277, wherein some of the plurality of spatially and spectrally distinct optical channels have a different field of view than some other of the plurality of spatially and spectrally distinct optical channels.
The imaging system or method of any of Embodiments 242-277, wherein some of the plurality of spatially and spectrally distinct optical channels have a field of view that is lower than the field of view of the system.
The imaging system or method of any of Embodiments 242-279, wherein at least two of the plurality of optical channels is in the long-wavelength infrared spectral range.
The imaging system or method of any of Embodiments 242-279, wherein a number of optical channels in the long-wavelength infrared spectral range is less than 50.
The imaging system or method of any of Embodiments 242-279, wherein at least two of the plurality of optical channels is in the mid-wavelength infrared spectral range.
The imaging system or method of any of Embodiments 242-279, wherein a number of optical channels in the mid-wavelength infrared spectral range is less than 50.
Details of one or more implementations of the subject matter described in this disclosure are set forth in the accompanying drawings and the description below. Other features, aspects, and advantages will become apparent from the description, the drawings and the claims. Note that the relative dimensions of the following figures may not be drawn to scale.
Like reference numbers and designations in the various drawings indicate like elements.
The following description is directed to certain implementations for the purposes of describing the innovative aspects of this disclosure. However, a person having ordinary skill in the art will readily recognize that the teachings herein can be applied in a multitude of different ways. The described implementations may be implemented in any device, apparatus, or system that can be configured to operate as an imaging system such as in an infra-red imaging system. The methods and systems described herein can be included in or associated with a variety of devices such as, but not limited to devices used for visible and infrared spectroscopy, multispectral and hyperspectral imaging devices used in oil and gas exploration, refining, and transportation, agriculture, remote sensing, defense and homeland security, surveillance, astronomy, environmental monitoring, etc. The methods and systems described herein have applications in a variety of fields including but not limited to agriculture, biology, physics, chemistry, defense and homeland security, environment, oil and gas industry, etc. The teachings are not intended to be limited to the implementations depicted solely in the Figures, but instead have wide applicability as will be readily apparent to one having ordinary skill in the art.
Various embodiments disclosed herein describe a divided-aperture infrared spectral imaging (DAISI) system that is structured and adapted to provide identification of target chemical contents of the imaged scene. The system is based on spectrally-resolved imaging and can provide such identification with a single-shot (also referred to as a snapshot) comprising a plurality of images having different wavelength compositions that are obtained generally simultaneously. Without any loss of generality, snapshot refers to a system in which most of the data elements that are collected are continuously viewing the light emitted from the scene. In contrast in scanning systems, at any given time only a minority of data elements are continuously viewing a scene, followed by a different set of data elements, and so on, until the full dataset is collected. Relatively fast operation can be achieved in a snapshot system because it does not need to use spectral or spatial scanning for the acquisition of infrared (IR) spectral signatures of the target chemical contents. Instead, IR detectors (such as, for example, infrared focal plane arrays or FPAs) associated with a plurality of different optical channels having different wavelength profiles can be used to form a spectral cube of imaging data. Although spectral data can be obtained from a single snapshot comprising multiple simultaneously acquired images corresponding to different wavelength ranges, in various embodiments, multiple snap shots may be obtained. In various embodiments, these multiple snapshots can be averaged. Similarly, in certain embodiments multiple snap shots may be obtained and a portion of these can be selected and possibly averaged. Also, in contrast to commonly used IR spectral imaging systems, the DAISI system does not require cooling. Accordingly, it can advantageously use uncooled infrared detectors. For example, in various implementations, the imaging systems disclosed herein do not include detectors configured to be cooled to a temperature below 300 Kelvin. As another example, in various implementations, the imaging systems disclosed herein do not include detectors configured to be cooled to a temperature below 273 Kelvin. As yet another example, in various implementations, the imaging systems disclosed herein do not include detectors configured to be cooled to a temperature below 250 Kelvin. As another example, in various implementations, the imaging systems disclosed herein do not include detectors configured to be cooled to a temperature below 200 Kelvin.
Implementations disclosed herein provide several advantages over existing IR spectral imaging systems, most if not all of which may require FPAs that are highly sensitive and cooled in order to compensate, during the optical detection, for the reduction of the photon flux caused by spectrum-scanning operation. The highly sensitive and cooled FPA systems are expensive and require a great deal of maintenance. Since various embodiments disclosed herein are configured to operate in single-shot acquisition mode without spatial and/or spectral scanning, the instrument can receive photons from a plurality of points (e.g., every point) of the object substantially simultaneously, during the single reading. Accordingly, the embodiments of imaging system described herein can collect a substantially greater amount of optical power from the imaged scene (for example, an order of magnitude more photons) at any given moment in time especially in comparison with spatial and/or spectral scanning systems. Consequently, various embodiments of the imaging systems disclosed herein can be operated using uncooled detectors (for example, FPA unit including an array of microbolometers) that are less sensitive to photons in the IR but are well fit for continuous monitoring applications. For example, in various implementations, the imaging systems disclosed herein do not include detectors configured to be cooled to a temperature below 300 Kelvin. As another example, in various implementations, the imaging systems disclosed herein do not include detectors configured to be cooled to a temperature below 273 Kelvin. As yet another example, in various implementations, the imaging systems disclosed herein do not include detectors configured to be cooled to a temperature below 250 Kelvin. As another example, in various implementations, the imaging systems disclosed herein do not include detectors configured to be cooled to a temperature below 200 Kelvin. Imaging systems including uncooled detectors can be capable of operating in extreme weather conditions, require less power, are capable of operation during day and night, and are less expensive. Some embodiments described herein can also be less susceptible to motion artifacts in comparison with spatially and/or spectrally scanning systems which can cause errors in either the spectral data, spatial data, or both.
An aperture of the system 100 associated with the front objective lens system 124 is spatially and spectrally divided by the combination of the array of optical filters 130 and the array of imaging lenses 128. In various embodiments, the combination of the array of optical filters 130 and the array of imaging lenses 128 can be considered to form a spectrally divided pupil that is disposed forward of the optical detector array 136. The spatial and spectral division of the aperture into distinct aperture portions forms a plurality of optical channels 120 along which light propagates. In various embodiments, the array 128 of re-imaging lenses 128a and the array of spectral filters 130 which respectively correspond to the distinct optical channels 120. The plurality of optical channels 120 can be spatially and/or spectrally distinct. The plurality of optical channels 120 can be formed in the object space and/or image space. In one implementation, the distinct channels 120 may include optical channels that are separated angularly in space. The array of spectral filters 130 may additionally include a filter-holding aperture mask (comprising, for example, IR light-blocking materials such as ceramic, metal, or plastic). Light from the object 110 (for example a cloud of gas), the optical properties of which in the IR are described by a unique absorption, reflection and/or emission spectrum, is received by the aperture of the system 100. This light propagates through each of the plurality of optical channels 120 and is further imaged onto the optical detector array 136. In various implementations, the detector array 136 can include at least one FPA. In various embodiments, each of the re-imaging lenses 128a can be spatially aligned with a respectively-corresponding spectral region. In the illustrated implementation, each filter element from the array of spectral filters 130 corresponds to a different spectral region. Each re-imaging lens 128a and the corresponding filter element of the array of spectral filter 130 can coincide with (or form) a portion of the divided aperture and therefore with respectively-corresponding spatial channel 120. Accordingly, in various embodiment an imaging lens 128a and a corresponding spectral filter can be disposed in the optical path of one of the plurality of optical channels 120. Radiation from the object 110 propagating through each of the plurality of optical channels 120 travels along the optical path of each re-imaging lens 128a and the corresponding filter element of the array of spectral filter 130 and is incident on the detector array (e.g., FPA component) 136 to form a single image (e.g., sub-image) of the object 110. The image formed by the detector array 136 generally includes a plurality of sub-images formed by each of the optical channels 120. Each of the plurality of sub-images can provide different spatial and spectral information of the object 110. The different spatial information results from some parallax because of the different spatial locations of the smaller apertures of the divided aperture. In various embodiments, adjacent sub-images can be characterized by close or substantially equal spectral signatures. The detector array (e.g., FPA component) 136 is further operably connected with a processor 150 (not shown). The processor 150 can be programmed to aggregate the data acquired with the system 100 into a spectral data cube. The data cube represents, in spatial (x, y) and spectral (X) coordinates, an overall spectral image of the object 110 within the spectral region defined by the combination of the filter elements in the array of spectral filters 130. Additionally, in various embodiments, the processor or processing electronics 150 may be programmed to determine the unique absorption characteristic of the object 110. Also, the processor/processing electronics 150 can, alternatively or in addition, map the overall image data cube into a cube of data representing, for example, spatial distribution of concentrations, c, of targeted chemical components within the field of view associated with the object 110.
Various implementations of the embodiment 100 can include an optional moveable temperature-controlled reference source 160 including, for example, a shutter system comprising one or more reference shutters maintained at different temperatures. The reference source 160 can include a heater, a cooler or a temperature-controlled element configured to maintain the reference source 160 at a desired temperature. For example, in various implementations, the embodiment 100 can include two reference shutters maintained at different temperatures. The reference source 160 is removably and, in one implementation, periodically inserted into an optical path of light traversing the system 100 from the object 110 to the detector array (e.g., FPA component) 136 along at least one of the channels 120. The removable reference source 160 thus can block such optical path. Moreover, this reference source 160 can provide a reference IR spectrum to recalibrate various components including the detector array 136 of the system 100 in real time. The configuration of the moveable reference source 160 is further discussed below.
In the embodiment 100, the front objective lens system 124 is shown to include a single front objective lens positioned to establish a common field-of-view (FOV) for the imaging lenses 128a and to define an aperture stop for the whole system. In this specific case, the aperture stop substantially spatially coincides with and/or is about the same size as or slightly larger than, the plurality of smaller limiting apertures corresponding to different optical channels 120. As a result, the positions for spectral filters of the different optical channels 120 coincide with the position of the aperture stop of the whole system, which in this example is shown as a surface between the lens system 124 and the array 128 of the imaging lenses 128a. In various implementations, the lens system 124 can be an objective lens 124. However, the objective lens 124 is optional and various embodiments of the system 100 need not include the objective lens 124. In various embodiments, the objective lens 124 can slightly shift the images obtained by the different detectors in the array 136 spatially along a direction perpendicular to optical axis of the lens 124, thus the functionality of the system 100 is not necessarily compromised when the objective lens 124 is not included. Generally, however, the field apertures corresponding to different optical channels may be located in the same or different planes. These field apertures may be defined by the aperture of the imaging lens 128a and/or filters in the divided aperture 130 in certain implementations. In one implementation, the field apertures corresponding to different optical channels can be located in different planes and the different planes can be optical conjugates of one another. Similarly, while all of the filter elements in the array of spectral filters 130 of the embodiment 100 are shown to lie in one plane, generally different filter elements of the array of spectral filter 130 can be disposed in different planes. For example, different filter elements of the array of spectral filters 130 can be disposed in different planes that are optically conjugate to one another. However, in other embodiments, the different filter elements can be disposed in non-conjugate planes.
In contrast to the embodiment 100, the front objective lens 124 need not be a single optical element, but instead can include a plurality of lenses 224 as shown in an embodiment 200 of the DAISI imaging system in
In one implementation, the front objective lens system such as the array of lenses 224 is configured as an array of lenses integrated or molded in association with a monolithic substrate. Such an arrangement can reduce the costs and complexity otherwise accompanying the optical adjustment of individual lenses within the system. An individual lens 224 can optionally include a lens with varying magnification. As one example, a pair of thin and large diameter Alvarez plates can be used in at least a portion of the front objective lens system. Without any loss of generality, the Alvarez plates can produce a change in focal length when translated orthogonally with respect to the optical beam.
In further reference to
The embodiment 400 has several operational advantages. It is configured to provide a spectrally known object within every image (e.g., sub-image) and for every snapshot acquisition which can be calibrated against. Such spectral certainty can be advantageous when using an array of IR FPAs like microbolometers, the detection characteristics of which can change from one imaging frame to the next due to, in part, changes in the scene being imaged as well as the thermal effects caused by neighboring FPAs. In various embodiments, the field reference array 438 of the embodiment 400 can be disposed within the Rayleigh range (approximately corresponding to the depth of focus) associated with the front objective lenses 424, thereby removing unusable blurred pixels due to having the field reference outside of this range. Additionally, the embodiment 400 of
In various embodiments, the multi-optical FPA unit of the IR imaging system can additionally include an FPA configured to operate in a visible portion of the spectrum. In reference to
Optical Filters.
The optical filters, used with an embodiment of the system, that define spectrally-distinct IR image (e.g., sub-image) of the object can employ absorption filters, interference filters, and Fabry-Perot etalon based filters, to name just a few. When interference filters are used, the image acquisition through an individual imaging channel defined by an individual re-imaging lens (such as a lens 128a of
The optical filtering configuration of various embodiments disclosed herein may advantageously use a bandpass filter defining a specified spectral band. Any of the filters 0a through 3a, the transmission curves of which are shown in
Referring again to
In one implementation, the LP and SP filters can be combined, in a spectrally-multiplexed fashion, in order to increase or maximize the spectral extent of the transmission region of the filter system of the embodiment.
The advantage of using spectrally multiplexed filters is appreciated based on the following derivation, in which a system of M filters is examined (although it is understood that in practice an embodiment of the invention can employ any number of filters). As an illustrative example, the case of M=7 is considered. Analysis presented below relates to one spatial location in each of the images (e.g., sub-images) formed by the differing imaging channels (e.g., different optical channels 120) in the system. A similar analysis can be performed for each point at an image (e.g., sub-image), and thus the analysis can be appropriately extended as required.
The unknown amount of light within each of the M spectral channels (corresponding to these M filters) is denoted with f1, f2, f3, . . . fM, and readings from corresponding detector elements receiving light transmitted by each filter is denoted as g1, g2, g3 . . . gM, while measurement errors are represented by n1, n2, n3, . . . nM. Then, the readings at the seven FPA pixels each of which is optically filtered by a corresponding band-pass filter of
g1=f1+n1,
g2=f2+n2,
g3=f3+n3,
g4=f4+n4,
g5=f5+n5,
g6=f6+n6,
g7=f7+n7,
These readings (pixel measurements) gi are estimates of the spectral intensities fi. The estimates gi are not equal to the corresponding fi values because of the measurement errors ni. However, if the measurement noise distribution has zero mean, then the ensemble mean of each individual measurement can be considered to be equal to the true value, i.e. (gi)=fi. Here, the angle brackets indicate the operation of calculating the ensemble mean of a stochastic variable. The variance of the measurement can, therefore, be represented as:
((gi−fi)2)=(ni2)=σ2
In embodiments utilizing spectrally-multiplexed filters, in comparison with the embodiments utilizing band-pass filters, the amount of radiant energy transmitted by each of the spectrally-multiplexed LP or SP filters towards a given detector element can exceed that transmitted through a spectral band of a band-pass filter. In this case, the intensities of light corresponding to the independent spectral bands can be reconstructed by computational means. Such embodiments can be referred to as a “multiplex design”.
One matrix of such “multiplexed filter” measurements includes a Hadamard matrix requiring “negative” filters that may not be necessarily appropriate for the optical embodiments disclosed herein. An S-matrix approach (which is restricted to having a number of filters equal to an integer that is multiple of four minus one) or a row-doubled Hadamard matrix (requiring a number of filters to be equal to an integer multiple of eight) can be used in various embodiments. Here, possible numbers of filters using an S-matrix setup are 3, 7, 11, etc and, if a row-doubled Hadamard matrix setup is used, then the possible number of filters is 8, 16, 24, etc. For example, the goal of the measurement may be to measure seven spectral band f intensities using seven measurements gi as follows:
Optical transmission characteristics of the filters described above are depicted in
These {circumflex over (f)}i are unbiased estimates when the ni are zero mean stochastic variables, so that ({circumflex over (f)}i−fi)=0. The measurement variance corresponding to ith measurement is given by the equation below:
From the above equation, it is observed that by employing spectrally-multiplexed system the signal-to-noise ratio (SNR) of a measurement is improved by a factor of √16/7=1.51.
For N channels, the SNR improvement achieved with a spectrally-multiplexed system can be expressed as (N+1)/(2√{square root over (N)}). For example, an embodiment employing 12 spectral channels (N=12) is characterized by a SNR improvement, over a non-spectrally-multiplexed system, comprising a factor of up to 1.88.
Two additional examples of related spectrally-multiplexed filter arrangements 0c through 3c and 0d through 3d that can be used in various embodiments of the imaging systems described herein are shown in
The use of microbolometers, as detector-noise-limited devices, in turn not only can benefit from the use of spectrally multiplexed filters, but also does not require cooling of the imaging system during normal operation. In contrast to imaging systems that include highly sensitive FPA units with reduced noise characteristics, the embodiments of imaging systems described herein can employ less sensitive microbolometers without compromising the SNR. This result is at least in part due to the snap-shot/non-scanning mode of operation.
As discussed above, an embodiment may optionally, and in addition to a temperature-controlled reference unit (for example temperature controlled shutters such as shutters 160; 460a, 460b), employ a field reference component (e.g., field reference aperture 338 in
In particular, and in further reference to
Indeed, the process of calibration of an embodiment of the imaging system starts with estimating gain and offset by performing measurements of radiation emanating, independently, from at least two temperature-controlled shutters of known and different radiances. The gain and offset can vary from detector pixel to detector pixel. Specifically, first the response of the detector unit 456 to radiation emanating from one shutter is carried out. For example, the first shutter 460a blocks the FOV of the detectors 456 and the temperature T1 is measured directly and independently with thermistors. Following such initial measurement, the first shutter 460a is removed from the optical path of light traversing the embodiment and another second shutter (for example, 460b) is inserted in its place across the optical axis 226 to prevent the propagation of light through the system. The temperature of the second shutter 460b can be different than the first shutter (WTI). The temperature of the second shutter 460b is also independently measured with thermistors placed in contact with this shutter, and the detector response to radiation emanating from the shutter 460b is also recorded. Denoting operational response of FPA pixels (expressed in digital numbers, or “counts”) as gi to a source of radiance Li, the readings corresponding to the measurements of the two shutters can be expressed as:
g1=γL1(T1)+goffset
g2=γL2(T2)+goffset
Here, goffset is the pixel offset value (in units of counts), and γ is the pixel gain value (in units of counts per radiance unit). The solutions of these two equations with respect to the two unknowns goffset and γ can be obtained if the values of g1 and g2 and the radiance values L1 and L2 are available. These values can, for example, be either measured by a reference instrument or calculated from the known temperatures T1 and T2 together with the known spectral response of the optical system and FPA. For any subsequent measurement, one can then invert the equation(s) above in order to estimate the radiance value of the object from the detector measurement, and this can be done for each pixel in each FPA within the system.
As already discussed, and in reference to
Because each FPA's offset value is generally adjusted from each frame to the next frame by the hardware, comparing the outputs of one FPA with another can have an error that is not compensated for by the static calibration parameters goffset and γ established, for example, by the movable shutters 160, 460a, 460b. In order to ensure that FPAs operate in radiometric agreement over time, it is advantageous for a portion of each detector array to view a reference source (such as the field reference 338 in
Among the FPA elements in an array of FPAs in an embodiment of the imaging system, one FPA can be selected to be the “reference FPA”. The field reference temperature measured by all the other FPAs can be adjusted to agree with the field reference temperature measured by the reference as discussed below. The image obtained by each FPA includes a set of pixels obscured by the field reference 338. Using the previously obtained calibration parameters goffset and γ (the pixel offset and gain), the effective blackbody temperature Ti of the field reference as measured by each FPA is estimated using the equation below:
Ti=mean{(g+Δgi+goffset/γ}=mean{(g−goffset)/γ}+ΔT,
Using the equation above, the mean value over all pixels that are obscured by the field reference is obtained. In the above equation Δgi is the difference in offset value of the current frame from Δgoffset obtained during the calibration step. For the reference FPA, Δgi can be simply set to zero. Then, using the temperature differences measured by each FPA, one obtains
Ti−Tref=mean{(g+Δgi+goffset/γ}+ΔTi−mean{(g−goffset)/γ}=ΔTi
Once ΔTi for each FPA is measured, its value can be subtracted from each image in order to force operational agreement between such FPA and the reference FPA. While the calibration procedure has been discussed above in reference to calibration of temperature, a procedurally similar methodology of calibration with respect to radiance value can also be implemented.
Examples of Methodology of Measurements.
Prior to optical data acquisition using an embodiment of the IR imaging system as described herein, one or more, most, or potentially all the FPAs of the system can be calibrated. For example, greater than 50%, 60%, 70%, 80% or 90% of the FPAs 336 can be initially calibrated. As shown in
To achieve at least some of these goals, a spectral differencing method may be employed. The method involves forming a difference image from various combinations of the images from different channels. In particular, the images used to form difference images can be registered by two or more different FPAs in spectrally distinct channels having different spectral filters with different spectral characteristics. Images from different channels having different spectral characteristics will provide different spectral information. Comparing (e.g., subtracting) these images, can therefore yield valuable spectral based information. For example, if the filter element of the array of spectral filters 130 corresponding to a particular FPA 336 transmits light from the object 110 including a cloud of gas, for example, with a certain spectrum that contains the gas absorption peak or a gas emission peak while another filter element of the array of spectral filters 130 corresponding to another FPA 336 does not transmit such spectrum, then the difference between the images formed by the two FPAs at issue will highlight the presence of gas in the difference image.
A shortcoming of the spectral differencing method is that contributions of some auxiliary features associated with imaging (not just the target species such as gas itself) can also be highlighted in and contribute to the difference image. Such contributing effects include, to name just a few, parallax-induced imaging of edges of the object, influence of magnification differences between the two or more optical channels, and differences in rotational positioning and orientation between the FPAs. While magnification-related errors and FPA-rotation-caused errors can be compensated for by increasing the accuracy of the instrument construction as well as by post-processing of the acquired imaging, parallax is scene-induced and is not so easily correctable. In addition, the spectral differencing method is vulnerable to radiance calibration errors. Specifically, if one FPA registers radiance of light from a given feature of the object as having a temperature of 40° C., for example, while the data from another FPA represents the temperature of the same object feature as being 39° C., then such feature of the object will be enhanced or highlighted in the difference image (formed at least in part based on the images provided by these two FPAs) due to such radiance-calibration error.
One solution to some of such problems is to compare (e.g., subtract) images from the same FPA obtained at different instances in time. For example, images can be compared to or subtracted from a reference image obtained at another time. Such reference image, which is subtracted from other later obtained images, may be referred to as a temporal reference image. This solution can be applied to spectral difference images as well. For example, the image data resulting from spectral difference images can be normalized by the data corresponding to a temporal reference image. For instance, the temporal reference images can be subtracted from the spectral difference image to obtain the temporal difference image. This process is referred to, for the purposes of this disclosure, as a temporal differencing algorithm or method and the resultant image from subtracting the temporal reference image from another image (such as the spectral difference image) is referred to as the temporal difference image. In some embodiments where spectral differencing is employed, a temporal reference image may be formed, for example, by creating a spectral difference image from the two or more images registered by the two or more FPAs at a single instance in time. This spectral difference image is then used as a temporal reference image. The temporal reference image can then be subtracted from other later obtained images to provide normalization that can be useful in subtracting out or removing various errors or deleterious effects. For example, the result of the algorithm is not affected by a prior knowledge of whether the object or scene contains a target species (such as gas of interest), because the algorithm can highlight changes in the scene characteristics. Thus, a spectral difference image can be calculated from multiple spectral channels as discussed above based on a snap-shot image acquisition at any later time and can be subtracted from the temporal reference image to form a temporal difference image. This temporal difference image is thus a normalized difference image. The difference between the two images (the temporal difference image) can highlight the target species (gas) within the normalized difference image, since this species was not present in the temporal reference frame. In various embodiments, more than two FPAs can be used both for registering the temporal reference image and a later-acquired difference image to obtain a better SNR figure of merit. For example, if two FPAs are associated with spectral filters having the same spectral characteristic, then the images obtained by the two FPAs can be combined after they have been registered to get a better SNR figure.
While the temporal differencing method can be used to reduce or eliminate some of the shortcomings of the spectral differencing, it can introduce unwanted problems of its own. For example, temporal differencing of imaging data is less sensitive to calibration and parallax induced errors than the spectral differencing of imaging data. However, any change in the imaged scene that is not related to the target species of interest (such as particular gas, for example) is highlighted in a temporally-differenced image. Thus such change in the imaged scene may be erroneously perceived as a location of the target species triggering, therefore, an error in detection of target species. For example, if the temperature of the background against which the gas is being detected changes (due to natural cooling down as the day progresses, or increases due to a person or animal or another object passing through the FOV of the IR imaging system), then such temperature change produces a signal difference as compared to the measurement taken earlier in time. Accordingly, the cause of the scenic temperature change (the cooling object, the person walking, etc.) may appear as the detected target species (such as gas). It follows, therefore, that an attempt to compensate for operational differences among the individual FPAs of a multi-FPA IR imaging system with the use of methods that turn on spectral or temporal differencing can cause additional problems leading to false detection of target species. Among these problems are scene-motion-induced detection errors and parallax-caused errors that are not readily correctable and/or compensatable. Accordingly, there is a need to compensate for image data acquisition and processing errors caused by motion of elements within the scene being imaged. Various embodiments of data processing algorithms described herein address and fulfill the need to compensate for such motion-induced and parallax-induced image detection errors.
In particular, to reduce or minimize parallax-induced differences between the images produced with two or more predetermined FPAs, another difference image can be used that is formed from the images of at least two different FPAs to estimate parallax effects. Parallax error can be determined by comparing the images from two different FPAs where the position between the FPAs is known. The parallax can be calculated from the known relative position difference. Differences between the images from these two FPAs can be attributed to parallax, especially, if the FPA have the same spectral characteristics, for example have the same spectral filter or both have no spectral filters. Parallax error correction, however, can still be obtained from two FPAs that have different spectral characteristics or spectral filters, especially if the different spectral characteristics, e.g., the transmission spectra of the respective filters are known and/or negligible. Use of more than two FPAs or FPAs of different locations such as FPAs spaced farther apart can be useful. For example, when the spectral differencing of the image data is performed with the use of the difference between the images collected by the outermost two cameras in the array (such as, for example, the FPAs corresponding to filters 2 and 3 of the array of filters of
Another capability of the embodiments described herein is the ability to perform the volumetric estimation of a gas cloud. This can be accomplished by using (instead of compensating or negating) the parallax induced effects described above. In this case, the measured parallax between two or more similar spectral response images (e.g., two or more channels or FPAs) can be used to estimate a distance between the imaging system and the gas cloud or between the imaging system and an object in the field of view of the system. The parallax induced transverse image shift, d, between two images is related to the distance, z, between the cloud or object 110 and the imaging system according to the equation z=−sz′/d. Here, s, is the separation between two similar spectral response images, and z′ is the distance to the image plane from the back lens. The value for z′ is typically approximately equal to the focal length f of the lens of the imaging system. Once the distance z between the cloud and the imaging system is calculated, the size of the gas cloud can be determined based on the magnification, m=f/z, where each image pixel on the gas cloud, Δx′, corresponds to a physical size in object space Δx=Δx′/m. To estimate the volume of the gas cloud, a particular symmetry in the thickness of the cloud based on the physical size of the cloud can be assumed. For example, the cloud image can be rotated about a central axis running through the cloud image to create a three dimensional volume estimate of the gas cloud size. It is worth noting that in the embodiments described herein only a single imaging system is required for such volume estimation. Indeed, due to the fact that the information about the angle at which the gas cloud is seen by the system is decoded in the parallax effect, the image data includes the information about the imaged scene viewed by the system in association with at least two angles.
When the temporal differencing algorithm is used for processing the acquired imaging data, a change in the scene that is not caused by the target species can inadvertently be highlighted in the resulting image. In various embodiments, compensation for this error makes use of the temporal differencing between two FPAs that are substantially equally spectrally sensitive to the target species. In this case, the temporal difference image will highlight those pixels the intensity of which have changed in time (and not in wavelength). Therefore, subtracting the data corresponding to these pixels on both FPAs, which are substantially equally spectrally sensitive to the target species, to form the resulting image, excludes the contribution of the target species to the resulting image. The differentiation between (i) changes in the scene due to the presence of target species and (ii) changes in the scene caused by changes in the background not associated with the target species is, therefore, possible. In some embodiments, these two channels having the same or substantially similar spectral response so as to be substantially equally spectrally sensitive to the target species may comprise FPAs that operate using visible light. It should also be noted that, the data acquired with a visible light FPA (when present as part of the otherwise IR imaging system) can also be used to facilitate such differentiation and compensation of the motion-caused imaging errors. Visible cameras generally have much lower noise figure than IR cameras (at least during daytime). Consequently, the temporal difference image obtained with the use of image data from the visible light FPA can be quite accurate. The visible FPA can be used to compensate for motion in the system as well as many potential false-alarms in the scene due to motion caused by people, vehicles, birds, and steam, for example, as long as the moving object can be observed in the visible region of the spectra. This has the added benefit of providing an additional level of false alarm suppression without reducing the sensitivity of the system since many targets such as gas clouds cannot be observed in the visible spectral region. In various implementations, an IR camera can be used to compensate for motion artifacts.
Another method for detection of the gases is to use a spectral unmixing approach. A spectral unmixing approach assumes that the spectrum measured at a detector pixel is composed of a sum of component spectra (e.g., methane and other gases). This approach attempts to estimate the relative weights of these components needed to derive the measurement spectrum. The component spectra are generally taken from a predetermined spectral library (for example, from data collection that has been empirically assembled), though sometimes one can use the scene to estimate these as well (often called “endmember determination”). In various embodiments, the image obtained by the detector pixel is a radiance spectrum and provides information about the brightness of the object. To identify the contents of a gas cloud in the scene and/or to estimate the concentration of the various gases in the gas cloud, an absorption/emission spectrum of the various gases of interest can be obtained by comparing the measured brightness with an estimate of the expected brightness. The spectral unmixing methodology can also benefit from temporal, parallax, and motion compensation techniques.
In various embodiments, a method of identifying the presence of a target species in the object includes obtaining the radiance spectrum (or the absorption spectrum) from the object in a spectral region indicative of the presence of the target species and calculating a correlation (e.g., a correlation coefficient) by correlating the obtained radiance spectrum (or the absorption spectrum) with a reference spectrum for the target species. The presence or absence of the target species can be determined based on an amount of correlation (e.g., a value of correlation coefficient). For example, the presence of the target species in the object can be confirmed if the amount of correlation or the value of correlation coefficient is greater than a threshold. In various implementations, the radiance spectrum (or the absorption spectrum) can be obtained by obtaining a spectral difference image between a filtered optical channel and/or another filtered optical channel/unfiltered optical channel or any combinations thereof.
For example, an embodiment of the system configured to detect the presence of methane in a gas cloud comprises optical components such that one or more of the plurality of optical channels is configured to collect IR radiation to provide spectral data corresponding to a discrete spectral band located in the wavelength range between about 7.9 μm and about 8.4 μm corresponding to an absorption peak of methane. The multispectral data obtained in the one or more optical channels can be correlated with a predetermined absorption spectrum of methane in the wavelength range between about 7.9 and 8.4 μm. In various implementations, the predetermined absorption spectrum of methane can be saved in a database or a reference library accessible by the system. Based on an amount of correlation (e.g., a value of correlation coefficient), the presence or absence of methane in the gas cloud can be detected.
Examples of Practical Embodiments and Operation
The embodiment 300 of
Due to the specifics of operation in the IR range of the spectrum, the use of the so-called noise-equivalent temperature difference (or NETD) is preferred and is analogous to the SNR commonly used in visible spectrum instruments. The array of microbolometer FPAs 336 is characterized to perform at NETD≤72 mK at an f-number of 1.2. Each measurement was carried out by summing four consecutive frames, and the reduction in the NETD value expected due to such summation would be described by corresponding factor of √4=2. Under ideal measurement conditions, therefore, the FPA NETD should be about 36 mK.
It is worth noting that the use of optically-filtered FPAs in various embodiments of the system described herein can provide a system with higher number of pixels. For example, embodiments including a single large format microbolometer FPA array can provide a system with large number of pixels. Various embodiments of the systems described herein can also offer a high optical throughput for a substantially low number of optical channels. For example, the systems described herein can provide a high optical throughput for a number of optical channels between 4 and 50. By having a lower number of optical channels (e.g., between 4 and 50 optical channels), the systems described herein have wider spectral bins which allows the signals acquired within each spectral bin to have a greater integrated intensity.
An advantage of the embodiments described herein over various scanning based hyperspectral systems that are configured for target species detection (for example, gas cloud detection) is that, the entire spectrum can be resolved in a snapshot mode (for example, during one image frame acquisition by the FPA array). This feature enables the embodiments of the imaging systems described herein to take advantage of the compensation algorithms such as the parallax and motion compensation algorithms mentioned above. Indeed, as the imaging data required to implement these algorithms are collected simultaneously with the target-species related data, the compensation algorithms are carried out with respect to target-species related data and not with respect to data acquired at another time interval. This rapid data collection thus improves the accuracy of the data compensation process. In addition, the frame rate of data acquisition is much higher. For example, embodiments of the imaging system described herein can operate at video rates from about 5 Hz and higher. For example, various embodiments described herein can operate at frame rates from about 5 Hz to about 60 Hz or 200 Hz. Thus, the user is able to recognize in the images the wisps and swirls typical of gas mixing without blurring out of these dynamic image features and other artifacts caused by the change of scene (whether spatial or spectral) during the lengthy measurements. In contradistinction, scanning based imaging systems involve image data acquisition over a period of time exceeding a single-snap-shot time and can, therefore, blur the target gas features in the image and inevitably reduce the otherwise achievable sensitivity of the detection. This result is in contrast to embodiments of the imaging system described herein that are capable of detecting the localized concentrations of gas without it being smeared out with the areas of thinner gas concentrations. In addition, the higher frame rate also enables a much faster response rate to a leak of gas (when detecting such leak is the goal). For example, an alarm can trigger within fractions of a second rather than several seconds.
To demonstrate the operation and gas detection capability of the imaging systems described herein, a prototype was constructed in accordance with the embodiment 300 of
The same prototype of the system can also demonstrate the dynamic calibration improvement described above by imaging the scene surrounding the system (the laboratory) with known temperature differences. The result of implementing the dynamic correction procedure is shown in
Dynamic Calibration Elements and References
In
As discussed above, in some embodiments, the reference sources 972a and 972b are imaged onto the detector array 1 and detector array 9, without much blur such that the reference sources 972a and 972b are focused. In contrast, in other embodiments, the image of reference sources 972a and 972b formed on the detector array 1, and detector array 9 are blurred such that the reference sources 972a and 972b are defocused, and thereby provide some averaging, smoothing, and/or low pass filtering. The reference sources 972a and 972b may comprise a surface of known temperature and may or may not include a heater or cooler attached thereto or in thermal communication therewith. For example, the reference source 972a and 972b may comprises heaters and coolers respectively or may comprise a surface with a temperature sensor and a heater and sensor respectively in direct thermal communication therewith to control the temperature of the reference surface. In various implementations, the reference sources 972a and 972b can include a temperature controller configured to maintain the reference sources 972a and 972b at a known temperature. In some implementations, the reference sources 972a and 972b can be associated with one or more sensors that measure the temperature of the reference sources 972a and 972b and communicate the measured temperature to the temperature controller. In some implementations, the one or more sensors can communicate the measured temperature to the data-processing unit. In various implementations, the reference sources 972a and 972b may comprise a surface of unknown temperature. For example, the reference sources may comprise a wall of a housing comprising the imaging system. In some implementations, the reference sources 972a and 972b can comprise a surface that need not be associated with sensors, temperature controllers. However, in other implementations, the reference sources 972a and 972b can comprise a surface that can be associated with sensors, temperature controllers.
In
In the implementations depicted in
The temperature of the reference sources 972b, 972a can be different. For example, the reference source 972a can be at a temperature TA, and the reference source 972b can be at a temperature TB lower than the temperature TA. A heater can be provided under the temperature-controlled element 972a to maintain it at a temperature TA, and a cooler can be provided underneath the temperature-controlled element 972b to maintain it at a temperature TB. In various implementations, the embodiments illustrated in
The reference sources 972a and 972b can be coated with a material to make it behave substantially as a blackbody (for which the emission spectrum is known for any given temperature). If a temperature sensor is used at the location of each reference source, then the temperature can be tracked at these locations. As a result, the regions in the image of each camera (e.g., on the detector arrays 1 and 9) in which the object has such known temperature (and, therefore, spectrum) can be defined. A calibration procedure can thus be used so that most of the cameras (if not every camera) so operated agrees, operationally, with most or every other camera, for objects at the temperatures represented by those two sources. Calibrating infrared cameras using sources at two different temperatures is known as a “two-point” calibration, and assumes that the measured signal at a given pixel is linearly related to the incident irradiance. Since this calibration can be performed during multiple, more, or even every frame of a sequence, it is referred to as a “dynamic calibration”.
An example of the dynamic calibration procedure is as follows. If there is a temperature sensor on the reference sources or reference surface, then the temperature measurements obtained by these temperature sensors can be used to determine their expected emission spectra. These temperature measurements are labeled as TA[R], TB[R], and Tc[R] for the “reference temperatures” of sources/surfaces A, B, and C. These temperature measurements can be used as scalar correction factors to apply to the entire image of a given camera, forcing it to agree with the reference temperatures. Correcting the temperature estimate of a given pixel from T to T′ can use formulae analogous to those discussed below in reference to
In the configuration illustrated in
A “static” calibration (a procedure in which the scene is largely blocked with a reference source such as the moving shutters 960 in
This design is an enhancement to the systems 300 and 400 shown in
The imaging elements in the system 1000 (shown as mirrors in
In various implementations, a heater can be provided underneath, adjacent to, or in thermal communication with reference source/surface A to give it a higher temperature TA, and a cooler can be provided underneath, adjacent to, or in thermal communication with reference source B to give it a lower temperature TB. In various implementations, the embodiments illustrated in
The dynamic calibration is used to obtain a corrected temperature T′ from the initial temperature T estimated at each pixel in a camera using the following formulae:
T′[x,y,c]=(T[x,y,c]−TA[R])G[c]+TA[R]
where is TA[R] is a dynamic offset correction factor, and,
is a dynamic gain correction factor. The term c discussed above is a camera index that identifies the camera whose data is being corrected.
Various embodiments of the divided aperture infrared spectral imager (DAISI) disclosed herein (e.g., embodiments illustrated in
For example, the detector arrays 136, 236, 336 and/or 456 can include at least one mid-wave infrared (MWIR) FPA configured to detect infrared radiation in the wavelength range between about 3 microns and about 7 microns and at least one long wave infrared (LWIR) FPA configured to detect infrared radiation in the wavelength range between about 7 microns and about 14 microns. As another example, the detector arrays 136, 236, 336 and/or 456 can include one MWIR FPA configured to detect infrared radiation in the wavelength range between about 3 microns and about 7 microns and a plurality of LWIR FPAs configured to detect infrared radiation in the wavelength range between about 7 microns and about 14 microns. In various implementations, the MWIR FPA can be cooled and/or uncooled. In various implementations, one or more of the LWIR FPAs can be cooled and/or uncooled. The MWIR and/or the LWIR FPAs can be cooled to temperatures below room temperature. For example, the MWIR and/or the LWIR FPAs can include coolers that maintain the FPAs at a temperature between about 200 degree Kelvin and about 273 degree Kelvin, a temperature between about 150 degree Kelvin and about 200 degree Kelvin, a temperature between about 100 degree Kelvin and about 150 degree Kelvin or a temperature between about 50 degree Kelvin and about 150 degree Kelvin. In various implementations, the coolers employed to maintain the MWIR and/or LWIR FPAs at a desired temperature can be a cryogenic cooler. In various implementations, the coolers employed to maintain the MWIR and/or LWIR FPAs at a desired temperature can include pulse tube coolers available from Thales Cryogenics or Canberra Industries, Inc. In some implementations, the coolers employed to maintain the MWIR and/or LWIR FPAs at a desired temperature can include high operating temperature (HOT) coolers that can maintain the MWIR and/or LWIR FPAs at a temperature between about 110 degree Kelvin and 150 degree Kelvin. For example, in some implementations, the HOT coolers can maintain the at least one MWIR and/or LWIR FPAs at a temperature of about 135 degree Kelvin. The use of HOT coolers can facilitate detector arrays comprising novel semiconductor materials that have reduced cooling requirements as compared to detector arrays comprising InSb or Mercury Cadmium Telluride (MCT). The HOT coolers can additionally extend the lifetime of the detector arrays beyond 25000 hours, such as, for example up to 90,000 hours. The coolers can include cryogenic coolers and/or thermo-electric coolers.
Implementations of the dual band DAISI including MWIR and LWIR FPAs can have increased detection sensitivity to chemicals/gases that have stronger spectral features in the mid wave infrared wavelength range than the long wave infrared wavelength range. Additionally, since implementations of the dual band DAISI including MWIR and LWIR FPAs are capable of obtaining spectral information in both the mid-wave infrared wavelength range and the long wave infrared wavelength range, such implementations can have enhanced chemical/gas identification capabilities as compared to implementations that operate only in the mid-wave infrared wavelength range or the long wave infrared wavelength range.
Furthermore, since implementations of the dual band DAISI including MWIR and LWIR FPAs are capable of obtaining an image of a scene by combining spectral information in the mid-wave infrared wavelength range and the long wave infrared wavelength range, the accuracy of detecting various gases/chemicals in the imaged scene can be improved. Additionally, since the implementations of the dual band DAISI including MWIR and LWIR FPAs obtain information in the mid-wave infrared wavelength range and the long wave infrared wavelength range using solar as well as thermal sources for signal, they can be used in different weather conditions (e.g., on sunny days, cloudy days, etc.) and at various times of the day (e.g., during the day or night).
Various implementations of the dual band DAISI system can be used in continuous monitoring of explosive hydrocarbons or hazardous gas leaks and/or any fugitive gas emission. Implementations of the dual band DAISI system used in such applications can provide video imagery identifying the species, size and direction, and concentration of any detected gas cloud. Other implementations of the dual band DAISI system can also be used for exploration, standoff chemical detection, explosive detection, bio-imaging, medical imaging, gas cloud imaging, surveillance, food inspection, and remote sensing applications and other applications including but not limited to biological warfare, gas leaks at refineries, rigs and petroleum plants, imaging for the purpose of reconnaissance, underwater applications, space application, telecommunications and/or optical computing.
Various configurations of detector arrays including MWIR and LWIR FPAs are shown in
Although the embodiment illustrated in
The MWIR FPA can have a similar size and weight as the LWIR FPA. In some implementations, the MWIR FPA can have a smaller size as compared to the LWIR FPA.
The one or more MWIR FPAs can be statically and dynamically calibrated using a procedure similar to the static and dynamic calibration of the one or more LWIR FPAs described above. For example, the one or more MWIR FPAs can image the moveable temperature-controlled reference source 160, the field reference 338, the field reference array 438, the moveable temperature-controlled shutters 460a and 460b, temperature-controlled shutters 960 and/or the reference sources 972a and 972b described above. The static and dynamic calibration procedures described above can maintain agreement among all the FPAs including the LWIR and MWIR FPAs when viewing the same radiant energy. Additionally, the static and dynamic calibration procedures described above can aid in differentiating between thermally-induced signal and solar-reflection-induced signal which can affect the detection capabilities of a MWIR FPA. For example, if a MWIR FPA (or camera) sees a change in signal within the scene, but none in the reference source, then the user can conclude that the change was induced by a change in the scene illumination or a change in the object temperature. The user can confirm with certain degree of confidence that the change was not induced by changes in the response of the detector which would indicate that the detector calibration is in need of adjustment.
One or more optical filters can be disposed in the optical path of each MWIR and LWIR FPA in the detector array similar to the embodiments disclosed in
Cooled MWIR FPAs can advantageously operate at a higher frame rate than an uncooled MWIR/LWIR FPA. Accordingly, for any video frame rate, cooled MWIR FPAs can average data over more frames than an uncooled MWIR/LWIR FPA. This can result in an increase in the signal to noise ratio of the images obtained by the uncooled MWIR FPA.
The increased signal to noise ratio provided by cooled MWIR FPAs can be advantageous in systems that employ one or more uncooled LWIR FPAs. For example, the reduced signal to noise ratio of images obtained from uncooled LWIR FPAs can result in reduced absorption strength of spectral features present in the images obtained from one or more uncooled LWIR FPAs. Accordingly, imaging systems employing one or more uncooled LWIR FPAs only can result in false positive detection of the presence or absence of one or more chemical species. For example, imaging systems employing one or more uncooled LWIR FPAs only can indicate the presence of a chemical species when it is absent in truth or indicate the absence of a chemical species when it is present in truth. Thus, differentiating between true and false chemical species detections can be challenging when imaging systems employing only one or more uncooled LWIR FPAs are used.
A cooled MWIR FPA providing images with increased signal to noise ratio, and utilizing the stronger absorption features of various chemical species in mid-wavelength infra-red spectral range, can be used to primarily detect the presence or absence of a chemical species in a scene while the uncooled LWIR FPAs can be used to remove false positive detections and/or to identify the chemical species.
The one or more MWIR and LWIR FPAs can be configured to output one or more image frames. The one or more image frames can be output at video frame rates. For example, the output of the one or more MWIR and LWIR FPAs can be between 5 image frames/second and 120 image frames/second. The output image frames can be analyzed using hyperspectral video analytics using spectra-temporal algorithms to obtain spectral features of various chemical species that may be present in the scene. The spectral features obtained from the image data output from one or more of the MWIR elements and LWIR elements can be cross-correlated with known spectra of various chemical species that are stored in a reference library in a database to identify the chemical species that may be present in the gas cloud 1505.
In various embodiments, image frames output from the one or more MWIR and LWIR FPAs at the beginning of the measurement period can be used to estimate background features of the scene and image frames acquired later in the measurement period can be used to detect and speciate various chemical species of interest that may be present in the scene. Thus, the one or more MWIR and LWIR FPAs can be used to not only detect and speciate various chemical species of interest that may be present in the scene but to also estimate the dynamic properties of those chemical species. For example, the one or more MWIR and LWIR FPAs can be used to estimate movement of the gas cloud 1505 and/or the region 1507 over time in addition to detecting and speciating various chemical species that may be present in the gas cloud 1505.
The results associated with detecting and speciating various chemical species that may be present in the gas cloud 1505, such as, for example, spectral features, concentration, data regarding movement of the gas cloud, etc. 1505 and/or 1507 can be obtained within 1 second from the start of imaging the scene. The associated results can be updated after every frame thereafter. In various implementations, the results associated with detecting and speciating various chemical species that may be present in the gas cloud 1505 can be obtained in sufficiently real time (e.g., between about 0.01 millisecond and about 1 millisecond from the start of imaging the scene, between about 0.01 millisecond and about 10 milliseconds from the start of imaging the scene, between about 0.01 millisecond and about 50 milliseconds from the start of imaging the scene, between about 0.01 millisecond and about 100 milliseconds from the start of imaging the scene, between about 0.01 millisecond and about 500 milliseconds from the start of imaging the scene, between about 0.01 millisecond and about 1 second from the start of imaging the scene, between about 0.01 millisecond and about 10 seconds from the start of imaging the scene, between about 0.01 millisecond and about 30 seconds from the start of imaging the scene, between about 0.01 millisecond and about 1 minute from the start of imaging the scene, between about 0.01 millisecond and about 5 minutes from the start of imaging the scene, or between about 0.01 millisecond and about 10 minutes from the start of imaging the scene).
The ability of one or more MWIR FPAs (e.g., cooled MWIR FPAs) to aid the detection capabilities of the one or more LWIR FPAs can be enhanced when the image frames output from the one or more MWIR FPAs are synchronized with the image frames output from the one or more LWIR FPAs. In various implementations, a feedback system can be employed to synchronize the image frames output from the one or more LWIR FPAs and the image frames output from the one or more MWIR FPAs.
The inclusion of one or more MWIR FPAs in an array of LWIR FPAs also provides improved chemical speciation due to the extension in the spectral range over which data is collected. For example, sulfur dioxide (SO2) produces a signature in the long-wavelength infra-red spectral range that can be confused with the signature for propylene (C3H6) at very low spectral resolution. However, the spectral features of SO2 and C3H6 do not overlap in the mid-wavelength infra-red spectral range. Thus, the information provided by one or more MWIR FPAs when used in combination with information provided by one or more LWIR FPAs can provide additional information to help differentiate the chemical species that are detected.
The information provided from the one or more MWIR and LWIR FPAs can also provide information that cannot be obtained by either the MWIR or the LWIR FPAs individually. For example, the combined information from MWIR and LWIR FPAs can be used to estimate the illumination of a scene. The LWIR FPAs are more sensitive to thermally emitted radiation and less sensitive to reflected solar radiation, since solar light is weaker in the long-wavelength infra-red spectral range in comparison with terrestrial sources that are much closer. The MWIR FPAs are sensitive to both thermal and reflected solar radiation. Thus, if LWIR FPAs observing a scene record only small and gradual changes in signal, whereas the MWIR FPAs observing the same scene record much sharper changes then it can be concluded with a degree of confidence that the illumination of the scene is changing.
This effect is illustrated in
Some of the plurality of spatially and spectrally distinct optical channels in various implementations of the dual band DAISI including MWIR and LWIR FPAs can obtain information from spatially distinct portions of an object. Some of the plurality of spatially and spectrally distinct optical channels in various implementations of the dual band DAISI including MWIR and LWIR FPAs can have a field of view (FOV) that is different from the field of view (FOV) of some other of the plurality of spatially and spectrally distinct optical channels. Some of the plurality of spatially and spectrally distinct optical channels in various implementations of the dual band DAISI including MWIR and LWIR FPAs can have a field of view (FOV) that is lower than the field of view (FOV) of the entire system.
Various embodiments disclosed herein can be used to detect and identify hydrogen sulfide (H2S) gas, including concentrations of hydrogen sulfide that are less than (or greater than) a lethal dose for a human. Many conventional imaging systems are not able to detect hydrogen sulfide gas because (1) the hydrogen sulfide gas absorption signature is weak in many of the spectral bands where measurements are commonly performed, (2) in those locations where the H2S gas absorption signature is strong, water absorption is also strong enough to obscure the signal for H2S, and (3) in the LWIR and MWIR wavelength ranges, the H2S absorption signature is obscured by the larger absorption of other hydrocarbon gases. Advantageously, the embodiments disclosed herein can detect H2S gas and can indeed provide an entire video of the hydrogen sulfide gas plume. The hydrogen sulfide imaging camera (HSIC) disclosed herein can image the gas leak coming from the source and can track the gas leak across its field-of-view until the gas leak leaves or is reduced in concentration below the detection threshold. In this way, the HSIC can have a very large coverage area and can detect the gas leak at its source without having to wait for the gas cloud to come in physical contact with the HSIC or to cross a single optical path. Thus, the HSIC disclosed herein can image H2S gas leaks better than other systems, such as electrochemical or semiconductor devices, or open path sensors.
Another advantage of the HSIC is that it is easy for the user to verify alarms based on video streams and/or images of the gas cloud. Other devices (e.g., point sensors) can be used in conjunction with the HSIC's video feed to locate the cloud. Further, with the HSIC, the user can easily find and fix the gas leaks because the user can see the source of the gas leak. Such an imaging system can provide complete information for an effective response for a H2S gas leak. By contrast, point sensors can only indicate whether the sensor detects a particular chemical. Point sensors do not provide the user with a visual image to monitor the location of the leak over time. Moreover, the systems disclosed herein can include processing electronics that are configured to identify H2S and estimate the concentration of H2S in the gas cloud. Additional details of methods for estimating gas concentrations may be found in U.S. patent application Ser. No. 14/792,477, published as US 2016/0097713, filed Jul. 6, 2015, the contents of which are incorporated by reference herein in their entirety and for all purposes. Because the HSIC can self-calibrate, the HSIC has relatively low maintenance costs. The HSIC can also be manufactured to have a small size, low weight, and low power consumption. For example, the HSIC may be miniaturized in some embodiments to fit within the size, weight, and/or power parameters disclosed in U.S. patent application Ser. No. 14/700,791, published as US 2015/0316473, filed Apr. 30, 2015, the contents of which are incorporated by reference herein in their entirety and for all purposes.
Thus, the embodiments of the HSIC disclosed herein represent a new type of device for detecting explosive and/or toxic gas plumes with a particular focus on hydrogen sulfide. The HSIC can comprise a passive infrared spectral imaging camera that can visualize a hydrogen sulfide gas cloud at levels below the lethal concentration when the gas cloud is exposed to ambient sunlight. The HSIC can also be combined with external light sources, such as infrared light emitting diodes, to enhance detection during the night but which may also help during the day (e.g., during periods of insufficient sunlight and/or cloud cover).
The HSIC can detect the absorption signature of hydrogen sulfide in the short wave infrared region (1-1.7 microns) and/or the longwave infrared region (7-14 microns). The advantage of these spectral regions is that there are absorption bands where hydrogen sulfide has a dominating absorption signal compared to other common gases found in the presence of hydrogen sulfide, such as water (H2O), carbon dioxide (CO2), sulfur dioxide (SO2) and methane (CH4). Two of these spectral absorption regions that are particularly useful are the wavelength bands having wavelengths of 1.5-1.7 microns and 8.8-9.0 microns.
With a relatively small number of spectral bands (e.g., about 1-20 bands) in each of these regions, the HSIC can identify hydrogen sulfide from other gases present based on its unique absorption spectral shape. The HSIC can also determine the concentration of the gas plume based on the strength of the absorption spectral signal.
The HSIC can be used for fixed-installation continuous monitoring, and can also be mounted to a land or aerial vehicle (manned or unmanned) for mobile monitoring of sites. The HSIC can also be used as a hand-held video HSIC, or as a wearable gas monitoring device, similar to existing point sensors that are attached to clothing. For example, in some embodiments, the HSIC may be embodied in a mobile system, which may be similar to the mobile DAISI systems disclosed in U.S. patent application Ser. No. 14/700,791, published as 2015/0316473, filed Apr. 30, 2015, the contents of which are incorporated by reference herein in their entirety and for all purposes.
Various experimental data has been developed to verify the efficacy of the HSIC system. For example, a detailed optical radiometric model for hydrogen sulfide (H2S) gas and other commonly present gases found with H2S including water, methane, CO2 and SO2, has been developed. The radiometric model was validated through experimentation with similarly absorbing gases (primarily methane). Performance parameters for the HSIC were then extrapolated from the model to demonstrate the viability of the HSIC system. Additional details of this model are described herein.
The absorption spectra described herein were derived from digitizing HITRAN data (available at https://www.cfa.harvard.edu/hitran/), following a Lorentzian assumption for line broadening (pressure and temperature). The linewidths may be approximated as being the weighted average of the foreign linewidth with the self linewidth, where “self” and “foreign” are the fractions of concentration of that same molecule in the atmosphere versus that of all others. For example, at 5% humidity at standard temperature and pressure (STP), the concentration of water molecules in air is 1.56%. Thus the linewidth γtotal can be formed by taking
γtotal=0.9844γforeign+0.0156γself
The above linewidth γtotal may give the appearance that the self-broadening term is not significant, but, typically, γself is a lot larger than γforeign, which means that the self-broadening term is significant when the self-concentration fraction is not negligible. For water vapor, the self-broadening width is about 5 times larger than that of the foreign broadening term. At 5% humidity, the self-broadening term only increases the linewidth by about 6%, but at 50% humidity, the water concentration increases to 15.6% of the air, and then the self-broadening increases the linewidth by 58%. However, for all of the spectra below, self-broadening has been ignored. Additional details for this subject matter may be found at least in Chapter 2 of Craig F. Bohren and Eugene E. Clothiaux, Fundamentals of Atmospheric Radiation (published by Wiley, printed in Darmstadt, 2006).
The above HITRAN conversion model does not calculate the shifting of the line strengths with changes in temperature. (The molecular orbital population ratios shift as temperature changes.) As long as the temperatures remain close to 300 K, such shifting of line strengths should be small for gas constituents in atmosphere.
Tables 1 and 2 below summarize several of the spectral regions in which the absorption of H2S is strong compared to that of the other chemicals. For example, Tables 1 and 2 below compare the mean absorption of the various gases over the spectral regions of interest.
Accordingly, in various embodiments, it can be advantageous to analyze light reflected from an object, such as a gas cloud at a site of interest that includes wavelengths in the SWIR. Although various embodiments disclosed herein may include filtering and/or analyzing light in the SWIR, in other embodiments, light in the MWIR and/or LWIR wavelengths may be filtered and/or analyzed to identify H2S.
SWIR Detection and SNR
The radiometric model for detection of H2S generally assumes that the sun is the primary light source, with reflected sunlight as the background against which to view the gas. For this model, a Planck blackbody curve can be constructed for the spectral radiance of the sun, assuming a blackbody temperature of 5777 K. The sunlight passes through the atmosphere, so the blackbody spectral radiance Lbb can be multiplied by the estimated spectral transmission of the atmosphere Tatm. (The atmospheric transmission data was obtained from [http://www.gemini.edu/sciops/telescopes-and-sites/observing-condition-constraints/ir-transmission-spectra].). Next an integration over the solid angle of the sun, Ω, can be conducted to obtain the spectral irradiance at the object, which can be multiplied by the cosine of the object tilt angle θ1 to get the projected irradiance:
The projected irradiance can be multiplied by the object's reflectivity Robj and by the cosine of the surface's tilt angle with respect to the viewer θ2. Assuming a Lambertian reflector, dividing the result by π converts the reflected emittance into a reflected radiance:
The HSIC equation can be used to convert an object radiance into an irradiance at the focal plane, using the transmission of the optics Topt and the lens f-number f #:
The object irradiance can be multiplied by the pixel area p2, the detector quantum efficiency spectrum η(λ), and the exposure time t to convert the irradiance into a photoelectron count:
N(λ)=tp2η(Δ)Ifpa(λ)
The photoelectron count can be integrated over the spectral width of the measurement collected by that pixel:
N=∫λ
As one example, a 12 nm full-width half maximum (FWHM) optical filter can be placed behind the lens of a SWIR HSIC and the following parameters shown in Table 3 can be inserted into the measurement model, together with an integration time of 10 ms, and a lens with an estimated f-number of 1.5. The resulting spectrum produced by the model for the 1.4-1.6 μm spectral range is shown in
When imaging a high-reflectivity object under solar illumination, the saturation point of the HSIC (1.9 million photoelectrons) can be reached when integrating for just 5 ms. Also, for this example measurement, fluctuations in the measured signal due to Poisson noise can be seen at about sqrt(1900000)=1378 photelectrons, which is equivalent to an SNR of 1378. This means that changes in intensity within a pixel should be seen if the change is more than one part in 1378.
The column density of H2S gas needed to produce a change in the background measurement equivalent to the noise level (i.e. the “noise equivalent column density”, NECD) can also be calculated. If it is assumed that the optical filter used for measuring H2S gas is centered at 1590 nm, and has a FWHM of 3 nm, then the absorption cross-section a of the gas is 8.34e-27 m′. Using Beer's Law, the absorption A can be calculated as:
A=1−exp(−σnl),
where n is the molecular number density of the gas (2.687e25 molecules/m3 at standard atmospheric temperature and pressure), and f is the path length through the gas. For a weak gas, this equation is accurately approximated in the weak absorption regime as
A=σnl,
which can be solved as:
nl=A/σ.
If the number density is represented in terms of “ppm” (parts per million), this becomes:
nl=(2.687e25 ppm m3)×A/σ.
For the problem of detecting H2S gas, the minimum observable absorption value is 1/1378, which can be substituted into this equation to give the noise-equivalent column density as
NECD(ppm.m)=nl=(2.687e25 ppm m3)/(1378σ)=3238 ppm.m.
The NECD is the single-pixel detection value. If multiple pixels are used for detection (such as by binning the detector array or by smoothing), or average multiple frames of data, then this NECD value can be reduced substantially by at least an order of magnitude, which places the NECD at 323.8 ppm.m which is below the lethal limit for hydrogen sulfide.
Experimental Demonstration
Experiments have been performed to demonstrate that these numbers are approximately correct. For the example experiment, rather than releasing H2S (which is a dangerous gas to humans), methane gas was used, as it is much safer to humans. Thus, the measurement used a different optical filter than the H2S measurement would use, and the absorption cross-section is also different. However, once the demonstration result for methane has been obtained, then the measured SNR can be scaled by the expected difference between the experiment measurement and an equivalent one for detecting H2S. For the experiment, a filter centered at 1650 nm with a FWHM of 12 nm was chosen. For this filter, the absorption cross-section for methane should be 4.0e-26 m2. Thus, the absorption cross-section will be 4.78 times more than it would be for the H2S measurement.
The measurements of
Since the above measurement used pure methane, the column density can be estimated using the known geometry of the hose. For pure gas (1000000 ppm) and a path length of 7 mm, the resulting column density is 7000 ppm.m. Thus, the lower limit of H2S measurement sensitivity can be estimated as about:
7000 ppm.m×(4.78/10)=3350 ppm.m
The estimated sensitivity for H2S is for the gas that is exiting just outside the release point. One feature that has consistently been observed when imaging hydrocarbon gases is that the release point often produces the weakest signal for measurement. When the gas collects at some distance from the leak, the gas tends to produce local regions of higher column density that improve the measurement signal. This phenomena is what was observed in the methane experiment above as well. Note that the source used in this experiment was generated using a lamp rather than the sun, which reduces the initial signal level affecting the final measurement.
Based on these experiments and the optical model, the primary spectral band for detection of H2S in the SWIR can be about 1590 nm with a full width half maximum (FWHM) of 3 nm. A single filter imaging system at this spectral band may have sufficient differentiation for operators to identify and fix H2S leaks. A multi-filter imaging system (e.g., two or more spectral bands) can allow for automatic detection of H2S. These bands can be adjacent to this primary spectral band or separate. One potential spectral band to include is wavelengths at or near 1650 nm with a bandwidth of approximately 12 nm, which is a dominate absorption region for methane gas. H2S and methane are commonly found together and these two spectral bands can be used to isolate each gas from one another. More spectral bands may provide additional capabilities for separation of other gases and/or removal of common false alarms such as steam, people, clouds, solar glints, etc. In some embodiments, the HSIC disclosed herein can also detect the presence of CO2, SO2, etc.
For example, in some embodiments, the HSIC 2300 shown in
One advantage of the embodiment shown in
In
As explained above, the HSIC 2300, 2400, 2500 can include a processing unit 2322, 2422 comprising processing electronics configured to process image data detected by the detector arrays. The processing electronics can be configured to identify whether the detected image data comprises infrared spectra representative of H2S gas. Moreover, the processing electronics can be configured to detect the concentration of H2S gas in the scene. Advantageously, the HSIC can include processing electronics that detects H2S gas at a wide range of concentrations, including concentrations below and above levels that are dangerous to humans. For example, in various embodiments, the HSICs disclosed herein can beneficially detect H2S gas at concentrations in a range of 1 ppm.m to 5,000 ppm.m, e.g., in a range of 100 ppm.m to 2,500 ppm.m, or more particularly, in a range of 500 ppm.m to 1500 ppm.m, e.g., in a range of 800 ppm.m to 1200 ppm.m (such as 1000 ppm.m).
Filter Design for H2S Gas Detection
As explained above, the absorption spectrum for H2S gas may at least partially overlap and/or may include peaks that are close to the absorption spectra for other chemicals that are commonly found at sites to be monitored (such as oil wells, petroleum refinement facilities, etc.). For example, as illustrated at least in
Thus, it can be advantageous to select an optical filter for an H2S gas detection channel (such as the filters 2303-2304, 2404, 2504) that passes wavelengths of light that have high H2S gas absorption and low absorption of the other chemicals (e.g., CO2, CH4, SO2, and H2O). In addition, a background radiation channel can be defined to include a filter that minimizes or reduces the absorption signal from chemicals typically present at sites to be monitored (e.g., the signals of H2S, CO2, CH4, H2O, SO2). In some embodiments, a methane gas detection channel can be defined to include a filter that increases or maximizes the methane signal and reduces or minimizes the absorption signals of the other chemicals (e.g., the signals of H2S, CO2, H2O, SO2). In some embodiments, a carbon dioxide gas detection channel can be defined to include a filter that increases or maximizes the CO2 signal and reduces or minimizes the signals of the other chemicals (e.g., the signals of H2S, CH4, H2O, SO2).
The filters described herein can comprise transmissive or reflective filters having a passband with a center frequency and a width (FWHM). The filter can comprise a filter element, such as an interference filter or any other suitable type of filter. For example, the filter can comprise one or more layers of transmissive material formed or otherwise coupled together. For example, in embodiments that utilize an interference filter, the filter can comprise a plurality of transmissive layers with a thickness selected to pass a band of wavelengths and selectively attenuate other wavelengths. In some embodiments, it should be appreciated that the optical filter disclosed herein can include one or more filter components. For example, in some embodiments, the optical filter can comprise a single passband filter component configured to selectively pass the wavelengths of interest. In other embodiments, the optical filter can comprise multiple filter components constructed such that the multiple components selectively pass the wavelengths of interest. For example, in some embodiments, the optical filter can comprise a combination of a high pass filter component and a low pass filter component that, taken in combination, pass a band of desired wavelengths. Moreover, although the optical filters disclosed herein may be described as selectively passing light within a band of wavelengths and attenuating light outside the band, the optical filters may have multiple passbands at different wavelength ranges.
As shown in
As explained herein, the HSIC can detect H2S gas in the presence of the other chemicals (such as CH4, CO2, H2O, SO2), even where the concentrations and/or signals of the H2S gas are small compared with the concentrations and/or signals of the other gases. As explained herein with respect to Table 7 below, for example, if the HSIC detects other gases (such as CO2 or CH4) even within the H2S channel C1, the processing electronics of the HSIC can estimate the contribution of the signal in the H2S channel C1 from CH4 and/or CO2 (and, indeed, other commonly found chemicals) based on the in-band signal for the H2S channel C1 of CH4 and/or CO2. For example, the processing electronics can use a weighted average of the in-band signals of CH4 and/or CO2 for the H2S channel C1 to compare how much of the signal in the H2S channel C1 is representative of each of H2S, CO2, CH4, etc. A similar analysis can be performed by the processing electronics for the other channels. Thus, the embodiments disclosed herein can beneficially detect H2S gas in the presence of large concentrations of other chemicals.
To at least partially define the respective optical channels, the filters (such as filters 2303-2304, 2404, 2504) can be selected to have passbands which selectively pass radiation in which the absorption of the chemical to be detected (e.g., H2S) is larger than the absorption of the other commonly-found chemicals (e.g., CO2, CH4, SO2, H2O). The passband for each filter can define a band of wavelengths centered at a central wavelength. To determine a desirable passband for the H2S filter, a filter transmission model was developed and convolved with the highest-resolution H2S absorption spectra available, to prevent or mitigate any errors due to sampling.
The resolution of the plots in
From
In various embodiments, if the bandpass range of the filter is widened, more light is collected, which enables the measurement of the H2S gas when the scene brightness is relatively low. However, a wide filter also means that the average absorption through the filter decreases, such that discriminating between H2S and CO2 is more difficult (lower ratio R) than for narrower passband ranges. If a narrower passband filter is selected, then both absorption sensitivity and the ratio R can be relatively high. The measurement model indicates that a 3 nm wide filter may be close to saturating the detector when measuring a high reflectivity object in direct sunlight, which risks a decrease in measurement sensitivity. Measurements also indicate that a 3 nm wide filter at a central wavelength of 1.5905 μm exhibits good detection sensitivity and discrimination, while a 10 nm wide filter centered at 1.5922 μm exhibits good light collection and manufacturing tolerances used for manufacturing the filter.
As shown in Table 5, the sensitivity and the discrimination are both lower for the wider 10 nm wide filter. Even though the wider 10 nm filter enables the ability to work at lower light levels, the higher sensitivities for the narrow 3 nm filter may be beneficial. From these results, example filter design parameters are shown in Table 6 below, where the “OFF” filter indicates the filter that passes background radiation and attenuates the other common gases.
In some embodiments, as shown in Table 6, the filter widths can be the same to ensure that the different filters saturate at about the same integration time, in order to simplify operation of the HSIC. In addition, as explained herein, a narrowband filter may shift to shorter wavelengths as the angle of incidence increases. Thus, if the filter is placed in front of the lens (in collimated space), the center of the field may be at normal incidence to the filter, but the edges of the field may have their passbands shifted to shorter wavelengths than those for the center. The magnitude of the passband shift can be estimated by:
where θ is the angle of incidence and neff is a unique index of refraction for the filter. For the filters disclosed herein, neff=1.7, which produces the wavelength shift shown in
An additional consideration is the degree to which the filter blocks or transmits out-of-band signals. In the model, the out-of-band transmission value can be treated as a constant to determine the amount of the other signals (e.g., signals representing H2S, CO2, etc.) that are transmitted, which can be compared with the in-band filter signal. For the H2S filter and the background radiation filter (denoted “OFF”), the signal can be collected for the in-band filter and for two different out-of-band Tout transmission values as shown in Table 7 below.
Thus, as shown in Table 7, to prevent detection of CO2 or CH4 when the system is trying to detect H2S, the HCIS can ensure that the out-of-band signal for CO2 and CH4 are less than 1/100 or 1/1000 that of the H2S in-band signal. For example, in some embodiments, the H2S channel (i.e., the channel configured to detect H2S) can be configured such that the signals for CO2 and CH4 within the same wavelength band are each in a range of 0.01% to 5% of the H2S signal, in a range of 0.01% to 1% of the H2S signal, in a range of 0.01% to 0.5% of the H2S signal, in a range of 0.05% to 0.5% of the H2S signal, or in a range of 0.1% to 0.5% of the H2S signal. For CH4, the HCIS can achieve 1/100 of the H2S signal when the out-of-band transmission is 1.0e-4, but 1/1000 of the H2S when Tout<1.0e-5. However, the in-band contribution due to CO2 is at 1%, so there may be little benefit to have the ability to discriminate H2S from CO2 by going from Tout=1.0e-4 to Tout=1.0e-5.
In some embodiments, the systems disclosed herein can include a CH4 channel configured to detect CH4 gas. In such an arrangement, the CH4 channel can be configured such that the signals for CO2 and H2S within the same wavelength band are each in a range of 0.01% to 5% of the CH4 signal, in a range of 0.01% to 1% of the CH4 signal, in a range of 0.01% to 0.5% of the CH4 signal, in a range of 0.05% to 0.5% of the CH4 signal, or in a range of 0.1% to 0.5% of the CH4 signal. Further, in some embodiments, the systems disclosed herein can include a CO2 channel configured to detect CO2 gas. In such an arrangement, the CO2 channel can be configured such that the signals for CH4 and H2S within the same wavelength band are each in a range of 0.01% to 5% of the CO2 signal, in a range of 0.01% to 1% of the CO2 signal, in a range of 0.01% to 0.5% of the CO2 signal, in a range of 0.05% to 0.5% of the CO2 signal, or in a range of 0.1% to 0.5% of the CO2 signal.
In various embodiments, the OFF channel, which can correspond to a channel configured to pass broadband background radiation, can pass low amounts of the chemicals commonly found at sites to be monitored, such as CO2, H2S, CH2, SO2, and H2O. For example, the OFF channel can comprise a band of IR wavelengths in which the transmission of signals representative of background light (such as sunlight) is in a range of 50 to 10,000 times, 50 to 5,000 times, 100 to 10,000 times, 500 to 10,000 times, or 200 to 1,000 times the signal representative of CO2, H2S, CH2, SO2, and/or H2O. As explained herein, in some embodiments, the HSIC can comprise processing electronics that detects or identifies H2S gas by subtracting the signal of the OFF channel (background radiation) from the signal of the H2S channel.
Addition of Active Illumination
Passive detection of H2S in the SWIR spectral region can be highly dependent on the amount of solar illumination available. The solar illumination for any given scene can vary drastically, e.g., by up to 20×-30× from overcast to sunny weather conditions. The variation from day to night can be even more extreme. To compensate for these fluctuations, it can be beneficial to add an active illumination source to the camera system, such as the light sources 2320, 2420 shown in
Advantageously, the HSIC systems disclosed herein can distinguish between H2S gas and other chemicals (CO2, CH4, SO2, H2O) commonly found at sites to be monitored, even at relatively low concentrations. As explained herein, the imaging system can include an optical detector and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector. Within the band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of H2S gas can be greater than a convolution of the optical filter with an absorption spectrum of one or more of carbon dioxide (CO2), methane (CH4), sulfur dioxide (SO2), and water (H2O). The optical filter can selectively pass light having a wavelength in a range of 1585 nm to 1595 nm (e.g., 1590 nm) while selectively attenuating light at wavelengths above 1600 nm and below 1580 nm. For example, the optical filter can selectively block or attenuate light at wavelengths below 1585 nm and above 1595 nm. In various embodiments, within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least ten times greater than the convolution of the optical filter with the absorption spectrum of of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O). Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least fifty times greater than the convolution of the optical filter with the absorption spectrum of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O). Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least one hundred times greater than the convolution of the optical filter with the absorption spectrum of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O). In some embodiments, within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least fifty times (or at least 100 times or 300 times) greater than the convolution of the optical filter with the absorption spectrum of CO2. In some embodiments, within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least fifty times (or at least 100 times or 300 times) greater than the convolution of the optical filter with the absorption spectrum of CH4. For example, in some embodiments, within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least twenty times to at least 500 times greater than the convolution of the optical filter with the absorption spectrum of CH4 or the convolution of the optical filter with the absorption spectrum of CO2. In some embodiments, within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least forty times to at least 300 times greater than the convolution of the optical filter with the absorption spectrum of CH4 or the convolution of the optical filter with the absorption spectrum of CO2. In various embodiments, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least 100 times to 10,000 times, in a range of 200 times to 5,000 times, or in a range of 500 times to 5,000 times the convolution of the optical filter with the absorption spectrum of each of CO2, CH4, H2O, and SO2.
In some embodiments, the optical filter can have a center wavelength in a range of 1580 nm to 1600 nm (e.g., 1590 nm) with a band pass between 1-4 nm FWHM. In some embodiments, the optical filter can have a band pass between 3-6 nm FWHM. In some embodiments, the optical filter can have a band pass between 5-10 nm FWHM. A majority of the light impinging on the optical detector can be at wavelengths in a range of 1580 nm to 1600 nm. In some embodiments, the system can include at least one beamsplitter configured to split incoming radiation into different wavelength bands.
The HSICs disclosed herein can include a processor or processing electronics configured to process image data detected by the detector, the processor configured to identify whether the image data comprises infrared spectra representative of hydrogen sulfide gas. The processing electronics can also be configured to quantify the concentration of H2S gas. In some arrangements, the HSIC can include an optical focal plane array (FPA) unit along a plurality of spatially and spectrally different optical channels such that infrared (IR) radiation is transferred from a scene towards the optical FPA unit. Each optical channel can be positioned to transfer a portion of the IR radiation incident on the imaging system from the scene towards the optical FPA unit. A first channel of the plurality of optical channels can comprise a first optical filter configured to pass a signal representative of H2S gas, and a second channel of the plurality of optical channels can comprise a second optical filter configured to pass broadband background radiation within a second band of IR wavelengths. In some embodiments, a processing unit can be configured to identify hydrogen sulfide (H2S) gas in the scene based at least in part on a difference between a first signal transferred from the optical filter and a second signal transferred from the second optical filter.
In some embodiments, a third channel of the plurality of optical channels can include a third optical filter that selectively transfers light within a third band of IR wavelengths to the optical detector. Within the third band of IR wavelengths, a convolution of the third optical filter with an absorption spectrum of carbon dioxide (CO2) gas can be greater (e.g., at least five times greater, at least ten times greater, at least fifty times greater, at least 100 times greater, at least 500 times greater) than a convolution of the third optical filter with an absorption spectrum of hydrogen sulfide (H2S), or methane (CH4), or sulfur dioxide (SO2), or water (H2O). In some embodiments, within the third band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of CO2 gas can be in a range of at least twenty times to at least 500 times greater than the convolution of the optical filter with the absorption spectrum of CH4 or the convolution of the optical filter with the absorption spectrum of H2S. In some embodiments, within the third band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of CO2 gas can be in a range of at least forty times to at least 300 times greater than the convolution of the optical filter with the absorption spectrum of CH4 or the convolution of the optical filter with the absorption spectrum of H2S.
In some embodiments, a fourth channel of the plurality of optical channels can include a fourth optical filter that selectively transfers light within a fourth band of IR wavelengths to the optical detector. Within the fourth band of IR wavelengths, a convolution of the fourth optical filter with an absorption spectrum of methane (CH4) gas can be greater (e.g., at least five times greater, at least ten times greater, at least fifty times greater, at least 100 times greater, at least 500 times greater) than a convolution of the fourth optical filter with an absorption spectrum of hydrogen sulfide (H2S), or carbon dioxide (CO2), or sulfur dioxide (SO2), or water (H2O). In some embodiments, the HSIC can include an illumination source configured to provide illumination to a scene to be imaged. In some embodiments, within the fourth band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of CH4 gas can be in a range of at least twenty times to at least 500 times greater than the convolution of the optical filter with the absorption spectrum of CO2 or the convolution of the optical filter with the absorption spectrum of H2S. In some embodiments, within the third band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of CH4 gas can be in a range of at least forty times to at least 300 times greater than the convolution of the optical filter with the absorption spectrum of CO2 or the convolution of the optical filter with the absorption spectrum of H2S.
In some embodiments, the HSIC comprises an infrared (IR) imaging system for detecting a gas. The imaging system can include an optical filter that selectively passes light having a wavelength in a range of 1500 nm to 1700 nm while attenuating light at wavelengths above 1700 nm and below 1500 nm. The imaging system can include an optical detector array sensitive to light having a wavelength of 1590 nm that is positioned rear of the optical filter. The optical filter can selectively pass light having a wavelength in a range of 1550 nm to 1650 nm, while attenuating light at wavelengths above 1650 nm and below 1550 nm. The optical filter can selectively pass light having a wavelength in a range of 1575 nm to 1625 nm, while attenuating light at wavelengths above 1625 nm and below 1575 nm. The optical filter can selectively pass light having a wavelength in a range of 1580 nm to 1620 nm, while attenuating light at wavelengths above 1625 nm and below 1575 nm. In some embodiments, the optical filter can selectively pass light having a wavelength in a range of 1585 nm to 1595 nm, while attenuating light at wavelengths above 1600 nm and below 1580 nm. The optical filter can pass light having a wavelength of 1590 nm. In some embodiments, the optical filter can attenuate light at wavelengths below 1580 nm and above 1600 nm.
Light at wavelengths in a range of 1580 nm to 1600 nm can be in a range of 50 to 10,000 times light at wavelengths below 1580 nm and above 1600 nm. Light at wavelengths in the range of 1580 nm to 1600 nm can be in a range of 150 to 5,000 times light at wavelengths below 1580 nm and above 1600 nm. In some embodiments, the optical filter can have a band pass between 1-4 nm FWHM, between 3-6 nm FWHM, or between 5-10 nm FWHM. The imaging system can also include an imaging lens. In some embodiments, a majority of the light impinging on the optical detector array can be at wavelengths in a range of 1580 nm to 1600 nm. For example, a majority of the light impinging on the optical detector array can be at wavelengths in a range of 1585 nm to 1595 nm.
The imaging system can include at least one beamsplitter configured to split incoming radiation into different wavelength bands. The imaging system can include processing electronics configured to process image data detected by the detector array, the processing electronics configured to identify whether the image data comprises infrared spectra representative of hydrogen sulfide gas. The imaging system can include an optical focal plane array (FPA) unit and a plurality of spatially and spectrally different optical channels to transfer infrared (IR) radiation from a scene towards the optical FPA unit, each optical channel positioned to transfer a portion of the IR radiation incident on the imaging system from the scene towards the optical FPA unit. In some embodiments, at least one of the plurality of optical channels is in the short-wavelength infrared spectral range. At least one of the plurality of optical channels can be in the long-wavelength infrared spectral range. At least one of the plurality of optical channels can be in the mid-wavelength infrared spectral range.
The imaging system can be configured to acquire a first video image of the scene in the mid-wavelength infrared spectral range and a second video image of the scene in the long-wavelength infrared spectral range. The imaging system can be configured to acquire a first video image of the scene in the short-wavelength infrared spectral range and a second video image of the scene in the long-wavelength infrared spectral range. The imaging system can be configured to acquire a first video image of the scene in the mid-wavelength infrared spectral range and a second video image of the scene in the short-wavelength infrared spectral range. In some embodiments, at least one of the plurality of optical channels passes broadband background radiation. In some embodiments, an illumination source can be configured to provide illumination to a scene to be imaged.
In one embodiment, an infrared (IR) imaging system for imaging a scene is disclosed. The imaging system can include an optical system comprising an optical focal plane array (FPA) unit and a plurality of spatially and spectrally different optical channels to transfer IR radiation from the scene towards the optical FPA unit, each optical channel positioned to transfer a portion of the IR radiation incident on the optical system from the scene towards the optical FPA unit. At least one of the plurality of optical channels can be in the short-wavelength infrared (SWIR) spectral range. The imaging system can be configured to acquire a first video image of the scene in the short-wavelength infrared spectral range.
The imaging system can comprise an optical filter in said shortwave infrared (SWIR) channel that selectively passes light having a wavelength of 1590 nm while attenuating light at wavelengths above 1600 nm and below 1580 nm. The optical filter can attenuate light at wavelengths below 1585 nm and above 1595 nm. Light at wavelengths in a range of 1585 nm to 1595 nm can be in a range of 50 to 10,000 times light at wavelengths below 1580 nm and above 1600 nm.
In various embodiments, the optical filter can have a band pass between 1-4 nm FWHM, between 3-6 nm FWHM, or between 5-10 nm FWHM. In some embodiments, the band pass can be centered at 1590 nm. The imaging system can include a plurality of imaging lenses. Moreover, the imaging system can include at least one beamsplitter configured to split incoming radiation into different wavelength bands. The imaging system can include processing electronics configured to process image data detected by the optical FPA unit, the processing electronics configured to identify whether the image data comprises infrared spectra representative of hydrogen sulfide gas. The system can include a second optical filter which transmits light at a wavelength of 1650 nm with a FWHM bandwidth of 12 nm or less. In some embodiments, at least one of the plurality of optical channels can be in the long-wavelength infrared (LWIR) spectral range. The system can include a LWIR optical filter configured to pass light having a wavelength in the long wavelength infrared range (LWIR). The LWIR optical filter can be configured to pass light having a wavelength in a range of 7 microns to 14 microns. In various embodiments, at least one of the plurality of optical channels can be in the mid-wavelength infrared (MWIR) spectral range. The system can also include a MWIR optical filter configured to pass light having a wavelength in the mid-wavelength infrared range (MWIR). The MWIR optical filter can be configured to pass light having a wavelength in a range of 3 microns to 7 microns. In various embodiments, the optical FPA unit can include a plurality of optical FPAs. A majority of the light that passes through the optical filter can be at wavelengths in a range of 1580 nm to 1600 nm. In some embodiments, the system can include an illumination source configured to provide illumination to a scene to be imaged.
In one embodiment, an infrared (IR) imaging system for detecting hydrogen sulfide (H2S) gas is disclosed. The imaging system can an optical detector array, and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector array. Within the band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of H2S gas can be greater than a convolution of the optical filter with an absorption spectrum of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O).
The optical filter can selectively pass light having a wavelength in a range of 1585 nm to 1595 nm while attenuating light at wavelengths above 1600 nm and below 1580 nm. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least ten times greater than the convolution of the optical filter with the absorption spectrum of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O). Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least fifty times greater than the convolution of the optical filter with the absorption spectrum of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O). Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least one hundred times greater than the convolution of the optical filter with the absorption spectrum of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O). Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least twice the convolution of the optical filter with the absorption spectrum of carbon dioxide (CO2), or methane (CHO, or sulfur dioxide (SO2), or water (H2O). Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least fifty times greater than the convolution of the optical filter with the absorption spectrum of CO2. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least 2 to at least 10,000 times greater than the convolution of the optical filter with the absorption spectrum of CO2. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least fifty times to at least 10,000 times greater than the convolution of the optical filter with the absorption spectrum of CO2. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least 100 times to at least 10,000 times greater than the convolution of the optical filter with the absorption spectrum of CO2. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least 500 times to at least 10,000 times greater than the convolution of the optical filter with the absorption spectrum of CO2. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least 500 times to at least 1,000 times greater than the convolution of the optical filter with the absorption spectrum of CO2. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least twenty times to at least 500 times greater than the convolution of the optical filter with the absorption spectrum of CO2. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be at least fifty times greater than the convolution of the optical filter with the absorption spectrum of CH4. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least two times to at least 500 times greater than the convolution of the optical filter with the absorption spectrum of CH4. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least twenty times to at least 500 times greater than the convolution of the optical filter with the absorption spectrum of CH4. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least fifty times to at least 10,000 times greater than the convolution of the optical filter with the absorption spectrum of CH4. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least 100 times to at least 10,000 times greater than the convolution of the optical filter with the absorption spectrum of CH4. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least 500 times to at least 10,000 times greater than the convolution of the optical filter with the absorption spectrum of CH4. Within the band of IR wavelengths, the convolution of the optical filter with the absorption spectrum of H2S gas can be in a range of at least 500 times to at least 1,000 times greater than the convolution of the optical filter with the absorption spectrum of CH4.
In various embodiments, the optical filter can have a band pass between 1-4 nm FWHM, between 3-6 nm FWHM, or between 5-10 nm FWHM. The system can also include an imaging lens. In some embodiments, a majority of the light impinging on the optical detector array can be at wavelengths in a range of 1580 nm to 1600 nm. At least one beamsplitter can be configured to split incoming radiation into different wavelength bands. The system can include processing electronics configured to process image data detected by the detector, the processing electronics configured to identify whether the image data comprises infrared spectra representative of hydrogen sulfide gas. The system can include an optical focal plane array (FPA) unit and a plurality of spatially and spectrally different optical channels to transfer infrared (IR) radiation from a scene towards the optical FPA unit, each optical channel positioned to transfer a portion of the IR radiation incident on the imaging system from the scene towards the optical FPA unit. A first channel of the plurality of optical channels can include the optical filter and a second channel of the plurality of optical channels can include a second optical filter, the second optical filter configured to pass broadband background radiation within a second band of IR wavelengths. The system can include processing electronics, the processing electronics configured to identify hydrogen sulfide (H2S) gas in the scene based at least in part on a difference between a first signal transferred from the optical filter and a second signal transferred from the second optical filter.
In various embodiments, a third channel of the plurality of optical channels can include a third optical filter that selectively transfers light within a third band of IR wavelengths to a third optical detector, wherein, within the third band of IR wavelengths, a convolution of the third optical filter with an absorption spectrum of carbon dioxide (CO2) gas is greater than a convolution of the third optical filter with an absorption spectrum of hydrogen sulfide (H2S), or methane (CH4), or sulfur dioxide (SO2), or water (H2O). A fourth channel of the plurality of optical channels can include a fourth optical filter that selectively transfers light within a fourth band of IR wavelengths to a fourth optical detector, wherein, within the fourth band of IR wavelengths, a convolution of the fourth optical filter with an absorption spectrum of methane (CH4) gas is greater than a convolution of the fourth optical filter with an absorption spectrum of hydrogen sulfide (H2S), or carbon dioxide (CO2), or sulfur dioxide (SO2), or water (H2O). The system can further comprise an illumination source configured to provide illumination to a scene to be imaged.
In one embodiment, an infrared (IR) imaging system for detecting hydrogen sulfide (H2S) gas is disclosed. The system can include an optical detector array and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector. The optical filter can comprise a passband that transmits within the passband a first signal representative of hydrogen sulfide (H2S) and a second signal representative of one of carbon dioxide (CO2), methane (CH4), water (H2O), and sulfur dioxide (SO2), the first signal comprising a first peak within the passband, the second signal comprising a second peak within the passband, wherein a ratio of the first peak to the second peak is in a range of 1 to 100,000. In some embodiments, the ratio of the first peak to the second peak can be in a range of 50 to 10,000, 50 to 5,000, in a range of 100 to 10,000, in a range of 200 to 10,000, in a range of 200 to 1,000, in a range of 200 to 5,000, or in a range of 500 to 5,000. In other arrangements, however, the ratio of the first peak (representative of H2S) to the second peak may be less than 1, e.g., in situations in which there are high concentrations of other gases. For example, in some arrangements, the ratio may be between 0.001 and 1, between 0.01 and 1, or between 0.1 and 1. Beneficially, even at low relative H2S concentrations as compared with other chemicals, the embodiments disclosed herein can identify H2S gas from amongst other gases such as CO2, CH4, H2O, and SO2. The wavelength ranges passed by the filter can be any of the suitable wavelengths described above.
In one embodiment, an optical filter is disclosed. The optical filter can include a filter element comprising a passband that selectively passes light within a band of infrared (IR) wavelengths. The passband can transmit a first signal representative of hydrogen sulfide (H2S) and a second signal representative of one of carbon dioxide (CO2), methane (CH4), water (H2O), and sulfur dioxide (SO2), the first signal comprising a first peak within the passband, the second signal comprising a second peak within the passband, wherein a ratio of the first peak to the second peak is in a range of 1 to 100,000. In various embodiments, the ratio of the first peak to the second peak can be in a range of 50 to 10,000, 50 to 5,000, in a range of 100 to 10,000, in a range of 200 to 10,000, in a range of 200 to 1,000, in a range of 200 to 5,000, or in a range of 500 to 5,000. In other arrangements, however, the ratio of the first peak (representative of H2S) to the second peak may be less than 1, e.g., in situations in which there are high concentrations of other gases. For example, in some arrangements, the ratio may be between 0.001 and 1, between 0.01 and 1, or between 0.1 and 1. Beneficially, even at low relative H2S concentrations as compared with other chemicals, the embodiments disclosed herein can identify H2S gas from amongst other gases such as CO2, CH4, H2O, and SO2. As above, the wavelength ranges passed by the filter can be any of the suitable wavelengths described above.
In one embodiment, an optical filter is disclosed. The optical filter can include a filter element comprising a passband that selectively passes light within a band of infrared (IR) wavelengths. Within the band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of H2S gas can be greater than (e.g., at least twice) a convolution of the optical filter with an absorption spectrum of carbon dioxide (CO2), or methane (CH4), or sulfur dioxide (SO2), or water (H2O).
In some embodiments, the filter element can comprise a transmissive filter element. The convolution of the optical filter with an absorption spectrum of H2S gas can be in a range of 2 to 10,000 times greater than a convolution of the optical filter with an absorption spectrum of CO2. The convolution of the optical filter with an absorption spectrum of H2S gas can be in a range of 50 to 10,000 times greater than a convolution of the optical filter with an absorption spectrum of CO2. The convolution of the optical filter with an absorption spectrum of H2S gas can be in a range of 100 to 5,000 times greater than a convolution of the optical filter with an absorption spectrum of CO2. The convolution of the optical filter with an absorption spectrum of H2S gas can be in a range of 500 to 5,000 times greater than a convolution of the optical filter with an absorption spectrum of CO2. The convolution of the optical filter with an absorption spectrum of H2S gas can be in a range of 50 to 10,000 times greater than a convolution of the optical filter with an absorption spectrum of CH4. The convolution of the optical filter with an absorption spectrum of H2S gas can be in a range of 100 to 5,000 times greater than a convolution of the optical filter with an absorption spectrum of CH4. The convolution of the optical filter with an absorption spectrum of H2S gas can be in a range of 500 to 5,000 times greater than a convolution of the optical filter with an absorption spectrum of CH4.
In one embodiment, an infrared (IR) imaging system for imaging a scene is disclosed. The imaging system can include an optical system comprising an optical focal plane array (FPA) unit and a plurality of spectrally different optical channels to transfer IR radiation from the scene towards the optical FPA unit. Each optical channel can be positioned to transfer a portion of the IR radiation incident on the optical system from the scene towards the optical FPA unit. The plurality spectrally different optical channels can be coupled to or integrally formed with the optical FPA unit. At least one of the plurality of optical channels can be in the short-wavelength infrared (SWIR) spectral range. The imaging system can be configured to acquire a first video image of the scene in the short-wavelength infrared spectral range.
In various embodiments disclosed herein, the optical filter can selectively pass light having a wavelength in a range of 1550 nm to 1650 nm, while attenuating light at wavelengths above 1650 nm and below 1550 nm. For example, the optical filter can selectively pass light having a wavelength in a range of 1200 nm to 1400 nm, while attenuating light at wavelengths above 1400 nm and below 1200 nm. The optical filter can selectively pass light having a wavelength in a range of 1230 nm to 1330 nm, while attenuating light at wavelengths above 1330 nm and below 1230 nm. The optical filter can selectively pass light having a wavelength in a range of 1900 nm to 2100 nm, while attenuating light at wavelengths above 2100 nm and below 1900 nm. The optical filter can selectively pass light having a wavelength in a range of 1950 nm to 2100 nm, while attenuating light at wavelengths above 2100 nm and below 1950 nm.
In one embodiment, an infrared (IR) imaging system for detecting carbon dioxide (CO2) gas is disclosed. The imaging system can include an optical detector array and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector array. Within the band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of CO2 gas can be greater than (e.g., at least twice) a convolution of the optical filter with an absorption spectrum of hydrogen sulfide (H2S), or methane (CH4), or sulfur dioxide (SO2), or water (H2O). The convolution of the optical filter with the absorption spectrum of CO2 gas can be in a range of 2 to 10,000 times greater than the convolution of the optical filter with the absorption spectrum of hydrogen sulfide (H2S). The convolution of the optical filter with the absorption spectrum of CO2 gas can be in a range of 50 to 10,000 times greater than the convolution of the optical filter with the absorption spectrum of hydrogen sulfide (H2S). The convolution of the optical filter with the absorption spectrum of CO2 gas can be in a range of 100 to 5,000 times greater than the convolution of the optical filter with the absorption spectrum of hydrogen sulfide (H2S). The convolution of the optical filter with the absorption spectrum of CO2 gas can be in a range of 500 to 5,000 times greater than the convolution of the optical filter with the absorption spectrum of hydrogen sulfide (H2S). In various embodiments, the optical filter can selectively pass light having a wavelength in a range of 1400 nm to 1500 nm while attenuating light at wavelengths above 1500 nm and below 1400 nm. The optical filter can selectively pass light having a wavelength in a range of 1900 nm to 2100 nm while attenuating light at wavelengths above 2100 nm and below 1900 nm. The optical filter can selectively pass light having a wavelength in a range of 9 microns to 10 microns while attenuating light at wavelengths above 10 microns and below 9 microns.
In one embodiment, an infrared (IR) imaging system for detecting carbon dioxide (CO2) gas is disclosed. The imaging system can comprise an optical detector array and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector. The optical filter can comprise a passband that transmits within the passband a first signal representative of carbon dioxide (CO2) and a second signal representative of one of hydrogen sulfide (H2S), methane (CH4), water (H2O), and sulfur dioxide (SO2), the first signal comprising a first peak within the passband, the second signal comprising a second peak within the passband, wherein a ratio of the first peak to the second peak is in a range of 1 to 100,000. The ratio of the first peak to the second peak can be in a range of 50 to 10,000, in a range of 50 to 5,000, in a range of 100 to 10,000, in a range of 200 to 10,000, in a range of 200 to 1,000, in a range of 200 to 5,000, or in a range of 500 to 5,000. The optical filter can selectively pass light having a wavelength in a range of 1400 nm to 1500 nm while attenuating light at wavelengths above 1500 nm and below 1400 nm. The optical filter can selectively pass light having a wavelength in a range of 1900 nm to 2100 nm while attenuating light at wavelengths above 2100 nm and below 1900 nm. In other arrangements, however, the ratio of the first peak to the second peak may be less than 1, e.g., in situations in which there are high concentrations of other gases. For example, in some arrangements, the ratio may be between 0.001 and 1, between 0.01 and 1, or between 0.1 and 1. The optical filter can selectively pass light having a wavelength in a range of 9 microns to 10 microns while attenuating light at wavelengths above 10 microns and below 9 microns
In one embodiment, an infrared (IR) imaging system for detecting methane (CH4) gas is disclosed. The system can include an optical detector array and an optical filter that selectively transfers light within a band of IR wavelengths to the optical detector array. Within the band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of CH4 gas can be greater than (e.g., at least twice) a convolution of the optical filter with an absorption spectrum of hydrogen sulfide (H2S), or carbon dioxide (CO2), or sulfur dioxide (SO2), or water (H2O). The convolution of the optical filter with the absorption spectrum of CH4 gas can be in a range of 50 to 10,000 times greater than the convolution of the optical filter with the absorption spectrum of hydrogen sulfide (H2S). The convolution of the optical filter with the absorption spectrum of CH4 gas can be in a range of 2 to 10,000 times greater than the convolution of the optical filter with the absorption spectrum of hydrogen sulfide (H2S). The convolution of the optical filter with the absorption spectrum of CH4 gas can be in a range of 100 to 5,000 times greater than the convolution of the optical filter with the absorption spectrum of hydrogen sulfide (H2S). The convolution of the optical filter with the absorption spectrum of CH4 gas can be in a range of 500 to 5,000 times greater than the convolution of the optical filter with the absorption spectrum of hydrogen sulfide (H2S). The optical filter can selectively pass light having a wavelength in a range of 1600 nm to 1700 nm while attenuating light at wavelengths above 1700 nm and below 1600 nm.
In various embodiments disclosed herein, the optical filter can include a plurality of filter components (e.g., a bandpass filter can include a high pass filter component and a low pass filter component). In various embodiments, the optical filter can pass or transfer light by reflection. In various embodiments, the optical filter can pass or transfer light by transmission. In the embodiments disclosed herein, the system can include processing electronics configured to detect H2S gas in a scene comprising a plurality of gases. The processing electronics can be configured to determine a concentration of the detected H2S gas. The system can also include imaging optics to form a spectral image on the optical detector array. The imaging optics can include one or more lenses.
In various embodiments for detecting H2S, the optical filter can selectively pass light in a passband with a center wavelength in a range of 1580 nm to 1620 nm. The optical filter can selectively pass light in a passband with a center wavelength in a range of 1585 nm to 1595 nm. The optical filter can attenuate light at wavelengths above 1620 nm and below 1580 nm. In some embodiments, the optical filter can attenuate light at wavelengths between 1500 nm to 1580 nm and between 1620 nm and 1700 nm. The optical filter can attenuate light at wavelengths in a range of 1400 nm to 1500 nm and in a range of 1700 nm to 1800 nm. The optical filter can attenuate light at wavelengths in a range of 1350 nm to 1550 nm and in a range of 1650 nm to 1850 nm. The filter element can selectively pass light having a wavelength in a range of 1500 nm to 1700 nm while attenuating light at wavelengths above 1700 nm and below 1500 nm. The filter element can selectively pass light having a wavelength in a range of 1550 nm to 1650 nm, while attenuating light at wavelengths above 1650 nm and below 1550 nm. The filter element can selectively pass light having a wavelength in a range of 1575 nm to 1625 nm, while attenuating light at wavelengths above 1625 nm and below 1575 nm. The filter element can selectively pass light having a wavelength in a range of 1580 nm to 1620 nm, while attenuating light at wavelengths above 1625 nm and below 1575 nm. The filter element can selectively pass light having a wavelength in a range of 1585 nm to 1595 nm, while attenuating light at wavelengths above 1600 nm and below 1580 nm. The filter element can attenuate light at wavelengths in a range of 1400 nm to 1500 nm and in a range of 1700 nm to 1800 nm. The filter element can attenuate light at wavelengths in a range of 1450 nm to 1550 nm and in a range of 1650 nm to 1850 nm.
In one embodiment, an infrared (IR) imaging system for detecting hydrogen sulfide (H2S) gas is disclosed. The system can include an optical system comprising an optical detector array and one or more optical channels that transfer infrared radiation to the optical detector array. The system can include processing electronics configured to process image data received by the optical detector array, the processing electronics configured to detect H2S gas based on the captured image data.
In some embodiments, the processing electronics can be configured to distinguish H2S gas from one or more of carbon dioxide (CO2), methane (CH4), sulfur dioxide (SO2), or water (H2O). in an imaged object. For example, the processing electronics can be configured to distinguish H2S gas from CO2 gas in an imaged object. For example, the processing electronics can be configured to distinguish H2S gas from CH4 gas in an imaged object. In various embodiments, a first channel of the one or more optical channels can comprise an H2S channel that transfers a first signal to the optical detector array that is representative of H2S gas. A second channel of the one or more optical channels can comprise a background channel that transfers a second signal to the optical detector array that is representative of broadband background radiation. The processing electronics can be configured to identify H2S gas based at least in part on a difference between the first signal and the second signal. In some embodiments, a third channel of the one or more optical channels can selectively transfer a third signal to the optical detector array that is representative of carbon dioxide (CO2), methane (CH4), sulfur dioxide (SO2), or water (H2O). The processing electronics can be configured to identify at least one of carbon dioxide (CO2), methane (CH4), sulfur dioxide (SO2), and water (H2O) based at least in part on a difference between the third signal and the second signal. The processing electronics can be configured to compare the difference between the first signal and the second signal with the difference between the third signal and the second signal to determine an amount of H2S gas relative to an amount of at least one of carbon dioxide (CO2), methane (CH4), sulfur dioxide (SO2), and water (H2O). An illumination source can be configured to provide illumination to a scene to be imaged. In various embodiments, the one or more optical channels can include one or more imaging lenses. In various embodiments, the one or more optical channels can include an optical filter. Furthermore, within a band of IR wavelengths, a convolution of the optical filter with an absorption spectrum of H2S gas can be greater than (e.g., at least twice, at least ten times, at least 100 times, at least 1000 times) a convolution of the optical filter with an absorption spectrum of carbon dioxide (CO2), or methane (CHO, or sulfur dioxide (SO2), or water (H2O).
In the embodiments described herein, the processing electronics can configured to detect H2S gas at concentrations in a range of 1 ppm.m to 5,000 ppm.m, in a range of 100 ppm.m to 2,500 ppm.m, in a range of 500 ppm.m to 1500 ppm.m, or in a range of 800 ppm.m to 1200 ppm.m.
The HSIC systems disclosed herein can be used for many applications including gas leaks at refineries, rigs, and petroleum plants, remote sensing applications, biological warfare, deep to mid IR imaging (reconnaisance), missile/target identifier, underwater applications, space applications, telecommunications switch, optical computing, detection of toxic gases like hydrogen sulfide, flare efficiency monitoring, pipe blockages from hot & cold spots, and gas leak detection and quantification. The HSIC can be made explosion-proof by using existing HSIC explosion proof housings designed for visible light surveillance systems, including for the shortwave infrared version of the system.
References throughout this specification to “one embodiment,” “an embodiment,” “a related embodiment,” or similar language mean that a particular feature, structure, or characteristic described in connection with the referred to “embodiment” is included in at least one embodiment of the present invention. Thus, appearances of the phrases “in one embodiment,” “in an embodiment,” and similar language throughout this specification may, but do not necessarily, all refer to the same embodiment. It is to be understood that no portion of disclosure, taken on its own and in possible connection with a figure, is intended to provide a complete description of all features of the invention.
In the drawings like numbers are used to represent the same or similar elements wherever possible. The depicted structural elements are generally not to scale, and certain components are enlarged relative to the other components for purposes of emphasis and understanding. It is to be understood that no single drawing is intended to support a complete description of all features of the invention. In other words, a given drawing is generally descriptive of only some, and generally not all, features of the invention. A given drawing and an associated portion of the disclosure containing a description referencing such drawing do not, generally, contain all elements of a particular view or all features that can be presented is this view, for purposes of simplifying the given drawing and discussion, and to direct the discussion to particular elements that are featured in this drawing. A skilled artisan will recognize that the invention may possibly be practiced without one or more of the specific features, elements, components, structures, details, or characteristics, or with the use of other methods, components, materials, and so forth. Therefore, although a particular detail of an embodiment of the invention may not be necessarily shown in each and every drawing describing such embodiment, the presence of this detail in the drawing may be implied unless the context of the description requires otherwise. In other instances, well known structures, details, materials, or operations may be not shown in a given drawing or described in detail to avoid obscuring aspects of an embodiment of the invention that are being discussed. Furthermore, the described single features, structures, or characteristics of the invention may be combined in any suitable manner in one or more further embodiments.
The features recited in claims appended to this disclosure are intended to be assessed in light of the disclosure as a whole, including features disclosed in prior art to which reference is made.
At least some elements of a device of the invention can be controlled—and at least some steps of a method of the invention can be effectuated, in operation—with a programmable processor governed by instructions stored in a memory. The memory may be random access memory (RAM), read-only memory (ROM), flash memory or any other memory, or combination thereof, suitable for storing control software or other instructions and data. Those skilled in the art should also readily appreciate that instructions or programs defining the functions of the present invention may be delivered to a processor in many forms, including, but not limited to, information permanently stored on non-writable storage media (e.g. read-only memory devices within a computer, such as ROM, or devices readable by a computer I/O attachment, such as CD-ROM or DVD disks), information alterably stored on writable storage media (e.g. floppy disks, removable flash memory and hard drives) or information conveyed to a computer through communication media, including wired or wireless computer networks. In addition, while the invention may be embodied in software, the functions necessary to implement the invention may optionally or alternatively be embodied in part or in whole using firmware and/or hardware components, such as combinatorial logic, Application Specific Integrated Circuits (ASICs), Field-Programmable Gate Arrays (FPGAs) or other hardware or some combination of hardware, software and/or firmware components.
While examples of embodiments of the system and method of the invention have been discussed in reference to the gas-cloud detection, monitoring, and quantification (including but not limited to greenhouse gases such as Hydrogen Sulfide, as well as Carbon Dioxide, Carbon Monoxide, Nitrogen Oxide as well as hydrocarbon gases such as Methane, Ethane, Propane, n-Butane, iso-Butane, n-Pentane, iso-Pentane, neo-Pentane, Sulfur Hexafluoride, Ammonia, Benzene, p- and m-Xylene, Vinyl chloride, Toluene, Propylene oxide, Propylene, Methanol, Hydrazine, Ethanol, 1,2-dichloroethane, 1,1-dichloroethane, Dichlorobenzene, Chlorobenzene, to name just a few), embodiments of the invention can be readily adapted for other chemical detection applications. For example, detection of liquid and solid chemical spills, biological weapons, tracking targets based on their chemical composition, identification of satellites and space debris, opthalmological imaging, microscopy and cellular imaging, endoscopy, mold detection, fire and flame detection, and pesticide detection are within the scope of the invention.
As used herein, a phrase referring to “at least one of” a list of items refers to any combination of those items, including single members. As an example, “at least one of: a, b, or c” is intended to cover: a, b, c, a-b, a-c, b-c, and a-b-c.
If implemented in software, the functions may be stored on or transmitted over as one or more instructions or code on a computer-readable medium. The steps of a method or algorithm disclosed herein may be implemented in a processor-executable software module which may reside on a computer-readable medium. Computer-readable media includes both computer storage media and communication media including any medium that can be enabled to transfer a computer program from one place to another. A storage media may be any available media that may be accessed by a computer. By way of example, and not limitation, such computer-readable media may include RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium that may be used to store desired program code in the form of instructions or data structures and that may be accessed by a computer. Also, any connection can be properly termed a computer-readable medium. Disk and disc, as used herein, includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk, and blu-ray disc where disks usually reproduce data magnetically, while discs reproduce data optically with lasers. Combinations of the above also may be included within the scope of computer-readable media. Additionally, the operations of a method or algorithm may reside as one or any combination or set of codes and instructions on a machine readable medium and computer-readable medium, which may be incorporated into a computer program product.
Various modifications to the implementations described in this disclosure may be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other implementations without departing from the spirit or scope of this disclosure. Thus, the claims are not intended to be limited to the implementations shown herein, but are to be accorded the widest scope consistent with this disclosure, the principles and the novel features disclosed herein.
Certain features that are described in this specification in the context of separate implementations also can be implemented in combination in a single implementation. Conversely, various features that are described in the context of a single implementation also can be implemented in multiple implementations separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a subcombination or variation of a subcombination.
This application claims priority to U.S. Provisional Patent Application No. 62/168,620, filed May 29, 2015, the contents of which are hereby incorporated by reference herein in their entirety and for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
3841763 | Lewis | Oct 1974 | A |
3849005 | Girard et al. | Nov 1974 | A |
4134683 | Goetz et al. | Jan 1979 | A |
4464789 | Sternberg | Aug 1984 | A |
4933555 | Smith | Jun 1990 | A |
4963963 | Dorman | Oct 1990 | A |
4965448 | Morse et al. | Oct 1990 | A |
5127742 | Fraden | Jul 1992 | A |
5136421 | Sagan | Aug 1992 | A |
5157258 | Gunning, III et al. | Oct 1992 | A |
5354987 | MacPherson | Oct 1994 | A |
5550373 | Cole et al. | Aug 1996 | A |
5559336 | Kosai et al. | Sep 1996 | A |
5604346 | Hamrelius et al. | Feb 1997 | A |
5822222 | Kaplinsky et al. | Oct 1998 | A |
5877500 | Braig et al. | Mar 1999 | A |
5920066 | DiRenzo et al. | Jul 1999 | A |
5926283 | Hopkins | Jul 1999 | A |
5973844 | Burger | Oct 1999 | A |
5994701 | Tsuchimoto et al. | Nov 1999 | A |
6023061 | Bodkin | Feb 2000 | A |
6097034 | Weckström | Aug 2000 | A |
6184529 | Contini | Feb 2001 | B1 |
6268883 | Zehnder et al. | Jul 2001 | B1 |
6456261 | Zhang | Sep 2002 | B1 |
6465785 | McManus | Oct 2002 | B1 |
6556853 | Cabib et al. | Apr 2003 | B1 |
6680778 | Hinnrichs et al. | Jan 2004 | B2 |
6700527 | Martin et al. | Mar 2004 | B1 |
7109488 | Milton | Sep 2006 | B2 |
7119337 | Johnson et al. | Oct 2006 | B1 |
7242478 | Dombrowski et al. | Jul 2007 | B1 |
7315377 | Holland et al. | Jan 2008 | B2 |
7321119 | King | Jan 2008 | B2 |
7364697 | McFarland et al. | Apr 2008 | B2 |
7433042 | Cavanaugh et al. | Oct 2008 | B1 |
7606484 | Richards et al. | Oct 2009 | B1 |
7634157 | Richards et al. | Dec 2009 | B1 |
7750802 | Parish et al. | Jul 2010 | B1 |
7835002 | Muhammed et al. | Nov 2010 | B2 |
7888624 | Murguia et al. | Feb 2011 | B1 |
8027041 | Mitchell et al. | Sep 2011 | B1 |
8153980 | Brady et al. | Apr 2012 | B1 |
8159568 | Ahdoot | Apr 2012 | B2 |
8212213 | Myrick et al. | Jul 2012 | B2 |
8373757 | Nguyen | Feb 2013 | B1 |
8629930 | Brueckner et al. | Jan 2014 | B2 |
8653461 | Benson et al. | Feb 2014 | B1 |
8654328 | Tkaczyk et al. | Feb 2014 | B2 |
8686364 | Little et al. | Apr 2014 | B1 |
9225913 | Ekdahl | Dec 2015 | B2 |
9395516 | Katsunuma et al. | Jul 2016 | B2 |
9562849 | Kester et al. | Feb 2017 | B2 |
9599508 | Kester et al. | Mar 2017 | B2 |
9625318 | Kester et al. | Apr 2017 | B2 |
9641772 | Yujiri | May 2017 | B2 |
9644562 | Fujita | May 2017 | B2 |
9756263 | Kester et al. | Sep 2017 | B2 |
10084975 | Kester et al. | Sep 2018 | B2 |
10254166 | Kester et al. | Apr 2019 | B2 |
10267686 | Kester et al. | Apr 2019 | B2 |
10375327 | Kester | Aug 2019 | B2 |
20010040216 | Knauth et al. | Nov 2001 | A1 |
20020015151 | Gorin | Feb 2002 | A1 |
20020121370 | Kurkjian et al. | Sep 2002 | A1 |
20020159101 | Alderson et al. | Oct 2002 | A1 |
20030102435 | Myers et al. | Jun 2003 | A1 |
20030134426 | Jiang et al. | Jul 2003 | A1 |
20030183756 | Huniu | Oct 2003 | A1 |
20040093167 | Braig et al. | May 2004 | A1 |
20040111232 | Butler et al. | Jun 2004 | A1 |
20040252300 | Slater | Dec 2004 | A1 |
20050029453 | Allen et al. | Feb 2005 | A1 |
20050057366 | Kadwell et al. | Mar 2005 | A1 |
20050103989 | Watson et al. | May 2005 | A1 |
20060044562 | Hagene et al. | Mar 2006 | A1 |
20060183241 | Lehmann | Aug 2006 | A1 |
20060232675 | Chamberlain et al. | Oct 2006 | A1 |
20060279632 | Anderson | Dec 2006 | A1 |
20070018105 | Grimberg | Jan 2007 | A1 |
20070075888 | Kelly et al. | Apr 2007 | A1 |
20070108385 | Mantese et al. | May 2007 | A1 |
20070170359 | Syllaios et al. | Jul 2007 | A1 |
20070170363 | Schimert et al. | Jul 2007 | A1 |
20080170140 | Silver et al. | Jul 2008 | A1 |
20080204744 | Mir et al. | Aug 2008 | A1 |
20080231719 | Benson et al. | Sep 2008 | A1 |
20080251724 | Baliga et al. | Oct 2008 | A1 |
20090015824 | Shubinsky et al. | Jan 2009 | A1 |
20090252650 | Lakshmanan | Oct 2009 | A1 |
20100162206 | Roth et al. | Jun 2010 | A1 |
20100171866 | Brady et al. | Jul 2010 | A1 |
20100211333 | Pruet et al. | Aug 2010 | A1 |
20100309467 | Fox et al. | Dec 2010 | A1 |
20110176577 | Bandara et al. | Jul 2011 | A1 |
20110185048 | Yew et al. | Jul 2011 | A1 |
20110261321 | Ramella-Roman et al. | Oct 2011 | A1 |
20120273680 | Furry | Nov 2012 | A1 |
20130181836 | Cardoso et al. | Jul 2013 | A1 |
20130206990 | Hsu et al. | Aug 2013 | A1 |
20130228887 | Wehner et al. | Sep 2013 | A1 |
20130235256 | Kodama | Sep 2013 | A1 |
20130250124 | Furry | Sep 2013 | A1 |
20130307991 | Olsen et al. | Nov 2013 | A1 |
20130321806 | Kester et al. | Dec 2013 | A1 |
20130341509 | Nelson et al. | Dec 2013 | A1 |
20130342680 | Zeng et al. | Dec 2013 | A1 |
20140002639 | Cheben et al. | Jan 2014 | A1 |
20140139643 | Högasten et al. | May 2014 | A1 |
20140320843 | Streuber et al. | Oct 2014 | A1 |
20150069239 | Kester et al. | Mar 2015 | A1 |
20150136981 | Kester et al. | May 2015 | A1 |
20150136982 | Kester et al. | May 2015 | A1 |
20150138534 | Tidhar | May 2015 | A1 |
20150144770 | Choi | May 2015 | A1 |
20150226613 | Bauer et al. | Aug 2015 | A1 |
20150288894 | Geelen et al. | Oct 2015 | A1 |
20150292948 | Goldring et al. | Oct 2015 | A1 |
20150316473 | Kester et al. | Nov 2015 | A1 |
20160037089 | Silny et al. | Feb 2016 | A1 |
20160041095 | Rothberg et al. | Feb 2016 | A1 |
20160097713 | Kester et al. | Apr 2016 | A1 |
20160097714 | Zeng | Apr 2016 | A1 |
20160238454 | Pillans | Aug 2016 | A1 |
20160245698 | Pau et al. | Aug 2016 | A1 |
20160313181 | Golub et al. | Oct 2016 | A1 |
20160356702 | Hinnrichs | Dec 2016 | A1 |
20160380014 | Ganapathi et al. | Dec 2016 | A1 |
20170026588 | Kester et al. | Jan 2017 | A1 |
20170234761 | Augusto | Aug 2017 | A1 |
20170248517 | Scherer et al. | Aug 2017 | A1 |
20170347037 | Hall et al. | Nov 2017 | A1 |
20170350758 | Kester et al. | Dec 2017 | A1 |
20170356802 | Kester et al. | Dec 2017 | A1 |
20180188163 | Kester et al. | Jul 2018 | A1 |
20180191967 | Kester | Jul 2018 | A1 |
20190003984 | Kester et al. | Jan 2019 | A1 |
20190137388 | Mallery et al. | May 2019 | A1 |
20190273875 | Kester et al. | Sep 2019 | A1 |
Number | Date | Country |
---|---|---|
2 365 866 | Sep 2000 | CA |
2 787 303 | Jul 2011 | CA |
2 870 419 | May 2015 | CA |
0 837 600 | Apr 1998 | EP |
2 871 452 | May 2015 | EP |
2 942 615 | Nov 2015 | EP |
2 955 496 | Dec 2015 | EP |
3 040 706 | Jul 2016 | EP |
1014769 | Dec 1965 | GB |
2518224 | Mar 2015 | GB |
2013-128185 | Jun 2013 | JP |
WO 2004097389 | Nov 2004 | WO |
WO 2007008826 | Jan 2007 | WO |
WO 2008109183 | Sep 2008 | WO |
WO 2009094782 | Aug 2009 | WO |
WO 2010053979 | May 2010 | WO |
WO 2012078417 | Jun 2012 | WO |
WO 2012082366 | Jun 2012 | WO |
WO 2013173541 | Nov 2013 | WO |
WO 2015108236 | Jul 2015 | WO |
WO 2016196224 | Dec 2016 | WO |
WO 2017201194 | Nov 2017 | WO |
WO 2018075957 | Apr 2018 | WO |
WO 2018075964 | Apr 2018 | WO |
WO 2018156795 | Aug 2018 | WO |
WO 2019094639 | May 2019 | WO |
Entry |
---|
US 10,113,914 B2, 10/2018, Kester et al. (withdrawn) |
Cossel KC, Adler F, Bertness KA, Thorpe MJ, Feng J, Raynor MW, Ye J. Analysis of trace impurities in semiconductor gas viaácavity-enhanced direct frequency comb spectroscopy. Applied Physics B. Sep. 1, 2010;100(4):917-24 (Year: 2010). |
Allen et al., “Measurements of Methane Emissions at Natural Gas Production Sites in the United States”, PNAS, Oct. 29, 2013, vol. 110, No. 44, pp. 7. |
Alvarez et al., “Greater Focus Needed on Methane Leakage from Natural Gas Infrastructure”, PNAS, Apr. 24, 2012, vol. 109, No. 17, pp. 12. |
Bedard et al., “Image Mapping Spectrometry: Calibration and Characterization”, Optical Engineering, Nov. 2012, vol. 51, No. 11, pp. 111711-1-111711-13. |
Ben-David et al., “Probability Theory for 3-Layer Remote Sensing Radiative Transfer Model: Univariate Case,” Optics Express, Apr. 2012, vol. 20, No. 9, pp. 10004-10033. |
Ben-David et al., “Probability Theory for 3-Layer Remote Sensing Radiative Transfer Model: Errata,” Optics Express, May 20, 2013, vol. 21, No. 10, pp. 11852. |
Brady et al., “Multiscale Lens Design”, Optics Express, Jun. 22, 2009, vol. 17, No. 13, pp. 10659-10674. |
Brochure provided by Lofty Designs to Rebellion Photonics on Oct. 31, 2012 as noted from the email. Subsequent to that date brochure was used in connection with potential customers. |
Caulton et al., “Toward a Better Understanding and Quantification of Methane Emissions from Shale Gas Development”, PNAS, Apr. 29, 2014, vol. 111, No. 17, pp. 7. |
Chen et al., “Quantitative Sectioning and Noise Analysis for Structured Illumination Microscopy: Erratum”, Optics Express, Oct. 19, 2015, vol. 23, No. 21, pp. 27633-27634. |
Chidley et al., “Flow-Induced Birefringence: The Hidden PSF Killer in High Performance Injection-Molded Plastic Optics”, Endoscopic Microscopy, Proceedings of SPIE vol. 6082, 2006, pp. 11. |
Chu et al., “The NIST Quantitative Infrared Database”, Journal of Research of the National Institute of Standards and Technology, Jan.-Feb. 1999, vol. 104, No. 1, pp. 59-81. |
DiPietro et al., “Hyperspectral Matched Filter with False-Alarm Mitigation”, Optical Engineering, Jan. 2012, vol. 51, No. 1, pp. 016202-1-016202-7. |
“Directed Inspection and Maintenance at Gas Processing Plants and Booster Stations,” United States Environmental Protection Agency Air and Radiation (6202J), EPA430-B-03-018, Oct. 2003 available at https://www3.epa.gov/gasstar/documents/II_dimgasproc.pdf. |
Eriksson et al., “Radiative Cooling Computed for Model Atmospheres”, Applied Optics, Dec. 1, 1982, vol. 21, No. 23, pp. 4381-4388. |
Flanigan, “Detection of Organic Vapors with Active and Passive Sensors: A Comparison,” Applied Optics, 1986, vol. 25, No. 23, pp. 4253-4260. |
Gålfalk et al., “Making Methane Visable”, Nature Climate Change, Apr. 2016, vol. 6, pp. 426-430. |
Gålfalk et al., “Making Methane Visable”, Supplementary Information, Nature Climate Change, 2015, pp. 1-14. |
Gallagher et al., “Error Analysis for Estimation of Trace Vapor Concentration Pathlength in Stack Plumes”, Applied Spectroscopy, 2003, vol. 57, No. 6, pp. 614-621. |
Gallagher et al., “Estimation of Trace Vapor Concentration-Pathlength in Plumes for Remote Sensing Applications from Hyperspectral Images”, Analytica Chimica Acta, 2003, vol. 490, pp. 139-152. |
Gao et al., “Compact Image Slicing Spectrometer (ISS) for Hyperspectral Fluorescence Microscopy”, Optics Express, Jul. 20, 2009, vol. 17, No. 15, pp. 12293-12308. |
Gao et al., “Depth-Resolved Image Mapping Spectrometer (IMS) with Structured Illumination”, Optics Express, Aug. 29, 2011, vol. 19, No. 18, pp. 17439-17452. |
Gao et al., “Optical Design of a Snapshot High-Sampling Image Mapping Spectrometer (IMS) for Hyperspectral Microscopy”, Three-Dimensional and Multidimensional Microscopy:Image Acquisition and Processing XVII, Proceedings of SPIE vol. 7570, 2010, pp. 1-7. |
Gao et al., “Quantitative Comparison Between Full-Spectrum and Filter-Based Imaging in Hyperspectral Fluorescence Microscopy”, Journal of Microscopy, 2012, vol. 246, No. 2, pp. 113-123. |
Gao et al., “Snapshot Image-Mapping Spectrometer for Hyperspectral Fluorescence Microscopy”, Optics and Photonics News, Nov. 2010, vol. 21, No. 12, p. 50. |
Gao et al., “Snapshot Image Mapping Spectrometer (IMS) with High Sampling Density for Hyperspectral Microscopy”, Optics Express, Jul. 5, 2010, vol. 18, No. 4, pp. 14330-14344. |
Gerhart et al., “Detection and Tracking of Gas Plumes in LWIR Hyperspectral Video Sequence Data,” Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XIX, 2013, SPIE Proceedings vol. 8743, pp. 1-14. |
Gittins, Christopher M., “Detection and Characterization of Chemical Vapor Fugitive Emissions by Nonlinear Optimal Estimation: Theory and Simulation”, Applied Optics, Aug. 10, 2009, vol. 48, No. 23, pp. 4545-4561. |
Goldberg et al., “Dual Band MWIR/LWIR Focal Plane Array Test Results,” Army Research Lab, Adelphi, MD, Aug. 1999, pp. 18. |
Golowich et al., “Performance Limits of LWIR Gaseous Plume Quantification”, Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XVII, 2011, Proceedings of SPIE vol. 8048, pp. 1-12. |
Griffin et al., “The Herschel—SPIRE 1-15 Instrument and its In-Flight Performance,” Astronomy and Astrophysics, Jul. 1, 2010, vol. 518, pp. 7. |
Gross et al., “Remote Identification and Quantification of Industrial Smokestack Effluents via Imaging Fourier-Transform Spectroscopy”, Environmental Science & Technology, 2010, vol. 44, No. 24, pp. 9390-9397. |
Gupta et al., “Miniature Snapshot Multispectral Imager,” Optical Engineering, 2011, vol. 50, pp. 033203-1-033203-9. |
Hadlington, Simon, “New Camera Makes Methane Visible”, Chemistry World, http://web.archive.org/web/20160305234907/http://www.rsc.org/chemistryworld/2015/12/methane-camera-infrared-greenhouse-gas, Dec. 14, 2015, pp. 2. |
Hagen et al., “Analysis of Computed Tomographic Imaging Spectrometers. I. Spatial and Spectral Resolution”, Applied Optics, Oct. 1, 2008, vol. 47, No. 28, pp. F85-F95. |
Hagen et al., “Coded Aperture DUV Spectrometer for Standoff Raman Spectoscopy”, Next-Generation Spectroscopic Technologies II, Proceedings of SPIE vol. 7319, 2009, pp. 1-10. |
Hagen et al., “Compound Prism Design Principles, I”, Applied Optics, Sep. 1, 2011, vol. 50, No. 25, pp. 4998-5011. |
Hagen et al., “Compound Prism Design Principles, II: Triplet and Janssen Prisms”, Applied Optics, Sep. 1, 2011, vol. 50, No. 25, pp. 5012-5022. |
Hagen et al., “Compound Prism Design Principles, III: Linear-in-Wavenumber and Optical Coherence Tomography Prisms”, Applied Optics, Sep. 1, 2011, vol. 50, No. 25, pp. 5023-5030. |
Hagen et al., “Fourier Methods of Improving Reconstruction Speed for CTIS Imaging Spectrometers”, Imagin Spectrometry XII, Proceedings of SPIE vol. 6661, 2007, pp. 11. |
Hagen et al., “Foveated Endoscopic Lens”, Journal of Biomedical Optics, Feb. 2012, vol. 17, No. 2, pp. 021104-1-021104-6. |
Hagen et al., “Gaussian Profile Estimation in One Dimension”, Applied Optics, Aug. 1, 2007, vol. 46, No. 22, pp. 5374-5383. |
Hagen et al., “Gaussian Profile Estimation in Two Dimension”, Applied Optics, Dec. 20, 2008, vol. 47, No. 36, pp. 6842-6851. |
Hagen et al., “Quantitative Sectioning and Noise Analysis for Structured Illumination Microscopy”, Optics Express, Jan. 2, 2012, vol. 20, No. 1, pp. 403-413. |
Hagen et al., “Quantitative Sectioning and Noise Analysis for Structured Illumination Microscopy: Errata”, Optics Express, Feb. 27, 2012, vol. 20, No. 5, pp. 5343. |
Hagen et al., “Real-Time Quantitative Hydrocarbon Gas Imaging with the Gas Cloud Imager (GCI)”, Proceedings of SPIE, vol. 8358, Chemical, Biological, Radiological, Nuclear, and Explosives (CBRNE) Sensing XIII, May 1, 2012, pp. 7. |
Hagen et al., “Review of Snapshot Spectral Imaging Technologies”, Optical Engineering, Sep. 2013, vol. 52, No. 9, pp. 090901-1-090901-23. |
Hagen et al., “Snapshot Advantage: A Review of the Light Collection Improvement for Parallel High-Dimensional Measurement Systems,” Optical Engineering, Jun. 13, 2012, vol. 51, No. 11, p. 111702-1-111702-7. |
Hagen et al., “Snapshot Mueller Matrix Spectropolarimeter” Optics Letters, Aug. 1, 2007, vol. 32, No. 15, pp. 2100-2102. |
Hagen et al., “Spectrally-Resolved Imaging of Dynamic Turbid Media”, Multimodal Biomedical Imaging VI, Proceedings of SPIE vol. 7892, 2011, pp. 1-7. |
Hagen et al., “Video-Rate Spectral Imaging of Gas Leaks in the Longwave Infrared,” Chemical, Biological, Radiological, Nuclear, and Explosives (CBRNE) Sensing XIV, May 29, 2013, SPIE Proceedings vol. 8710, pp. 7. |
Harley et al., “Remote Quantification of Smokestack Effluent Mass Flow Rates Using Imaging Fourier Transform Spectrometry,” Chemical, Biological, Radiological, Nuclear, and Explosives (CBRNE) Sensing XII, Apr. 25-29, 2011, SPIE Proceedings vol. 8018, pp. 1-13. |
Hayden et al., “Determination of Trace-Gas Amounts in Plumes by the Use of Orthogonal Digital Filtering of Thermal-Emission Spectra”, Applied Optics, Jun. 1, 1996, vol. 35, No. 16, pp. 2802-2809. |
Hirsch et al., “Detection of Gaseous Plumes in IR Hyperspectral Images Using Hierarchical Clustering”, Applied Optics, Sep. 1, 2007, vol. 46, No. 25, pp. 6368-6374. |
Johnston et al., “A Real-Time FPGA Implementation of a Barrel Distortion Correction Aglorithm”, Projects, 2003, vol. 10, pp. 91-96. |
Karion et al., “Methane Emissions Estimate from Airborne Measurements Over a Western United States Natural Gas Field”, Geophysical Research Letters, 2013, vol. 40, pp. 4393-4397. |
Keshava et al., “A Survey of Spectral Unmixing Algorithms”, Lincoln Laboratory Journal, 2003, vol. 14, No. 1, pp. 55-78. |
Kester et al., “A Real-Time Gas Cloud Imaging Camera for Fugitive Emission Detection and Monitoring”, Imaging and Applied Optics Technical Digest, 2012, pp. 3. |
Kester et al., “Development of Image Mappers for Hyperspectral Biomedical Imaging Applications”, Applied Optics, Apr. 1, 2010, vol. 49, No. 10, pp. 1886-1899. |
Kester et al., “High Numerical Aperture Microendoscope Objective for a Fiber Confocal Reflectance Microscope”, Optics Express, Mar. 5, 2007, vol. 15. No. 5, pp. 2409-2420. |
Kester et al., “Low Cost, High Performance, Self-Aligning Miniature Optical Systems”, Applied Optics, Jun. 20, 2009, vol. 48, No. 18, pp. 3375-3384. |
Kester et al., “Real-Time Snapshot Hyperspectral Imaging Endoscope”, Journal of Biomedical Optics, May 2011, vol. 16, No. 5, pp. 056005-1-056005-12. |
Kudenov et al., “Fourier Transform Channeled Spectropolarimetry in the MWIR”, Optics Express, Oct. 1, 2007, vol. 15, No. 20, pp. 12792-12805. |
Kudenov et al., “Snapshot Imaging Mueller Matrix Polarimeter Using Polarization Gratings”, Optics Letters, Apr. 15, 2012, vol. 37, No. 8, pp. 1367-1369. |
Landau et al., “Design and Evaluation of an Ultra-Slim Objective for in-vivo Deep Optical Biopsy”, Optics Express, Mar. 1, 2010, vol. 18, No. 5, pp. 4758-4775. |
Levi, Michael A., “Comment on ‘Hydrocarbon Emissions Characterization in the Colorado Front Range: A Pilot Study’ by Gabrielle Pétron et al.”, Journal of Geophysical Research, 2012, vol. 117, No. D21203, pp. 1-5. |
Levi, Michael A., “Reply to “‘Reply to ‘Comment on ‘Hydrocarbon Emissions Characterization in the Colorado Front Range—A Pilot Study’ by Michael A. Levi”’ by Gabrielle Pétron et al.”, Journal of Geophysical Research: Atmospheres, 2013, vol. 118, pp. 3044-3046. |
Low et al., “Remote Sensing and Characterization of Stack Gases by Infrared Spectroscopy. An Approach by Using Multiple-Scan Interferometry”, Environmental Science & Technology, Jan. 1967, vol. 1, No. 1, pp. 73-74. |
Luo et al., “Fast Processing of Imaging Spectrometer Data Cube Based on FPGA Design”, MIPPR 2007: Multispectral Image Processing, Proceedings of SPIE vol. 6787, pp. 7. |
Manolakis et al., “Long-Wave Infrared Hyperspectral Remote Sensing of Chemical Clouds”, IEEE Signal Processing Magazine, Jul. 2014, vol. 31, No. 4, pp. 120-141. |
Mathews, “Design and Fabrication of a Low-Cost, Multispectral Imaging System,” Applied Optics, 2008, pp. F71-F76, vol. 47. |
Naranjo et al., “IR Gas Imaging in an Industrial Setting,” Thermosense XXXII, Published in SPIE Proceedings vol. 7661, May 4, 2010, pp. 1-8. |
Nguyen et al., “Snapshot 3D Optical Coherence Tomography System using Image Mapping Spectrometer”, Biomedical Optics and 3D Imaging OSA, 2012, pp. 3. |
Niu et al., “New Approach to Remote Gas-Phase Chemical Quantification: Selected-Band Algorithm”, Optical Engineering, Feb. 2014, vol. 53, No. 2, pp. 021111-1-021111-10. |
“Oil and Natural Gas Sector Leaks”, U.S. EPA Office of Air Quality Planning and Standards (OAQPS), Review Panel, Apr. 2014, pp. 63. |
Pétron et al., “Hydrocarbon Emissions Characterization in the Colorado Front Range: A Pilot Study”, Journal of Geophysical Research, 2012, vol. 117, No. D04304, pp. 1-19. |
Pétron et al., “Reply to Comment on ‘Hydrocarbon Emissions Characterization in the Colorado Front Range—A Pilot Study’ by Michael A. Levi”, Journal of Geophysical Research:Atmospheres, 2013, vol. 118, pp. 236-242. |
Pisano et al., “Thermal Illuminators for Far-Infrared and Submillimeter Astronomical Instruments,” Applied Optics, Jun. 1, 2005, vol. 44, No. 16, pp. 3208-3217. |
Polak et al., “Passive Fourier-Transform Infrared Spectroscopy of Chemical Plumes: An Algorithm for Quantitiative Interpretation and Real-Time Background Removal”, Applied Optics, Aug. 20, 1995, vol. 34, No. 24, pp. 5406-5412. |
Rebellion Photonics, “Gas Cloud Imaging Camera: A Breakthrough in Leak Monitoring for the Rig & Refinery Safety Market”, Presentation at SPIE Defense Security and Sensing, 28 pages, Apr. 29-May 3, 2013. |
Sandsten et al., “Volume Flow Calculations on Gas Leaks Imaged with Infrared Gas-Correlation,” Optics Express, 2012, vol. 20, No. 18, pp. 20318-20329. |
Shogenji et al., “Multispectral Imaging Using Compact Compound Optics,” Optics Express, Apr. 19, 2004, vol. 12, No. 8, pp. 1643-1655. |
Telops, “Hyper-Cam”, http://web.archive.org/web/20160608180941/http://www.telops.com/en/hyperspectral-cameras/hyper-cam as archived Jun. 8, 2016 in 2 pages. |
Telops, “Innovative Infrared Imaging”, http://web.archive.org/web/20160603212729/http://www.telops.com/en/ as archived Jun. 3, 2016 in 2 pages. |
Walter Jr., et al., “Detection of Atmospheric Pollutants: a Correlation Technique”, Applied Optics, Jun. 1975, vol. 14, No. 6, pp. 1423-1428. |
Williams et al., “Dual-Band MWIR/LWIR Radiometer for Absolute Temperature Measurements,” SPIE Thermosense Conference XXVIII—Manuscript 6205-23, Apr. 18, 2006, pp. 13. |
Young et al., “An In-Scene Method for Atmospheric Compensation of Thermal Hyperspectral Data”, Journal of Geophysical Research, 2002, vol. 107, No. D24, pp. 14-1-14-20. |
Zheng et al., “A Static Multiplex Fabry-Perot Spectrometer”, Sensors, Cameras, and Systems for Industrial/Scientific Applications X, Proceedings of SPIE-IS&T Electronic Imaging, SPIE vol. 7249, 2009, pp. 8. |
Zheng et al., “Analytic-Domain Lens Design with Proximate Ray Tracing”, Journal of the Optical Society of America A, Aug. 2010, vol. 27, No. 8, pp. 1791-1802. |
Preliminary Amendment as filed in U.S. Appl. No. 14/538,827 dated Jan. 28, 2015 in 6 pages. |
Office Action received in U.S. Appl. No. 14/538,827 dated Jun. 30, 2015 in 8 pages. |
Non-Final Office Action Response as filed in U.S. Appl. No. 14/538,827 dated Dec. 28, 2015 in 11 pages. |
Notice of Allowance received in U.S. Appl. No. 14/538,827 dated Feb. 1, 2016 in 18 pages. |
Corrected Notice of Allowance received in U.S. Appl. No. 14/538,827 dated Feb. 10, 2016 in 4 pages. |
Corrected Notice of Allowance received in U.S. Appl. No. 14/538,827 dated Feb. 22, 2016 in 4 pages. |
Request for Continued Examination and Response to Correct Application Papers as filed in U.S. Appl. No. 14/538,827 dated Apr. 29, 2016 in 14 pages. |
Notice of Allowance received in U.S. Appl. No. 14/538,827 dated May 26, 2016 in 9 pages. |
Office Action received in U.S. Appl. No. 14/543,692 dated Nov. 3, 2015 in 7 pages. |
Interview Summary received in U.S. Appl. No. 14/543,692 dated Feb. 17, 2016 in 5 pages. |
Response to Office Action as filed in U.S. Appl. No. 14/543,692 dated May 2, 2016 in 9 pages. |
Office Action received in U.S. Appl. No. 14/543,692 dated Jun. 1, 2016 in 18 pages. |
Office Action received in U.S. Appl. No. 14/539,899 dated Mar. 26, 2015 in 6 pages. |
Non-Final Office Action Response as filed in U.S. Appl. No. 14/539,899 dated Aug. 26, 2015 in 8 pages. |
Final Office Action received in U.S. Appl. No. 14/539,899 dated Dec. 11, 2015 in 9 pages. |
Amendment as filed in U.S. Appl. No. 14/539,899 dated Jun. 9, 2016 in 6 pages. |
Notice of Allowance received in U.S. Appl. No. 14/539,899 dated Jun. 21, 2016 in 17 pages. |
Extended European Search Report received in European Application No. 14192862.2 dated Mar. 30, 2015 in 10 pages. |
Official Communication received in European Application No. 14192862.2 dated Apr. 19, 2016 in 6 pages. |
International Search Report in PCT Application No. PCT/US2013/041278 dated Aug. 27, 2013 in 4 pages. |
International Preliminary Report on Patentability in PCT Application No. PCT/US2013/041278 dated Nov. 27, 2014 in 10 pages. |
Preliminary Amendment as filed in U.S. Appl. No. 14/700,791 dated Jul. 13, 2015 in 8 pages. |
Notice of Allowance received in U.S. Appl. No. 14/700,791 dated Jun. 9, 2016 in 11 pages. |
Extended European Search Report received in European Application No. 15165877.0 dated Oct. 8, 2015 in 12 pages. |
Preliminary Amendment as filed in U.S. Appl. No. 14/700,567 dated Jul. 10, 2015 in 6 pages. |
Publication Request as filed in U.S. Appl. No. 14/700,567 dated Aug. 24, 2016 in 237 pages. |
Extended European Search Report received in European Application No. EP 15165880.4 dated Nov. 24, 2015 in 8 pages. |
Preliminary Amendment as filed in U.S. Appl. No. 14/792,477 dated Dec. 21, 2015 in 7 pages. |
Adams, et al., “Advances in Detectors: Hot IR sensors improve IR camera size, weight, and power”, Laser Focus World, vol. 50, Issue 01, Jan. 17, 2014, 6 pages. Available at http://www.ircameras.com/articles/advances-detectors-hot-ir-sensors-improve-ir-camera-size-weight-power/. |
International Search Report in PCT Application No. PCT/US2016/034455 dated Oct. 24, 2016 in 12 pages. |
Notice of Allowance received in U.S. Appl. No. 14/538,827 dated Sep. 19, 2016 in 9 pages. |
Notice of Allowance received in U.S. Appl. No. 14/539,899 dated Oct. 31, 2016 in 10 pages. |
Notice of Allowance received in U.S. Appl. No. 14/700,791 dated Sep. 30, 2016 in 19 pages. |
Sandsten et al., “Development of Infrared Spectroscopy Techniques for Environmental Monitoring”, Doctoral Thesis, Aug. 2000, pp. 123. |
Sandsten et al., “Real-Time Gas-Correlation Imaging Employing Thermal Background Radiation”, Optics Express, Feb. 14, 2000, vol. 6, No. 4, pp. 92-103. |
Wikipedia entry https://en.wikipedia.org/wiki/Mobile_computing, last modified on Dec. 30, 2016; retrieved from the internet on Feb. 2, 2017 in 6 pages. |
Amendment as filed in U.S. Appl. No. 14/538,827 dated Dec. 16, 2016 in 9 pages. |
Official Communication received in U.S. Appl. No. 15/418,532 dated Jun. 23, 2017 in 7 pages. |
Amendment as filed in U.S. Appl. No. 15/418,532 dated Nov. 22, 2017 in 8 pages. |
Official Communication received in U.S. Appl. No. 15/418,532 dated Dec. 11, 2017 in 21 pages. |
Notice of Allowance received in U.S. Appl. No. 15/418,532 dated Jun. 15, 2018 in 12 pages. |
Corrected Notice of Allowance received in U.S. Appl. No. 15/418,532 dated Jul. 6, 2018 in 3 pages. |
Response to Final Action as filed in U.S. Appl. No. 14/543,692 dated Nov. 30, 2016 in 12 pages. |
Notice of Allowance received in U.S. Appl. No. 14/543,692 dated Dec. 9, 2016 in 12 pages. |
Amendment after Allowance as filed in U.S. Appl. No. 14/543,692 dated Mar. 3, 2017 in 6 pages. |
Notice of Allowance received in U.S. Appl. No. 14/543,692 dated Mar. 17, 2017 in 4 pages. |
Preliminary Amendment as filed in U.S. Appl. No. 15/471,398 dated Oct. 6, 2017 in 6 pages. |
Notice of Allowance received in U.S. Appl. No. 14/571,398 dated Oct. 18, 2017 in 8 pages. |
Notice of Allowance received in U.S. Appl. No. 14/571,398 dated Feb. 7, 2018 in 20 pages. |
Notice of Allowance received in U.S. Appl. No. 14/571,398 dated Jul. 2, 2018 in 8 pages. |
Notice of Allowance received in U.S. Appl. No. 14/571,398 dated Oct. 24, 2018 in 7 pages. |
Amendment as filed in U.S. Appl. No. 14/539,899 dated Jan. 27, 2017 in 5 pages. |
Official Communication received in U.S. Appl. No. 15/462,352 dated Sep. 28, 2017 in 6 pages. |
Amendment as filed in U.S. Appl. No. 15/462,352 dated Feb. 28, 2018 in 5 pages. |
Notice of Allowance received in U.S. Appl. No. 15/462,352 dated Jul. 17, 2018 in 25 pages. |
Notice to File Corrected Application Papers received in U.S. Appl. No. 15/462,352 dated Aug. 8, 2018 in 3 pages. |
Response to Notice to File Corrected Application Papers filed in U.S. Appl. No. 15/462,352 dated Oct. 8, 2018 in 3 pages. |
Notice of Allowance received in U.S. Appl. No. 15/462,352 dated Oct. 31, 2018 in 9 pages. |
Official Communication received in European Application No. 13732285.5 dated Jul. 26, 2018 in 6 pages. |
Official Communication received in European Application No. 14192862.2 dated May 2, 2018 in 3 pages. |
Notice of Allowance received in U.S. Appl. No. 14/700,791 dated Feb. 21, 2017 in 20 pages. |
Comments on Allowance filed in U.S. Appl. No. 14/700,791 dated May 19, 2017 in 2 pages. |
Notice of Allowance received in U.S. Appl. No. 14/700,791 dated Jul. 10, 2017 in 24 pages. |
Preliminary Amendment as filed in U.S. Appl. No. 15/623,942 dated Dec. 7, 2017 in 6 pages. |
Notice of Allowance received in U.S. Appl. No. 15/623,942 dated Jan. 24, 2018 in 22 pages. |
Notice of Allowance received in U.S. Appl. No. 15/623,942 dated May 24, 2018 in 23 pages. |
Comments on Allowance filed in U.S. Appl. No. 15/623,942 dated Aug. 23, 2018 in 2 pages. |
Official Communication received in European Application No. 15165877.0 dated Jan. 3, 2017 in 9 pages. |
Official Communication received in U.S. Appl. No. 14/700,567 dated Jun. 14, 2017 in 29 pages. |
Amendment as filed in U.S. Appl. No. 14/700,567 dated Dec. 13, 2017 in 12 pages. |
Official Communication received in U.S. Appl. No. 14/700,567 dated Mar. 5, 2018 in 38 pages. |
Amendment as filed in U.S. Appl. No. 14/700,567 dated Jul. 5, 2018 in 10 pages. |
Office Action as filed in U.S. Appl. No. 14/700,567 dated Aug. 27, 2018 in 36 pages. |
Official Communication received in U.S. Appl. No. 14/792,477 dated Jan. 27, 2017 in 10 pages. |
Response to Restriction Requirement submitted in U.S. Appl. No. 14/792,477 dated May 8, 2017 in 6 pages. |
Official Communication received in U.S. Appl. No. 14/792,477 dated Jul. 19, 2017 in 20 pages. |
Amendment as filed in U.S. Appl. No. 14/792,477 dated Jan. 18, 2018 in 10 pages. |
Notice of Allowance received in U.S. Appl. No. 14/792,477 dated Apr. 19, 2018 in 13 pages. |
Notice of Allowance received in U.S. Appl. No. 14/792,477 dated Sep. 20, 2018 in 14 pages. |
International Preliminary Report on Patentability in PCT Application No. PCT/US2016/034455 dated Dec. 5, 2017 in 8 pages. |
Preliminary Amendment as filed in U.S. Appl. No. 15/789,811 dated Mar. 20, 2018 in 6 pages. |
Official Communication received in U.S. Appl. No. 15/789,811 dated Jul. 27, 2018 in 22 pages. |
Invitation to Pay Additional Fees in PCT Application No. PCT/US2017/057725 dated Dec. 14, 2017 in 3 pages. |
International Search Report in PCT Application No. PCT/US2017/057725 dated Feb. 14, 2018 in 14 pages. |
Preliminary Amendment as filed in U.S. Appl. No. 15/789,829 dated Mar. 20, 2018 in 8 pages. |
Official Communication received in U.S. Appl. No. 15/789,829 dated Jun. 5, 2018 in 16 pages. |
Invitation to Pay Additional Fees in PCT Application No. PCT/US2017/057712 dated Jan. 10, 2018 in 2 pages. |
International Search Report in PCT Application No. PCT/US2017/057712 dated Mar. 6, 2018 in 12 pages. |
Preliminary Amendment as filed in U.S. Appl. No. 15/902,336 dated Sep. 20, 2018 in 9 pages. |
International Search Report in PCT Application No. PCT/US2018/019271 dated Jun. 27, 2018 in 15 pages. |
ARPA-E, “Portable Methane Detection System”, dated Dec. 16, 2014 (including innovation update from May 2018) in 2 pages https://arpa-e.energy.gov/?q=slick-sheet-project/portable-methane-detection-system. |
ARPA-E, “Wearable, Continuously Monitoring Methane Imagers”, as updated Jan. 15, 2018 in 2 pages https:/arpa-e.energy.gov/sites/default/files/Rebellion-MONITOR-May1.pdf. |
Weldon et al., “H2S and CO2 gas sensing using DFB laser diodes emitting at 1.57 μm”, Sensors and Actuators B: Chemical, Oct. 1995, vol. 29, Issues 1-3, pp. 101-107. |
Amendment after Allowance as filed in U.S. Appl. No. 15/418,532 dated Sep. 14, 2018 in 6 pages. |
Notice of Allowance received in U.S. Appl. No. 15/418,532 dated Dec. 5, 2018 in 11 pages. |
Amendment After Allowance as filed in U.S. Appl. No. 15/471,398 dated Jan. 24, 2019 in 5 pages. |
Notice of Allowance received in U.S. Appl. No. 14/571,398 dated Feb. 27, 2019 in 14 pages. |
Notice of Allowance received in U.S. Appl. No. 14/571,398 dated Mar. 6, 2019 in 5 pages. |
Notice of Allowance received in U.S. Appl. No. 15/462,352 dated Feb. 12, 2019 in 9 pages. |
Amendment as filed in U.S. Appl. No. 15/462,352 dated Apr. 30, 2019 in 5 pages. |
Notice of Allowance received in U.S. Appl. No. 15/462,352 dated May 23, 2019, 2019 in 10 pages. |
Official Communication received in Canadian Application No. 2,873,989 dated Mar. 21, 2019 in 6 pages. |
Preliminary Amendment as filed in U.S. Appl. No. 16/138,823 dated May 23, 2019 in 5 pages. |
Notice of Allowance received in U.S. Appl. No. 16/138,823 dated Jun. 14, 2019 in 10 pages. |
Notice of Allowance received in U.S. Appl. No. 14/792,477 dated Jan. 30, 2019 in 11 pages. |
Notice of Allowance received in U.S. Appl. No. 14/792,477 dated Jun. 21, 2019 in 10 pages. |
Extended European Search Report received in European Application No. EP 16804077.2 dated Jan. 8, 2019 in 8 pages. |
International Preliminary Report on Patentability in PCT Application No. PCT/US2017/057725 dated May 2, 2019 in 10 pages. |
Amendment as filed in U.S. Appl. No. 15/789,829 dated Dec. 4, 2018 in 9 pages. |
Notice of Allowance received in U.S. Appl. No. 15/789,829 dated Feb. 25, 2019 in 28 pages. |
Amendment as filed in U.S. Appl. No. 15/789,829 dated May 24, 2019 in 7 pages. |
Notice of Allowance received in U.S. Appl. No. 15/789,829 dated Jul. 19, 2019 in 14 pages. |
International Preliminary Report on Patentability in PCT Application No. PCT/US2017/057712 dated May 2, 2019 in 9 pages. |
International Preliminary Report on Patentability in PCT Application No. PCT/US2018/019271 dated Sep. 6, 2019 in 11 pages. |
Official Communication received in U.S. Appl. No. 16/185,399 dated Apr. 2, 2019 in 24 pages. |
Amendment as filed in U.S. Appl. No. 16/185,399 dated Jul. 2, 2019 in 7 pages. |
Notice of Allowance received in U.S. Appl. No. 16/185,399 dated Jul. 26, 2019 in 9 pages. |
International Search Report in PCT Application No. PCT/US2018/059890 dated Jan. 23, 2019 in 10 pages. |
Number | Date | Country | |
---|---|---|---|
20160349228 A1 | Dec 2016 | US |
Number | Date | Country | |
---|---|---|---|
62168620 | May 2015 | US |