a illustrates a first set of embodiments of a first aspect of a range imaging LIDAR system;
b illustrates a transverse cross-section of a first embodiment of a beam of light;
c illustrates a transverse cross-section of a second embodiment of a beam of light;
d illustrates a second aspect of a Fabry-Pérot interferometer of a range imaging LIDAR system;
a illustrates a plot of signal intensity as a function of image distance of the fringe pattern illustrated in
b illustrates a plot of signal intensity as a function of range from the LIDAR system to the interaction region, corresponding to the plot illustrated in
a and 7b illustrate a circular image compression process operating on a fringe pattern from a Fabry-Pérot interferometer;
a illustrates a flow chart of a first aspect of a circular binning process;
b illustrates an alternate decision block of the first aspect of a circular binning process illustrated in
a illustrates a radial cross-section of an intensity distribution of a set of fringes from a Fabry-Pérot interferometer;
b illustrates fringes from the Fabry-Pérot interferometer from two scattered signals associated with different velocities;
c illustrates a fringe associated with a scatter signal channel processed by the Fabry-Pérot etalon, wherein the fringe comprises aerosol (Mie), molecular (Rayleigh) and background signal components;
a illustrates a first embodiment of a third aspect of an associated detection system of a range imaging LIDAR system;
b illustrates a plan view of a digital micromirror device (DVD) used in the embodiments illustrated in
a-e illustrate radial cross-sections through the complementary reflection patterns illustrated in
a illustrates a first embodiment of an eighth aspect of a range imaging LIDAR system incorporating a first aspect of the associated mask system and the first aspect of an associated detection system;
b illustrates a transverse cross-section of an expanded reference beam of light associated with the range imaging LIDAR system illustrated in
c illustrates a transverse cross-section of the expanded reference beam of light after passing through a mask associated with the range imaging LIDAR system illustrated in
d illustrates an image that would be produced by a Fabry-Pérot interferometer of the range imaging LIDAR system illustrated in
e illustrates an image from the Fabry-Pérot interferometer of the range imaging LIDAR system illustrated in
a illustrates a transverse cross-section of an expanded reference beam of light after passing through a mask associated with a first aspect of plural fringe patterns generated by a first variation of the eighth aspect of a range imaging LIDAR system used to process light signals from plurality of associated regions of interest;
b illustrates an image from the Fabry-Pérot interferometer of the range imaging LIDAR system associated with the image illustrated in
a illustrates a transverse cross-section of an expanded reference beam of light after passing through a mask associated with a second aspect of plural fringe patterns generated by a second variation of the eighth aspect of a range imaging LIDAR system used to process light signals from plurality of associated regions of interest;
b illustrates an image from the Fabry-Pérot interferometer of the range imaging LIDAR system associated with the image illustrated in
a illustrates a transverse cross-section of an expanded reference beam of light after passing through a mask associated with a third aspect of plural fringe patterns generated by a third variation of the eighth aspect of a range imaging LIDAR system used to process light signals from plurality of associated regions of interest;
b illustrates an image from the Fabry-Pérot interferometer of the range imaging LIDAR system associated with the image illustrated in
a illustrates a second embodiment of the eighth aspect of a range imaging LIDAR system incorporating the first aspect of the associated mask system and the third aspect of an associated detection system;
b-e illustrate various images associated with the second embodiment of the eighth aspect of a range imaging LIDAR system corresponding to corresponding images of
a-e illustrates a third embodiment of the eighth aspect of a range imaging LIDAR system incorporating a second aspect of the associated mask system and the third aspect of an associated detection system;
Referring to
A set of receiver optics 24, for example, a telescope 24′, laterally offset from the beam of light 18, provides for imaging a portion of the beam of light 18 onto an intermediate image plane 26, so as to provide for a one-to-one mapping of volumetric elements 28 within the beam of light 18 and corresponding associated regions or points 30 in the intermediate image plane 26. More particularly, the beam of light 18 illuminates molecules 16′ or aerosols 16″ of the atmosphere 16, or a combination thereof, within the interaction region 22, which in turn scatter the monochromatic light 14 of the beam of light 18. The resulting scattered light 32 within the field-of-view 34 of the receiver optics 24 is collected thereby and imaged onto the intermediate image plane 26. The receiver optics 24 is laterally offset from and points towards the beam of light 18, so that the optic axis 36 of the receiver optics 24 is inclined relative to the optic axis 38 of the beam of light 18 at an associated parallax angle θ. Accordingly, each volumetric element 28 of the beam of light 18 imaged onto a corresponding region or point 30 on the intermediate image plane 26 corresponds to a different nominal range R from the intermediate image plane 26 to a point 40 on the optic axis 38 of the beam of light 18 associated with the corresponding volumetric element 28. Accordingly, each region or point 30 on the intermediate image plane 26, corresponding to the volumetric element 28 of the beam of light 18 within the field-of-view 34 of the receiver optics 24, corresponds to a different nominal range R. Accordingly, different regions or points 30 of the intermediate image 42 in the intermediate image plane 26 correspond to different nominal ranges R to the beam of light 18, and therefore correspond to different nominal ranges R to the associated volumetric elements 28 thereof within the interaction region 22. For example, as illustrated in
The range imaging LIDAR system 10, 10i further comprises a Fabry-Pérot interferometer 44 having an input focal plane 44.1 and an output focal plane 44.2. The input focal plane 44.1 is collocated with the intermediate image plane 26 so as to receive scattered light 32 therefrom, which is then processed by the Fabry-Pérot interferometer 44 and imaged onto a detection system 46 located at the output focal plane 44.2. Between the input 44.1 and output 44.2 focal planes, the Fabry-Pérot interferometer 44 comprises a collimating lens 48, a Fabry-Pérot etalon 50, and imaging optics 52 spaced along an associated common optic axis 53, wherein the input focal plane 44.1 is a focal plane of the collimating lens 48, the output focal plane 44.2 is a focal plane of the imaging optics 52, and scattered light 32 at the input focal plane 44.1 entering the collimating lens 48 is substantially collimated thereby, then processed by the Fabry-Pérot etalon 50, and finally imaged onto the detection system 46 by the imaging optics 52. The Fabry-Pérot etalon 50 of the Fabry-Pérot interferometer 44 comprises first 54 and second 56 partially-reflective surfaces that are parallel to one another and separated by a fixed gap 58. The angles at which the scattered light 32 is passed through the Fabry-Pérot etalon 50 is dependent upon the optical frequency of the scattered light 32 and the length of the gap 58, resulting in an associated scatter fringe pattern 60 comprising a plurality of concentric arcuate fringes 62′—also known as Haidinger fringes—in the output focal plane 44.2 of the Fabry-Pérot interferometer 44, for example, as illustrated in
For example, in one embodiment, the Fabry-Pérot etalon 50 comprises a pair of planar optical windows 70—for example, constructed of either optical glass or fused quartz—aligned parallel to and facing one another, and spaced apart from one another by the gap 58, wherein, for example, the first 54 and second 56 partially-reflective surfaces—e.g. partially-silvered surfaces or other partially-reflective surfaces—are on separate facing surfaces of the planar optical windows 70. Alternatively, the first 54 and second 56 partially-reflective surfaces could be on the outside opposing faces of the planar optical windows 70, or one of the first 54 and second 56 partially-reflective surfaces could be on a inner facing surface of one of the planar optical windows 70, and the other of the first 54 and second 56 partially-reflective surfaces could be on a outer facing surface of the other of the planar optical windows 70. In one embodiment, the gap 58 is substantially fixed, whereas in other embodiments, the gap 58 is moveable, e.g. adjustable, for example, using an etalon control actuator 72 responsive to a controller 74 operatively associated with or a part of the data processor 66, so as to provide for a tunable Fabry-Pérot etalon 50.
Referring to
Referring to
For example, referring to
Absent the Fabry-Pérot etalon 50, the associated collimating lens 48 and imaging optics 52 provide for imaging the intermediate image plane 26 onto the output focal plane 44.2 that is detected by the detection system 46. Accordingly, the first 30.1 and second 30.2 regions or points on the intermediate image plane 26—corresponding to the closest 28.1 and farthest 28.2 volumetric elements of the beam of light 18 within the field-of-view 34 of the receiver optics 24—are imaged as corresponding first 82.1 and second 82.2 regions or points on the output focal plane 44.2. More generally, there is a one-to-one correspondence between regions or points 82 on the output focal plane 44.2 and corresponding volumetric elements 28 of the beam of light 18, and therefore, there is a one-to-one correspondence between regions or points 82 on the output focal plane 44.2 and the corresponding nominal range R thereto, given the parallax angle θ between the optic axes 36, 38 of the receiver optics 24 and the beam of light 18, respectively, so that the nominal range R associated with any region or point 82 on the output focal plane 44.2—or in the associated corresponding scatter electronic image signal 64 detected by the detection system 46—can be inferred from the location of that region or point 82 on the output focal plane 44.2. With the Fabry-Pérot etalon 50 present, the arcuate fringes 62′ of the scatter fringe pattern 60 are present for those regions or points 82 for which the associated frequency or wavelength of the associated scattered light 32 in cooperation with the gap 58 of the Fabry-Pérot etalon 50 provide for constructive interference, whereas arcuate nulls 84 in the scatter fringe pattern 60 are present for those regions or points 82 for which the associated frequency or wavelength of the associated scattered light 32 in cooperation with the gap 58 of the Fabry-Pérot etalon 50 provide for destructive interference. Locations of the arcuate fringes 62′ are determined by the frequency or wavelength of the associated scattered light 32, the gap 58 of the Fabry-Pérot etalon 50 and the angle of incidence in the Fabry-Pérot etalon 50.
For example, for the conditions described hereinabove for
The locations of the arcuate fringes 62′ and associated arcuate nulls 84 can be changed by either changing the gap 58 of the Fabry-Pérot etalon 50, for example, by the etalon control actuator 72 responsive to a signal from the controller 74, or by tilting the Fabry-Pérot etalon 50. For example, the gap 58 of the Fabry-Pérot etalon 50 could be repeatedly scanned by the etalon control actuator 72 responsive to a signal from the controller 74 so as to repeatedly generate associated sets of scatter fringe pattern 60 collectively having arcuate fringes 62′ associated with all nominal ranges R to the beam of light 18 within the field-of-view 34 of the receiver optics 24, so as to directly provide for associated atmospheric data 68 at any particular nominal range R within the range of associated nominal ranges R from RMIN to RMAX.
The range imaging LIDAR system 10, 10i provides for directly detecting light scattered off of either molecules 16′ of the atmosphere, aerosols 16″ in the atmosphere, or a combination of the two, and provides for directly measuring the density and temperature of the atmosphere 16, and the velocity thereof in the direction of the optic axis 36 of the receiver optics 24. For example, relatively short wavelength light is scattered by molecules 16′ of the atmosphere in accordance with Rayleigh scattering. Light can also be scattered by aerosols 16″ in the atmosphere in accordance with Mie scattering. Rayleigh scattering generally refers to the scattering of light by either molecules or particles having a size less than about 1/10th the wavelength of the light, whereas Mie scattering generally refers to scattering of light by particles greater than 1/10th the wavelength of the light. Being responsive to Rayleigh scattering, the range imaging LIDAR system 10, 10i is therefore responsive to the properties—e.g. velocity, density and temperature—of those molecules 16′ in the atmosphere giving rise to the associated scattering of the light detected by the range imaging LIDAR system 10, 10i. Furthermore, the range imaging LIDAR system 10, 10i can provide for operation in clean air, i.e. in an atmosphere with no more than a negligible amount of aerosols 16″, depending substantially upon only molecular scatter. If scattered from a moving molecule 16′ or aerosol 16″, the frequency scattered light 32 is Doppler shifted, which for a given gap 58 in the associated Fabry-Pérot etalon 50 thereby causes the associated arcuate fringes 62′ of the scatter fringe pattern 60 from the Fabry-Pérot interferometer 44 to be shifted to a location for which an associated constructive interference condition is satisfied for the corresponding rays of scattered light 32 entering the Fabry-Pérot interferometer 44 at a given angle from a corresponding given nominal range R. Accordingly, the Doppler shift in the frequency of the scattered light 32 will depend upon the local velocity of the atmosphere 16 within the interaction region 22 interacting with the beam of light 18, and for different amounts of Doppler shift, arcuate fringes 62′ associated with corresponding different nominal ranges R will be generated by the Fabry-Pérot interferometer 44, thereby causing the arcuate fringes 62′ to shift within the scatter fringe pattern 60—possibly relative to one another depending upon the distribution of velocity of the atmosphere 16 within the interaction region 22.
The range imaging LIDAR system 10, 10i further incorporates a filter system 88 to filter the scattered light 32 received by the receiver optics 24 so as to prevent background light from being detected by the detection system 46. For example, referring to
The Fabry-Pérot interferometer 44 is subject to mechanical defects and thermally induced drift that can be compensated through calibration using a reference beam portion 90 of the substantially monochromatic light 14 extracted from the light source 12 with a beam splitter optic 92 and then input to the Fabry-Pérot interferometer 44 at the intermediate image plane 26 as a reference source 94. For example, referring to
The light source 12 provides for generating a sufficient amount of sufficiently narrow-band monochromatic light 14 so as to provide for a sufficient amount of scattered light 32 so that the resulting scatter fringe pattern 60 is detectable by the detection system 46 with a sufficient signal-to-noise ratio (SNR) so that the resulting atmospheric data 68 determined therefrom is accurate within a given accuracy threshold and provides for an information temporal bandwidth that is within a given temporal bandwidth threshold. For example, the light source 12 could comprise one or more lasers, light emitting diodes (LEDs), flash lamps, for example, xenon flash lamps, sodium lamps or mercury lamps. The light source 12 may be either continuous or pulsed, and need not necessarily be coherent. If the spectral bandwidth of the light source 12 is not inherently substantially less than the expected minimum Doppler shifts to be measured, then the output of the light source 12 may be filtered with a filter 108 so as to provide for generating sufficiently monochromatic light 14 so as to enable Doppler shifts in the scattered light 32 to be measured sufficiently accurately so as to provide for resolving velocity sufficiently accurately, i.e. less than a given threshold. The particular operating wavelength of the range imaging LIDAR system 10, 10i is not limiting. For example, any optical wavelength that interacts with that which is being sensed in the associated interaction region 22 may be used.
For example, in one embodiment, the monochromatic light 14 comprises ultraviolet (UV) laser light at a wavelength of about 266 nm that is generated using a laser light source 12. A wavelength of about 266 nm, being invisible to the human eye and substantially absorbed by the atmosphere, is beneficial for its stealth, eye safety and molecular scattering properties. There is relatively little natural background light at this frequency due to absorption of most natural 266 nm light by ozone and molecular oxygen. Ultraviolet light at about 266 nm is readily absorbed by glass and plastic, such as used in aircraft wind screens, which provides for improved eye safety. The particular operating wavelength of the range imaging LIDAR system 10 is not limiting, and it should be understood that any optical wavelength that interacts with that which is being sensed in the associated interaction region 22 may be used.
For example, a Nd:YAG laser 12.1 can operate at relatively high power levels so as to provide sufficiently intense illumination so as to provide for relatively long range atmospheric sensing applications. An Nd:YAG laser 12.1 has a fundamental wavelength of 1064 nm, from which shorter wavelengths/higher frequencies may be generated using one or more harmonic generators operatively associated with or a part of the Nd:YAG laser 12.1. For example, a second-harmonic generator could be used to convert the fundamental 1064 nm light to second-harmonic 532 nm light which could then be transformed with either a third- or fourth-harmonic generator to generate associated 355 nm or 266 nm light respectively. For example, these second-, third- and/or fourth-harmonic generators may be either incorporated in, free-space coupled to, or coupled with a fiber optic to the Nd:YAG laser 12.1. Accordingly, alternative embodiments of the range imaging LIDAR system 10, 10i incorporating a Nd:YAG laser 12.1 may be operated at frequencies other than 266 nm, for example, at either the second or third harmonics, respectively, for example, as described in U.S. Pat. No. 7,495,774 which is incorporated herein by reference.
The arcuate fringes 62′, 62″ of the scatter 60 and reference 104 fringe patterns are circumferentially (φ) or transversely (Y) integrated either optically prior to detection, or electronically or by software during or following detection by the detection system 46, so as to provide for corresponding detected image signals I(X) and I0(X), respectively, that representing the total radiometric counts as a function of radial distance through the corresponding scatter 60 and reference 104 fringe patterns. The resulting detected image signals I(X) and I0(X) are then processed by the data processor 66 as described hereinbelow so as to generate one or more measures of the atmosphere 16 as a function of nominal range R, or at a particular nominal range R, within the interaction region 22.
Generally, depending upon how the resulting scatter 64 and reference 106 electronic image signals are processed, in accordance with a first aspect, the detection system 46 may comprise either one- or two-dimensional photodetector arrays, for example, either charge-coupled devices (CCDs) or charge injection devices (CIDs); or corresponding arrays of individual photodetectors, for example, photo-conductive, photo-voltaic, photo-emissive, bolometer, or thermopile photodetectors, i.e. generally any device that converts photons to a corresponding electrical signal. The particular detection system 46 may be adapted in cooperation with the associated light source 12 so as to provide for increasing the associated signal-to-noise ratio (SNR). For example, in cooperation with a continuous light source 12, a relatively high-sensitivity, low-noise, low-bandwidth detectors can be used, so as to provide for a higher signal-to-noise ratio (SNR) than possible with corresponding relatively higher-bandwidth detectors, so as to provide for relatively more precise associated measurements. Alternatively, the detection system 46 could comprise a camera with at least one array of concentric circular-segment photodetectors for each of the images being processed.
For example, in one embodiment, to process the scatter fringe pattern 60, the range imaging LIDAR system 10, 10i incorporates a linear photodetector array or a linear array of photodetectors, wherein, referring to
As another example, in another embodiment, to process both the scatter 60 and reference 104 fringe patterns, the range imaging LIDAR system 10, 10i incorporates a two-dimensional photodetector array or a two-dimensional array of photodetectors that provide for generating associated two-dimensional scatter 64 and reference 106 electronic image signals that, for example, can then be integrated either electronically; or by a process in the data processor 66, for example, as described hereinbelow.
Scattered light signal 32′ from each of the associated interaction regions 22 are substantially simultaneously processed together with a reference light signal 105 from the reference fringe pattern 104 so as to provide for calibrating, and maintaining the calibration of, the range imaging LIDAR system 10, and so as to provide for determining the associated air data products such as the speed, temperature and density of the atmosphere 16. This provides for an inherent self-calibration of the associated measurements or quantities derived therefrom. If wavelength drift of the light source 12 is not otherwise accounted for in the data, then errors can arise when making a measurement of the Doppler shift and resulting wavelength shift of the scattered light signal 32′. The range imaging LIDAR system 10 provides for automatically compensating for wavelength drift of the light source 12 from the data because each measurement from a scattered light signal 32′ is corrected using a corresponding measurement from the reference light signal 105 associated with the reference source 94.
In accordance with a first aspect, the associated detection system 46.1 provides for capturing an image 114 of the scatter 60 and reference 104 fringe patterns in the output focal plane 44.2 of the Fabry-Pérot interferometer 44. For example, in one embodiment, the detection system 46.1 comprises an electronic camera, for example, a CCD detection system 46.1′.
Referring to
Each pixel 118 is read from the CCD detection system 46.1′ and converted by an A/D conversion process. The ratio of signal to read noise can be enhanced by increasing the exposure time of the CCD detection system 46.1′ between read cycles, although at the cost of reduced dynamic frequency response of the associated resulting air data products. After identifying the center 122 of the circular fringe patterns 80, the circular binning algorithm sums up the CCD charges (i.e. pixel values) for each pixel 118 at a particular radius from the center 122, for a particular circular fringe pattern 80, for each of the circular fringe patterns 80, 60.1, 60.2, 60.3, 60.4, 104.1, 104.2, 104.3, 104.4, so as to provide a respective associated linear set of binned pixels 116 for each of the respective circular fringe patterns 80, 60.1, 60.2, 60.3, 60.4, 104.1, 104.2, 104.3, 104.4.
Referring to
Referring to
x
j
=j·α
X
−x
0
y
i
=i·α
y
−y
0 (1)
wherein αX and αY are the distances per pixel in the X and Y directions, respectively, and x0 and y0 are the coordinates of the center 122 relative to Pixel(1,1) at the lower left corner of the image 114. Then, in step (912), the Cartesian coordinates (xj, yi) from step (910) are transformed to cylindrical coordinates (R, θ), as follows:
Then, in step (914), if the angle θ is within a region of interest (ROI) 126.1-126.8, the associated region of interest ROI 126 is identified, and in step (916), the radial bin index k is given by:
where β is the distance per pixel in the radial direction, and k0 is the number of pixels 118 between the center 122 and the closest portion of the circular fringe pattern 80 closest thereto. Then, in step (918), the associated value Pixel(i,j) of the associated pixel 118 is added to the bin element BIN(k,ROI) of the bin array BIN(*,NROI) as follows:
BIN(k,ROI)=BIN(k,ROI)+Pixel(i,j) (4)
Then, or otherwise from step (914), in step (920), if all of the pixels 118 have been circumferentially binned, then, in step (922), the circumferentially-binned values for each of the circular fringe patterns 80 are returned in the associated bin array BIN(*,NROI). Otherwise, the process repeats with steps (904) and (906) for each of the rows and columns of pixels 118 until all of the circular fringe patterns 80 are binned.
Referring to
Referring to
BIN(k(m,ROI),ROI)=BIN(k(m,ROI),ROI)+Pixel(i(m,ROI),j(m,ROI)) (5)
Then, in step (1014), if all of the pixels m in the particular region of interest (ROI) 126.1′-126.8′ have not been binned, then the process continues with step (1008). Otherwise, in step (1016), if all of the regions of interest (ROI) 126.1′-126.8′ have not been binned, then the process continues with step (1002). Otherwise, in step (1018), the circumferentially-binned values for each of the circular fringe patterns 80 are returned in the associated bin array BIN(*,NROI).
In one embodiment, the circular fringe patterns 80, 104.1, 104.2, 104.3, 104.4 associated with the reference fringe pattern 104 are binned into a single common linear reference fringe pattern 104L, whereas in other embodiments the circular fringe patterns 80, 104.1, 104.2, 104.3, 104.4 associated with the reference fringe pattern 104 are either binned into separate associated linear reference fringe pattern 104L, 104.1L, 104.2L, 104.3L, 104.4L, or partially combined into a fewer number of associated linear reference fringe patterns 104L.
As yet another example, in yet another embodiment, the range imaging LIDAR system 10, 10i incorporates a plurality of circle-to-line interferometer optic (CLIO) elements 128 that provide for optically integrating the scatter 60 and reference 104 fringe patterns so as to generate corresponding linearly distributed associated fringe patterns that can then be detected with corresponding linear photodetector arrays or linear arrays of photodetectors, for example, as described hereinabove. For example, a separate circle-to-line interferometer optic (CLIO) element 128 would be used for each circular fringe patterns 80, 60.1, 60.2, 60.3, 60.4, 104.1, 104.2, 104.3, 104.4 on diametrically opposing portions of the Fabry-Pérot interferometer 44 relative to the optic axis 53, wherein each circle-to-line interferometer optic (CLIO) element 128 may be constructed and operated in accordance with the teachings of U.S. Pat. No. 4,893,003, which is incorporated herein by reference in its entirety, and in accordance with the teachings of U.S. Pat. No. 7,495,774, from line 22 at column 8 through line 50 at column 10 with reference to
As yet another example, in yet another embodiment, the range imaging LIDAR system 10, 10i incorporates a holographic optical element 128′ adapted to transform the arcuate fringes 62′, 62″ into corresponding linear distributions of light, for example, in accordance with the teachings of U.S. Pat. No. 6,313,908, which is incorporated herein by reference in its entirety, but adapted so that the arcuate fringes 62′ associated with the scatter fringe pattern 60 are transformed to a first linear distribution of light and the arcuate fringes 62″ associated with the reference fringe pattern 104 are transformed to a second linear distribution of light, wherein the first and second linear distributions are distinct, and detected by corresponding first and second linear photodetector arrays or linear arrays of photodetectors of the associated detection system 46, for example, as described hereinabove.
The reference 106 and scatter 64 electronic image signals are transmitted to the data processor 66, which processes the reference electronic image signal 106 to characterize the Fabry-Pérot etalon 50, and which then determines one or more range-dependent measures of the atmosphere 16—at one or more given ranges, or as a function of range—from the scatter electronic image signal 64 associated with arcuate fringes 62′, wherein each arcuate fringes 62′ corresponds to a different associated nominal range R and is analyzed separately. More particularly, the scatter electronic image signal 64 provides the information sufficient to determine the following measures of the atmosphere 16: aerosol counts A, molecular counts M, velocity u, temperature t, and background counts B, wherein molecular counts M provides for generating a measure of atmospheric density. As described more fully hereinbelow, data from each arcuate fringe 62′ is analyzed separately, so as to determine one or more of the measures: aerosol counts A, molecular counts M, velocity u, temperature t, and background counts B either at a given nominal range R or set of nominal ranges R, or as a function of nominal range R. The measures are determined by non-linearly fitting the measured reference electronic image signal 106 with a parameterized model of the Fabry-Pérot etalon 50, parameterized with respect to the measures so as to characterize the Fabry-Pérot etalon 50, and then non-linearly fitting the measured scatter electronic image signal 64 associated with different arcuate fringes 62′ to the parameterized model of the Fabry-Pérot etalon 50, parameterized with respect to the measures to be determined, i.e. with respect to aerosol counts A, molecular counts M, velocity u, temperature t, and background counts B, so as to determine values for those measures at the nominal range R associated with that particular arcuate fringe 62′.
A radial plot of the intensity of the circular fringe pattern 80 is illustrated in
The spectral shape of the scattered light signal 32′ processed by the Fabry-Pérot etalon 50, for a single associated fringe to be modeled, has a qualitative form illustrated in
The range imaging LIDAR system 10 provides for directly detecting laser energy scattered off of either molecules 16′ of the atmosphere, aerosols 16″ in the atmosphere, or a combination of the two, provides for directly measuring the associated velocity and direction, density, and temperature of the atmosphere, and provides for deriving other measurements therefrom, for example, a set of air data products. For example, relatively short wavelength laser energy is scattered by molecules of the atmosphere in accordance with Rayleigh scattering. Laser energy can also be scattered by aerosols in the atmosphere in accordance with Mie scattering. Rayleigh scattering generally refers to the scattering of light by either molecules or particles having a size less than about 1/10th the wavelength of the light, whereas Mie scattering generally refers to scattering of light by particles greater than 1/10th the wavelength of the light. Being responsive to Rayleigh scattering, the range imaging LIDAR system 10 is therefore responsive to the properties—e.g. velocity, density and temperature—of those molecules in the atmosphere giving rise to the associated scattering of the light detected by the range imaging LIDAR system 10. Accordingly, the range imaging LIDAR system 10 provides for operation in clean air, i.e. in an atmosphere with no more than a negligible amount of aerosols 16″, depending substantially only upon molecular scatter.
Referring to
where L is the loss per plate (absorption and scattering), R is the plate reflectivity, and M is the order of interference. Equation (6) describes a periodic transmission function, which is illustrated in
Equation (7) is a useful form of the Airy function since it provides for relatively easy convolutions with broadening functions.
The order of interference M is given by:
M=2 μtv cos θ (8)
where μ is the index of refraction of the material between the first 54 and second 56 partially-reflective surfaces, t is the effective gap 58, 58.1, v is the wavenumber of light, and θ is the angle of incidence in the Fabry-Pérot etalon 50 which is responsive to the focal length of the imaging optics 52 and the size of the detection system 46. Perturbations of t, v and θ from a set of standard conditions and normal incidence, can be modeled as follows:
The order of interference can then be written as follows:
where only the first order terms have been retained, and can be further expressed as follows:
The quantity ½ μt0 is the change in wavenumber required to change the order of interference by one, and is defined as the free spectral range, ΔvFSR, which results in:
Without loss of generality M0 can be an integer and therefore T(M)=T(ΔM).
Real instruments have defects which influence the behavior thereof and can be accounted for by broadening functions in the models used to characterize the device. These broadening functions are well known and are represented by a set of probability functions which can be convolved with the basic Fabry-Pérot Airy function to give the general result:
wherein the broadening function Dn filters the transmission T depending upon the magnitude of the defect or broadening process, and is calculated from the following product:
wherein dnq is the nth element of the convolution of the qth broadening function Gq—described hereinbelow—with the instrument model of equation (7). The convolution integral is defined as follows:
d
n
q=∫−∞∞Gq(δ′)*T(M(n)−δ′)dδ′ (19)
where T(M(n)−δ′) is the Fabry-Perot infinite series term.
A simplified notation can be used to provide for a more compact representation, wherein
so that the Airy function can be written as follows:
The broadening functions Gq account for broadening resulting from each of Doppler shift, laser width, scattering broadening, and turbulent motion, respectively, as given hereinbelow, for Nq=3 in equation (18).
Doppler Broadening The Doppler shift due to the mean air motion is given by:
where Δv is the Doppler shift, vl is the laser wavenumber, Uh is the horizontal wind speed in the direction of viewing, and φ is the angle from the zenith made by the beam of light 18 as it passes through the atmosphere 16, wherein Uh sin φ is the line-of-sight relative wind velocity U. Accordingly, equation (22) provides the relationship between line-of-sight relative wind velocity U and the Doppler shift Δv.
Laser Spectral Width Broadening: The spectral shape of the laser is assumed to be of Gaussian form, as follows:
where Δvl is the 1/e width of the laser, wherein the shorter the duration a laser pulse, the broader the associated broadening function, which results in lowered finesse for the Fabry-Pérot etalon 50.
Scattering Broadening: The affect on the transmission T of a Fabry-Pérot interferometer 44 due to broadening induced by molecular scattering is different from that induced by aerosol scattering. Accordingly, different broadening functions Gq are used to account for molecular and aerosol scattering, respectively, in respective corresponding models for the molecular TMol and aerosol TAero components of transmission T of the Fabry-Pérot interferometer 44.
The molecular scattering media broadens the signal due to associated random motions. The molecules have a Gaussian broadening function, as follows:
where ΔVG is the 1/e width and is given by:
where k is Boltzmann's constant, m is the mean mass of a molecule in the atmosphere, Temp is the static absolute temperature in degrees Kelvin, and
The aerosol broadening function has a Lorentzian form as follows, for example, as described in Fiocco, G., and DeWolf, J. B., “Frequency spectrum of laser echoes from atmospheric constituents and determination of aerosol content of air,” Journal of Atmospheric Sciences, v.25, n3, May 1968, pp. 488-496; and Benedetti-Michelangeli, G., Congeduti, F., and Fiocco, G., “Measurement of aerosol motion and wind velocity in the lower troposphere by Doppler optical radar,” Journal of the Atmospheric Sciences, v.29, n5, July 1972, pp. 906-910, both of which references are incorporated herein by reference:
where the half width αA is given by:
The spectral width of the aerosol-induced broadening component is extremely narrow compared to the molecular-induced broadening component, and in most cases are much narrower than the laser pulse, so that aerosol scattering essentially acts as a delta function and is not dependent on temperature.
Turbulent Motion Broadening: In addition to random motions of molecules and aerosols, the model allows for random motions of bulk parcels, i.e. turbulence, wherein this broadening is represented by a relatively simple Gaussian shape, as follows:
and UT is a characteristic turbulent velocity, which is a predefined constant that is independent of the line-of-sight relative wind velocity U. In some embodiments, this term is ignored because it is indistinguishable from temperature, so that the affects of equations (24) and (29) are indistinguishable from one another.
Other broadening functions Gq can also be utilized in addition to those described hereinabove, for example, so as to account for a defocus of the imaging optics 52.
The values of the linear sets of binned pixels 116 for the reference light signal 105 and scattered light signals 32′, respectively, provide a corresponding transmission measure T′ of the Fabry-Pérot interferometer 44 for the corresponding reference light signal 105 and scattered light signals 32′, respectively. Each transmission measure T′ is an N-element vector, wherein each element n of the vector corresponds to a different wavelength or corresponding order of interference. The element values are in units of measurement counts; for example, with one measurement count being equal to one photo-electron captured by the detection system 46. The transmission measure T′ is a measure of data from the Fabry-Pérot interferometer 44 that can be modeled as described hereinabove in accordance with equations (6) through (30), as represented by
T=T
Mol(Temp,U)·MolCounts+TAero(U)·AeroCounts+TBack·BackCounts (31)
where TMol(Temp,U)·MolCounts is the component of transmission T of the Fabry-Pérot interferometer 44 resulting from molecular scatter, which is a function of temperature and line-of-sight relative wind velocity U; TAero(U)·AeroCounts is the component of transmission T of the Fabry-Pérot interferometer 44 resulting from aerosol scatter, which is not affected by temperature but is dependent upon the line-of-sight relative wind velocity U; and TBack·BackCounts is the component of transmission T of the Fabry-Pérot interferometer 44 resulting from stray light and background wherein TBack is the continuum distribution or illumination profile through the instrument that is measured during calibration of the instrument from the response of the Fabry-Pérot interferometer 44 with the laser seeder 208 turned off, which is representative of the associated spectral distribution from the Fabry-Pérot interferometer 44 that would result from background illumination. During operation of the range imaging LIDAR system 10, the continuum distribution TBack is obtained from pre-measured values that are stored in memory, and the components TMol and TAero are calculated from equation (21) using the appropriate associated broadening terms. Each of the above-described components of transmission T of the Fabry-Pérot interferometer 44 is in units of counts resulting from the charge collected by the elements of the detection system 46. The distributions TMol(Temp, U), TAero(U) are evaluated with equation (21) using broadening functions that are appropriate for the molecular and aerosol components of scatter, respectively. In practice, when evaluating equation (21), the associated infinite series is truncated to ignore higher-order terms of relatively insignificant value, wherein the level of truncation is either predetermined, or determined during the accumulation of the elements of the series.
Accordingly, the transmission T of the Fabry-Pérot interferometer 44 is modeled with a non-linear model of equation (31) that is parameterized by a first set (or vector) of parameters P that characterize a particular measurement, i.e. which characterize a particular transmission measure T′; and a second set of parameters Q which are assumed constant during operation of the Fabry-Pérot interferometer 44, the values of which are determined during calibration. Referring to
The observables P can be determined as the values of the parameters P that minimize the following χ2 merit function:
using, for example, a Levenberg-Marquardt method of a non-linear least square process which provides for varying smoothly between an inverse-Hessian method and a steepest descent method, as described, along with other suitable non-linear methods, by W. H. Press, S. A. Teukolsky, W. T Veterling, and B. P. Flannery in Numerical Recipes in C, The Art of Scientific Computing, Second Edition, Cambridge University Press, 1992, pp. 656-661 and 681-706 which is incorporated herein by reference. In equation (32), T′(n) is the value of the nth binned pixel 116′, and T(M(n),P,Q) is the value of the transmission model T from equation (31).
Accordingly, for the range imaging LIDAR system 10, the transmission model T is overdetermined in the sense that the number of elements N of the detection system 46, i.e. the number of binned pixels per channel, is of a higher dimension than the number of observables P. For the range imaging LIDAR system 10 embodiment described herein, there are 5 observables P.
In the inverse Hessian method, the gradient of χ2 is given by:
and the Hessian is approximated by:
where k=1 to 5 for the 5 observables.
The observables are then solved by solving the set of linear equations:
where δP1 is an vector increment that is to be added to a current approximation for the observable vector P1. This system of equations can be represented as:
A·δP=B (36)
where A is the Hessian matrix, δP is a vector of increments to the observables that are to be added to a current approximation for the observable P, and B is the gradient vector. This system of equations can be solved as follows:
δP=A−1·B (37)
where A−1 is the inverse Hessian matrix.
The inverse Hessian method is suitable when the χ2 merit function can be locally approximated by a quadratic form. If a quadratic form is a relatively poor local approximation, then the steepest descent formula can be used to find the increment δP of the observable P as follows:
δP1=constant×βk (38)
The Levenberg-Marquardt method provides for a combination of the inverse Hessian and steepest descent methods, wherein the Hessian matrix in equation (35) is replaced with:
α′kk=αkk·(1+λ)
α′jk=αjk(j≠k) (39)
and both equations (35) and (38) are replaced with the following:
the solution of which is given by:
δP=A′−1·B (41)
where the elements of A′ are given by α′jk.
The Levenberg-Marquardt method commences with an initial guess for the observable vector P, after which χ2(P,Q) is calculated, and an initial value of λ is chosen (e.g. λ=0.001). An iterative process then commences with the solution for δP of equation (41), and the evaluation of χ2(P+δP,Q). If χ2(P+δP,Q)≧χ2(P,Q), then λ, is increased, e.g. by a factor of 10, and the iteration is repeated. Otherwise, if χ2(P+δP,Q)<χ2(P,Q), then λ is decreased, e.g. by a factor of 10, and the iteration is repeated. The iterations on the observable vector P are continued until a stopping criteria is satisfied, for example, on the first or second occasion when χ2 decreases by a negligible amount, and with the final solution, the method converses towards the inverse Hessian method.
The components of the gradient of the transmission model T used in calculating the gradient of χ2 and the Hessian matrix are given as follows, and are calculated numerically:
When processing the reference light signal 105, the observables MolCounts and BackCounts are assumed to be zero valued, and the partial derivatives with respect to MolCounts, BackCounts and Temp of equations (43), (46) and (45), respectively, are also assumed to be zero.
The σ2(n) weighing term in the χ2 merit function is the associated variance of the nth measurement channel (i.e. interference order or wavelength), which includes variance of the collected signal in combination with the variance associated with the noise from the detection system 46. The collected photons exhibit Poisson noise statistics. Accordingly, for Signal(n) photons/counts/photo-electrons collected on a single channel, the associated variance is equal to the signal level, as follows:
σSignal2(n)=Signal(n) (47)
wherein Signal(n) is the sum of the molecular, aerosol and background components, i.e.:
Signal(n)=Molecular(n)+Aerosol(n)+Background(n) (48)
so that Signal(n) is the predicted value from equation (31). The total variance is the combination of the signal variance and the variance of the detector, as follows:
σ2(n)=Signal(n)+NoiseDetector(n)2 (49)
wherein, for a CCD detection system 46.1, the detector noise is the associated read noise on each detector channel.
Alternatively, the observables P could be estimated using other non-linear modeling or non-linear programming techniques, or other techniques such as non-linear estimation or Kalman filtering.
Referring to
The substantially monochromatic light 14 from the laser 12′ is divided by a beam splitter optic 92 into a reference source 94 and the beam of light 18, the latter of which in some embodiments may be further divided into a plurality of beams of light 18 by beam steering optics 210, for example, incorporating beam splitting mirrors, prisms, a combination thereof, or some other type of beam splitter, each different beam of light 18 directed in a different direction into the atmosphere 16. The scattered light signals 32′ and reference source 94 are each first collimated by a collimator 212, e.g. a collimating lens 48, then filtered by a filter system 88 as described hereinabove, and then processed by an associated Fabry-Pérot etalon 50, the output of which is imaged by associated imaging optics 52 as associated circular fringe patterns 80 onto the associated detection system 46. The associated optical components are adapted for the frequency and power levels of operation. For example, for a range imaging LIDAR system 10 incorporating a Nd:YAG laser 12.1 operating at 355 nanometers, the optical elements would incorporate UV-grade fused silica substrates and standard anti-reflection coatings tuned for 355 nanometers.
The geometry of the circular fringe patterns 80 from the Fabry-Pérot etalon 50 is responsive to the operative gap 58, 58.1 thereof, which would vary with temperature if the associated material or materials controlling the length of the gap 58, 58.1 were to exhibit a non-zero coefficient of thermal expansion. Although the reference source 94 simultaneously processed by the Fabry-Pérot etalon 50 provides for compensating for thermal drift affecting all portions of the Fabry-Pérot etalon 50 equally, it is beneficial if the temperature of the Fabry-Pérot etalon 50 can be controlled or maintained at a constant level so as to prevent a thermal expansion or contraction thereof during the operation thereof. Accordingly, in accordance with one aspect of the range imaging LIDAR system 10, the Fabry-Pérot etalon 50 is thermally stabilized by enclosure in a thermally-controlled enclosure 214 so as to prevent thermally-induced drift of the circular fringe pattern 80.
In accordance with one aspect, the thermally-controlled enclosure 214 is passive, for example, with the Fabry-Pérot etalon 50 enclosed, i.e. thermally insulated or isolated, using a material or materials with a very low thermal conductance to increase the thermal time constant and to prevent any substantial thermal shock from reaching the Fabry-Pérot etalon 50. In accordance with another embodiment, or in combination therewith, the thermally-controlled enclosure 214 is constructed from a combination of materials adapted so that there is negligible net coefficient of thermal expansion in the portions of the structure surrounding the Fabry-Pérot etalon 50 that affect the length of the gap 58, 58.1.
Referring to
The inner 238 and outer 246 enclosures are assembled together to form a core assembly 268, as follows. The solid optical element 76 Fabry-Pérot etalon 50 is bonded inside a bore 270 of the etalon mount 222 with a thermal epoxy which provides for thermal conduction therebetween, wherein the inside diameter of the bore 270 is adapted so as to provide for a non-interfering fit with the solid optical element 76. The flange 240 of the etalon mount 222 is attached with fasteners 244 to the first faces 242 of the three heat sink segments 224 assembled around the outside surface 228 of the etalon mount 222. Three thermo-electric heat pumps 236 are sandwiched between respective recesses 232, 252 in a corresponding outer face 230 of each heat sink segment 224 and a corresponding inside face 250 of each outer ring segment 248, so that the first 234 and second 254 surfaces of the thermo-electric heat pumps 236 abut and are in thermal communication with the corresponding associated heat sink segment 224 and outer ring segment 248 respectively. The core assembly 268 further comprises a plurality, e.g. three, temperature sensors 216, e.g. thermistors, resistive temperature devices, or thermocouples—each of which is inserted in a corresponding hole 272 in a second face 274 of each heat sink segment 224, so as to provide for monitoring the temperature thereof, and so as to provide in cooperation with the associated temperature controller 218 and the associated thermo-electric heat pump 236, for controlling the temperature thereof.
The core assembly 268 is inserted in the outer shell 262 so that the flanges 240 of the outer ring segments 248 mate with the corresponding internal grooves 260 of the outer shell 262, and the outer ring retainer wedges 266 are inserted in the gaps 276 between the facing sides 264 of the flanges 240 so as to wedge the opposing sides 258 of the flanges 240 against associated internal grooves 260 of the outer shell 262, thereby providing for retaining the core assembly 268 within the outer shell 262, and providing for thermal communication therebetween. The ends 278 of the outer shell 262 are closed with associated end cap assemblies 280 secured thereto with associated fasteners 282 and sealed therewith associated seals 284, e.g. gaskets or o-rings. The end cap assemblies 280 incorporate associated window assemblies 286 fastened thereto and incorporating optical windows 288, e.g. constructed from UV grade fused silica substrates with standard anti-reflection coatings, which provide for transmission of the associated scattered 32′ and reference 105 light signals. The resulting assembly constitutes a thermally stabilized etalon assembly 290 incorporating a thermally-controlled enclosure 214. The thermally stabilized etalon assembly 290 further comprises a plurality of electrical connectors 292 therein which provide for connecting the thermo-electric heat pumps 236 and the temperature sensors 216 with the associated temperature controller 218. The temperature controller 218 uses the temperature sensors 216 to monitor the temperature of the core assembly 268, and controls the heating or cooling thereof relative to the environment using the associated thermo-electric heat pumps 236 so as to maintain the temperature of the core assembly 268 at a specified set-point. The outer enclosure 246 in thermal communication with the outer shell 262 provides for either supplying heat to or rejecting heat from the inner enclosure 238 responsive to the thermal effort of the thermo-electric heat pumps 236 as needed to maintain a particular set-point temperature. For example, in one embodiment, the set-point temperature is adapted so as to minimize the energy needed to maintain that temperature, while also maintaining a sufficient offset so as to operate the thermo-electric heat pumps 236 most efficiently. For example, for a thermo-electric heat pump 236 that operates most efficiently when heating, the set-point temperature might be 5 to 10 degrees Celsius above the nominal environmental temperature, e.g. 5 to 10 degrees Celsius above room temperature.
Referring to
The range imaging LIDAR system 10 can take advantage of aerosols when present, but does not rely upon their presence. The reference light signal 105 and the scattered light signals 32′ of the range imaging LIDAR system 10 can be used to directly measure velocity, true airspeed, vertical speed, angle of attack, angle of sideslip, static density, static temperature, and aerosol to total scattering ratio (ASR). From these data products the following quantities can be directly calculated: calibrated airspeed, Mach number, static pressure, total pressure, dynamic pressure, pressure altitude, air density ratio, total temperature, angle of attack, pressure differential, and angle-of-sideslip pressure differential. Wind velocity, density, and temperature are directly calculated using the fringe data from the Fabry-Pérot interferometer 44. The other air data products are derived from these three basic measurements, in view of the knowledge of the associated geometry of the beam steering optics 210. The molecular signal yields a measure of air density that can be related to pressure. The aerosol to total scattering ratio is also directly derived from the results.
As used herein, the term relative wind is intended to refer to the relative motion between the atmosphere—included molecules and aerosols—and the range imaging LIDAR system 10. In addition to frequency—which, responsive to associated Doppler shift, provides for measuring associated velocity—the algorithm determines the contribution to the fringe pattern from molecular and aerosol scatter, the background radiation, and the temperature of the atmosphere 16 for each particular associated direction associated with each corresponding volumetric element 28 as viewed by the associated receiver optics 24.
For example, referring to
Referring to
Referring to
Referring to
More particularly, referring to
Referring to
A reference beam portion 90 of the substantially monochromatic light 14 from the light source 12 is reflected from a first beam splitter optic 92 so as to generate an associated reference source 94 which is coupled into an associated fiber optic 98 that routes the signal to where it is needed. The output from the fiber optic 98 is divergent and is subsequently collimated by an associated lens 134 and then combined with the scattered light 32 using a second beam splitter optic 136 that reflects a relatively small portion of the substantially monochromatic light 14 from the reference source 94 into the Fabry-Pérot interferometer 44 as the associated reference light signal 105 while transmitting a substantial portion of the scattered light 32 therethrough into the Fabry-Pérot interferometer 44 as the scattered light signal 32′.
The position of the fiber optic 98 in the image plane of the lens 134 determines where the associated image 114 of the reference light signal 105 will appear on the detection system 46. In one embodiment, the image 114 of the reference light signal 105 is positioned so as to not overlap the associated scattered light signal 32′ in the output focal plane 44.2 of the Fabry-Pérot interferometer 44. In another embodiment, in accordance with the eighth aspect of the range imaging LIDAR system 10, 10viii described more fully herein below, the image 114 of the reference light signal 105 is positioned so as to overlap the associated scattered light signal 32′, with the portion of the reference light signal 105 overlapping the scattered light signal 32′ blocked by an associated mask 138 between the lens 134 and the second beam splitter optic 136.
The associated optics can be designed so that the reference light signal 105 will be sufficient to determine the center of the interference pattern produced by the Fabry-Pérot interferometer 44 as well as the location of the associated arcuate fringes 62′, 62″.
Referring to
Referring to
The micromirrors 144 of the associated array of micromirrors 144 of the digital micromirror device (DMD) 142 in the first pixel mirror rotational state 148 cause first portions 160′ of either the scatter fringe pattern 60 or the reference fringe pattern 104 from the Fabry-Pérot interferometer 44 impinging thereupon to be reflected in a first direction 162 to an associated first objective lens 164, and to be directed thereby to the a first photomultiplier detector 154A′. Similarly, micromirrors 144 of the associated array of micromirrors 144 of the digital micromirror device (DMD) 142 in the second pixel mirror rotational state 150 cause second portions 160″ of either the scatter fringe pattern 60 or the reference fringe pattern 104 from the Fabry-Pérot interferometer 44 impinging thereupon to be reflected in a second direction 166 to an associated second objective lens 168, and to be directed thereby to the a second photomultiplier detector 154B′. Finally, micromirrors 144 of the associated array of micromirrors 144 of the digital micromirror device (DMD) 142 in the third pixel mirror rotational state 152 cause third portions 160′″ of either the scatter fringe pattern 60 or the reference fringe pattern 104 from the Fabry-Pérot interferometer 44 impinging thereupon to be reflected in a third direction 170 to the light block 172 that provides for absorbing light impinging thereupon. For example, in one embodiment, the third pixel mirror rotational state 152 corresponds to a state of substantially no rotation of the associated micromirrors 144, which may be achieved, for example, by applying a common voltage to the associated micromirror 144 and it associated mirror address electrodes and yoke address electrodes, so as to create an equal state of electrostatic repulsion between all associated pairs of electrodes associated with the micromirror 144, thereby maintaining the micromirror 144 in a substantially unrotated condition.
The micromirrors 144 of the digital micromirror device (DMD) 142 are relatively efficient, with overall efficiency approaching 90% in one set of embodiments. Accordingly, the digital micromirror device (DMD) 142 provides for digitally isolating light impinging thereupon into two disjoint sets for the portion of the light being analyzed, and for masking a remaining portion of the light. More particularly, the digital micromirror device (DMD) 142 is used to interrogate portions the scatter 60 and reference 104 fringe patterns from the Fabry-Pérot interferometer 44, and in cooperation with the associated first 154A′ and second 154B′ photomultiplier detectors, to provide for generating associated one or more pairs of associated complementary signals 156, 158, each responsive to the number of photons in the associated two disjoint sets of light reflected by the digital micromirror device (DMD) 142 resulting from a particular pattern of pixel mirror rotational states to which the associated array of micromirrors 144 of the digital micromirror device (DMD) 142 are set for a particular set of measurements, wherein the associated first 154A′ and second 154B′ photomultiplier detectors provide for counting the corresponding number of photons associated with each of the disjoint sets of light reflected by the digital micromirror device (DMD) 142.
For example, referring also to
For example, referring to
Commercial digital micromirror devices (DMD) 142 comprise arrays of micromirrors 144 ranging from an array of 640×480 micromirrors 144 containing approximately a half million micromirrors 144 in total, to an array of 2048×1080 micromirrors 144 containing over two million micromirrors 144 in total. Each micromirror 144 of the array represents one pixel 146 of a pattern 190 of associated pixel mirror rotational states 148, 150, 152, wherein each pixel is independently controllable or programmable responsive to a signal from the data processor 66.
The scattered light signal 32′ of the associated scattered light signal 32′ received from the interaction region 22 associated with the field-of-view 34 of the telescope 24′ is processed by the Fabry-Pérot interferometer 44 to generate an associated scatter fringe pattern 60 that is then separated by the digital micromirror device (DMD) 142 into disjoint portions 60′, 60″ that are then detected by the corresponding associated first 154A′ and second 154B′ photomultiplier detectors. The reference light signal 105 is processed by the same Fabry-Pérot interferometer 44, either simultaneously or sequentially, to generate an associated reference fringe pattern 104 that is then separated by the digital micromirror device (DMD) 142 or a separate corresponding digital micromirror device (DMD) (not illustrated) into disjoint portions 104′, 104″ that are then detected by the corresponding associated first 154A′ and second 154B′ photomultiplier detectors, or by a separate set of first and second photomultiplier detectors (not illustrated). The resulting complementary signals 156, 158 associated with the reference light signal 105 are used to provide for calibrating atmospheric measurements associated with the scattered light signal 32′. Accordingly, the range imaging LIDAR system 10 uses the Fabry-Pérot interferometer 44 to directly detect information from the scattered laser energy, wherein the scatter 32′ and reference 105 light signals are each detected separately, and information from the reference light signal 105 can then be used to calibrate the associated scattered light signal 32′. The detection process is responsive to an incoherent Doppler shift of the laser light scattered by molecules and aerosols in the atmosphere 16 responsive to Rayleigh and Mie scattering respectively.
The response of a Fabry-Pérot interferometer 44 is well documented in the literature, for example, as described by P. B. Hays and R. G. Roble in “A Technique for Recovering Doppler Line Profiles from Fabry-Perot Interferometer Fringes of very Low Intensity”, Applied Optics, 10,193-200, 1971, which is incorporated herein by reference. The ideal intensity distribution of a the fringe pattern for a single wavelength transmitted through a Fabry-Pérot interferometer 44 by a LIDAR system without optical defects is given by
wherein T is the transmissivity, R is the reflectivity, μ is the refractive index of the Fabry-Pérot etalon 50, d is the thickness of the gap 58, 58.1 of the Fabry-Pérot etalon 50, λ, is the wavelength of the source, θ is the angle of transmission through the Fabry-Pérot etalon 50, c is the speed of light, and u is the line-of-sight air velocity. Hence, the Doppler shift is 2 u/c. In the presence of a source distribution including many wavelengths and optical defects it is advantageous to use the Fourier cosine series expansion of the response. The distribution of intensity transmitted per molecular weight (of the scattering species) is given by:
where t is the atmospheric temperature, k is the Boltzmann constant, A0 is Avogadro's number, m is the molecular mass of the scattering species, and the convolution effects of the optical defects are represented by associated defect coefficients Dn,k.
If there were no optical defects, then each of the defect coefficients Dn,k would be identically equal to one. However, in a system with optical defects, these may be accounted for in various ways. For example, in accordance with a first method, the defect coefficients Dn,k are calibrated using a reference source 94 that does not interact with the atmosphere 16. As long as the range imaging LIDAR system 10 stays calibrated then these defect coefficients Dn,k may be used directly in the inversion of data to recover atmospheric state variables. As another example, in accordance with a second method, a signal from the reference light signal 105 is periodically collected together with one or more associated signals from the corresponding one or more scattered light signals 32′, and the effect of the defect coefficients Dn,k is computed by de-convolving the ideal signal, Hideal,—for example, Hideal as given by equation (50.1),—from the recovered data using the Fourier transform of the ideal signal, Hideal, for example, as given by equations (60.1), (60.2) and (61) described hereinbelow. The function G(t) approximates the effect of thermal broadening of a source by a low density gas, which effects are more precisely accounted for by Rayleigh-Brillouin scattering, although that level of detail is not essential to the practice of the range imaging LIDAR system 10.
For an atmosphere 16 containing both aerosols and molecules, and for the range imaging LIDAR system 10 adapted to sample the entire circular fringe pattern 80, the associated total response is given by:
where I is the total number of photons reaching the photodetector 154, A is the number of photons that have been scattered by aerosols, M is the number of photons that have been scattered by molecules, B is the number of background photons transmitted to the range imaging LIDAR system 10 by the ambient atmosphere 16, mA is the molecular mass of an aerosol particle (for example, a very large number on the order of 1.0e5), and mM is the molecular mass of air (about 28.92). Given this model, the sensitivity of the system to the atmospheric variables A, M, u, t and B is respectively given by respectively taking partial derivatives of equation (52) with respect to each respective variable, as follows:
For example,
The separate influence of molecules and aerosols is evident in the partial derivative of the total fringe response I with respect to velocity u illustrated in
Generally, the range imaging LIDAR system 10 provides for sampling, collecting and integrating separate portions, for example, disjoint portions 60′, 60″, 104′, 104″, of the scatter 60 and reference 104 fringe patterns, and then using the resulting associated signals, for example complementary signals 156, 158, for each of a set of different disjoint portions 60′, 60″, 104′, 104″, to determine the values of the variables or parameters characterizing the associated scatter fringe pattern 60. The scatter 60 and reference 104 fringe patterns are sampled by the digital micromirror device (DMD) 142, with the pixel mirror rotational states 148, 150, 152 of the associated micromirrors 144 controlled according to a particular pattern 190, so that the micromirrors 144 in the first pixel mirror rotational state 148 provide for reflecting light from a first disjoint portion 60′, 104′ of the scatter 60 or reference 104 fringe pattern to the first objective lens 164, which focuses the light onto the first photomultiplier detector 154A′ that provides for integrating the light from the first disjoint portion 60′, 104′ of the scatter 60 or reference 104 fringe pattern so as to generate a first complementary signal 156; and so that the micromirrors 144 in the second pixel mirror rotational state 150 provide for simultaneously reflecting light from a second disjoint portion 60″, 104″ of the scatter 60 or reference 104 fringe pattern to the second objective lens 168, which focuses the light onto the second photomultiplier detector 154B′ that provides for integrating the light from the second disjoint portion 60″, 104″ of the scatter 60 or reference 104 fringe pattern so as to generate a second complementary signal 158. This process is repeated for each different set of N different sets of disjoint portions 60′, 60″ of the scatter fringe pattern 60, and for one set of disjoint portions 104′, 104″ of the reference fringe pattern 104, so as to provide for generating N corresponding sets of complementary signals 156, 158, from which up to N different variables or parameters can be characterized.
For example, in accordance with a first aspect, the scatter fringe pattern 60 is characterized with respect to the following N=5 variables: aerosol counts A, molecular counts M, velocity u, temperature t, and background counts B as provided by equations (51.1), (51.2) and (52) hereinabove, using a corresponding N=5 different patterns 190 of pixel mirror rotational states 148, 150, 152 of the micromirrors 144 of the digital micromirror device (DMD) 142, wherein each of the associated patterns 190 is chosen in advance based upon the expected sensitivity of the optical response with respect to each of these variables. For example, in on embodiment, the pattern 190 of pixel mirror rotational states 148, 150, 152 for each of the N=5 variables are chosen responsive to the sign of the partial derivatives of the total fringe response I(φ) with respect to that variable, i.e. responsive to the sign of equations (53.1)-(53.5), subject to a fixed offset, respectively. For example,
More particularly,
It should be noted that the pattern 190, 190.1 of pixel mirror rotational states 148, 150, 152 used for the measure of aerosol counts A is a subset of the pattern 190, 190.2 of pixel mirror rotational states 148, 150, 152 used for the measure of molecular counts M, and that each of the patterns 190, 190.1-190.5 of pixel mirror rotational states 148, 150, 152 is mathematically independent of the others, so that none of these patterns 190, 190.1-190.5 may be constructed by superposition of the other patterns 190, 190.1-190.5 of pixel mirror rotational states 148, 150, 152. Accordingly, the five sets of complementary signals 156.1-156.5, 158.1-158.5 from the first 154A and second 154B photodetectors for the circular fringe pattern 80 from the scattered light signal 32′ provides sufficient information as necessary to determine aerosol counts A, molecular counts M, velocity u, temperature t, and background counts B therefrom.
Generally, any collection of patterns 190 of pixel mirror rotational states 148, 150, 152 that are spatially independent will work however, not all patterns 190 of pixel mirror rotational states 148, 150, 152 provide the same expected error. The optimum selection of patterns 190 of pixel mirror rotational states 148, 150, 152 depends on the variables of interest in the remote sensing problem at hand and also on the state of the solution being sought. In accordance with the first aspect, the patterns 190 of pixel mirror rotational states 148, 150, 152 are chosen in view of an associated model of the optical response of the range imaging LIDAR system 10, wherein the derivatives of the optical response provide for resulting associated complementary signals 156, 158 that are sensitive to changes in the associated variables of interest. From the partial derivatives of the total fringe response I with respect to aerosol counts A, molecular counts M, velocity u, temperature t, and background counts B as given by equations (53.1)-(53.5), the associated regions of interest are relatively broad and well defined. For example, referring to
In accordance with a second aspect, the patterns 190 may be adapted as with the first aspect, but with the use of an associated threshold when mapping the results of equations (53.1)-(53.5) to the corresponding patterns 190, wherein the patterns 190 are then given responsive whether or not the value of the associated derivative is either greater or less than a chosen threshold, for example, as shown in
The programmability of the digital micromirror device (DMD) 142 allows the regions being selected to be varied dynamically as the measurement conditions vary. For example: in the case of a LIDAR, the pattern 190.3 of pixel mirror rotational states 148, 150, 152 for velocity u is most sensitive when its divisions coincide with the fringe peaks (which move with velocity dependent Doppler shifts). Accordingly, real time accuracy can be improved if the pattern 190.3 of pixel mirror rotational states 148, 150, 152 for velocity u were adapted in real time to account for this shift. This ability to adapt the observations can be beneficial in a highly variable natural environment. Similarly, the temporal duration of exposure for each pattern 190 of pixel mirror rotational states 148, 150, 152 may be adjusted within a sample set, i.e. the duration of measurement may be different for different patterns 190 of pixel mirror rotational states 148, 150, 152, so as to provide for re-balancing the sensitivity of the range imaging LIDAR system 10 to increase accuracy in the state variable or state variables of greatest interest.
The choice of temporal exposure weighting and patterns 190 of pixel mirror rotational states 148, 150, 152 depend on the present environmental state and a ranking of the parameters of interest. One approach for examining potential systems is by a Monte-Carlo simulation. Another is by a non-linear optimization technique such as the Broyden-Fletcher-Goldfarb-Shanno (BFGS) method, a quasi-Newton, variable metric method, for example, as described by J. Nocedal and, S. Wright, Numerical Optimization, Springer-Verlag New York, Inc., 1999, pages 194-201, which is incorporated herein by reference. In these cases one may design a cost function based on the covariance of the minimum variance unbiased estimate—for example, as described by D. Luenberger in “Optimization by Vector Space Methods”, John Wiley & Sons, Inc. (1969) on page 15, which is incorporated herein by reference—using the system dynamics from the model response and expected environmental noise, for example, as given by equation (70) hereinbelow. At which point Monte-Carlo can be employed to understand how the distribution of solutions vary with respect to the system design, or descent-based schemes can by employed to find a best candidate according to ones rankings of state variable accuracy.
Once a scheme for generating patterns 190 of pixel mirror rotational states 148, 150, 152 is established, the associated thresholds and temporal weighting fractions can then be mathematically optimized. The resulting optimal set of parameters will be referred to as a solution to the optimization problem. Given a pattern 190 of pixel mirror rotational states 148, 150, 152, the system partial derivatives (Jacobean Matrix) and the expected measurement covariance, one can estimate the inversion errors that would occur in using that system. In particular the Jacobean derivative, J, is given
which allows the intensity at any phase point, φ, to be approximated as
I≈I0+J[ΔA,ΔM,Δu,Δt,B]T (55)
The expected covariance of the noise in intensity is given by Q. In the case of a shot noise limited system this covariance would be a diagonal matrix of the counts collected in each measurement. The matrix of dynamics, W, is formed by integrating the Jacobean over each pattern 190 of pixel mirror rotational states 148, 150, 152 and applying the corresponding temporal weighting factor. Let ΩA,ΩM,Ωu,Ωt,ΩB represent the patterns 190 of pixel mirror rotational states 148, 150, 152 that send light to the first photodetector, and {tilde over (Ω)}A,{tilde over (Ω)}M,{tilde over (Ω)}u,{tilde over (Ω)}t,{tilde over (Ω)}B be the complements of these patterns 190 of pixel mirror rotational states 148, 150, 152 which send light to the second photodetector, then one can form a 10×5 matrix where the kth row is given by cycling Ωk through the set {ΩA,{tilde over (Ω)}A, ΩM,{tilde over (Ω)}M, Ωu,{tilde over (Ω)}u, Ωt,{tilde over (Ω)}t, ΩB,{tilde over (Ω)}B} and similarly for the temporal weighting fractions pk through {pA,pA,pM,pM,pu,pu,pt,pt,pB,pB}.
This equation (56) is valid for any set of patterns 190 of pixel mirror rotational states 148, 150, 152 (such as those shown in
At this point one may compute the standard deviation of the errors expected in each measured parameter through the minimum variance unbiased estimator as
σ=√{square root over (diag([WTQ−1W]−1))} (58)
Each element of the σ vector represents the expected error in A, M, u, t, B respectively. With this ability to estimate the errors in each parameter of the system, one may perform a Monte-Carlo analysis to vary the associated thresholds and temporal weighting factors to see how the parameters affect the accuracy of the system, for example, in accordance with the Monte-Carlo procedure is illustrated in
The distribution of the solution space can be understood by viewing the Monte-Carlo results, for example, such as those shown in
Example Linear Cost Functional
Example Gaussian Multivariate Functional
Example Logarithmic Functional
J(σ)=log(ω,σn+γ) (59.3)
Alternatively, any number of schemes could be used to find patterns 190 of pixel mirror rotational states 148, 150, 152 which optimize a cost function. For example, in a Genetic algorithm procedure, the first step of
It is an interesting point that the patterns 190 of pixel mirror rotational states 148, 150, 152 used with the Fabry-Pérot interferometer 44 are not required to be generated without regard to the expected fringe pattern. In fact, the only requirement is that the patterns 190 of pixel mirror rotational states 148, 150, 152 are algebraically independent, such that no pattern 190 of pixel mirror rotational states 148, 150, 152 can be constructed as a linear combination of the other patterns 190 of pixel mirror rotational states 148, 150, 152 in the set
Referring to
Furthermore, the patterns 190 of pixel mirror rotational states 148, 150, 152 do not necessarily have to be radially symmetric. Although the information content of a Fabry-Pérot interferometer 44 is circularly symmetric, if circular symmetry of the selected patterns 190 of pixel mirror rotational states 148, 150, 152 is broken then one may consider the value of the pattern 190 of pixel mirror rotational states 148, 150, 152 for that specific radii to be the fraction (or probability) of pixels in either the first 148 or second 150 pixel mirror rotational states. Such a pattern 190 of pixel mirror rotational states 148, 150, 152 is shown in
The set of measurements of the complementary signals 156, 158 for the corresponding set of patterns 190 of pixel mirror rotational states 148, 150, 152 can then be used to estimate the parameters or measurements from the range imaging LIDAR system 10. All routines must account for the optical defects in the system as in equations (51.1-51.2). These defects typically have a convolution type response such as a defocus-blurring or an etalon wedge defect. In a Fabry-Pérot imaging system one can usually acquire a reference fringe pattern of the laser before it has interacted with the atmosphere. This response will contain all the information necessary to model the system's optical defects and any changes to the Fabry-Pérot etalon 50. For example changes in the temperature of a solid Fabry-Pérot etalon 50 will change its refractive index thereby changing the systems response to velocity and temperature. This information is readily accessible by comparing the Fourier Transform of the reference to the Fourier transform of the ideal signal. Term by term (i.e. per mode) division reveals the defect coefficients (in a noise free environment), for example, as described by T. L. Killeen and P. B. Hays in “Doppler line profile analysis for a multichannel Fabry-Perot interferometer,” Applied Optics 23, 612 (1984), which is incorporated herein by reference. These can be applied to the forward model of the Fabry-Perot response as discussed earlier. As such, the Fourier expansion of an ideal signal, Hideal=H0(φ), and the reference signal, Href(φ), is
where the Ĥ[n] terms are the Fourier coefficients of the normalized responses. The orthogonality of the cosine basis implies that the nth coefficient of the optical defects can be obtained from
These are the terms to be computed in the calibration of the instrument. The reference signal is also used to track the intensity of the beam and any phase shifts in response due to drift of the gap 58, 58.1 of the Fabry-Pérot etalon 50. The refractive index of the Fabry-Pérot etalon 50 may be obtained by independently monitoring the temperature of the Fabry-Pérot etalon 50. This tracking is accomplished in an iterative process using measurements akin to equation (56). Starting with the matrix of dynamics
and the vector of measurements
then the change in those measurements is expected to be driven by changes in the state of the system. Hence the measurements at time j+1 are given by the previous measurements, j, and the system dynamics existing at the time of the jth measurement:
M
j+1
=M
j
+W
j
δx (64)
where δx=[δA, δu, δB]T. Recall that the phase is given by
The velocity term should be zero, however changes in length d of the gap 58, 58.1 of the Fabry-Pérot etalon 50, will have a similar impact as velocity, namely δd=−2dδu/c. Because the reference signal has not been broadened its response is exactly the same as the scatter signal from aerosols. As such, the aerosol term will be used to track the change in laser power. Equation (64) is then solved for the updates [δA, δd, δB]T. These updates then define the normalization and phase changes necessary to consider for inversion of the total scatter signal. The reference state may be computed with each scattered signal, or as often as necessary to capture the rate at which the optical system changes (for example with temperature). If one can guarantee thermal stability via a temperature controlled Fabry-Pérot etalon 50 and housing then it may only be necessary to evaluate the reference periodically or on system initialization.
A similarly related technique is to divide the Fourier Transform coefficients of the reference fringe from the fringe pattern produced by the scattered atmospheric response. The remaining response reveals a phase shift (linearly correlated to the velocity via the expected Doppler shift) and broadening function related to the thermal effects. This method is very sensitive to noise in the collected data. More than the five patterns 190.1-190.5 of pixel mirror rotational states 148, 150, 152 already described would be used in order to recover the defect coefficients. One generally requires at least as many patterns 190 of pixel mirror rotational states 148, 150, 152 states as Fourier coefficients that one needs to faithfully represent the signal. In a rich aerosol environment this could be anywhere from 45 to 100 coefficients thus requiring the same number or more of independent measurements. One simple method gaining these measurements is to create a pattern 190 of pixel mirror rotational states 148, 150, 152 of rings which sweep outward from the center. These measurements may be made periodically within normal system operation and post-processed later to produce the analytical representation of the reference fringe. Alternatively, a large enough digital micromirror device (DMD) 142 could simultaneously image the atmospheric response with one set of patterns 190 of pixel mirror rotational states 148, 150, 152 and a reference fringe pattern with another set of patterns 190 of pixel mirror rotational states 148, 150, 152.
One method for estimating the parameters of the atmospheric state from the scattered signal is the classic Levenberg-Marquardt nonlinear least squares method which provides for varying smoothly between an inverse-Hessian method and a steepest descent method, as described, along with other suitable non-linear methods, by W. H. Press, S. A. Teukolsky, W. T Vetterling, and B. P. Flannery in Numerical Recipes in C, The Art of Scientific Computing, Second Edition, Cambridge University Press, 1992, pp. 656-661 and 681-706 which is incorporated herein by reference. This method works by iteratively minimizing the mean square error of a set of acquired samples against the output of a forward model (such as the model for the Fabry-Perot transmitted fringe pattern). It only requires the system dynamics equation given in equation (56) for any given state of the parameters. It operates by performing Quasi-Newton decent type steps toward the parameter state which minimizes the residual (mean square error of the difference between the data and the model). The algorithm works as follows:
Consider the measurements made with each pattern 190 of pixel mirror rotational states 148, 150, 152 to be the vector:
Let
be the estimates of return signal given the model described in equations (50-52). As described in equation (56), the Jacobean of this model is:
such that, given a state vector, x=(A,M,u,t,B) (and another nearby state, x0, the measured response is approximately:
Y(x)≈Y(x0)+W·(x−x0). (69)
One can form a cost functional for the mismatch of the model to the data:
F(x)=∥(Y(x)−M)∥σ2=Σk(Y[k])2/σk2=(Y−M)TQ−1(Y−M) (70)
Where σk is the standard deviation (in counts) of the kth measurement, namely √{square root over (M[k])} and Q is defined in equation (57).
One selects a candidate solution for x and then seeks to update it in a fashion that minimizes the cost functional. One method of minimizing this is via steepest descent iteration. A steepest descent step simply updates the guess using some fraction of the gradient, xj+1=xj−Δt·∇F(xj). The gradient of the cost functional given in equation (68) is simply
∇F(x)=WTQ−1(Y(x0)−M+W·(x−x0)) (71)
The Levenberg-Marquardt algorithm extends this to handle quasi Newton steps by adding a curvature dependent regularization term and iteratively solving:
(WTQ−1W+λ·diag(WTQ−1W))·δ=WTQ−1(M−Y(x0)) (72)
where
δ=(xj+1−xj), (73)
and the regularization parameter is updated via
In the case of a velocity only solution, one may correlate the phase shift of the acquired data against the response of the model. A normalized correlation operation will produce a maximum for the correct response when swept through a sequence of parameters. This may be efficiently implemented by Fast Fourier Transforms. Correlation has a long history of utilization in Radar applications. This concept may be extended to solve for temperature and aerosol and molecular density.
One advantage of the range imaging LIDAR system 10 is that the associated ring or pattern parameters can be reconfigured rapidly. The micromirrors 144 of the digital micromirror device (DMD) 142 can be reconfigured in about 10 microseconds. This allows the instrument to adapt as the environment changes. One other advantage of this type of system is that there is no read noise from the pixels like there is with an imaging photodetector such as a CCD. The only noise is from the first 154A′ and second 154B′ photomultiplier detectors which when cooled produces very low background signals. Also, the range imaging LIDAR system 10 uses the molecular response as well as the strong aerosol response which has a very high signal to noise ratio and effectively reduces the system error due to noise; the range imaging LIDAR system 10 can account for and exploit the known effects due to thermal broadening; the range imaging LIDAR system 10 can simultaneously measure velocity, temperature, aerosol and molecular components, and the range imaging LIDAR system 10 can adapt to the changing environment in order to always produce measurements based on the highest sensitivity.
However, this is subject to several limitations, the first being the relatively low quantum efficiency of the first 154A′ and second 154B′ photomultiplier detectors and the second being the fact that only two of the patterns 190 of pixel mirror rotational states 148, 150, 152 or “ring sets” are being monitored at any given time. However, there is need to cycle amongst all of the patterns 190 of pixel mirror rotational states 148, 150, 152 with equal temporal resolution. The knowledge of aerosol content might only be required infrequently to provide a reasonable measurement of the Ratio parameter. Temperature is not always required and again could be provided only at infrequent intervals. Accordingly, the basic advantage of the edge type of detection can be achieved with the range imaging LIDAR system 10, and most of the limitations associated with the simple edge detection can be eliminated.
The range imaging LIDAR system 10 can be employed utilized for any optical remote sensing scenario. Every remote sensing problem is solved by fitting a model for the system response to the data observed while accounting for the expected deviations in the data. In a Fabry-Pérot interferometer 44 system this response is a collection of fringes for which exists a wealth of phenomenological models. The range imaging LIDAR system 10 incorporates a digital micromirror device (DMD) 142 in cooperation with a Fabry-Pérot interferometer 44 to segment the optical response between two fast photodetectors. These segmented measurements are made using patterns 190 of pixel mirror rotational states 148, 150, 152 based on the derivatives of the model with respect to each parameter to be estimated thereby granting the highest sensitivity possible. An optimization with respect to segmentation thresholds and timing exposure resolution is performed to minimize the covariance of the minimum variance unbiased estimator of the system. Cost functions based on this covariance may be formed to allow trade-offs to be computed automatically with nonlinear optimization techniques such as BFGS or the Nelder-Mead Simplex algorithm. The ability to use fast photodetectors allows one to apply the range imaging LIDAR system 10 to problems where one wishes to measure state variable with a fine spatial resolution.
There are future possibilities for improving the range imaging LIDAR system 10 when digital micromirror devices (DMD) 142 become available having more than two programmable angle states. In this case one could step the digital micromirror device (DMD) 142 through a range of angles and, by using an array of photomultiplier detectors 154′, observe many more patterns 190 of pixel mirror rotational states 148, 150, 152 at one time. The patterns 190 of pixel mirror rotational states 148, 150, 152 producing these observations could be optimized in much the same way as described here by simply increasing the number of threshold states used for each derivative.
In operation of the third aspect of an associated detection system 46.3 of a range imaging LIDAR system 10 first calibrates the Fabry-Pérot etalon 50 by analyzing the reference fringe pattern 104, and then generates measures of aerosol counts A, molecular counts M, velocity u, temperature t, and background counts B from the scatter 60 and reference 104 fringe patterns at one or more particular nominal ranges R, or as a function of nominal range R, by parsing the scatter fringe pattern 60 in accordance with the process illustrated in
Referring to
In accordance with a first aspect of signal processing associated with the second embodiment of a third aspect of an associated detection system 46.3, 46.3″, the first 198′ and second 198″ portions are sequentially reflected using different associated pixel mirror rotational states 148, 150 of the associated array of micromirrors 144 of the digital micromirror device (DMD) 142 at different times, wherein the first 198′ and second 198″ portions are relatively disjoint as for the first embodiment the third aspect of the associated detection system 46.3, 46.3′, so that the resulting signals 200, 202 correspond to the complementary signals 156, 158 that would otherwise be sampled by the first embodiment the third aspect of the associated detection system 46.3, 46.3′. Accordingly, for each and every parameter, the micromirrors 144 of the digital micromirror device (DMD) 142 associated with the first disjoint portion 60′ of the scatter fringe pattern 60, or the first disjoint portion 104′ of the reference fringe pattern 104, within the region being processed are set to the first pixel mirror rotational state 148 at a first point in time to measure the first complementary signal 156, and the micromirrors 144 of the digital micromirror device (DMD) 142 associated with the second disjoint portion 60″ of the scatter fringe pattern 60, or the second disjoint portion 104″ of the reference fringe pattern 104, within the region being processed are set to the first pixel mirror rotational state 148 at a second point in time to measure the second complementary signal 158. During both the first and second points in time, the micromirrors 144 of the associated array of micromirrors 144 of the digital micromirror device (DMD) 142 outside of the region being processed are set to the second pixel mirror rotational state 150 so as to cause the remaining portion of either the scatter fringe pattern 60 or the reference fringe pattern 104 from the Fabry-Pérot interferometer 44 impinging thereupon to be reflected in the second direction 166 to a stray light block 172′ that provides for absorbing light impinging thereupon. An additional stray light block 172′ is provided to receive stray light reflected from the digital micromirror device (DMD) 142. This process is repeated for each of the parameters being detected. Accordingly, a total of 2N measurements are needed in order to identify N parameters using the first aspect of signal processing associated with the second aspect of the second embodiment of a third aspect of an associated detection system 46.3, 46.3″.
In accordance with a second aspect of signal processing associated with the second embodiment of a third aspect of an associated detection system 46.3, 46.3″, only N+1 measurements are needed within each region of the scatter 60 or reference 104 fringe patterns to identify N parameters associated with that region, wherein one of the measurements is of the light from the entire region, and the remaining N measurements are for one of the disjoint portions 60′, 104′ or 60″, 104″ associated with each of the parameters. Then, either the signals associated with the remaining disjoint portions 60″, 104″ or 60′, 104′ are then found for each parameter by subtracting the corresponding measurement for the one of the disjoint portions 60′, 104′ or 60″, 104″ from the corresponding measurement of the total signal 203 for the entire region, or the N parameters are identified by solving a system of equations based upon the N+1 measurements directly, rather than the corresponding 2N complementary signals.
Accordingly, the measurement of the total signal 203 for the entire region is made by setting the associated micromirrors 144 of the digital micromirror device (DMD) 142 to the first pixel mirror rotational state 148 at a first point to make a measurement of the total signal 203 from the light of that entire region as one of the first 200 and second 202 signals. Then, for each parameter, as corresponding distinct points in time, the micromirrors 144 of the digital micromirror device (DMD) 142 associated with either the first 60′, 104′ or second 60″, 104″ disjoint portion within the region being processed is set to the first pixel mirror rotational state 148 at that point in time to measure the other of the first 200 and second 202 signals corresponding to the first 156 or second 158 complementary signal. While these measurements are being made, the micromirrors 144 of the associated array of micromirrors 144 of the digital micromirror device (DMD) 142 outside of the region being processed are set to the second pixel mirror rotational state 150 so as to cause the remaining portion of either the scatter fringe pattern 60 or the reference fringe pattern 104 from the Fabry-Pérot interferometer 44 impinging thereupon to be reflected in the second direction 166 to a light block 172 that provides for absorbing light impinging thereupon. The remaining second 158 or first 156 complementary signal is then found by subtracting the measured first 156 or second 158 complementary signal from the total signal 203, for each of the N different parameters, or the first 200 and second 202 signals are used directly to solve for the N parameters.
The method of processing the disjoint portions 60′, 60″, 104′, 104″ of the associated scatter 60 and reference 104 fringe patterns, or one of the disjoint portions 60′, 60″, 104′, 104″ in combination with the corresponding total signal 203, can also be applied in cooperation with other systems that provide for generating the associated disjoint portions 60′, 60″, 104′, 104″ similar to that provided for by one or more digital micromirror devices (DMD) 142 as described hereinabove, but without requiring a digital micromirror device (DMD) 142.
For example, in one embodiment, a Liquid Crystal Device (LCD), could be used to generate the associated disjoint portions 60′, 60″, 104′, 104″ that are extracted from the associated underlying scatter 60 or reference 104 fringe pattern by controlling the pattern of transmission of associated pixels of the LCD provide for transmitting corresponding selected disjoint portions 60′, 60″, 104′, 104″ at any given time. For example, this can be accomplished by replace one of the polarizers normally used in the LCD with a polarization selective beam splitter, wherein the beam splitter provides for a transmission of one polarization while reflecting the other polarization. The output of the LCD would then consist of the selected disjoint pattern and its compliment, one transmitted and the other reflected.
As another example, a Holographic Optical Element (HOE), could be fabricated that would direct the light from disjoint regions onto individual areas. A Holographic Optical Element (HOE) could be constructed that would focus the light from a ring for example onto a single small area where a detector could be located. Separate disjoint areas would direct the light to different detectors which would then be used to detect the light in each disjoint pattern.
As yet another example, micro-machined mirrors could be fabricated to focus the light in a selected pattern onto a particular region. Detectors located at those regions would then convert the light to an electrical signal that would be measured and processed.
As yet another example, individual masks could be moved into position to generate the disjoint patterns. These masks could be configured around the edge of a disk and the individual masks rotated into position or the masks could be arranged in a linear or two dimensional array, and either a linear or a pair of linear actuators could be used to move the selected masks into position.
Alternatively, the disjoint portions 60′, 60″, 104′, 104″ can be extracted from an electronically captured image 114 of the scatter 60 or reference 104 fringe pattern that—or the corresponding regions thereof to be processed corresponding to the associated scattered 32′ and reference 105 light signals—is subsequently compressed by using electronic or software integration or binning as described hereinabove. For example, the image 114 may be captured using the first aspect of the associated detection system 46.1, for example, using an electronic camera, for example, a CCD detection system 46.1′, from which the corresponding linear scatter 60L and reference 104L fringe patterns are for example formed in accordance with the methodology described hereinabove and illustrated in
Referring to
Referring to
In one embodiment, the image 114 of the reference light signal 105 is positioned so as to not overlap the associated scattered light signal 32′ in the output focal plane 44.2 of the Fabry-Pérot interferometer 44. In another embodiment, in accordance with the eighth aspect of the range imaging LIDAR system 10, 10viii described more fully herein below, the image 114 of the reference light signal 105 is positioned so as to overlap the associated scattered light signal 32′, with the portion of the reference light signal 105 overlapping the scattered light signal 32′ blocked by an associated mask 138 between the lens 134 and the second beam splitter optic 136. In yet another embodiment, the light source 12 is pulsed, for example, a pulsed Nd:YAG laser 12.1, and the associated detection system 46—for example, using a fast CCD detection system 46.1′ instead of the relatively slower DVD-based detection system 46.3 as illustrated—is sampled in synchronism with the light source 12 so as to provide for initially capturing the reference light signal 105 prior to receiving the scattered light signal 32′, and to then receive the process the scattered light signal 32′ thereafter.
Referring to
Referring to
Referring to
The seventh aspect of the range imaging LIDAR system 10, 10vii comprises a pyramidal image combiner 312 that provides for separating the scattered light signals 32′ from one another in the image 114, for example, uniformly separating the scattered light signals 32′ from one another as illustrated in
Referring to
Referring to
The mask 138, 138.1 is configured and aligned so as to provide for masking all of the light from the uniform and diffuse reference beam 90′ for which the image thereof at the output focal plane 44.2 of the Fabry-Pérot interferometer 44 would otherwise overlap the corresponding image 114′ of the scattered light signal 32′. Accordingly, within the output focal plane 44.2 of the Fabry-Pérot interferometer 44, the light within the region 326 associated with the image 114′ of the scattered light signal 32′ is exclusively from the scattered light 32, and light associated with the remaining region 328 of the output focal plane 44.2 is exclusively from the uniform and diffuse reference beam 90′.
The reference illuminator 324 that provides for illuminating the mask 138 could be implemented in various ways. For example, in one embodiment, the rotating diffuser 308 may be replaced with a scanning mirror that would scan a narrow laser beam across the inside of the integrating sphere 310. In another embodiment, the integrating sphere 310 could be replaced by either single or multiple diffusers. In yet another embodiment, optics could be employed to provide for a uniform illumination of the mask 138.
Referring to
The range imaging LIDAR system 10, 10viii may be expanded with additional sets of receiver optics 24, either with one or more associated beams of light 18, in cooperation with a common Fabry-Pérot interferometer 44,—for example, similar to the fifth through seventh aspects of the range imaging LIDAR system 10, 10v-vii illustrated in
For example,
As another example,
As yet another example,
For each of the embodiments illustrated in
Referring to
Referring to
Referring to
For example, the aircraft 400, 400.1 and UAV 402 illustrated in
As another example, the aircraft 400, 400.1, 400.2, UAV 402, and balloon 404 illustrated in
As yet another example, the satellite 406 and the ground-based LIDAR system 408 illustrated in
As yet another example, the ground-based LIDAR system 408 and associated range imaging LIDAR system 10 may be operatively associated with a gimbal mechanism 410 comprising an azimuthally-rotatable platform 412 which is adapted to pivotally support associated beam steering optics 210 so as to provide for an elevational rotation thereof relative a base 414 to which the azimuthally-rotatable platform 412 is operatively associated. Accordingly, the azimuthally-rotatable platform 412 is adapted to rotate relative to the base 414, for example, responsive to an associated motor drive system, so as to define an associated azimuth angle of the beam steering optics 210, and the beam steering optics 210 is adapted to rotate relative to the azimuthally-rotatable platform 412, for example, responsive to an associated motor drive system, so as to define an associated elevation angle of the beam steering optics 210.
The range imaging LIDAR system 10 in accordance with any of the above-described aspects 10i-10viii can be employed utilized for any optical remote sensing scenario. For example, the range imaging LIDAR system 10 could be applied to the detection of Clear Air Turbulence, Optical Air Data systems, Atmospheric Aerosol Characterization, Smog detection and Chemical/Biological Agent detection. The range imaging LIDAR system 10 can be used to provide air data for Field Artillery Fire Direction Control, Small Arms Wind correction, Airport Turbulence Monitoring and Ship Navigation velocity/weather monitoring. The range imaging LIDAR system 10 can also be used to provide air data for predicting winds for any sporting events in which micro-scale airflow plays a significant role such as golf, football, baseball, etc. This range imaging LIDAR system 10 can also be used to provide air data for Wind Farm Site Prospecting, Assessment, and Optimization, Wind Farm Monitoring, Wake Effects Measurement and Analysis Wind Turbine Control, and Weather Forecasting for Wind Farms and Grid Management.
For example, in application to artillery, the range imaging LIDAR system 10 can be mounted on a vehicle or carried by an operator to a location from which artillery is to be fired. The range imaging LIDAR system 10 would then measure atmospheric parameters such as wind speed, wind direction, temperature, density, and pressure in the atmospheric volume through which the projectile will be fired. These are the standard inputs to contemporary fire direction control systems in use by the military, for example, as described in FM 6-40/MCWP 3-16.4 Tactics, Techniques, and Procedures for FIELD ARTILLERY MANUAL CANNON GUNNERY (Field Manual), which is incorporated herein by reference. By accounting for these atmospheric parameters along the projectile's flight path, the circular error probable (CEP) can be reduced and accuracy improved.
As another example, in application to sailing ships, the range imaging LIDAR system 10 can be used to provide measures of wind speed, wind direction, temperature, density, pressure, or the associated wind field around the ship, for ships that obtain their propulsion from the wind. For example, racing yachts such as used in the America's Cup, can benefit from knowing the winds near their ship as well as the winds near their competition. This information can be used to provide for trimming sails, deploying wings or aerodynamic propulsion devices, or planning trajectories so as to take maximum advantage of the current wind conditions. Recreational users can similarly use information about the winds blowing in the region near their craft.
As yet another example, in application to sporting events, the range imaging LIDAR system 10 can provide information about the local winds so as to enable participants to adapt accordingly. For example, a golf player can compensate for or take advantage of local winds, given information about how the wind is blowing over the entire flight path of the ball, or if a wind gust was approaching or would soon dissipate, so as to enable the golfer to either adjust their shot according, or to wait for better conditions. Even if the wind information is not available to the individual players, it would be of benefit to broadcasters in showing the viewing audience a graphic of the winds, a trajectory of the ball, and how the winds affected a particular shot. The range imaging LIDAR system 10 can also be of benefit in other sporting venues, such as baseball or football, for example, so as to enable broadcasters to illustrate how a baseball might have been held up by the winds in the stadium, or to show how winds had impacted a pass, punt or field goal in football, to as to enhance the viewing experience for fans. Given information about the winds in the stadium, players could adjust their actions accordingly, for example, when hitting a fly ball or kicking a field goal.
As yet another example, in application to the control of wind-induced building sway, the range imaging LIDAR system 10 can provide advance information about the wind field of a building so as to provide for wind-responsive or wind-anticipative control of tall buildings that are otherwise subject to sway in strong winds. Most modern tall buildings incorporate some form of damping to control how much the building sways in strong winds. The range imaging LIDAR system 10 can provide a predictive component (feed forward) to the associated control loops, so as to provide for improving the performance of these damping systems.
As yet another example, in application to road safety, the range imaging LIDAR system 10 can be used to monitor the wind fields that affect bridges, so as either to provide for an active control of the bridge structure responsive thereto, or to provide for controlling or limiting traffic over the bridge. Similarly, the range imaging LIDAR system 10 can be used to monitor wind conditions along roads in zones where high winds regularly pose a danger to travelers, and provide a real-time alert to motorists who are about to enter these zones. The range imaging LIDAR system 10 can be used to detect the presence of fog in fog-prone road zones, and to alert motorists of the presence of fog in advance of entering these zones.
As yet another example, in application to the control and/or dispersal of air pollution, the range imaging LIDAR system 10 can be used in a portable wind measuring system so as to enable responsible parties to more accurately predict where airborne pollution is headed as well as assisting in the assessment how much the pollution is being dispersed or diluted. Local wind mapping along with temperature and pressure measurements would provide input to models for prediction of the Nominal Hazard Zone even when there are no visible aerosols to define the plume.
As yet another example, the range imaging LIDAR system 10 can be used in a wind tunnel to provide for range resolved airflow measurements within the wind tunnel that can provide density and temperature as well as velocity of the air flow within the wind tunnel at a point, along a line, or within a volume of the wind tunnel, without perturbing the associated flow field, wherein the wind tunnel is used to measure how airflow interacts with the objects being tested therein.
As yet another example, the range imaging LIDAR system 10 can be used at an airport to enhance airport safety, for example, by providing for detecting clear air turbulence resulting from large aircraft taking off or landing, and to also provide measures of air temperature and density that can affect the lift, and hence performance, of aircraft operating at that airport.
As yet another example, the range imaging LIDAR system 10 can be used to enhance aircraft safety, for example, by providing for mapping the winds in the vicinity of an aircraft and thus providing the pilot with information that is difficult at best to obtain with other means. For example, in a roto-craft, the range imaging LIDAR system 10 can provide wind information outside of the rotor down wash so as to aid the pilot in maintaining hover in gusty wind conditions. In a conventional fixed-wing aircraft, the range imaging LIDAR system 10 can provide a measure of cross winds during landing or takeoff, and can be used to detect clear air turbulence during flight. In a sail-plane aircraft, the range imaging LIDAR system 10 can provide a measure of the wind field within which the aircraft is operating, and can provide assistance in locating updrafts in order to stay aloft. The range imaging LIDAR system 10 provides for measuring wind speed, air temperature and air density, which, for example, for purposes of landing, might not be otherwise be available at some airfields.
As yet another example, a range imaging LIDAR system 10 can be used support airdrops, for example, by either monitoring the wind field below from the aircraft making the drop so as to determine when to drop the payload, or by monitoring the wind field aloft with a range imaging LIDAR system 10 mounted on the payload so as to provide for adjusting the associated parachute during descent so as to provide for controlling the resulting drop location so that the payload is deposited closer to the desired drop zone than might otherwise be possible. Alternatively, the wind field could be monitored from above by an associated aircraft, and the resulting measurements could then be communicated to the payload to provide for controlling one or more associated parachutes or drag chutes accordingly so as to control the resulting drop location.
As yet another example, a range imaging LIDAR system 10 can be used to characterize the atmosphere. A range imaging LIDAR system 10 can be used to provide range resolved measures of velocity, temperature, and density of the atmosphere 16 that can be used by meteorologists and/or by atmospheric scientists, for example, so as to provide for predicting or analyzing the weather.
As yet another example, a range imaging LIDAR system 10 can be used on ocean and lake buoys and other ocean platforms, for example, site assessment and optimization for off-shore wind farms, oil drilling and production platforms, so as to provide range resolved measures of wind speed and direction, for example, to provide for landing helicopters, to control the location of the platform on the ocean, or to provide a warning for general platform operations in advance of the occurrence of high winds or wind gusts.
Referring to
It should be understood that the range imaging LIDAR systems 10 can be used with any fluid medium that provides for generating detectable scattered light 32 when illuminated with a beam of line 18, including, but not limited to, non-atmospheric gases flowing in a pipe and liquids flowing in pipes, channels or sprays. For example, the range imaging LIDAR systems 10 could also be used to measure water flow in pipes or channels, or to provide for measuring the speed of a marine vehicle or the associated conditions of the water upon which or within which the marine vehicle operates.
Although the range imaging LIDAR systems 10, 10i-10viii described herein have each incorporated a Fabry-Pérot interferometer 44, it should be understood that any type of interferometer could instead also be used, for example, including but not limited to either a Michelson interferometer and associated variations thereof, a Twyman-Green interferometer or a Fizeau interferometer.
Furthermore, although the range imaging LIDAR systems 10, 10i-10viii described herein have been illustrated with associated geometries that provide for detecting backscattered scattered light 32, it should be understood that a range imaging LIDAR system 10, 10i-10viii could also or alternatively incorporate an associated geometry that provides for detecting either transversely scattered light 32, or forward scattered light 32. Yet further, although the range imaging LIDAR systems 10, 10i-10viii described herein have been illustrated as providing for range-responsive measurements responsive to a range R along the optic axis 36 of the receiver optics 24, for example, a range R to the receiver optics 24 or the detection system 46, the range-responsive measurements could also be characterized with respect to a range measured along the optic axis 38 of the beam of light 18, or any other axis, by geometric transformation.
While specific embodiments have been described in detail in the foregoing detailed description and illustrated in the accompanying drawings, those with ordinary skill in the art will appreciate that various modifications and alternatives to those details could be developed in light of the overall teachings of the disclosure. It should be understood, that any reference herein to the term “or” is intended to mean an “inclusive or” or what is also known as a “logical OR”, wherein the expression “A or B” is true if either A or B is true, or if both A and B are true. Furthermore, it should also be understood that unless indicated otherwise or unless physically impossible, that the above-described embodiments and aspects can be used in combination with one another and are not mutually exclusive. Accordingly, the particular arrangements disclosed are meant to be illustrative only and not limiting as to the scope of the invention, which is to be given the full breadth of the appended claims, and any and all equivalents thereof.
The instant application claims benefit of priority from U.S. Provisional Patent Application Ser. No. 61/178,550, filed on May 15, 2009, and U.S. Provisional Patent Application Ser. No. 61/290,004, filed on Dec. 24, 2009. The entire content of each of the above-identified applications is incorporated herein by reference.
Number | Date | Country | |
---|---|---|---|
61178550 | May 2009 | US | |
61290004 | Dec 2009 | US |