Optical sensing, detecting, and tracking devices may comprise a lens system to project light rays from a segment of an environment onto a planar or other surface. One such device is shown schematically in
Depending on the application, other detecting and tracking devices may be based on non-visible electromagnetic radiation. Still other detecting and tracking devices may be based on non-electromagnetic phenomena.
However, many current devices using multiple sensors typically are configured so that the sensors only react or respond to inputs local to themselves. Inputs impinging on other nearby sensors are considered to be problems to be avoided or designed around. This limits the overall resolution of, for example, location detection.
Systems and methods are disclosed for object detection and tracking using multiple sensors, such as photo sensors or acoustic wave sensors, that are arranged to have overlapping output response profiles. The sensors' outputs can then be differenced or combined to produce a location accuracy at least, and often better, than what could be achieved by the same sensors without overlapping profiles. An arrangement of sensors having such overlapping output response profiles is termed an ommatidium (plural: ommatidia).
Hyperacuity is the ability of a detection or tracking device or system to discriminate a minimal angular displacement of an object moving in the system's field of view that is smaller than the sensor dimension or spacing. The hyperacuity of such systems can be limited only by the system electronics. In preferred embodiments, analog electronics are used to achieve faster results by avoiding limitations due to sampling rate; nevertheless, the methods and systems can still be implemented in digital electronics.
Multiple ommatidia can be organized into a single system to achieve various advantages, such as increased resolution and/or wider field of view. Also, a multiple ommatidia can be organized hierarchically to achieve zooming capability for detection and tracking.
Multiple ommatidia can be configured on surface to provide a wide field of view (FoV) for detection and tracking. The multiple ommatidia can use combined methods of detection and tracking, and may use IR, visible, or polarization detection.
I. Introduction
Systems and methods are disclosed in which multiple sensors are arranged to form a combined sensing unit, called an ommatidium. The sensors may be of any of a variety of physical phenomena, such as visible, infrared or other electromagnetic waves, acoustic waves, or light polarization.
In exemplary embodiments the sensors in an ommatidium have overlapping output response profiles. This overlapping response is typically non-linear, and is in fact used to achieve a higher location resolution (hyperacuity) than would a system using similar sensors without such overlap. Further, exemplary embodiments will make use of analog components, such as fast logarithmic amplifiers, to achieve fast detection and tracking that is not limited by a sampling rate, as in a digital system. The methods and systems can nevertheless be implemented with sampling and digital processing.
For simplicity of exposition, the exemplary embodiments disclosed herein will be of methods for, and systems of, photo sensors but it will be clear to one of skill in the art that the methods and systems are not restricted to photo sensor systems. The methods and systems described herein for optical (visual light) based sensing and tracking using photo sensors will be seen as readily adaptable to other frequency ranges (such as infrared, ultraviolet or x-ray) of electromagnetic radiation for corresponding sensors and imaging devices. Further, the methods and systems in fact will be seen as readily adaptable to location and tracking devices based on other physical phenomena, such as for sonar devices that sense acoustic waves.
The following sections describe different embodiments of the inventions. Part II discloses systems and methods based on a single ommatidium array of sensors. Methods of achieving hyperacuity include differencing methods that may use logarithmic amplification of sensor signals and analog electronics for real-time tracking. Other methods disclosed include orthogonal decomposition methods that use sinusoidal weighting of the sensor signals to form a pair of resultant signals which are used to achieve hyperacuity tracking Part III discloses systems and methods based on multiple ommatidia arrays. Such systems may use a single lens or multiple lenses. Part IV discloses embodiments directed to HYPRIS™ systems, which may use multiple ommatidia systems to achieve wide field of view, and other tracking and imaging results.
II. Single Ommatidia Systems and Methods
The methods and systems disclosed in this section combine multiple sensors to function as a single location or position detection (herein just “detection”) and location tracking (herein just “tracking”) device. Part A describes exemplary physical system structures that may be used to implement the systems and methods. Part B describes methods based on differencing, and Part C describes alternate methods based on orthogonal decomposition of the sensor signals. Part D describes how the methods and systems can implement trajectory tracking Part E describes methods to compensate when a sensor's output signal response to an image feature of a detected/tracked object in fact is darker than the output response to the imaged background. Part F describes ways to combine the various methods.
A. Single Ommatidium Systems
In
The described sensor response profile is to be a monotonically decreasing function from each sensor's center; these decaying functions include but are not limited to Gaussian, Airy, cosine power, sin(x)/x, sin c(x), (1−x2)n polynomials, etc. There are several ways to generate the nonlinear response profile, including nonlinear doping of the detector, optical termination, waveguide transmission, and intentional image feature blurring to achieve the desired response profile. Each sensor has an overlapping field of view with its adjacent neighbors.
B. Differencing Systems and Methods
This part describes differencing methods by which the exemplary arrangements of multiple individual sensors into ommatidia previously described can implement hyperacuity detection and tracking.
Embodiments of the differencing method will be disclosed by way of example using the seven sensor hexagonal ommatidium configuration of
For this example case, each sensor is assumed to have a Gaussian-type, axisymmetric response profile, as exemplified in
(x−x0)2+(y−y0)2=r02
(x−x2)2+(y−y2)2=r22
(x−x3)2+(y−y3)2=r32 (II.B.1)
The solution of this set of equations can be interpreted as the intersection of three circles at the point (x, y), the circles having radii r0, r2 and r3 centered at (x0, y0), (x2, y3), and (x3, y3), respectively. Taking the difference between the second and the first equation in (II.B.1), and then taking the difference between the third and the first in (II.B.1) yields
(x−x2)2+(y−y2)2−(x−x0)2−(y−y0)2=r22−r02
(x−x3)2+(y−y3)2−(x−x0)2−(y−y0)2=r32−r02 (II.B.2)
respectively. Letting the point (x0, y0)=(0, 0) be the origin, which coincides with the geometric center of the array, the set (II.B.2) then becomes
(−2x2)x+(−2y2)y=r22−r02−x22−y22
(−2x3)x+(−2y3)y=r32−r02−x32−y32 (II.B.3)
The right hand sides of (II.B.3) are constant values:
A=r22−r02−x22−y22
B=r32−r02−x32−y32 (II.B.4)
Using Cramer's rule one finds from (II.B.3) and (II.B.4) the unknowns x and y as
Once the x and y coordinates of the image-spot have been determined, the angle φ that the spot subtends with respect to the array's reference axis and the distance ρ to the origin and center of the array are given respectively by
The term ri in these equations, i.e. the distances from the image spot center to the center of the photoreceptors, is determined by analyzing the response profile of the ommatidium's sensors. The Gaussian sensor response profile to focused light stimuli has the form:
Ii=Îe−βr
Equation (II.B.7) represents a Gaussian-like radial-dependence response. In this equation Ii=I(ri) is the signal intensity or response of the ith sensor to an image spot illuminating the sensor at a distance ri from the sensor's axis of symmetry, β is a constant that represents the rate of decay of the response with the distance from the center of the sensor. One may relate β to the value σ of the Gaussians used in statistics, the conversion is β=1/(2σ2), and Î is the peak value of the sensor response for a given image spot illuminating the center of the sensor. A cylindrical coordinate system is used here for the sensor response function.
The square of the radial distance at which the image-spot is illuminating the sensor, relative to the sensor's center, can be obtained from (II.B.7) as
Equation (II.B.8) requires knowing Î to solve for ri since only Ii, the sensor signal response due to the image spot illuminating at a distance ri, is observed or measured. It will be shown below that the knowledge of this value is not needed, as relative measures of the sensor responses will only be required.
Referring back to
which becomes
As shown in (II.B.9), Î was canceled-out by a logarithmic property. Similarly,
Substituting (II.B.9) and (II.B.10) into the right-hand side of (4) produces
In (II.B.11), d is the constant distance from the center of the sensor array to each photoreceptor. Substituting A and B from (II.B.11) into (5) give the image spot coordinates (x, y) as
For the general case, one would have
Applying some algebra, (II.B.14) and (II.B.15) can be written as
Since the only variables in (II.B.16) and (II.B.17) are the signals I0, Ii, and Ij, these equations can be written, respectively, as
x=ηi ln(Ii)+ηj ln(Ij)+η0 ln(I0)+ηd (II.B.18)
y=γi ln(Ii)+γj ln(Ij)+γ0 ln(I0)+γd (II.B.19)
Equations (II.B.18) and (II.B.19) are two linear equations in ln(Ii), ln(Ij), and ln(I0). The coefficients, using Δ=2(xiyj−xjyi), are given by
For the hexagonally packed sensor array in
Substituting these into (II.B.20) gives
The coefficient values for the seven sensor hexagonal ommatidium of
The circuit yields six pairs of solutions 512 as given by the equations above.
While implementation using analog circuitry is a preferred embodiment, the digital sampling of the sensor responses could still be used to implement the solutions given in equations (II.B.18) and (II.B.19). For an ommatidium with an alternative configuration, a comparable analysis will yield an alternate set of coefficients.
That the method for this configuration yields six pairs of solutions implies that up to and including six independently moving objects, producing six separate image features on the ommatidium can be separately located. When the differencing detection method is applied over real time, the separate detections allow for tracking the six moving objects. Generalizing from this to other configurations of ommatidia further implies that if there are a total of n many sensors in an ommatidium with n−1 arranged around a central sensor, then such an ommatidium could be able to detect and track up to n−1 objects.
C. Orthogonal Decomposition Methods and Systems
The methods and systems described in this section can provide location, motion tracking, velocity and acceleration of an image feature traversing the ommatidium. As with the differencing methods and systems, in preferred embodiments the processing may use real time parallel analog processing. Other embodiments may process the sensor signals using sampling with digital circuitry. Exemplary methods generate an orthogonal decomposition of the ommatidium sensor output responses to generate continuous position signals and, through temporal derivatives of the position signals, continuous velocity and acceleration signals of the image feature traversing the ommatidium. As with previous methods and systems, hyperacuity can be achieved.
There are two aspects of the embodiments directed to the orthogonal decomposition methods. The first aspect is related to the creation of the orthogonal signals. The second aspect relates to the processing of the orthogonal signals to generate a radial distance (rho: ρ) and incident angle (phi: ϕ, sometimes also denoted φ) with respect to the ommatidium center of an image feature being tracked as it traverses the ommatidium. At both stages, velocity and acceleration of the image feature can be generated by taking the temporal derivatives of the location/position signals.
As an overview, in some embodiments output signal intensity of the individual ommatidium sensors are correlated with an orthogonal system to provide location and motion tracking of an image feature traversing the ommatidium field of view. The sensor outputs of the ommatidium sensors are weighted with sine and cosine functions and summed using an algorithm to affix position to the sensors within the ommatidium to provide a position of an image feature with respect to the center of the ommatidium. The sums of the weighted sensor outputs represent the real and imaginary mathematical parts of the orthogonal decomposition of the image feature position. For a moving objects, orthogonal intensity components are constantly changing. The ratio of these values can be used to calculate the direction of object motion. Therefore, a change in trajectory can be extracted in real time. The temporal derivative of the position signal yields the velocity of the image feature. The second derivative yields the image feature's acceleration.
Embodiments of the orthogonal decomposition methods and systems can be understood with reference to the particular embodiment shown in
The orthogonal decomposition of the ommatidium's sensor signals can be best described by means of complex functions and reference to
sk(Îk,rk)=η(λ)Îke−βr
In (II.C.1), sk(Îk,r) is the sensor response to an assumed thin beam of light of intensity Îk illuminating the kth-photoreceptor system normally at a distance rk from the center of the sensor, η(λ) is an efficiency factor that includes the photon efficiency of the sensor, the optical transmissivity of the lens system, and other losses which are wavelength (λ) dependent. The signals from the sensor onto which the light beam is impinging and that of the neighboring sensors can be added up to give a resultant system output signal S shown in
Significantly more useful information can be obtained by weighting each signal sk by the complex exponential
This complex function relates the angular positions of the sensors to the (x, y) frame of reference of the sensors arrangement or assembly for one ommatidium. These weights are indicated in
It is convenient to call U the real part or component, and V the imaginary part or component of this decomposition of the sensor signals, both of which are mutually orthogonal to each other.
Based on (II.C.3), one may calculate the magnitude and the phase of the complex signal sum and express the composite complex signal as
That is, one obtains the magnitude (U2+V2)112 and phase φ=tan−1 (V/U) which can be used to locate the position of the image light spot in the ommatidium.
It should be noted that the U and the V signals described here comprise the orthogonal components of the intensity of the ommatidium sensor signals and should not be confused with the I, Q phase decomposition that is utilized in communications and antenna systems. Even though the mathematics of the orthogonal decomposition is comparable, the underlying principle of using sensor intensity instead of phase is novel.
For locating the light spot relative to the ommatidium image space, both the angle φ subtended between the line joining the spot and the ommatidium center and the x-axis of the ommatidium reference frame, and the distance ρ from the center of the ommatidium to the spot location must be computed. The distance is computed as the two-norm of the normalized-by-Σ orthogonal signals U and V, scaled by a number Kσ that depends on the one-sigma (or beta-value) of the Gaussian curve utilized. That is,
The coefficient r in equation (II.C.5) can be taken as the one-sigma value of the Gaussian response profile of each sensor.
For larger area of an image light spot, the sensor signal is the response corresponding to the integral, over the spot's cross-sectional area, of the Gaussian profile. This is described next with the help of
The sensor response can now be computed as the integral of the Gaussian profile over the region R over the sensor surface 720 that the light spot covers. Thus, for the ith-sensor,
The integration region R consists of the circle defined by the image light spot of radius r0. The integration of the sensor's Gaussian response profile is obtained by Riemann sums. That is, the sum of the products of the sensor Gaussian value g(ρk) at distance ρk with respect to the ith-sensor center (xi, yi), times the annular region areas ΔSk over the light-spot region R. A three-dimensional depiction of the annular slices used for the numerical integration is depicted in the following
The integration is carried over by summing up the individual products of the sensor response (Gaussian profile) at distance ρk from the sensor's center and the surface area ΔSk of the annular strip of width Δρk bounded by the light spot circle. The distance between the center (xi, yi) of the sensor and the center (x0,y0) of the image light spot is given by d. Letting the arc length of the central fiber of the annular region be ak, the area of the annular strip within the region R of the light-spot is given by
ΔSk=ak·Δρ=ρk·(2θkc)·Δρ (II.C.7)
In (II.C.7), θkc is the angle subtended by the radial vector ρk when the tip of this vector touches either point P or P′ in Figures-3 and 4, where P and P′ are the intersections of the central annular sector fiber (dotted line in the annular region in
The product of the Gaussian value g(ρk) and the annular region area ΔSk is the “volume” Vk given by
Vk=g(ρk)·ΔSk (II.C.9)
Substituting (II.C.7) into (II.C.9) yields
Vk=g(ρk)·ρk·(2θkc)·Δρ. (II.C.10)
Summing all these and replacing (II.C.8) into (II.C.10) gives
The sum extends to all annular strips that fit in the region R from ρk=ρ0=d−r0 to ρk=ρn=d+r0. The number n of annular strips between these limits is controlled by the width Δρ of the annular strips that is selected for accuracy considerations. That is, the number of strips and the strip width are related as
Substituting (II.C.12) and the normalized Gaussian response function
into (II.C.11) yields
Substituting
into (II.C.14) and performing simplifications gives
This Riemann sum can be computed numerically in simulation software.
Referring now to
The orthogonal decomposition method can be adapted to ommatidia with other configurations.
D. Tracking and Trajectory Determinations
This section describes systems and methods for continuous and analog tracking of the position, velocity and acceleration of an image feature using real time parallel analog processing of orthogonally decomposed ommatidium sensor outputs into continuous signals representing the radial distance and incident angle relative to the ommatidium reference axes of an image feature traversing the ommatidium's field of view and the temporal derivatives of these position signals to provide continuous velocity and acceleration signals of the image feature.
Further analog processing of the orthogonal signals U and V enables the determination and continuous tracking of the radial distance (ρ) and the radial angle (φ) of an image feature relative to the ommatidium reference which is the center of the central sensor (R0) in the ommatidium using analog circuitry (see
The methods and systems will be described for the particular embodiment of the seven sensor hexagonal ommatidium configuration, but it will be clear to one skill in the art that the methods and systems can be adapted for other configurations. The values of U, V, ρ and φ and their temporal derivatives may also be obtained by digitizing the ommatidium's analog sensor outputs and numerically computing these values so the embodiments also cover this approach too, as would be apparent to one of skill in the art. The derivatives may also be obtained by sampling the ommatidium's output at discrete time instants.
This orthogonal decomposition enables the determination of the angle φ and radial distance ρ of an image feature traversing the ommatidium where ρ is, ρ=kσ·r·√{square root over (U2+V2)} where r is the sensor radius. The angle φ is determined by U and V, by the formula
Thus, with real-time analog processing, from the two continuous signals U and V the image feature position within the ommatidium is specified.
Referring now to
There are two methods indicated in
S=[ρ12+ρ22−2ρ1ρ2 cos(φ2−φi)]1/2, (II.D.1)
and
Equation (II.D.1) is obtained by straightforward application of the law of cosines, while (II.D.2) is obtained by applying the law of sines as follows:
from which
From
An alternative method to determine the trajectory from the two (φ, ρ) data values is as follows: if
a=ρ2 cos(φ2)−ρ1 cos(φ1), b=ρ2 sin(φ2)−ρ1 sin(φ1) (II.D.5)
then
This method appears to be simpler to implement as it does not require inverse trigonometric functions as in the Discrete Method-1. Implementation of these two methods can be achieved using analog circuitry employing of a time-delay line, or digitally from two (φ, ρ) samples.
Referring to
dS≈(ds2+dρ2)1/2=[(ρdφ)2+dρ2]1/2, (II.D.7)
from which
Since for most cases (dρ/φ2≈0, then (II.D.8) can be written as
One can write
becomes
It can be shown that all the time-derivatives in (II.D.10) can be obtained from the time derivatives of U and V, that is, from dU/dt and dV/dt, which are given by
Substituting (II.D.11) and (II.D.12) into (II.D.10) gives
Using ρ=kσr(U2+V2)1/2 and after some algebra one obtains
This is the velocity of the image spot over the ommatidium image space. Integration of (II.D.14) over time would give displacement, and the derivative of (II.D.14) with respect to time would give the acceleration. These can be achieved using an integrator and a differentiator components in an analog circuit.
The angle η(t) of the instantaneous trajectory shown in
Substituting (II.D.11) and (II.D.12) into (II.D.15) and using
yields
E. Dynamic Contrast Control
This section describes embodiments for systems and methods for real-time tracking of an image feature with respect to the background and automatically adjusting to contrast reversals between an image feature and background. Such systems and methods can be used, for example, when an ommatidium-based system is imaging or tracking a dark object across a bright sky. Light sensors of the ommatidium can then have the bright sky background imaged onto most of their surfaces, with only a relatively small dark image feature on the sensor. As some previous embodiments presume a bright object image and dark background, a potential problem or inaccuracy can be prevented by applying an inversion-biasing to the output signals of the sensors.
The theory of operation for dynamic contrast reversal can be understood with reference to
The first case is shown in 1410, in which the image feature 1412 is brighter than the background on the sensor. The graph for 1410 shows a cross section of the sensor response I. The sensor outputs due to the background IBk and the “target image” ITi satisfy IBk<ITi. Thereafter, in the third case 1420 in which the image feature 1422 is darker than the background IBk>ITi. The detection scheme will automatically switch between one mode (normal) or the other mode (inversion and biasing) according to the relative intensities between the object image feature and the background, as the object image feature traverses the ommatidium field of view. When the situation of 1420 is detected, the detection scheme applies inversion 1430 so that the image feature 1432 is brighter than the background, and follows this with biasing 1440 so that both outputs are now positive and the image feature 1442 is now brighter than the background.
F. Combined Methods and Systems
This section describes methods and systems that combine previously described methods and systems of locating or tracking an image feature of an object in a single ommatidium. Combining the differencing methods described in section II.B with the orthogonal decomposition methods of section II.C can produce improved accuracy in position, velocity and acceleration determinations.
There are several analytical advantages gained by expressing the orthogonal and differencing methods in compact form using matrix-vector notation besides clarity and notational economy. In what follows,
The orthogonal system is described here by using complex exponential notation. For the differencing method, T2 is defined to be the distance between sensor centers for any Tn system besides the structure it refers to.
For the orthogonal decomposition system, denoted
where
For the differencing system, denoted Δ(T2), one has
Δ(T2)=[
where
for
β=4 ln(2)/T22, Δ=2(xkyk+−xk+1yk).
The methods may be combined using proportional-integral-derivative (PID) methods:
where f(ρ,φ,
represents the polar to rectangular conversion operator, and
wherein δ is the Dirac delta function, and Γ is a diagonal kth-vector Select Matrix.
The matrix A=w·I2 is a 2 by 2 diagonal weight matrix, where 0≤w≤1 would be determined on specific tracking criteria and for particular cases. If w is set to 1, then the Orthogonal Decomposition Method would be used exclusively. On the other extreme, if w is set to 0, then the Differencing Method would be used exclusively. Intermediate values of w can be set by different tracking strategies or criteria. For example, if a large object is being observed, then the weight w could be set by the ratio of the object size or angular extent divided by the ommatidia FOV extent. Other criteria may involve contrasts, brightness, object speed, distance from the ommatidia tracking limit or from object's distance from the ommatidia center (where the differencing method will have all the positions converging). One can also set the weight w based on probability measures obtained from the tracking data, from a feedback loop of the tracking platform, or from an optimization tracking algorithm that compares the two methods alternatively by switching the weight w between 0 and 1 and compares the object's trajectory and noise, etc. In summary, the algorithm(s) for determining of the optimal value for the combining weight w would be defined for each specific application. If w is set to zero, then also F=I6 and multiple objects can be tracked simultaneously.
These combined methods and systems can be implemented in analog or digital circuitry, and may provide improved location detection and tracking of one or more independent image features.
III. Multiple Ommatidia Systems and Methods
This section describes methods and systems that use one or more arrays of sensors that have been organized into subarrays, with each subarray acting as a single ommatidium, to perform detection and/or tracking of one or more objects.
A. Multiple Ommatidia Systems
The methods and systems disclosed for a single ommatidia can be applied when the sensor array comprises more than one individual ommatidia. The methods for single ommatidia can be extended and adapted to combine the detection and tracking results of each constituent ommatidium. Characteristics of the object or of multiple objects can be tracked, including size, shape, brightness, relative contrast which is directly related to object detectability, and the rate of the object(s) traversing the field of view of the system.
Various embodiments of configurations that combine multiple ommatidia, each comprising multiple sensors, are shown in
Methods the can be used in multiple ommatidia systems, such as to determine image feature position, include: orthogonal decomposition methods, differencing methods, or the combination methods described previously. Other methods will be clear to one of skill in the art. The combination orthogonal-differencing methods can use weighted methods, such as by a PID.
A similar alternate configuration 1620 uses circular shaped sensors. The central sensor of each seven sensor hexagonal ommatidium is the indicated with a bold circle.
In
The configuration 1640 of
In
ρk={[xoi+ρi cos(φi)]2+[yoi+ρi sin(φi)]2}1/2 (III.A.1)
A method for the transformation from a local to the global frame can be implemented using the following pseudo-code as follows:
The normal-to-the-surface unit vector to the ith-photoreceptor or array is given, as can be derived from
ûi=ux,uy,uz=sin(Θi)·cos(Φi), sin(Θi)·sin(Φi), cos(Θi) (III.A.3)
This vector defines also the local horizon plane, a plane tangent to the hemispherical surface at the center of the ommatidium, or at the ommatidia array's center. The south-looking unit vector ŝi is perpendicular to ûi and is contained in the plane defined by ûi and the z-axis (the trace or intersection between this plane and the hemisphere is called the local meridian of east-longitude Φi). This local-frame unit-vector (or versor), ŝi, is defined by
ŝi=sx,sy,sz)=(cos(Θi)cos(Φi), cos(Θi)·sin(Φi),−sin(θi) (III.A.4)
Finally, the east-looking unit-vector êi forms with the other two a right-handed system. Hence, it is defined by
êi=ûi×ŝi=ex,ey,ez=sin(Φi), cos(Φi),0 (III.A.5)
The local frame given by the set {ûi, ŝi, êi} for the ith—ommatidia, or array of ommatidia, is required for mapping any object's trajectory with respect to this local frame first, and then transform or map the image coordinates (or image position vectors) to the system or hemisphere frame given by the unit direction vectors {îi, ĵi, {circumflex over (k)}i}. This system frame is assumed to be the body (the convex hull hemisphere of
In more compact form, equation (III.A.6) can be written as
In (III.A.7), {right arrow over (x)}0 is the position vector of the image spot relative to the global frame, {right arrow over (x)}i is the position of the image spot relative to the ith-ommatidia, R is a composite 3-axes Euler rotation matrix given by, R=Πk=13Rk(αk), and Δ
Multiple ommatidia systems can provide improved detection and tracking capabilities, as will now be described. Other embodiments and applications include providing zooming detection and wider fields of view, as will be disclosed in sections III.C and section IV.
B. Inter-Ommatidia Feature Tracking Methods and Systems
In detection and tracking using multiple ommatidia systems, it may occur that an image feature either is near a boundary between two (or more) constituent unit ommatidia. Methods are now described for maintaining detection and tracking with hyperacuity.
During the transition, that is, when the light spot of the image feature begins to appear on a neighboring ommatidia and begins to fade out of the present ommatidia, the local vectors would point to the center of mass of that portion of the image light spot that affect the particular ommatidia. Because of this, the transformed vectors of each ommatidia excited by the light spot would differ by some relatively small amount. If precision tracking is required, there are two methods that can be followed to obtain the best estimates for the global set (ρk, φk): (i) Perform a weighted average of the two sets obtained by the transformations (1) and (2), where the weights would be computed as the ratio of the intensities (equivalent to areas) of the parts of the spot on each ommatidia, divided the total light spot intensity. That is, if I is the total light spot intensity, and I1 and I2 are the respective spot intensities detected on each ommatidia, then the weights w1=I1/I, and w2=I2/I, respectively. Then, these weights are used to perform a weighted average in both rho's and thetas obtained in the two transformations; or (ii), utilize a dynamic state estimator/propagator (or state predictor).
C. Meta-Ommatidia Zooming Methods
This section describes methods and systems by which, in a multiple ommatidia system, the separate ommatidia can be used hierarchically to provide location zooming. In some of these embodiments a location obtained for an image feature with respect to global coordinates of the multiple ommatidia system is used to select a component ommatidium of the ommatidia system, and the signals of that component ommatidium are then used to provide a finer resolution location. These hierarchical configurations of multiple ommatidia may be generalized to include the ability to zoom using both the differencing and the orthogonal decomposition methods previously described.
First embodiments of zooming make use of the orthogonal decomposition approach of section II.C.
D. Dynamic Reconfiguration in Multiple Ommatidia Systems
In the multiple ommatidia systems and methods described so far, it was presumed that the individual sensors of the entire system were uniquely assigned to respective unit ommatidia. This section describes systems and methods by which the arrangement of the individual sensors of the entire system can be dynamically reconfigured “on the fly” to provide improved tracking and detection. Such methods and systems are described with respect to the particular embodiment shown in
IV. HYPRIS System Applications of Ommatidia Systems
The following describes further embodiments (referred to as HYPRIS™) that extend and apply the embodiments disclosed above. HYPRIS™ embodiments provide continuous kinematic target tracking and may be expanded to any arbitrary field of view by combining multiple ommatidia into a multi-layered hierarchy, much like a compound eye structure. Inter-ommatidia processing and tracking may be accomplished using analog or sampling with numerical processing. In this section the objects discussed above will also be referred to as targets.
HYPRIS™ embodiments are based on modularity of the multiple ommatidia methods and systems. This modularity facilitates the placement of ommatidia on a smooth curved surface (such as vehicle hulls, aircraft wings, or missiles) providing full-surround situational awareness without breaking aerodynamic contours. The ability to operate on a wingtip rather than an airframe centerline may be made possible by real time compensation for wing motion relative to the airframe. The compound eye structure is useful for small or large airframes as it does not require pan and tilt control or large volume aplanatic optics.
A. Wide FoV Methods with Multiple Ommatidia
HYPRIS's modular system is extensible up to a full spherical FoV. A basic single ommatidium module may operate over a limited field of view (FoV); in some case this may be approximately ten degrees. Parallel, analog processing of a single ommatidium produces the continuous analog signals representing the kinematic states of up to six targets for the seven sensor ommatidium embodiment used as examples above. (Recall also that a greater number of targets can be tracked with other configurations.) Two embodiments for creating a wide FoV using an ommatidium are as follows. The first method assumes that the kinematic state signals of each ommatidium, arranged on a compound surface to achieve the desired FoV, will be numerically sampled and processed using algorithms to facilitate inter-ommatidia tracking of targets across the entire FoV. Although numerical processing is employed to accomplish multi-segment tracking, HYPRIS may use embedded processors to greatly reduce the computational burden required by image-based kinematic analysis of targets.
A second approach uses the multi-layer, hierarchical processing architecture possible with multi-ommatidia systems. This replicates the same sensing and processing methods of a basic ommatidia module at successively higher meta levels in order to preserve the ability to detect and track multiple targets over a wide FoV on a continuous analog basis. Information is propagated from higher to lower layers to detect, isolate and track objects with increasingly greater spatial resolution all in the analog domain. Since HYPRIS's parallel analog processing is not constrained by the Nyquist digital sampling rate and the processing limitations of image based sensor systems, it allows for very high speed operation.
The methods and systems may implement the capacity to calculate and subtract self-motion of a imaging sensor platform already exists but requires a separate reference system such as an INS or GPS for absolute accuracy or a large computational burden if self-motion is estimated from image-based analysis by the sensor system. In further embodiments, the methods and systems may implement embodiments that can be used to estimate and extract self-motion using kinematic state signals, which already contain motion cues, with algorithms and processing at a greatly reduced computational burden.
B. IR and Polarization Detection
In some embodiments, the multi-ommatidia configurations may use IR or polarization sensitive sensors.
C. High Level Partition Methods
By implementing one or more multiple ommatidia systems, HYPRIS methods and systems can be configured to detect and track multiple targets over a wide field of view.
This application claims priority to U.S. Provisional Patent Application No. 62/077,095, filed Nov. 7, 2014, and also claims priority to U.S. Provisional Patent Application No. 62/079,688, filed Nov. 14, 2014, both of which are incorporated by reference herein in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
4181398 | Sick | Jan 1980 | A |
4286760 | Couderc et al. | Sep 1981 | A |
4479717 | Cornillault | Oct 1984 | A |
5196689 | Sugita et al. | Mar 1993 | A |
5206499 | Mantravadi et al. | Apr 1993 | A |
5214492 | LoBianco et al. | May 1993 | A |
5220398 | Horn et al. | Jun 1993 | A |
5541652 | Jackson et al. | Jul 1996 | A |
5642299 | Hardin et al. | Jun 1997 | A |
5742325 | Curry et al. | Apr 1998 | A |
5814524 | Walt et al. | Sep 1998 | A |
5847398 | Shahar et al. | Dec 1998 | A |
5905567 | Dewan | May 1999 | A |
5909967 | Dewan | May 1999 | A |
5917191 | Cheng | Jun 1999 | A |
6175107 | Juvinall | Jan 2001 | B1 |
6215113 | Chen et al. | Apr 2001 | B1 |
6529281 | Takeshita et al. | Mar 2003 | B2 |
6573488 | Takeshita et al. | Jun 2003 | B1 |
6765195 | Leviton | Jul 2004 | B1 |
6781694 | Nahum et al. | Aug 2004 | B2 |
6828559 | Griffin et al. | Dec 2004 | B2 |
7773143 | Feldman et al. | Aug 2010 | B2 |
8344342 | Barrett et al. | Jan 2013 | B2 |
20040042653 | Hu | Mar 2004 | A1 |
20040193670 | Langan et al. | Sep 2004 | A1 |
20040227109 | Storm et al. | Nov 2004 | A1 |
20050279917 | Barrett et al. | Dec 2005 | A1 |
20080179490 | Ohno | Jul 2008 | A1 |
20120001058 | Luke et al. | Jan 2012 | A1 |
20130208126 | Barrett et al. | Aug 2013 | A1 |
20130322753 | Lim | Dec 2013 | A1 |
20160175530 | Mide et al. | Jun 2016 | A1 |
Entry |
---|
International Preliminary Report on Patentability for PCT/US2015/059625 dated May 9, 2017, 7 pages. |
International Search Report and Written Opinion for PCT/US15/59625 dated Feb. 17, 2016, 8 pages. |
D. T. Riley. W. Harman, S. F. Barrett, and C. H. G. Wright, “Musca domestica inspired machine vision sensor with hyperacuity,” IOP Journal of Bioinspiration & Biomimetics, vol. 3, No. 2, Jun. 2008. |
Harman, W., et al., “Biomimetic Machine Vision System,” 42nd Annual Rocky Mountain Bioengineering Symposium 2005, Instrument Society of America, Apr. 2005, 8 pages. |
Madsen, R., et al., “Software Model of a Machine Vision System Based on the Common House Fly,” 42nd Annual Rocky Mountain Bioengineering Symposium 2005, Instrument Society of America, Apr. 2005, 7 pages. |
Wilcox, M. et al., “A Retina with Parallel Input and Pulsed Output, Extracting High-Resolution Information,” IEEE Transactions on Neural Networks, vol. 10, No. 3, May 1999, 10 pages. |
Wright, C., et al., “Using parallel Evolutionary development for a Biologically-Inspired Computer Vision System for Mobile Robots,” 42nd Annual Rocky Mountain Bioengineering Symposium 2005, Instrument Society of America, Apr. 2005, 8 pages. |
U.S. Appl. No. 11/051,850, filed Feb. 4, 2005, Office Action dated Feb. 22, 2013, 6 pages. |
U.S. Appl. No. 11/051,850, filed Feb. 4, 2005, Office Action dated Sep. 4, 2007, 7 pages. |
U.S. Appl. No. 11/051,850, filed Feb. 4, 2005, Advisory Action dated May 4, 2007, 3 pages. |
U.S. Appl. No. 11/051,850, filed Feb. 4, 2005, Final Office Action dated Feb. 20, 2007, 6 pages. |
U.S. Appl. No. 11/051,850, filed Feb. 4, 2005, Office Action dated Sep. 29, 2006, 8 pages. |
U.S. Appl. No. 11/051,850, filed Feb. 4, 2005, Requirement for Restriction/Election dated Jul. 20, 2006, 6 pages. |
U.S. Appl. No. 11/150,478, filed Jun. 11, 2005, BPAI Decision dated Jun. 22, 2012, 14 pages. |
U.S. Appl. No. 11/150,478, filed Jun. 11, 2005, Appeal Brief filed Feb. 27, 2009, 16 pages. |
U.S. Appl. No. 11/150,478, filed Jun. 11, 2005, Final Office Action dated Mar. 14, 2007, 8 pages. |
U.S. Appl. No. 11/150,478, filed Jun. 11, 2005, Office Action dated Oct. 16, 2006, 10 pages. |
U.S. Appl. No. 11/150,478, filed Jun. 11, 2005, Advisory Action dated Jul. 20, 2006, 3 pages. |
U.S. Appl. No. 11/150,478, filed Jun. 11, 2005, Final Office Action dated Apr. 25, 2006, 9 pages. |
U.S. Appl. No. 11/150,478, filed Jun. 11, 2005, Office Action dated Dec. 28, 2005, 8 pages. |
Number | Date | Country | |
---|---|---|---|
20160134806 A1 | May 2016 | US |
Number | Date | Country | |
---|---|---|---|
62077095 | Nov 2014 | US | |
62079688 | Nov 2014 | US |