Underwater acoustic communication is a technique of sending and receiving messages below water. There are several ways of employing such communication but the most common is using acoustic transducers and hydrophones. Under water communication is difficult due to factors like multi-path propagation, time variations of the channel, small available bandwidth and strong signal attenuation, especially over long ranges. In underwater communication there can be low data rates compared to terrestrial communication.
In association with the following detailed description, reference is made to the accompanying drawings, where like numerals in different figures can refer to the same element.
Systems and methods are described herein for wireless communications, such as underwater wireless communication. Among other things, reliable underwater communication can help prevent environmental disasters, e.g., oil flow from deep water sites into the ocean. The transition from wired to wireless communication has fundamentally changed how people interact and how industries operate. This technological revolution has had little impact on communication undersea. Radio waves used to carry information wirelessly above land typically propagate poorly in seawater. As a result, wireless communication technologies such as Global Positioning Satellites (GPS), Wi-Fi or cellular communication do not work below the ocean surface. Without reliable underwater wireless communication, industries and organizations that operate underwater use underwater communication that almost entirely done through wired links, e.g., a wire or a cable connects the sender to the receiver.
Additionally, underwater operations that rely on divers can be expensive, restricted to shallow waters, and put a human life at risk. The subsea industry relies on remotely operated vehicles (ROVs) for work performed in the deep ocean. An operator on the surface communicates with the machine through a bulky cable. A massive surface ship is required to safely deploy such a vehicle and handle its heavy cable to the sea floor. Even when winds are strong and waves are high, the surface ship needs to be capable to hold its position above the vehicle. Mooring or anchoring is not practical in deep water or above dense infrastructure at the sea bottom. So the ROV support ships are outfitted with expensive dynamic positioning systems that use GPS, inertial sensor and gyro compass readings to automatically control position and heading by means of active thrust. Such ships can be extremely costly.
If, instead of a cable, a wireless carrier is used to communicate with the ROV, the heavy cables can be omitted and the expensive surface ships are no longer needed. Subsea missions can be accomplished quicker, cheaper and with fewer personnel. Wireless links could eliminate the surface vessel and associated cost. The systems and methods can provide for reliable, high-speed wireless underwater communications for remote-control of subsea machinery, e.g., on the order of Mbps instead of kbps. The systems and methods can provide desirable bandwidth, data rates, range, security, and/or reliability, etc. For example, the systems and methods can allow the collection of data from underwater sensors in real-time. The exemplary embodiments described herein incorporate by reference the features and processes described in “Communication and Time Distortion” by Thomas Riedl published by the Graduate College of the University of Illinois at Urbana-Champaign, 2014.
Free-space optical communication underwater has received renewed interest from researchers due to recent improvement in laser and light emitting diode (LED) technology. LEDs are low-cost and power-efficient light sources and their light intensity and switching speed have been shown to accommodate wireless underwater communication at 1 Mbps over 100 m. Transmissions can be error free for ranges up to 100 m but the error rate can increase at ranges beyond 100 m. The error rate reaches 0.5 at about 140 m. This is still a step up from RF communication. Several issues, however, can limit the applicability of free-space optical communication in practice. For example, communication range is highly dependent upon water turbidity. The above values for light attenuation in water only hold for operation in pristine and transparent water. But near-shore and estuarine waters are typically highly turbid because of inorganic particles or dissolved organic matter from land drainage. Light attenuation is exponential in distance. If, for a given wavelength λ, I0(λ) is the light intensity at the source, the light intensity I (λ, z) at distance z from the source is described by the Beer-Lambert law:
I(λ,z)=I0e−c(λ)z (Equation 1)
The wavelength-dependent factor c(λ) is the extinction coefficient of the water through which the optical system operates. For the type of light best suited for optical communication, blue-green light with a wavelength of 480 nm, the extinction factor is about 0.16 m−1 for pristine ocean water and about 2.8 m−1 for typical coastal waters. For an extinction coefficient of 0.05 m−1 for clear water, according Equation 1, the attenuation is about 21.7 dB at 100 m distance. This suggests that in typical coastal water with an extinction coefficient of about 2.8 m−1, the system may only manage a range of about 1.8 m. The waters of many commercial interests, such as in the Gulf of Mexico or in the Irish Sea are highly turbid. Measurements in the Gulf of Mexico indicate that the extinction factor exceeds 3 m−1 at many sites and can be as high as 5.1 m−1. Turbidity is also high near underwater work and construction sites because sand and other particles are stirred up by operations. These are the spaces in which many underwater vehicles operate, and in which the need for wireless communication is great.
An issue of underwater optical communication is that different hardware is needed for the emission and reception of light—LEDs for emission and a photo-multiplier tube for reception, for example. This roughly doubles the footprint of the complete system. Further, available light emission hardware such as LEDs and lasers are highly directional and require the transmitter and receiver to be aligned with each other. This is an issue in mobile applications where the emitter needs to be constantly re-aimed as the mobile platform moves through the water. High sensitivity to water turbidity, bulkiness and tight alignment requirements are issues in free-space underwater optical communication and limit its applicability to cases where a clear line-of-sight path is available and alignment of transmitter and receiver is simple.
The above capacity calculations ignore multi-path effects and assumed line-of-sight communication between stationary platforms. In this case, the underwater acoustic channel is understood and can be modeled as a linear time-invariant (LTI) system with additive white Gaussian noise (AWGN). A line-of-sight between transmitter and receiver is often available underwater but in a mobile communication scenario the assumption of stationary communication platforms is invalid. There is no consensus on the statistical characterization of this type of time variability and, in the absence of good statistical models for simulation, experimental demonstration of candidate communication schemes remains a de facto standard.
A channel model is described herein for mobile acoustic communication that builds upon the physical principles of acoustic wave propagation and also derives communication algorithms from it to outperform existing acoustic modems by several orders of magnitude. Unlike in mobile radio systems on land, motion-induced Doppler effects is not neglected in acoustic communication systems. Remotely operated underwater vehicles (ROVs) typically move at speeds up to about 1.5 m/s, autonomous underwater vehicles (AUVs) can run at speeds greater than 3 m/s, modern submarines reach speeds greater than 20 m/s, and supercavitating torpedoes propel to speeds of up to 100 m/s. This leads to underwater acoustic Mach numbers v/c (v=vehicle velocity projected onto the signal path between transmitter and receiver, c=wave propagation speed in the medium) on the order of 10−2 and higher. In comparison, the world's fastest train in regular commercial service, the Transrapid magnetic levitation train, operates at a top speed of 430 km/h. At this speed, the radio communication channel experiences a Mach number of only 4*10−7, e.g., five orders of magnitude smaller. Relative motion between the transmitter and receiver manifests as time-varying temporal scaling of the received waveform. In radio channels, such Doppler effects are minimal and are correctable under the popular narrowband assumption, while in acoustic communications, they can be catastrophic if not compensated dynamically. Further, when acoustic communication signals have multiple interactions with scatterers underwater, such as the surface or the ocean bottom, harsh multi-path arises.
There are acoustic modems on the market that provide a transparent data link and can reach a net data rate of about 2.5 kbps over 1 km distance, but when they are mobile or multiple signal paths to the receiver exist due to reflective boundaries nearby, these modems perform poorly and only achieve a net data rate of about 100 bps. Multi-path effects are typically most severe when communication signals propagate through a wave guide or in shallow water where both the surface and the bottom reflect the acoustic signal multiple times. Note that horizontal long range communication occurs in a waveguide because waves are always refracted towards the horizontal layer of water at which the speed of sound is lowest. This phenomenon has been described as the Sound Fixing and Ranging (SOFAR) channel.
Acoustic communication uses acoustic waves to carry information. To communicate digital information acoustically, a digitized waveform is converted into an electrical signal by a suitable waveform generator circuit and this electrical signal is then amplified and delivered to an acoustic transducer. The electrical signal stimulates the transducer to vibrate. The resulting pressure fluctuations in the medium create an acoustic signal that radiates off the transducer and propagates through the water. The transducer is typically a piezo-electric ceramic encapsulated in plastic. This type of transducer can be used for both the transmission and the reception of acoustic signals. It converts electrical signals into acoustic signals and vice versa. When a transducer is used for transmission, it is often referred to as a projector. When it is used as a receiver, it is usually called a hydrophone. At some distance from projector, the hydrophone is stimulated by the incident pressure fluctuations and generates an electrical signal. The measured electrical signal is amplified and digitized by another suitable circuit.
Given a point of reference, the position and orientation of a transducer array is determined by a six dimensional vector describing the translation in three perpendicular axes combined with the rotation about three perpendicular axes, the six degrees of freedom (6DoF). A channel model described herein explicitly models these states for the transmit and the receive array. In
A model of the acoustic channel is established below that is sophisticated enough to capture the dominant physical effects but simple enough to allow computationally tractable inference. Beginning from principles of acoustic wave propagation, the acoustic signal path is considered starting at the projector array 500 and ending at the receive hydrophone array 502 as the communication channel, e.g. from sending node 600, e.g., sending vehicle, to receiving node 602, e.g., receiving vehicle. Assume that there is only one transducer element on the transmit array 500 and receive array 502 and that their positions are x1(t) and x2(t), respectively, which depend on the time t. The transmitter 500 emits the acoustic signal s (t) and the receiver 502 senses the acoustic signal {tilde over (r)}2(t). If these elements are operating in an ideal fluid, where energy was conserved and there was no absorption loss and no ambient noise, the acoustic wave equation describes the channel:
where p(x, t) is the sound pressure at position x and time t, c is speed of sound and Δ denotes the Laplace operator. Assuming there are no reflective boundaries and both transmitter and receiver move subsonically, the far field solution to this equation at position x2(t) is
where te is the unique solution to the implicit equation
The time te is often called the emission time or retarded time. Neglecting the near field component of the solution, set {tilde over (r)}2(t)=pFF(x2(t), t). This relationship completely describes the communication channel under the mentioned assumptions. Write
{tilde over (r)}
2(t)=h(t){tilde over (s)}1(te) (Equation 5)
and consider h(t) a time dependent channel gain factor. Taking a close look at Equation 3, notice that the gain h(t) is inversely proportional to the communication distance. Further the “Doppler factor”
is always positive, equal to unity when there is no motion, greater than unity when the source and receiver are moving towards each other and smaller than unity otherwise.
The solution te to Equation 4 can be interpreted as a fixed-point and can be computed by a fixed-point iteration algorithm.
In one embodiment, theorem 1 can be utilized. Assume there are two functions {dot over (x)}1(t):→3 and {dot over (x)}2(t): →3, and that {dot over (x)}1(t) is continuously differentiable and ∥{dot over (x)}1(t)∥<c. Determine the function
Then for any t and te[0], the sequence te [n], n=0, 1, 2, . . . with
t
e[n+1]=Ft(te[n]),n=0,1,2, . . . (Equation 7)
converges to a real number te(t). This number is the unique solution to the implicit equation te=Ft(te), which is equivalent to Equation 4.
Proof: (x)=∥x∥ is a continuous function and derive
for any t∈. The inequalities follow from the triangle inequality. So
The function Ft(te) is hence a contraction mapping in te. By the Banach fixed-point theorem, there exists an unique te that solves the equation Ft(te)=te and the sequence te [n], n=0, 1, 2, . . . converges to this solution. The implicit equation Ft(te)=te is equivalent to Equation 4.
The absence of absorption was assumed in the derivation of Equation 5. In reality, emitted acoustic signals experience attenuation due to spreading and absorption, e.g., thermal consumption of energy. The absorption loss of acoustic signals in sea water increases exponentially in distance and super exponentially in frequency. The loss due to spreading is in principle the same as in electromagnetics. The total attenuation of the signal power is given by
where f is the signal frequency, 1 is the transmission distance and {tilde over (S)}1(f) and {tilde over (R)}2(f) are the Fourier transforms of the signals {tilde over (s)}1(t) and r2(t), respectively. The exponent k models the spreading loss. If the spreading is cylindrical or spherical, k is equal to 1 or 2, respectively. Several empirical formulas for the absorption coefficient a(f) have been suggested. Marsh and Schulkin conducted field experiments and derived the following empirical formula to approximate 10 log10 a(f) in sea water at frequencies between 3 kHz and 0.5 MHz:
where A=2.34·10−6, B=3.38·10−6, S is salinity in promille, P is hydrostatic pressure [kg/cm2], f is frequency in kHz and
f
T=2.19·106−1520/(T+273) (Equation 19)
is a relaxation frequency [kHz], with T the temperature [° C.].
{tilde over (r)}
2(t)∫τth(t,τ){tilde over (s)}1(te(t)−τ)dτ (Equation 20)
Acoustic channel observations in reality also contains some noise. There is ambient noise and site-specific noise. Site-specific noise is for example caused by underwater machines or biologics. Ambient noise arises from wind, turbulence, breaking waves, rain and distant shipping. The ambient noise can be modeled as a Gaussian process but has a colored spectrum. At low frequencies (0.1-10 Hz), the main sources are earthquakes, underwater volcanic eruptions, distant storms and turbulence in the ocean and atmosphere. In the frequency band 50-300 Hz, distant ship traffic is the dominant noise source. In the frequency band 0.5-50 kHz the ambient noise is mainly dependent upon the state of the ocean surface (breaking waves, wind, cavitation noise). Above 100 kHz, molecular thermal noise starts to dominate. The power spectral density of the ambient noise can be measured and modeled. Researcher Coates breaks the overall noise spectrum N (f) up into a sum of four components: The turbulence noise Nt(f), the shipping noise Ns(f), surface agitation noise Nw (f) and the thermal noise Nth(f). These noise spectra are given in μPau2/Hz as a function of frequency in kHz:
10 log10 Nt(f)=17−30 log10(f) (Equation 21)
10 log10 Ns(f)=40+20(s−0.5)+26 log10(f)−60 log10(f+0.03) (Equation 22)
10 log10Nw(f)=50+7.5w1/2+20 log10(f)−40 log10(f+0.4) (Equation 23)
10 log10 Nth(f)=−15+20 log10(f) (Equation 24)
and sum up to give the total ambient noise N (f)
N(f)=Nt(f)+Ns(f)+Nw(f)+Nth(f). (Equation 25)
In this empirical expression, s is the shipping activity factor taking values between 0 and 1 and w is the wind speed in m/s.
{tilde over (r)}
2(t)∫τth(t,τ){tilde over (s)}1(te(t)−τ)dτ+{tilde over (v)}(t). (Equation 26)
So far the acoustic signal path starting at the projector and ending at the receive hydrophone has been considered as the channel. But in reality, the involved transducers and amplifiers also shape the signal and introduce noise. The notion of the communication channel to encompass the distortion effects of the involved amplifiers and transducers can be extended as well. The effect of any frequency response shaping can be absorbed into the kernel h(t, τ). But at the receiver also significant electronic noise is added. The voltage generated by a hydrophone in response to an incident acoustic signal is small and is pre-amplified to better match the voltage range of the digitizer. The electronic noise produced at the input stage of the preamplifier depends upon the capacitance of the hydrophone, but is usually so high that it dominates the acoustic ambient noise picked up by the hydrophone. The most sensitive high frequency hydrophones by market leading companies ITC and RESON introduce self-noise of at least 45 dB re μPa/√{square root over (Hz)} referred to input.
Positions and angles are given with respect to this reference system. Assume xi(t) and θi(t) are the three dimensional position and orientation vectors of the i-th transducer array. The total number of available arrays depends on the scenario but the model can start indexing them with the integer 1. Two types of arrays include: A trivial array with only one element and a non-trivial array with K elements and fixed geometry. There is a function T: 6→3×K that maps the position xi(t) and orientation θi(t) of the i-th array to the positions xi,j (t), j∈[K], of its omnidirectional elements.
The j-th transducer of the i-th array sends the signal {tilde over (s)}i,j (t) and receives {tilde over (r)}i,j (t). Assume there is no multiple access interference (MAI). So, in case there is no multi-path but only a line of sight, the received signals can be expressed as
{tilde over (r)}
l,m(t)=Σj∫τthi,j;l,m(t,τ){tilde over (s)}i,j(ti,j;l,m(t)−τ)dτ+{tilde over (v)}l,m(t) (Equation 27)
where hi,j;l,m(t, τ) denotes the time-varying signal attenuation kernel along the path from the j-th transducer of the i-th array to the m-th transducer of the 1-th array, ti,j;l,m(t) is the unique solution to the implicit equation
and the {tilde over (v)}l,m(t) are independent Gaussian noise processes with flat power spectral density in the band of interest. When there is multi-path, interpret each path as the line of sight path from a phantom source array at position xi;p(t) and orientation θi;p(t), p∈[Pi;l], that sends out the same signals. The integer Pi;l counts the number of paths present between array i and l.
{tilde over (r)}
l,m(t)=Σj∈[K],p∈P
where ti,j;p;l,m(t) is the unique solution to the implicit equation
xi;p(t), j∈[K], are the positions of the transducer elements on the p-th phantom array and hi,j;p;l,m(t, τ) denotes the time-varying signal attenuation kernel along the path from the j-th transducer of the p-th phantom of the i-th array to the m-th transducer of the 1-th array.
For signal design and sampling, as described herein waveforms that can be designed that are suitable for bandwidth efficient data communication and channel estimation. Standard single carrier source signals are well-suited for this task. It is possible to detect and track motion from the phase margin or lag with respect to the carrier (center frequency). Furthermore, modulation of the phase can be used to embed data. An approach to construction of such a communication signal is through varying the amplitude and phase of a collection of basis functions with limited bandwidth. Suppose the j-th transducer of the i-th array is to transmit length N+1 sequences of symbols si,j [n], n∈[0: N], from a finite set of signal constellation points A⊂. To this end, the sequence si,j [n] is mapped to a waveform si,j (t): →
s
i,j(t)=Σl∈[0:N]si,j[l]p(t−lT) (Equation 31)
by use of a basic pulse p(t) time shifted by multiples of the symbol period T. The pulse p(t) is typically assumed to have a bandwidth of no more than 1/T. If some of these symbols are unknown, they can usually be assumed to be i.i.d., either because the underlying symbols have been optimally compressed or randomly interleaved. This signal is then modulated to passband
{tilde over (s)}
i,j(t)=2Re{si,j(t)e2π√{square root over (−1)}f
at carrier frequency fCi. These frequencies are chosen such that there is no Multiple Access Interference (MAI), e.g., |fCi−fCi′|>1/T for all i≠i′.
At the receiving array, the signal {tilde over (r)}l,m(t) from Equation 29 is demodulated by fCi and low-pass filtered, which yields
where vl,m(t) denotes the demodulated and filtered noise processes. Motion-induced Doppler shifts compress and/or dilate, e.g., widen, the bandwidth of the received signal. If the low-pass filter had only a bandwidth of 1/T a significant fraction of the signal could be lost. Assume that vmax is the maximal experienced speed. The maximum frequency of the emitted signal is designed to be fCi+½T and a sinusoid with that frequency would then experience a Doppler shift of at most
Hence increase the cut-off frequency of the low-pass filter by fdi and sample the filtered signal at the increased frequency 1/Ti=1/T+2fdi. The samples may be stored in memory, in a circular buffer in hardware, etc. The sampled output equations read
where ti,j;p;l,m[n]=ti,j;p;l,m(nTi), vl,m[n] is the sampled noise process and
h
i,j;p;l,m[n,k]=Tihi,j;p;l,m(nTi,kTi)e2π√{square root over (−1)}f
is the demodulated and sampled kernel. The original noise process vl,m(t) was Gaussian and white in the band of interest and hence the noise samples vl,m[n] are i.i.d. Gaussian.
An objective is to communicate data sequences to the receiver, that is parts of the sequences si,j[l] are unknown and to estimate them from the available observations rl,m[n]. Unfortunately, the kernels hi,j;p;l,m[n, k] as well as the position and orientation vectors of the transmit and receive arrays are unknown as well. A possible approach to this is to model all these states probabilistically and then perform Bayesian estimation and estimate all these states jointly. The following probabilistic model of attenuation can be used.
The channel gains hi,j;p;l,m[n] are random and assume their evolution is described by the following state equations:
h
i,j;p;l,m[n+1,k]=λhi,j;p;l,m[n,k]+ui,j;p;l,m[n,k] (Equation 36)
where, for each choice of the indices i, j, p, l, m and k, the random variables ui,j;p;l,m[n, k] form an independent white Gaussian noise process in n with variance σu2. The parameter λ∈(0, 1) is the forgetting factor. More sophisticated a priori models for the evolution of these gains could be used. For a simpler model, neglect the dependence of the length and the attenuation of the involved signal propagation paths.
For probabilistic modeling of receiver motion, various motion models have been considered in the position tracking literature. There are discrete time and continuous time models. The channel observations rl,m[n] depend on transmitter and receiver motion only through the emission time ti,j;p;l,m[n] which, by definition, is the solution to the implicit Equation 30 for t=nTi. The emission time is only influenced by the values of the functions xi(t) and θl(t) where t=nTi, n=0, 1, 2, . . . , and hence model the evolution of the receiver position and orientation in discrete time. Among the commonly used discrete time motion models, the discrete d-th order white noise model is among the simplest. In this model, each coordinate xl;k(t) of the vector xl(t) is uncoupled and for each coordinate, k, the d-th derivative xl;k(d)(t) is right-continuous and constant between sampling instants and xl;k(d)[n]=xl;k(d)(nTi), n=0, 1, 2 . . . , is a white Gaussian noise process with variance σa2. Iterated integration of xl;k(d)(t) and sampling with period Ti yields the following linear discrete time state equations with Toeplitz transition matrix:
where an is an independent white Gaussian noise process with variance σa2 and d>1. Other more sophisticated motion models for example allow correlation across coordinates and take into account on-line information about the maneuver but postpone a more detailed modeling. Further, the orientation and position of an array are often correlated. Vehicles typically move into the direction of the orientation vector. For simplicity, postpone the modeling of this effect as well and assume orientation and position to evolve independently but to share the same probabilistic model.
For probabilistic modeling of transmitter motion, again, the channel observations rl,m[n] depend on transmitter motion only through the emission time ti,j;p;l,m[n]. If both the position xi;p(t) and the orientation θi;p(t) of the (phantom) transmit array are modeled by random processes with continuous sample paths and their speed is bounded by a sufficiently small value, then the positions xi,j;p(t), j∈[K], p∈[Pi;l], of its array elements also have continuous sample paths and their speed is less than the speed of sound. In that case, by Theorem 1, there is a unique solution ti,j;p;l,m[n] to the implicit Equation 30 for t=nTi and each array element j∈[K] and path p∈[Pi;l]. Note that the emission times ti,j;p;l,m[n], j∈[K], p E [Pi;l] can be viewed as hitting times
In one example, model each coordinate of the transmitter position xi;p(t) and orientation θi;p(t) as independent strong Markov processes. More specifically, model the evolution of each coordinate by a bidimensional random process, the first dimension is a speed process, modeled as a Brownian motion reflected off a symmetric two-sided boundary, and the second dimension is the position process, which is the integral of the first dimension. For this setup, conjecture that the vector (xi;p(t), {dot over (x)}i;p(t), θi;p(t), {dot over (θ)}i;p(t)) describes a Feller process and that the set of states
{(xi;p(t),{dot over (x)}i;p(t),θi;p(t),{dot over (θ)}i;p(t)),t=ti,j;p;l,m[n],j∈[K]} (Equation 39)
indexed by the discrete time variable n, form a Markov chain of order R for each p∈[Pi;l] given the receiver motion. The order R depends on the array geometry and the maximal speed of the above mentioned Brownian motion speed processes. This conjecture is proved for some special cases and it is discussed on how these proofs could be extended to cover the general case. The first special case look at is that of one dimensional motion on a line with one element transmit and receive arrays.
Let the random processes xi(t) and xl(t) denote the position of the transmitter and receiver on the real line at time t, respectively. The simple model presented in Equation 36 may not be sufficient when transmitter motion is allowed. It would allow the transmitter and receiver to get arbitrarily high velocities with non-zero probability, leading to supersonic speed and non-unique emission times. Transmitter speed needs to be bounded in order for Theorem 1 to guarantee unique emission times. Further, receiver speed needs to be bounded, in order for the emission times te[n] to form a strictly increasing sequence. This is a condition for the bidimensional process (xi(te[n]), {dot over (x)}i(te[n])) to be Markov in n. The following definitions and theorems make these points more precise and give an approximation of the transition kernel of the Markov chain (xi(te[n]), {dot over (x)}i(te[n])).
Drive the motion model by a Brownian motion.
Definition 1. A stochastic process B(t), t∈≥0, is called a Brownian motion if
1. B(0)=0
2. B(t) is continuous almost surely
3. B(t) has independent increments
4. B(t)−B(s)˜N (0, t−s) for 0≤s<t, where N (0, t−s) is the normal distribution with zero mean and variance t−s.
The following motion model uses Brownian motion as the speed process, gives continuous sample paths, is strongly Markov, has Gaussian distributed independent increments and its hitting time distribution.
Definition 2. (Integrated Brownian motion (IBM) model) For any non-negative time t∈>0, the position of the transmitter is given by
x(t)=x0+∫0t{dot over (x)}(τ)dτ, (Equation 40)
where the speed process {dot over (x)}(t) is given by
{dot over (x)}(t)={dot over (x)}o+αB(t), (Equation 41)
the values x0, {dot over (x)}0∈ and α∈>o are model parameters and B(t) is a Brownian motion as in Definition 1. For any negative time t, x(t)=x0+{dot over (x)}ot and {dot over (x)}(t)={dot over (x)}o. The bidimensional process ξ(t)=(x(t), {dot over (x)}(t)) determines the integrated Brownian motion (IBM) model.
A problem with this motion model is that speed is unbounded and hence emission times can become non-unique. A motion model is described herein that gives position sample paths whose speed is bounded by some value smaller than the speed of sound so that there is a unique emission time. The above speed process is reflected off a symmetric two sided boundary to ensure it is bounded almost surely.
Definition 3. (Integrated reflected Brownian motion (IRBM) model) For any non-negative time t∈>o, the position of the transmitter is given by
x(t)=x0+∫0t{dot over (x)}(τ)dτ, (Equation 42)
where the speed process {dot over (x)}(t) is given by
{dot over (x)}(t)=g({dot over (x)}0+αB(t)) (Equation 43)
the values x0, {dot over (x)}0∈ and α>o are model parameters and B(t) is a Brownian motion as in Definition 1. Therefore:
g({dot over (x)})=(−1)n({dot over (x)})({dot over (x)}−2{dot over (x)}maxn({dot over (x)})), (Equation 44)
where
the operator └·┐ denotes rounding to the nearest integer and {dot over (x)}max∈>0 bounds |{dot over (x)}(t)|. Both |{dot over (x)}0| and ±max are chosen to be smaller than the speed of sound c. For any negative time t, x(t)=x0+{dot over (x)}0t and {dot over (x)}(t)={dot over (x)}o. The bidimensional process J(t)=(x(t), {dot over (x)}(t)) determines the integrated reflected Brownian motion (IRBM) model.
Lemma 1. If g({dot over (x)}) and n({dot over (x)}) are the functions defined in Equation 44 in Definition 3 for some {dot over (x)}max>0, then
g((−1)m{dot over (x)}+2{dot over (x)}maxm)=g{dot over (x)} (Equation 45)
for any integer m.
Proof.
and hence
Remark 1. For applications of interest, a maximum platform speed and acceleration of the underwater vehicle is about 2 m/s and 0.3 m/s2, respectively. The parameter a in the above motion models determines the level of acceleration and is chosen such that the standard deviation of αB(Ti) is a third of 0.3Ti, e.g., α=0.1√{square root over (Ti)}. Further choose {dot over (x)}max=5 m/s.
The integrated reflected Brownian motion (IRBM) model ξ(t) defined in Definition 3 is no longer an independent increment process but its sample paths are continuous and it is a Feller process. This property may exploit the strong Markov property that follows from it.
Definition 4. (Markov Process) Let (Ω,,P) be a probability space and let (S,) be a measurable space. The S-valued stochastic process ξ=(ξ(t), t∈>0) with natural filtration (t,t∈≥0) is said to be a strong Markov process, if for each A∈, s>0 and any stopping time τ,
P(ξ(τ+s)∈A|τ)=P(ξ(τ+s)∈A|ξ(τ)) (Equation 54)
where
τ
={A∈
:A∩{τ≤t}∈
t for all t≥0} (Equation 55)
is the sigma algebra at the stopping time z. If Equation 54 only holds for the trivial stopping times τ=t for any t≥0, then the process is just called a Markov process. The Markov transition kernel μt,t+s(ξ0, A): ≥0×≥0×S×)→[0, 1] is a probability measure given any initial state ξ0∈S and any t, s>0 and further:
μt,t+s(ξ(t),A)=P(ξ(t+s)∈A|ξ(t)) (Equation 56)
for any A∈ and any t, s>0. A Markov process is homogeneous if for any initial state ξ0∈S, any A∈ and any t, s>0
μt,t+s(ξ0,A)=μ0,s(ξ0,A) (Equation 57)
For homogeneous Markov processes, use the notation
P
ξ0(ξ(s)∈A≡μ0,s(ξ0,A). (Equation 58)
When the expected value of some random variable G is computed with respect to this probability measure, write ξ0[G].
Definition 5. (Feller Process) Let ξ=(ξ(t), t∈≥0) be a homogeneous Markov process as defined in Definition 4. Then this process is called a Feller process, when, for all initial states ξ0:
1. for any t≥0, any event A∈S and any sequence of states ξ0∈S,
implies
2. for any
Theorem 2. The bidimensional random process ξ(t) from Definition 3 is a Feller process.
Proof. The sample paths of the process ξ(t) are continuous and hence Property 2 in Definition 5 holds. The process ξ(t) is a homogeneous Markov process and Property 1 in Definition 5 holds as well. Let tx and t{dot over (x)} be the natural filtrations of the processes x(t) and {dot over (x)}(t), respectively.
Establish from the definition of the function g({dot over (x)}) in Equation 44 that
αB(t)+{dot over (x)}0=g(αB(t)+{dot over (x)}0)(−1)n+2{dot over (x)}maxn (Equation 59)
where abbreviate the notation n(αB(t)+{dot over (x)}0) by n.
Further, note that
Equation 62 follows from Equation 59. Equation 64 follows from Lemma 1. The weighted difference B′(τ)=(−1)n(B(t+τ)−B(t)) is itself a Brownian motion and independent of tx and t{dot over (x)}
Next, take a look at the conditional moment-generating function of the bidimensional process ξ(t).
Equation 68 follows from Equation 65. Equation 69 follows from the Markov property of Brownian motion. Equation 71 follows from the fundamental theorem of calculus and Equation 43. So ξ(t) is a homogeneous Markov process. Now assume there are two sequences xn: +→ and {dot over (x)}m: Z+→ such that
Equation 74 follows from the dominated convergence theorem. Convergence of the moment-generating function implies convergence of the corresponding distribution and hence Property 1 in Definition 5 holds as well.
Now assuming that the motion model for the transmitter and receiver is as defined in Definition 3, transmitter speed is bounded and there is a unique solution te to the implicit equation
for any t by Theorem 1. For the sequence te[n], the solutions of the implicit Equation 76 for t=nTi, is strictly increasing in n.
Theorem 3. Assume both transmitter and receiver motion, ξi(t) and ξl(t), are as determined in Definition 3. If te[n] denotes the solution of the implicit Equation 76 for t=nTi, then
t
e[n+1]>te[n],∀n. (Equation 77)
Further,
Proof. Evaluating the implicit Equation 76 for t=nTi and t=(n+1)Ti gives
The theorem follows from iterated application of the triangle inequality. By a suitable zero-sum expansion,
Subtracting Equation 80 from Equation 79, yields
The first inequality follows from Equation 81. The second inequality follows from the fact that the involved motion processes have bounded speed. Hence write
and conclude
This proves the inequality of Equation 77 and the right-hand side inequality in Equation 78. If instead of expanding the argument of the right-hand side norm of Equation 83, the argument of the left-hand side norm of Equation 83 is expanded, get the inequality
and conclude
The fact that the emission times te[n] are strictly increasing allows to prove that ξi(te[n]) is Markov.
Theorem 4. Assume both transmitter and receiver motion, ξi(t) and ξl(t), are as determined in Definition 3, but that the receiver motion ξl(t) is given at the times nTi. Also, let the time te[n] denote the solution of the implicit Equation 76 for t=nTi. Then the sequence ξi(te[n]) is Markov, e.g., for any A∈(2),
Proof. The sequence of σ-algebras tξ
is the stopping time σ-algebra for the stopping time te[n]+s. Since ξi(t) is a time-homogeneous strong Markov process, by Theorem 2, for any s, τ≥0 and A∈(2) that
For any two σ-algebras and of subsets of Ω, σ{, } denotes the smallest σ-algebra that contains both and . Determine
The stopping times te[n] form a strictly increasing sequence in n by Theorem 3 and hence
By the tower property of conditional expectation and Equations 94, 95 and 97, for any A∈(2),
So the process ξi(t) renews itself after any stopping time te[n].
Let δte[n+1]=te[n+1]−te[n]. By the definition of the emission times te[n],
or equivalently
Note that δte[n+1] is independent of
given ξi(te[n]) and hence
where δte is as determined in Equation 93.
There may not be an exact solution to the kernel Pξ
Theorem 5. Assume both transmitter and receiver motion, ξi(t) and ξl(t), are as determined in Definition 3, but that the receiver motion ξl(t) is given at the times nTi. Further assume that the motion ξ′l(t) is as determined in Definition 2. The value ξi(0) is given, it is the initial condition for the motion processes ξi(t) and ξ′i(t)) and it is such that
Then, for an A∈(2),
Proof. For all δte≤δtmax. Inequality of Equation 103 ensures
and hence
Note that by Theorem 3 the inequality δte[n+1]≤δtmax holds almost surely.
Thus write
By the law of total probability,
By Definition 2 and 3 and Equation 111,
because given {|{dot over (x)}i(δt)|<{dot over (x)}max, 0≤δt≤δtmax}, the integrated Brownian motion model and the integrated reflected Brownian motion model coincide. Further by monotonicity
0≤Pξ
≤Pξ
And by the Fréchet inequalities,
The following is an expression and an upper bound for the last term. Determine the square wave
S
B
(b)=Σn=−∞∞(−1)n1{2n-1<b/B
for
This function is antisymmetric around Bmax and −Bmax. Let τ be the first time the Brownian motion B(t) hits either of those values. Then, by the reflection principle,
B′(t)=B(t)+1{t≥τ}2(B(τ)−B(t)) (Equation 122)
is also a Brownian motion. It follows:
1(τ>δt
Applying the expectation operator on both sides gives
where pδt
And hence
For large real η, the following asymptotic expansion of the complementary error function exists:
The realistic values {dot over (x)}max=5, Ti=10−5, |{dot over (x)}i(0)=2 and α=0.1√{square root over (Ti)}, yield
a η=2.1143×106. The corresponding error
2 erfc(η)−erfc(3η)<10−1012 (Equation 130)
and is negligible.
The following is an expression for the approximate transition probability Pξ
Theorem 6. Assume the transmitter motion ξ′i(t) is as determined in Definition 2, the receiver motion ξl(t) is given at the times nTi and ξ′i(0) is the initial condition for the motion process ξ′i(t). Let
The random variables δt′e [n+1] and B′ have the joint distribution
P
β,γ(δt′e[n+1]∈dt;B′∈dz)=|z|[pt(β,γ;0,z)−∫0t∫0∞m(s,−|z|,μ)pt-s(β,γ;0,−∈μ)dμds]1R(z)dzdt (Equation 137)
where R=[0, ∞] if β<0, R=(−∞, 0] if β>0, ∈=sgn(−β), the function
Proof. First, manipulate the equation in the definition of the hitting time 6te[n+1] in the theorem statement. This equation reads
Note that
{dot over (x)}′
i(τ)={dot over (x)}′i(0)+αB(τ) (Equation 141)
and that
B′(τ)=−sgn(xl((n+1)Ti)−x′i(0))B(τ) (Equation 142)
is again a Brownian motion. Equation 140 is hence equivalent to
and have
Determine the random variable B′=B′(δte[n+1]). The joint distribution of the random variables δte[n+1] and B′ is known. The function pt(u, v; x, y) in Equation 139 is the transition density of the bidimensional process (∫0tB′(τ)dτ, B′(t)).
Therefore, the above theorems show that the sampled bidimensional process (xi(te[n]), {dot over (x)}i(te[n])) is Markov in n and Theorem 6 gives an excellent approximation of the transition kernel of this Markov chain.
Bayesian State Inference
On a high level, the above sections introduced a prior distribution on all relevant system states: the transmitted symbols (Equation 31), the channel gains (Equation 36), the receiver motion (Equation 37) and the transmitter motion (Theorem 4 and 6). Further, the likelihood functions are determined of the observable data given these states (Equation 34). Theoretically, this is sufficient to deduce the a posteriori distribution of the states and hence obtain estimates according to any given cost function. But inference may be found to be only tractable in some special cases, when abstaining from trying to jointly estimate all states, but instead assume that some of the states are known. The following is a case of a stationary transmitter.
Stationary Transmitter
Assume array i rests in the origin and transmits and array l is mobile and receives. If the transmit array has only one element, assuming an isotropic spreading model the generated acoustic field is spherically symmetric and the receiver cannot uniquely determine its position and orientation. In fact, the locus of possible positions is a sphere. However, if there are at least three elements on the receive array and the receiver has access to a compass and a tilt sensor, this symmetry can be broken. Accelerometers are inexpensive and can determine tilt reliably. The accuracy of magnetic compasses can be compromised by a submarine's shielding ferric hull, but gyro-compasses do not have this problem and are well suited for this task, because they rely on the effect of gyroscopic precession instead of the Earth's magnetic field. Measurements from inertial sensors can be included for state inference as explained below. For now assume that the transmitter has more than three elements and hence circumvent this problem. Further assume that there is no multi-path and that the transmitted signals are known.
Given these assumptions, the sampled output equations from Equation 34 specialize to
where ti,j;l,m[n] can now be solved for explicitly
and the noise vl,m[n] is i.i.d. Model the channel gains hi,j;l,m[n, k] as described above and model the receiver position xl(nTi) and orientation θ1(nTi) as described above. The signals si,j(t) are assumed to be known.
Equations 36 and 37 determine a linear state space system driven by Gaussian noise and Equations 145 determine non-linear output equations. Several inference methods have been developed for such systems. The extended Kalman filter (EKF) can linearize the equations around the current estimate in each step and then applies the Kalman filter equations. This algorithm can work with navigation systems and GPS, or other position determining device, to provide position information to the algorithm. When the state equations or the output equations are highly non-linear as in Equation 145, the EKF can, however, give poor performance.
The application of the Kalman filter to a nonlinear system includes the computation of the first two moments of the state vector and the observations. This can be viewed as specific case of a more general problem: The calculation of the statistics of a random vector after a nonlinear transformation. The unscented transformation attacks this problem with a deterministic sampling technique. It determines a set of points (called sigma points) that accurately capture the true mean and covariance of the sampled random vector. The nonlinear transformation is then applied on each of these points which results in samples of the transformed random vector and a new sample mean and covariance can be computed. It can be shown analytically that the resulting unscented Kalman filter (UKF) is superior to the EKF but has the same computational complexity. Developing the Taylor series expansions of the posterior mean and covariance shows that sigma points capture these moments accurately to the second order for any nonlinearity. For the EKF, only the accuracy of the first order terms can be guaranteed.
An UKF can be implemented for inference on the model presented here. In one example, a known signal (a 100 Hz wide pulse at a center frequency of 25 kHz) was played from a speaker and fed the UKF with the measurements from a moving microphone. For this simple one dimensional setup, it was verified that this approach yields position estimates with a precision of less than a millimeter.
Inertial sensors provide additional information about the trajectory to be tracked. Accelerometers, for example, provide noisy observations of the acceleration that the sensor experiences and gyroscopes measure experienced angular velocity. When using such sensors on the mobile receiver, the generated observations and measurements are taken into account by adding additional output equations to the state space system describing the position and orientation of the receiver array. This combination of sensory data is called sensor fusion. Measurements al;k[n] of the acceleration values xl;k(2)(nTi) can for example be incorporated by adding the output equations
a
l;k[n]=xl;k(2)(nTi)+uk[n], (Equation 147)
where uk [n] is assumed to be white Gaussian noise.
Ideally, not only track the receiver position and orientation, but also communicate data. As described above, broadband transmission signals si,j(t) can be used of the form
s
i,j(t)=Σl∈[0:N]si,j[l]p(t−lT). (148)
In order to communicate information from the transmitter to the receiver, assume some of the symbols si,j[l] to be unknown, i.i.d. random variables with a uniform distribution over the possible constellation points and then estimate those unknown symbols jointly with the channel attenuation and motion states. However, joint estimation of all these states may be difficult and hard to implement for several reasons.
The pulse p(t) should be band limited because as discussed above the channel is band limited. But in order for p(t) to have most of its energy in a finite band, the pulse length needs to be large and hence, for any time t, the value of si,j(t) depends on many symbols si,j[l]. The signals si,j(t) are sampled at the random times ti,j;l,m[n]−kTi in Equation 145 and
s
i,j(ti,j;l,m[n]−kTi)=Σl∈[N]0si,j[l]p(ti,j;l,m[n]−kTi−lT). (Equation 149)
The computational complexity of the EKF or the UKF is quadratic in the dimension of the state vector and both methods require that a state space model for the states to be estimated is available. The only state space system for the unknown symbols in the sequence, si,j [l], is a trivial one with very large dimensionality:
s
i,j[l]=si,j[l],∀j∈[K] and l∈Su (Equation 150)
where the set Su contains the indices of the unknown symbols. This would make the complexity of each EKF or UKF step quadratic in the size of Su, which is impractical. Another idea is run a particle filter on this high dimensional state space system and then to only update those indices in S, in each step, which are in the vicinity of [ti,j;l,m[n]/Ti−k]. The following describes a low complexity deterministic approach for joint data and channel estimation.
Deterministic Inference
The systems and methods can facilitate reliable communication over the underwater acoustic channel and at the same time be computationally light enough to allow for an implementation on modem embedded computing platforms. Assume that array i is mobile and transmits while array l is stationary and receives. Array i is trivial and only carries one transducer. Array l carries K transducers. Account for multi-path effects but assume that the Doppler is the same on all paths. This is a good approximation when all phantom sources are near each other, as is the case in long range shallow water channel for example. Since the transmit array is assumed trivial, no index is needed to enumerate its elements. Without loss of generality assume the parameters i and l fixed. In what follows there is no ambiguity as to which of the two arrays are being referring to and hence the indices i and j are dropped for the sake of notational simplicity.
Send a signal s(t) of the form described above and choose the symbols s[l] from an q-ary QAM constellation. Some of these symbols are known and used for training. Some are unknown and used for data communication.
Under the above assumptions the demodulated received signals from Equation 33 simplify to
r
m(t)=∫τthm(t,τ)e2π√{square root over (−1)}f
where m indexes the receiving transducers and the emission time tm(t) solves the implicit equation
The sent signal s(t) has a bandwidth of 1/T and can hence represent the integral in Equation 150 as a sum:
r
m(t)=Σkhm;k(t)e2π√{square root over (−1)}f
where
h
m;k(t)=Thm(t,kT)e−2π√{square root over (−1)}f
is the demodulated and sampled kernel.
Determine the sequence of arrival times tm−1 [n] as the solutions to the implicit equation
Abbreviate x(nT) by x[n] and the derivative of x(t) at time nT by {dot over (x)}[n]. Since the receiver was assumed stationary, solve for tm−1(nT) explicitly
The arrival times tm−1[n] are the inverse of the function tm(t) evaluated at the times nT. They specify when an hypothetical impulse sent from the transmitter at time nT, would arrive at the m-th receiving transducer.
If sample the signal from Equation 153 at t=tm−1[n], then
And if further multiply both sides of this equation by e−2π√{square root over (−1)}f
where hm [n, k]=hm;k (tm−1[n]) and vm[n] is some noise sequence.
These equations motivate a direct equalization estimator for the symbols s[n] of the following form:
where w[n, m, k] are the complex-valued equalizer weights. Assume that k ranges from −MA to MC for some positive integers MA and MC and that M=MA+MC+1. To reduce the number of parameters of this estimator, determine the function
and substitute tm−1[n−k] by tm;n,k−1(x[n], {dot over (x)}[n]) in Equation 159. The resulting estimator is ŝn(θ[n]), where the parameter vector θ[n]∈2MK+6 is such that its components θz[n] satisfy
This notation formalizes that the equalizer weights w[n, m, k−MA], k∈[0:M−1], m∈[K], the position x[n] and the velocity {dot over (x)}[n] are concatenated into one real-valued parameter vector, the vector θ[n]. The function ŝn (θ) reads
Choose the parameter vector θ[n] such that the following objective function is minimized:
for some number of known training symbols s[n],n=0, . . . , N. The vector {circumflex over (θ)} is the initial guess about the parameter vector θ[0] and C is a covariance matrix specifying how much confidence in this guess. The scalar −σs2 is a weighting factor and the matrix Q−1 is a weighting matrix.
The matrix T is a transition matrix with TO[n] specifying what θ[n+1] likely looks like. Allow for some error (θ[n+1]−Tθ[n]) in this plant model. Choose a simple transition matrix T with components Tz,u such that
The 6×6 sub matrix on the bottom right of T is the transition matrix for the position x[n] and the velocity {dot over (x)} [n] according to the motion model presented above for d=2.
The extended Kalman filter is known to find an approximate solution to the least squares problem in Equation 163 when run on the state space system
θ[n+1]=Tθ[n]+vn+1 (Equation 165)
and the output equations
s[n]=ŝn(θ[n])+vn (Equation 166)
for n≥0. The noise values vn are independent, mean-zero, circular symmetric complex Gaussian random variables with variance σs2. Denote the estimate of θ[n+1] given the symbols {s[l], l∈[0: n]} by {circumflex over (θ)}[n+1, n]. The initial state estimate {circumflex over (θ)}[0, −1] is a Gaussian random vector with mean {circumflex over (θ)} and covariance C. The random vectors vn are independent and mean-zero. Each vector vn is Gaussian with covariance Q. The covariance matrix Q is chosen such that its components Qz,u satisfy
for some variances σw2 and σa2. The 6×6 sub matrix on the bottom right of Q is the covariance matrix for the position x[n] and the velocity x[n] according to the motion model presented above for d=2. The 2MK×2MK sub matrix on the top left of Q is diagonal and hence renders the evolution of all equalizer weights independent. The equalizer weights are assumed to be independent of the position and velocity of the transmitter.
Perform a method akin to decision directed equalization to obtain estimates of the symbols s[n] that are unknown and used for communication. In total N+1 symbols s[n] are sent. Assume the first Npre symbols {s[l], lE[0: Npre−1]} and also a fraction of the subsequent symbols to be known. Let the set Su⊂[0: N] contain the indices of the unknown symbols. First run the extended Kalman filter on the known first Npre symbols {s[l], l∈[0: Npre−1]} and obtain {circumflex over (θ)}[Npre, Npre−1]. Now if Npre∈Su, then find the point in the symbol constellation A that is closest to ŝn({circumflex over (θ)}[Npre, Npre−1]) and declare that point to be s[Npre]. The operation of mapping a complex number to its nearest constellation point is called slicing. Now regardless whether Npre∈Su, the symbol s[Npre] is available. So the extended Kalman filter can be updated and the next prediction {circumflex over (θ)}[Npre+1, Npre]) can be computed. Now check again if Npre+1∈Su and, if so, slice ŝn({circumflex over (θ)}[Npre+1, Npre]) and declare the slicer output to be the symbol s[Npre+1]. Iterate the Kalman update and prediction steps and the conditional slicing operation until the last symbol s[N] is reached. At a high level, the estimator ŝn(λ) first resamples the received waveforms to undo any timing distortions and then filters the resampled signal to remove any frequency selectivity present in the channel. This estimator can be referred to as a resampling equalizer (RE). Algorithm 1 describes the operation of the equalizer in pseudocode. The function slice(·) performs the slicing operation. The extended Kalman filter uses the values of the partial derivatives
evaluated at {circumflex over (θ)}[n, n−1], n∈[0: N]. Approximate those numerically as shown in Algorithm 2.
Of course, it is not guaranteed that the slicer actually recovers the original symbol each time but as long as it does so most of the time, the Kalman filter remains stable. The rate at which the slicer misses is called the symbol error rate (SER). Each of the unknown QAM symbols {s[l], l∈Su} corresponds to a bit pattern. The receiver maps the sliced symbols back to their corresponding bit pattern and ideally the resulting bit sequence agrees with the bit sequence that was sent originally.
Data: The transition matrix T, the covariance matrices Q and C, the variance σs2 and the initial estimate {circumflex over (θ)} are given. Further, the set Su⊂[0: N] and the values of s[n] for n∉Su, are given.
Result: The sequence of symbol estimates ŝn, n∈[0, N], and the sequence of hard decisions
In most cases, however, there will be bit errors and the rate at which these occur is called the bit error rate (BER). If the BER at the equalizer output is too high for a given application, channel coding can be used at the transmitter to reduce the BER at the expense of the rate the sequence of information bits is transmitted. Channel coding adds redundancy to the sequence of information bits that is to be communicated. The enlarged bit sequence is mapped to QAM symbols. These symbols are unknown at the receiver and call them information symbols. The equalizer may need training before any unknown symbols can be estimated, and hence add in some known QAM symbols into this stream of information symbols. The resulting sequence carries information symbols at the indices n∈Su and known symbols at the other indices.
Data: The state vector {circumflex over (θ)} is given. The constants ∈, δ>0 are some small real numbers.
Result: The vector g∈1×2MK+6 that approximates the gradient
At the receiver the bit stream from the slicer output is fed into a channel decoder that uses the added redundancy to reduce the BER on the sequence of sent information bits. The amount of redundancy depends on the equalizer output BER and the maximal permissible BER on the sequence of information bits. BER performance can be improved significantly if the equalizer and the channel decoder collaborate. There is literature on the field of iterative equalization and decoding (also known as turbo equalization) that describes how this collaboration can be furnished. For these results to apply, the equalizer is capable of leveraging soft information from the decoder and further to produce soft output instead of sliced hard decisions. There are standard methods available to extend direct equalizers. When used in the setting of turbo equalization, refer to the equalizer as a turbo resampling equalizer (TRE).
Let x[n+1, n] and {dot over (x)} [n+1, n] denote the position estimate {circumflex over (θ)}2MK+[3] [n+1, n] and the velocity estimate {circumflex over (θ)}2MK+3+[3] [n+1, n], respectively. In some examples, it was found that, in order for the Kalman filter to converge, the initial estimates of the transmitter position x[0, −1] and velocity {dot over (x)}[0, −1] are accurate enough such that tm;0,0−1(x[0, −1], {dot over (x)}[0, −1]) deviates from tm;0,0−1(x[0], {dot over (x)}[0]) by at most about one symbol period T, for all m∈[K]. The trilateration method can be used to obtain estimates of x[0] and {dot over (x)} [0]. Transmit two chirps before any QAM symbols are sent and then measure when each of these two chirps arrives at the receive transducers. Trilateration computes two estimates of the transmitter position from these arrival time measurements—one estimate for each transmitted chirp. If it is assumed that the first chirp was sent at time t=tC1 and that the second chirp was sent at a later time t=tC2, then this method obtains estimates of the positions x(tC1) and x(tC2). The difference quotient (x(tC2)−x(tC1))/(tC2−tC1) gives the average velocity between the two times t=tC1 and t=tC2. Set {dot over (x)}[0, −1] equal to this average velocity and further set x[0, −1]=x(tC2)−tC2{dot over (x)}[0, −1].
For the derivation above, assume that the receiver array is stationary. If this assumption does not hold, still use the introduced equalizer for communication and accept that the states x[n] and {dot over (x)}[n] no longer correspond to the position and velocity of the transmitter with respect to a fixed cartesian frame of reference.
Experimental Results
The turbo resampling equalizer (TRE) demonstrated unprecedented communication performance in US Navy sponsored field tests and simulations. Some of the real data stems from the Mobile Acoustic Communications Experiment (MACE) conducted south of Martha's Vineyard, Mass. The depth at the site is approximately 100 m. A mobile V-fin with an array of transmit projectors attached was towed along a “race track” course approximately 3.8 km long and 600 m wide. The maximum tow speed was 3 kt. (1.5 m/s) and the tow depth varied between 30 and 60 m.
For interaction and discussions with the subsea oil and gas industry, focus can be on communication over shorter distances while scaling up bandwidth and data rate. In a 1.22 m×1.83 m×49 m wave-tank, an example experiment is conducted with a set of ITC-1089D transducers, which have around 200 kHz of bandwidth at a center frequency of around 300 kHz. The systems and methods can achieve about 1.2 Mbps over a distance of 12 m using this experimental setup. A 64-QAM constellation was employed and the equalizer output BER was about 10−3. In a smaller tank, rates reach about 120 Mbps over distances of less than 1 m. For the experiment, high frequency ultrasound transducers were repurposed with a bandwidth of 20 MHz and a center frequency of 20 MHz and again transmitted 64-QAM symbols. The BER at the equalizer output was about 2×10−2.
The Table shows examples to compare performance of the TRE method with competing approaches both from academia and industry. Speed values are maximum values with BER<10−9. The LinkQuest modem is representative of commercially available acoustic modems. The LinkQuest modem uses some proprietary spread spectrum (SS) method for communication. The WHOI modem uses frequency shift keying (FSK) for its robust 80 bps mode. Both of these methods handle motion but only provide low data rates. For their high data rate experiments, WHOI uses a combination of a phase-locked loop and standard linear decision feedback equalization (DFE). This method yields higher data rates than their FSK method but may require both transmitter and receiver to be near stationary. The at-sea experiments shows that at a carrier frequency of 15 kHz this method tolerates phase variations up to about 2 rad/s which corresponds to a speed of only 0.0318 m/s. The TRE method is robust to all levels of Doppler that was able to simulate in laboratory experiments and at-sea tests so far (>1.5 m/s) and still reliably obtains the highest data rates ever recorded for acoustic underwater communication. The ultrasound equipment used for the 100 Mbps experiment did not allow the transmitter or receiver to move so only the stationary case could be tested.
The systems and methods describe a sample-by-sample, recursive resampling technique, in which time-varying Doppler is modeled, tracked and compensated. Integrated into an iterative turbo equalization based receiver, the Doppler compensation technique can demonstrate unprecedented communication performance. In one example, field data stems from the MACE10 experiment conducted in the waters 100 km south of Martha's Vineyard, Mass. Under challenging conditions (harsh multi-path, ranges up to 7.2 km, SNRs down to 2 dB and relative speeds up to 3 knots) the algorithms sustained error-free communication over the period of three days at a data rate of 39 kbps at 2.7 km distance and a data rate of 23.4 kbps at 7.2 km distance using a 185 dB source. Using a 9.76 kHz of acoustic bandwidth lead to bandwidth efficiencies of 3.99 bps/Hz and 2.40 bps/Hz, respectively. Compared to frequency-shift keying with frequency-hopping (FH-FSK) with a bandwidth efficiency of 0.02 bps/Hz, which is the only existing acoustic communication method robust enough to handle these conditions, this provides an improvement of two orders of magnitude in data rate and bandwidth efficiency.
Other implementations can include applications of interest for the subsea oil and gas industry with a focus on communication over shorter distances while scaling up bandwidth and data rate. In one example, in a 1.22 m×1.83 m×49 m wave-tank, a set of ITC-1089D transducers was used which have around 200 kHz of bandwidth at a center frequency of around 300 kHz. About 1.2 Mbps was achieved over a distance of 12 m using this experimental setup. In a smaller tank, rates of 120 Mbps are reached over distances of less than 1 m. The underwater acoustic channel remains one of the most difficult communication channels.
In one embodiment, a system can include a receiver configured to receive a communication signal from a transmitter; a motion determining unit connected with the receiver configured to provide information about a motion of the receiver relative to the transmitter; and an adaptive equalizer connected with the transmitter, the adaptive equalizer configured to use the information about the motion to undo effects of time variation in the communication signal. In one embodiment, the system can include a Kalman filter connected with the receiver, where the Kalman filter estimates the motion between the transmitter and the receiver. In one embodiment, the system can include a motion determining unit configured to dynamically track compression and dilation of the received communication signal. In one embodiment, the system can include an adaptive equalizer configured to jointly handle both Doppler and time dispersion effects. In one embodiment, the system can include the adaptive equalizer configured to slice symbols from the communication signal back to a corresponding bit pattern that agrees with a bit sequence that was sent originally. In one embodiment, the system can include channel coding to reduce bit error rate. In one embodiment, the system can include the transmitter and/or the receiver being located underwater. In another embodiment, the transmitter and/or the receiver can be located at the surface of water. In one embodiment, the communications described herein can be applied to biomedical applications, such as having one of the transmitter or the receiver within the body and the other of the transmitter or the receiver outside of the body.
In one embodiment, a method can include receiving a communication signal from a transmitter; providing information about a motion of the receiver relative to the transmitter; and undoing effects of time variation in the communication signal based on the information about the motion. In one embodiment, the method can include performing transmit beam-forming based upon a location of the receiver to mitigate multi-path in short range channels. As an example, an array of transmit elements, and/or an array of receive elements can be utilized. These arrays could be used for beamforming on the transmitter, beamforming on the receiver, and/or using a vector-based receiver. In one embodiment, the method can include sending the receiver estimates of arrival times of the communication signal. In one embodiment, the method can include dynamically tracking compression and dilation of the received communication signal. In one embodiment, the method can include jointly handling both Doppler and time dispersion effects. In one embodiment, the method can include slicing symbols from the communication signal back to a corresponding bit pattern (or transmitted symbol) that agrees with a bit sequence (or transmitted symbol) that was sent originally. In one embodiment, the method can include channel coding to reduce bit error rate. In another embodiment, equalization and decoding can be performed jointly, such as via turbo equalization.
Referring to
At 1810, a sampling time can be updated. At 1812, carrier phase correction can be updated. At 1814, the sampling time and carrier phase correction can be used for compensation for Doppler effects in order to recover a Doppler corrected sample. At 1816, if the method is not operating at the symbol boundary then method 1800 can return to 1810 and 1812 for updating the sampling time and updating the carrier phase correction. If on the other hand, the method is operating at the symbol boundary then at 1818 corrected samples can be provided to the equalizer and a current received symbol can be estimated.
At 1820, if there is training data available then at 1822 the training data can be utilized to estimate a symbol error. If on the other hand, there is no training data available then at 1824 the output of the equalizer can be utilized to estimate the symbol error directly. At 1826, the estimated symbol error can be utilized to update the equalizer and Doppler compensator state vectors. At 1828, if this is the last symbol of the communications then method 1800 can end at 1830, otherwise, the method returns to 1810 and 1812 for updating the sampling time and updating the carrier phase correction.
Referring to
At 1910, a sampling time can be updated. At 1912, carrier phase correction can be updated. At 1914, re-sampling the received signal at the updated sampling time and applying updated carrier phase correction can be performed to compensate for Doppler effects to recover a Doppler corrected sample. At 1916, if the method is not operating at the symbol boundary then method 1900 can return to 1910 and 1912 for updating the sampling time and updating the carrier phase correction. If on the other hand, the method is operating at the symbol boundary then at 1918 corrected samples can be provided to the equalizer and a current received symbol can be estimated.
At 1920, if there is training data available then at 1922 the training data can be utilized to estimate a symbol error. If on the other hand, there is no training data available then at 1924 equalizer output can be quantized to a symbol constellation and the symbol error can be estimated from a tentative symbol decision. At 1926, the estimated symbol error can be utilized to update the equalizer and Doppler compensator state vectors. At 1928, if this is the last symbol of the communications then method 1900 can end at 1930, otherwise, the method returns to 1910 and 1912 for updating the sampling time and updating the carrier phase correction.
Referring to
At 2010, a sampling time can be updated. At 2012, carrier phase correction can be updated for each of the M received signals. At 2014, re-sampling each of the M received signals at the updated sampling time for that corresponding signal and applying updated carrier phase correction can be performed to compensate for Doppler effects to recover a Doppler corrected sample for each of the M received signals.
At 2016, if the method is not operating at the symbol boundary then method 2000 can return to 2010 and 2012 for updating the sampling time and updating the carrier phase correction. If on the other hand, the method is operating at the symbol boundary then at 2018 corrected samples for each of the M received signals can be provided to the equalizer. A current received symbol can be estimated jointly using all of the M received signals.
At 2020, if there is training data available then at 2022 the training data can be utilized to estimate a symbol error. If on the other hand, there is no training data available then at 2024 equalizer output can be quantized to a symbol constellation and the symbol error can be estimated from a tentative symbol decision. At 2026, the estimated symbol error can be utilized to update the equalizer and Doppler compensator state vectors. At 2028, if this is the last symbol of the communications then method 2000 can end at 2030, otherwise, the method returns to 2010 and 2012 for updating the sampling time and updating the carrier phase correction.
Referring to
At 2108, an estimate can be made of the number L of arrival paths from a source to the receiver. This estimate can include determining a main arrival path. In one embodiment, at 2110 state vectors can be initialized for each arrival path for equalizer and Doppler compensators.
At 2112, a sampling time can be updated for each arrival path. At 2114, carrier phase correction can be updated for each arrival path. At 2116, compensation can be performed for Doppler effects along the main arrival path using the sampling time and the carrier phase correction corresponding to the main arrival path. Compensation can be performed for Doppler effects along the remaining arrival paths using the sampling times and the carrier phase corrections corresponding to each of the remaining arrival paths. Re-sampling can be performed for the remaining path signals onto a time-scale of the main arrival path and then the remaining path signals can be subtracted off.
At 2118, if the method is not operating at the symbol boundary then method 2100 can return to 2112 and 2114 for updating the sampling time and updating the carrier phase correction for each of the arrival paths. If on the other hand, the method is operating at the symbol boundary then at 2120 corrected samples can be provided to the equalizer and a current received symbol can be estimated.
At 2122, if there is training data available then at 2124 the training data can be utilized to estimate a symbol error. If on the other hand, there is no training data available then at 2126 equalizer output can be quantized to a symbol constellation and the symbol error can be estimated from a tentative symbol decision. At 2128, the estimated symbol error can be utilized to update the equalizer and Doppler compensator state vectors. At 2130, if this is the last symbol of the communications then method 2100 can end at 2132, otherwise, the method returns to 2112 and 2114 for updating the sampling time and updating the carrier phase correction for the arrival paths.
Referring to
At 2208, an estimate can be made of the number L of arrival paths from a source to the receiver. This estimate can include determining a main arrival path. In one embodiment, at 2210 state vectors can be initialized for each arrival path for equalizer and Doppler compensators.
At 2212, a sampling time can be updated for each arrival path. At 2214, carrier phase correction can be updated for each arrival path. At 2216, compensation can be performed for Doppler effects along the main arrival path by resampling the received signal using the sampling time and the carrier phase correction for the main arrival path. The transmitted signal can be reconstructed from past symbol decisions and L−1 versions can be generated, where each version is time distorted and phase corrected using the sampling times and phase corrections for the L−1 remaining paths.
At 2218, if the method is not operating at the symbol boundary then method 2200 can return to 2212 and 2214 for updating the sampling time and updating the carrier phase correction for each of the arrival paths. If on the other hand, the method is operating at the symbol boundary then at 2220 the corrected received signal and the L-1 versions of the reconstructed transmitted signal can be provided to the equalizer and estimate current received symbol can be estimated.
At 2222, if there is training data available then at 2224 the training data can be utilized to estimate a symbol error. If on the other hand, there is no training data available then at 2226 equalizer output can be quantized to a symbol constellation and the symbol error can be estimated from a tentative symbol decision. At 2228, the estimated symbol error can be utilized to update the equalizer and Doppler compensator state vectors. At 2230, if this is the last symbol of the communications then method 2200 can end at 2232, otherwise, the method returns to 2212 and 2214 for updating the sampling time and updating the carrier phase correction for the arrival paths.
The systems and methods described above may be implemented in many different ways in many different combinations of hardware, software firmware, or any combination thereof. In one example, the systems and methods can be implemented with a processor and a memory, where the memory stores instructions, which when executed by the processor, causes the processor to perform the systems and methods. The processor may mean any type of circuit such as, but not limited to, a microprocessor, a microcontroller, a graphics processor, a digital signal processor, a graphics processing unit, a central processing unit, or another processor, etc. The processor may also be implemented with discrete logic or components, or a combination of other types of analog or digital circuitry, combined on a single integrated circuit or distributed among multiple integrated circuits. All or part of the logic described above may be implemented as instructions for execution by the processor, controller, or other processing device and may be stored in a tangible or non-transitory machine-readable or computer-readable medium such as flash memory, random access memory (RAM) or read only memory (ROM), erasable programmable read only memory (EPROM) or other machine-readable medium such as a compact disc read only memory (CDROM), or magnetic or optical disk. A product, such as a computer program product, may include a storage medium and computer readable instructions stored on the medium, which when executed in an endpoint, computer system, or other device, cause the device to perform operations according to any of the description above. The memory can be implemented with one or more hard drives, and/or one or more drives that handle removable media, such as diskettes, compact disks (CDs), digital video disks (DVDs), flash memory keys, and other removable media.
The processing capability of the system may be distributed among multiple system components, such as among multiple processors and memories, optionally including multiple distributed processing systems. Parameters, databases, and other data structures may be separately stored and managed, may be incorporated into a single memory or database, may be logically and physically organized in many different ways, and may implemented in many ways, including data structures such as linked lists, hash tables, or implicit storage mechanisms. Programs may be parts (e.g., subroutines) of a single program, separate programs, distributed across several memories and processors, or implemented in many different ways, such as in a library, such as a shared library (e.g., a dynamic link library (DLL)). The DLL, for example, may store code that performs any of the system processing described above.
Many modifications and other embodiments set forth herein can come to mind to one skilled in the art having the benefit of the teachings presented in the foregoing descriptions and the associated drawings. Although specified terms are employed herein, they are used in a generic and descriptive sense only and not for purposes of limitation.
Dedicated hardware implementations including, but not limited to, application specific integrated circuits, programmable logic arrays and other hardware devices can likewise be constructed to implement the methods described herein. Application specific integrated circuits and programmable logic array can use downloadable instructions for executing state machines and/or circuit configurations to implement embodiments of the subject disclosure. Applications that may include the apparatus and systems of various embodiments broadly include a variety of electronic and computer systems. Some embodiments implement functions in two or more specific interconnected hardware modules or devices with related control and data signals communicated between and through the modules, or as portions of an application-specific integrated circuit. Thus, the example system is applicable to software, firmware, and hardware implementations.
In accordance with various embodiments of the subject disclosure, the operations or methods described herein are intended for operation as software programs or instructions running on or executed by a computer processor or other computing device, and which may include other forms of instructions manifested as a state machine implemented with logic components in an application specific integrated circuit or field programmable gate array. Furthermore, software implementations (e.g., software programs, instructions, etc.) including, but not limited to, distributed processing or component/object distributed processing, parallel processing, or virtual machine processing can also be constructed to implement the methods described herein. It is further noted that a computing device such as a processor, a controller, a state machine or other suitable device for executing instructions to perform operations or methods may perform such operations directly or indirectly by way of one or more intermediate devices directed by the computing device.
Tangible computer-readable storage mediums are an example embodiment that includes a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions for performing all or some of the steps described herein (and may perform other steps as well). The term “tangible computer-readable storage medium” shall also be taken to include any non-transitory medium that is capable of storing or encoding a set of instructions for execution by the machine and that cause the machine to perform any one or more of the methods of the subject disclosure. The term “non-transitory” as in a non-transitory computer-readable storage includes without limitation memories, drives, devices and anything tangible but not a signal per se.
The term “tangible computer-readable storage medium” shall accordingly be taken to include, but not be limited to: solid-state memories such as a memory card or other package that houses one or more read-only (non-volatile) memories, random access memories, or other re-writable (volatile) memories, a magneto-optical or optical medium such as a disk or tape, or other tangible media which can be used to store information. Accordingly, the disclosure is considered to include any one or more of a tangible computer-readable storage medium, as listed herein and including art-recognized equivalents and successor media, in which the software implementations herein are stored.
The illustrations of embodiments described herein are intended to provide a general understanding of the structure of various embodiments, and they are not intended to serve as a complete description of all the elements and features of apparatus and systems that might make use of the structures described herein. Many other embodiments will be apparent to those of skill in the art upon reviewing the above description. The exemplary embodiments can include combinations of features and/or steps from multiple embodiments. Other embodiments may be utilized and derived therefrom, such that structural and logical substitutions and changes may be made without departing from the scope of this disclosure. Figures are also merely representational and may not be drawn to scale. Certain proportions thereof may be exaggerated, while others may be minimized. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.
Although specific embodiments have been illustrated and described herein, it should be appreciated that any arrangement which achieves the same or similar purpose may be substituted for the embodiments described or shown by the subject disclosure. The subject disclosure is intended to cover any and all adaptations or variations of various embodiments. Combinations of the above embodiments, and other embodiments not specifically described herein, can be used in the subject disclosure. For instance, one or more features from one or more embodiments can be combined with one or more features of one or more other embodiments. In one or more embodiments, features that are positively recited can also be negatively recited and excluded from the embodiment with or without replacement by another structural and/or functional feature. The steps or functions described with respect to the embodiments of the subject disclosure can be performed in any order. The steps or functions described with respect to the embodiments of the subject disclosure can be performed alone or in combination with other steps or functions of the subject disclosure, as well as from other embodiments or from other steps that have not been described in the subject disclosure. Further, more than or less than all of the features described with respect to an embodiment can also be utilized.
Less than all of the steps or functions described with respect to the exemplary processes or methods can also be performed in one or more of the exemplary embodiments. Further, the use of numerical terms to describe a device, component, step or function, such as first, second, third, and so forth, is not intended to describe an order or function unless expressly stated so. The use of the terms first, second, third and so forth, is generally to distinguish between devices, components, steps or functions unless expressly stated otherwise. Additionally, one or more devices or components described with respect to the exemplary embodiments can facilitate one or more functions, where the facilitating (e.g., facilitating access or facilitating establishing a connection) can include less than every step needed to perform the function or can include all of the steps needed to perform the function.
In one or more embodiments, a processor (which can include a controller or circuit) has been described that performs various functions. It should be understood that the processor can be multiple processors, which can include distributed processors or parallel processors in a single machine or multiple machines. The processor can be used in supporting a virtual processing environment. The virtual processing environment may support one or more virtual machines representing computers, servers, or other computing devices. In such virtual machines, components such as microprocessors and storage devices may be virtualized or logically represented. The processor can include a state machine, application specific integrated circuit, and/or programmable gate array including a Field PGA. In one or more embodiments, when a processor executes instructions to perform “operations”, this can include the processor performing the operations directly and/or facilitating, directing, or cooperating with another device or component to perform the operations.
The Abstract of the Disclosure is provided with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, it can be seen that various features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separately claimed subject matter.
This application is a continuation-in-part of U.S. application Ser. No. 13/844,543, filed Mar. 15, 2013, which claims the benefit of U.S. Provisional Application Ser. No. 61/731,406, filed Nov. 29, 2012. This application also claims the benefit of U.S. Provisional Application Ser. No. 61/977,699, filed Apr. 10, 2014. The disclosures of all of the aforementioned applications are hereby incorporated by reference in their entirety.
This invention was made with government support under contract numbers ONR MURI N00014-07-1-0738 and ONR MURI N00014-07-1-0311 awarded by the Office of Naval Research. The government has certain rights in the invention.
Number | Date | Country | |
---|---|---|---|
61977699 | Apr 2014 | US | |
61731406 | Nov 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14681584 | Apr 2015 | US |
Child | 16123598 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13844543 | Mar 2013 | US |
Child | 14681584 | US |