FIELD
The present disclosure concerns the field of acoustic-optical imaging, particularly for medical applications.
BACKGROUND
In a biological medium such as a human or animal tissue, visible light propagates according to a diffusion regime. In such a regime, the photons follow multiple optical paths, so that their relative arrangements are modified all along the propagation of light through the medium. It is then impossible, with a standard image sensor, to acquire a faithful image of the medium crossed by the light.
To obtain information of local absorption in the medium, so-called acoustic-optical imaging techniques have already been provided, where the observed medium is illuminated by a coherent light beam, typically a laser beam, and a portion of the observed medium is submitted to an acoustic wave, typically an ultrasound acoustic wave. The vibration of the diffusers along the propagation path of the acoustic wave, also called acoustic marking area or also acoustic marking column, and then modulates the phase of the diffused photons. At the output of the medium, one can find photons carrying the acoustic modulation, called marked photons, corresponding to the photons which have crossed the acoustic marking area, and photons which do not carry the acoustic modulation, called unmarked photons, corresponding to the photons which have not crossed the acoustic marking area. An interference between the light beam originating from the medium, called object beam, and a reference beam originating from the same light source but which has not crossed the diffusing medium, is then caused. This enables to demodulate the object beam and to take the pulse of the marked photons down to the frequency of the acoustic wave. An image sensor is then used to measure the optical signal modulated at the acoustic frequency and to deduce therefrom information relative to the absorption of light in the marking area.
It would be desirable to have an acoustic-optical imaging system, this system overcoming all or part of the disadvantages of known acoustic-optical imaging systems.
SUMMARY
For this purpose, an embodiment provides an imaging system, comprising:
a coherent light source delivering an object beam and a reference beam;
a modulation device capable of modulating all or part of the object beam with a modulation signal;
an image sensor arranged to receive an interference pattern resulting from a combination of the object beam and of the reference beam; and
a device for demodulating the modulated portion of the object beam,
the system being configured to, during a measurement phase:
apply to the modulation signal, via the modulation device, a first pseudo-random sequence of jumps of a parameter selected among the phase, the frequency, and the amplitude; and
simultaneously apply to the modulated portion of the object beam, via the demodulation device, a second pseudo-random sequence of jumps of said parameter,
wherein the first and second sequences of jumps of said parameter are non-correlated.
According to an embodiment, the system is configured to, during an acquisition phase, implement M successive measurement phases, M being an integer greater than or equal to 2, and, at each measurement phase, acquire, via the image sensor, a value representative of the complex field of the modulated portion of the object beam.
According to an embodiment, the system comprises a processing device configured to implement a step of reconstruction, based on the M values acquired during the acquisition phase, of a set of L values representative of the light absorption at L different positions on the optical path of the modulated portion of the object beam, L being an integer greater than or equal to 2.
According to an embodiment, L is greater than M.
According to an embodiment, the modulation device is an acoustic-optical modulation device.
According to an embodiment, the modulation device comprises an ultrasound transducer, the modulation signal being an ultrasound wave applied by the ultrasound transducer to a portion of a sample to be analyzed, placed on the optical path of the object beam.
According to an embodiment, the modulation device comprises a plurality of ultrasound transducers arranged in an array, linear or not, and, during each measurement phase, the ultrasound modulation wave is emitted in fractions by the different transducers according to a pseudo-random function γ.
According to an embodiment, the modulation device comprises an acoustic-optical modulator placed on the optical path of the object beam upstream of a scene to be analyzed.
According to an embodiment, the demodulation device is an electronic demodulation device integrated to the image sensor.
According to an embodiment, the demodulation device comprises an acoustic-optical modulator placed on the optical path of the reference beam, upstream of the image sensor.
BRIEF DESCRIPTION OF THE DRAWINGS
The foregoing features and advantages, as well as others, will be described in detail in the following description of specific embodiments given by way of illustration and not limitation with reference to the accompanying drawings, in which:
FIG. 1 schematically shows an example of an acoustic-optical imaging system;
FIG. 2 is a diagram illustrating the operation of the circuit of FIG. 1;
FIG. 3 schematically shows another example of an acoustic-optical imaging system;
FIG. 4 is a diagram illustrating the operation of the circuit of FIG. 3;
FIG. 5 schematically shows an example of an acoustic-optical imaging system according to a first embodiment;
FIG. 6 schematically shows an example of an acoustic-optical imaging system according to a second embodiment;
FIG. 7 schematically shows an example of an acoustic-optical imaging system according to a third embodiment;
FIG. 8 illustrates the operation of an alternative embodiment of an acoustic-optical imaging system; and
FIG. 9 illustrates in further detail the operation of the variant of FIG. 8.
DETAILED DESCRIPTION OF THE PRESENT EMBODIMENTS
Like features have been designated by like references in the various figures. In particular, the structural and/or functional features that are common among the various embodiments may have the same references and may dispose identical structural, dimensional and material properties.
For the sake of clarity, only the steps and elements that are useful for an understanding of the embodiments described herein have been illustrated and described in detail. In particular, the imaging applications capable of being implemented based on the described sensors have not been detailed, the described embodiments being compatible with usual acoustic-optical imaging applications in the medical field or in other fields.
Unless indicated otherwise, when reference is made to two elements connected together, this signifies a direct connection without any intermediate elements other than conductors, and when reference is made to two elements coupled together, this signifies that these two elements can be connected or they can be coupled via one or more other elements.
In the following description, when reference is made to terms qualifying absolute positions, such as terms “front”, “rear”, “top”, “bottom”, “left”, “right”, etc., or relative positions, such as terms “above”, “under”, “upper”, “lower”, etc., or to terms qualifying directions, such as terms “horizontal”, “vertical”, etc., unless otherwise specified, it is referred to the orientation of the drawings, it being understood that, in practice, the described devices may be oriented differently.
Unless specified otherwise, the expressions “around”, “approximately”, “substantially” and “in the order of” signify within 10%, and preferably within 5%.
FIG. 1 schematically shows an example of an acoustic-optical imaging system.
The system of FIG. 1 comprises a light source 101 (L) capable of generating a coherent light beam of frequency fL, for example, a laser source. The system further comprises a splitter 103, for example, a beam splitter 103 placed with a 45-degree angle with respect to the emission direction of source 101, enabling to divide the beam generated by source 101 into an object beam O and a reference beam R. The system is arranged so that object beam O illuminates an object or a sample 105 to be analyzed, and so that reference beam R does not cross object 105. The reference beam and the object beam reflected or transmitted by object 105 are then projected onto an image sensor 109 comprising an array of pixels (not detailed), to generate an interference pattern in the acquisition plane of the sensor.
In the system of FIG. 1, a portion of the object or sample 105 is excited by a sine acoustic wave of frequency FUS, for example in the range from 1 to 15 MHz, for example generated by an ultrasound transducer 120. As a result, the frequency of the rays of the object beam crossing the excited portion of the sample is shifted to value fL+FUS or fL−FUS, while the frequency of the rays of the object beam crossing the non-excited portion of the sample remains at value fL. Such a frequency shift enables to “mark” the photons of the object beam having crossed the excited portion of the sample, called marking area, with respect to the other photons of the object beam. It is then desired to measure the energy carried by the marked photons, to deduce therefrom information relative to light absorption in the marking area.
FIG. 2 is a diagram illustrating in further detail the operation of the system of FIG. 1.
The diagram of FIG. 2 comprises a curve 201 showing the variation over time t (in abscissas) of the light intensity I (in ordinates) received by a pixel of sensor 109 of the system of FIG. 1, when sensor 109 is illuminated by the combination of the object beam and of the reference beam.
In operation, sensor 109 sees an interference pattern also called speckle pattern, formed of speckles resulting from the constructive or destructive interferences of the diffuse object beam. Each speckle has a relatively constant modulus over the total extent of the speckle, first depending on the photon absorption in the marking area, and a random phase depending on the optical paths followed by light in the medium. The spatial sampling of the speckles is preferably such that each speckle covers at least one pixel. The interference with the reference beam results in the creation of interference fringes. The angle of incidence of the reference beam on the sensor may be controlled to obtain the desired spatial period between fringes. The interference pattern seen by sensor 109 has a beat frequency equal to acoustic modulation frequency FUS. Thus, as shown in FIG. 2, the intensity I received by each pixel varies in sinusoidal fashion around an average value or DC component a0, with a peak amplitude of value a1, at a frequency equal to the acoustic modulation frequency FUS of the system. The light intensities received by different pixels of sensor 109 may have different DC components a0 and/or different amplitudes a1, but all fluctuate in sinusoidal fashion at beat frequency FUS, with specific phase shifts.
Each pixel of sensor 109 comprises a photodiode (not shown) supplying a current representative of the light intensity received by the pixel.
In this example, the measurement of the marked portion of the object beam is performed during an acquisition period Tacq of duration N*TUS, where TUS designates the period of acoustic modulation of the system, equal to 1/FUS, and where N is an integer, greater than or equal to 1, for example, in the range from 100 to 10,000.
Each pixel comprises an integration circuit configured to, at each period of acoustic modulation TUS of acquisition period Tacq, integrate the photocurrent delivered by the photodiode successively in K=4 different capacitors C0, C1, C2, and C3 (not shown). Such a four-phase integration and sampling cycle is successively repeated N times during period Tacq, it being understood that the integration capacitors C0, C1, C2, and C3 of the pixel are not reset between the beginning and the end of period Tacq. The accumulation of the samples over a high number N of periods of the acoustic modulation signal enables to improve the signal-to-noise ratio of the measurement. Number N is preferably selected so that the duration of acquisition phase Tacq remains shorter than a characteristic duration of decorrelation of the medium, for example, in the range from 0.1 to 5 ms, for example, in the order of 1 ms, corresponding to the time beyond which the amplitude and the phase of the measured signal risk varying due to physiological variations of the observed medium (for example, due to a blood platelet movement).
In FIG. 2, for each of the N periods TUS of acquisition phase Tacq, references i0p, i1p, i2p, and i3p, designate the four successive periods of integration of the photocurrent delivered by the pixel photodiode, respectively in the four capacitors C0, C1, C2, and C3 of the pixel, where p is an integer in the range from 1 to N designating the rank of the acoustic modulation period TUS considered in acquisition phase Tacq. In this example, integration periods i0p, i1p, i2p, and i3p all substantially have the same duration, in the order of TUS/4.
At the end of acquisition phase Tacq, each pixel of the sensor delivers four values I0, I1, I2, and I3 respectively representative of the voltages across capacitors C0, C1, C2, and C3 of the pixel.
Based on these four values, the complex field EOM of the portion of the object beam marked by the acoustic modulation may be determined, for each pixel of the sensor, by formula EOM=(I0−I2)+i(I1−I3), i being the imaginary unit. Component I0-I2 corresponds to the real part of field EOM, and component I1-I3 corresponds to the imaginary part of field EOM. The phase POM and the amplitude AOM of the marked object beam received by each pixel of the sensor may be determined by the following formulas:
and
AOM=√{square root over ((I0−I2)2+(I1−I3)2)} [Eq. 2]
Knowing the amplitude AOM of the marked object beam enables, in particular, to obtain information relative to light absorption in the acoustic marking area, for example, to detect the possible presence of an absorbing body, for example, a tumor, in the marking area.
It should be noted that the acquisition technique described in relation with FIGS. 1 and 2 only enables to determine the average absorption within the marking area, but does not provide information relative to the absorption in local regions of the marking area.
FIG. 3 schematically shows another example of an acoustic-optical imaging system, enabling to determine the absorption in local regions of the marking area.
The system of FIG. 3 differs from the system of FIG. 1 mainly in that, in the system of FIG. 3, a sequence Φ of phase jumps with a random or pseudo-random pattern, called modulation sequence, is applied to the acoustic marking wave. As an example, at each new acoustic modulation period TUS, a phase jump of value 0 or π, randomly or pseudo-randomly selected, is applied to the acoustic marking wave. At the level of sensor 109, after the interference of the object beam with the reference beam, the signal is sampled according to the four-phase demodulation method discussed in relation with FIG. 2, adapted to take into account the sequence Φ of phase jumps applied to the acoustic marking wave.
FIG. 4 is a diagram illustrating in further detail the operation of the system of FIG. 3.
The diagram of FIG. 4 comprises a curve 301 representing the variation, over time t, in abscissas, of the contribution I(Ztarget), in ordinates, of a local position Ztarget of the marking area, to the object optical signal received by sensor 109 after the interference with the reference signal. In this example, position Ztarget corresponds to a position in the propagation axis of the acoustic modulation wave.
Intensity I(Ztarget) varies in sinusoidal fashion around an average value or DC component a0, with a peak amplitude of value a1, at a frequency equal to the acoustic modulation frequency FUS of the system. In this example, at each new period TUS of acquisition phase Tacq, intensity I(Ztarget) is submitted to a 0 or π phase jump. The sequence Φd of phase jumps applied to intensity I(Ztarget) corresponds to the sequence (of phase jumps applied to the acoustic marking wave, delayed by a delay Δt=Ztarget/υUS, υUS designating the propagation speed of the acoustic marking wave in the observed medium.
In each pixel of the sensor, the integration circuit is configured to, at each acoustic modulation period TUS of acquisition period Tacq, integrate the photocurrent delivered by the photodiode successively in the four capacitors C0, C1, C2, and C3 (not shown), taking into account the delayed phase jump pattern Φd=Φ(t−Δt), called modulation sequence, to determine the order in which capacitors C0, C1, C2, and C3 are connected to the photodiode. More particularly, in this example, at each phase jump of value π in sequence Φd, the order of integration of the photocurrent in capacitors C0 and C2 on the one hand and C1 and C3 on the other hand is inverted.
At the end of acquisition phase Tacq, each pixel of the sensor delivers four values I0, I1, I2, and I3 respectively representative of the voltages across capacitors C0, C1, C2, and C3 of the pixel. Values I0, I1, I2, and I3 are mainly representative of the contribution to the object optical signal received by the pixel, of the position of the marking area for which phase jump pattern Φd is correlated to the acoustic modulation signal, that is, position Ztarget=υUS*Δt.
Based on these four values, the complex field EOM of the portion of the object beam marked by the acoustic modulation at position Ztarget of the marking area may be determined by formula EOM=(I0−I2)+j(I1−I3). Component I0-I2 corresponds to the real part of field EOM, and component I1-I3 corresponds to the imaginary part of field EOM. The phase POM and the amplitude AOM of the marked object beam received from position Ztarget may be determined by the same formulas 1 and 2 as in the example of FIGS. 1 and 2.
The measurement may be repeated by modifying the delay Δt between the phase jump sequence Φ applied to the acoustic modulation signal and the sequence Φd applied to the sensor during acquisition phase Tacq, to measure the field EOM originating from another target position Ztarget in the marking area. More particularly, calling L the desired axial resolution, that is, the desired number of measurement points having different positions Ztarget in the acoustic marking target (that is, along the main propagation axis of the acoustic marking beam), the measurement should be repeated L times.
FIG. 5 schematically shows an example of an acoustic-optical imaging system according to a first embodiment.
The system of FIG. 5 comprises the same elements as the system of FIG. 3, arranged substantially in the same way, and differs from the system of FIG. 3 mainly by its operating mode.
In the embodiment of FIG. 5, it is provided to reconstruct an absorption profile of the marking column, of axial resolution L, based on a series of M measurements, L and M being integers greater than or equal to 2. During each measurement, a first random or pseudo-random phase jump sequence Φ, called modulation sequence, is applied to the acoustic marking signal, identically or similarly to what has been described in the example of FIGS. 3 and 4. During acquisition phase Tacq, each pixel of sensor 109 applies to the measured signal a second random or pseudo-random phase jump sequence Ω, called demodulation sequence. As an example, at each new acoustic modulation period TUS of acquisition phase Tacq, a phase jump of value 0 or π is applied to the measured signal. The N phase jump values applied during acquisition phase Tacq define sequence Ω. In each pixel of sensor, the integration circuit is configured to, at each acoustic modulation period TUS of acquisition phase Tacq, integrate the photocurrent delivered by the photodiode successively in the four capacitors C0, C1, C2, and C3, taking into account sequence Ω to determine the order in which capacitors C0, C1, C2, and C3 are connected to the photodiode. More particularly, in this example, at each phase jump of value π in sequence Ω, the order of integration of the photocurrent in capacitors C0 and C2 on the one hand and C1 and C3 on the other hand is inverted.
Conversely to the example of FIGS. 3 and 4 where the modulation sequence Φ of the phase of the acoustic marking beam and the demodulation sequence Φd controlling the sampling phases applied by sensor 109 during acquisition phase Tacq have a correlation peak defined for a specific position in the marking column, in the example of FIG. 5, the modulation and demodulation sequences Φ and Ω are non-correlated. Sequences Φ and Ω for example correspond to two independent random or pseudo-random drawings of binary sequences, for example, random drawings according to an equiprobable Bernoulli distribution, where a state of the binary sequence corresponds to a modulation period TUS. Two new random or pseudo-random drawings independent from the previous drawing, that is, non-correlated with the previous drawing, may be applied to each of the M measurements of the phase of acquisition of the absorption profile of the sample (that is, 2*M distinct drawings which are not correlated together for the acquisition of an absorption profile of L values). Each measurement then results from a variable contribution of the different points of the marking column of the sample.
The inventors have shown that an acquisition model may be constructed, enabling to model the measurements as a projection of the desired absorption profile through a transition matrix.
The light signal s(t) seen at the level of a pixel of sensor 109, resulting from the interference between the object beam (signal Es(t)) and the reference beam (signal Er(t)) is provided by the following equation:
s(t)=|Er(t)+Es(t)|2 [Eq. 3]
with:
Er(t)=AReiωLt [Eq. 4]
and:
Es(t)=ADeiwLt+∫zaM(z)ei(ωLt+ωUS(t−z/υUS)+πΦ(t−z/υUS+δ))dz+∫zaM(z)ei(ωLt−ωUS(t−z/υUS)−πΦ(t−z/υUS+δ))dz [Eq. 5]
where AR, AD, and aM are the complex amplitudes of the optical fields respectively of the reference arm, unmarked photons of the object arm, and marked photons of the object arm, t is a time variable, z designates the position along the acoustic marking column of sample (aM(z) corresponds to the absorption of the sample at position z of the marking column), νUS designates the propagation speed of the acoustic wave in the sample, ωL and ωUS respectively designate the pulse of the light wave emitted by source 101 and the pulse of the acoustic marking wave, Φ designates the binary sequence of modulation of the phase of the acoustic marking wave, and δ is a term reflecting the fact that the phase of the modulating signal may be non-zero at the origin.
Signal s(t) may then be expressed as follows:
and then:
s(t)=AD+AR+2∫zaM(z)cos(ωUS(t−z/υUS)+πΦ(t−z/υUS+δ))dz|2 [Eq. 7]
For each of the M measurements of the phase of acquisition of the absorption profile, a term I=I0−I2 and a term Q=I1−I3 is measured at the sensor level, by the above-discussed four-phase demodulation technique, with:
where Ω designates the binary demodulation sequence applied by sensor 109.
Term Q can be expressed as follows:
where Re and Im respectively designate the real part and the imaginary part of a complex number.
The terms which do not comprise the component AR of the reference arm may be neglected, the energy provided by this arm being much greater than the energy provided by the marked and unmarked portions of the object arm.
Equation 11 can then be simplified as follows:
Similarly, term I may be expressed as follows:
Terms Q and I may be re-expressed as follows:
Calling Qj, respectively Ij, the term Q, respectively I, measured by the pixel at each of the M measurements of the phase of acquisition of the absorption profile, and assigning an index j to all the terms varying from one measurement to the other, j being an integer in the range from 1 to M, the equations may be rewritten, for each measurement, as follows:
Qj=∫z Re(aM(z))Cj(z)dz−∫z Im(aM(z))Dj(z)dz [Eq. 19]
and:
Ij=∫z Re(aM(z))Ej(z)dz−∫z Im(aM(z))Fj(z)dz [Eq. 20]
with, knowing that Im(G(t,z))=0:
All the obtained measurements may be written in the form of a vector Y of M imaginary numbers Yj, with Yj=Ij+iQj, and:
where Δz represents a fraction of the wavelength λUS=TUS*υUS of the acoustic pulse signal ωUS, defining the desired axial measurement resolution (L*Δz corresponding to the thickness of the analyzed medium, that is, the length of the marking column having its absorption profile desired to be acquired). The absorption of the medium in a section of length Δz of the marking column is considered as constant along the section length Δz.
In this example, to construct the transition matrix of the system, terms I and Q are rewritten by considering, for each position z along the marking column, a predetermined reference contribution aMref(z) to the marked portion of the object beam received by the sensor, enabling for example to model the acoustic marking efficiency along the column (for example to take into account a possible lack of uniformity of the acoustic power density along the column). Term aMref is a predetermined vector of L values, defined according to the properties of the imaging system, and particularly to the properties of the transducer used to generate the acoustic marking wave. On can then write, for each of the M measurements:
and then:
with.
One then rewrites:
Having identified the relation between the measurements and the absorption profile according to equation 35, the following matrix relation can be established:
Y=RE Re(aM)+Im Im(aM) [Eq. 36]
where ARe and AIm are two matrices each having M rows and L columns, and aM is a vector of L rows and one column having its modulus corresponding to the desired absorption profile.
Term ARe j,k designates the value of the row of rank j and of the column of rank k in matrix ARe, and term AIm j,k designates the value of the row of rank j and of the column of rank k in matrix AIm.
The values ARe j,k and AIm j,k of matrices ARe and AIm can be expressed as follows:
Re j,k=4 Re(AR)Re(aM Ref(k·Δz))∫(k−1)k·Δz(S13j(z)+iS24j(z)dz [Eq. 37]
and:
=4 Im(AR)Im(aM Ref(k·Δz))∫(k−1)·Δzk·Δz(S13j(z)+iS24j(z))dz [Eq. 38]
Matrices ARe j,k and AIm j,k are the transitions matrices of the system. Considering M measurements (j varies from 1 to M), the matrices are each formed of M rows and L columns, L corresponding to the number of points of the absorption profile. Conversely to the solution discussed in relation with FIGS. 3 and 4, in the embodiment of FIG. 4, the number M of performed measurements may be smaller than the number L of points of the absorption profile which is desired to be reconstructed.
One thus as a model such that:
Y=Fct(aM)=ε [Eq. 39]
where Y is a vector of M imaginary numbers Yj respectively corresponding to the M measurements performed by the sensor, Fct is a function of absorption profile aM, and □ is an error term which will be desired to be minimized.
An estimate of absorption profile aM is thus desired, which is noted as follows:
[Eq. 40]
For this purpose, it is desired to solve the following error minimization problem:
This problem having an infinite number of solutions, a regularization term enabling to solve it is added.
Known transition matrices ARe j,k and AIm j,k, the reconstruction of the absorption profile defined by vector aM may be performed by minimizing the following cost function:
where ψ is a function defining a reconstruction constraint on the modulus of vector aM, corresponding to supposed properties of the signal to be reconstructed. Coefficient λ enables to weight the contribution of the reconstruction constraint in the global cost function. In this example, the constraint is defined from standard 1 of function ψ, multiplied by the estimate of absorption profile aM. As an example, function ψ is a filter bank formed from a differential operator along the propagation axis of the acoustic, duplicated at different scales for each of positions z. More generally, any other method of synthesis of the reconstruction constraint, enabling to take into account supposed properties of the desired signal, may be used.
Based on this equation, a vector aM having its module corresponding to the desired absorption profile is reconstructed.
The resolution of the minimization problem defined by equation 42 is for example implemented by an electronic processing circuit (not shown in the drawing), for example comprising a microprocessor. The processing circuit may be integrated to sensor 109, or external to sensor 109.
It should be noted that in the embodiment of FIG. 5, transition matrix ARe j,k can be expressed as follows:
The combination of Π and Γ of F provides a Rademacher law, similar to a Bernoulli law, having its sum tending, according to the Moivre Laplace theorem, towards a random Gaussian distribution variable. The coefficients of transition matrix ARe j,k are asymptotically related to drawings from a Gaussian distribution, allowing a reconstruction of compressive acquisition type.
A similar line of reasoning applies concerning transition matrix AIm j,k.
Although an example of operation based on a demodulation with K=4 phases of the modulated light signal has been described hereabove, the embodiment of FIG. 5 is not limited to this specific case. As a variation, the demodulation may be performed on any other number K of phases greater than or equal to 2. It will be within the abilities of those skilled in the art to accordingly adapt the pixel circuits, particularly to acquire a number of samples different from four at each modulation period TUS of acquisition phase Tacq. The formulas of reconstruction of the real and imaginary parts of the complex field of the object beam may be adapted accordingly.
Further, although, in the example described hereabove in relation with FIG. 5, a phase of the acoustic marking wave which is zero at the origin has been considered, the described embodiments are not limited to this specific case. As a variation, the above-described calculations may be resumed with the addition of a non-zero phase φ such that:
Es(t)=ADeiωLt+∫zaM(z)e−iωUSz/υUSei((ωL+ωUS)t+Φ(t−z/υUS+δφ)dz+∫zaM(z)eiωUSz/υUSei((ωL−ωUS)t−φ(t−z/υUS++δ)+φ)dz [Eq. 46]
FIG. 6 schematically shows an example of an acoustic-optical imaging system according to a second embodiment.
The system of FIG. 6 comprises, as in the previous examples, a light source 101 (L) capable of generating a coherent light beam of frequency fL, and a splitter 103 enabling to divide the beam generated by source 101 into an object beam O and a reference beam R. In the system of FIG. 6, at the output of separator 103, object beam O is modulated by a first acoustic-optical modulator MAO1 shifting its frequency by a value fMAO1, and reference beam R is modulated by a second acoustic-optical modulator MAO2 shifting its frequency by a value fMAO2 different from value fMAO1. The system is arranged so that the object beam, of frequency fL+fMAO1, illuminates the object or sample 105 to be analyzed, and that the reference beam, of frequency fL+fMAO2, does not cross sample 105.
As in the previous examples, the reference beam and the object beam reflected or transmitted by object 105 are then projected onto an image sensor 609, to generate an interference pattern in the acquisition plane of the sensor.
Further, as in the previous examples, a portion of sample 105 is excited by a focused ultrasound wave of frequency FUS. As a result, the frequency of the rays of the object beam running through the excited portion of the sample is shifted to value fL+fMAO1+FUS or fL+fMAO1−FUS, while the frequency of the rays of the object beam running through the non-excited portions of the sample remains at value fL+fMAO1. As previously, it is desired to measure the energy carried by the photons marked by the acoustic wave of frequency FUS, to deduce therefrom information relative to light absorption in the marking area.
In operation, sensor 609 sees an interference pattern having a beat frequency fB equal to the absolute value of difference fMAO1+FUS−fMAO2.
Conversely to the previously-described sensor 109, comprising an electronic demodulation device, for example in the form of an integration circuit with K switched capacitors per pixel, synchronously rated at a frequency equal to K times the frequency FUS of the acoustic marking wave, in the embodiment of FIG. 6, sensor 609 is a standard sensor, for example comprising an integration circuit with a single capacitor per pixel. In this example, the frequency FC of image acquisition by sensor 609 is equal to K times the beat frequency fB of the system. The assembly with two acoustic-optical modulators MAO1 and MAO2 of FIG. 6 enables to use a sensor 609 having a relatively low acquisition frequency FC, for example, in the range from 20 Hz to 20 kHz, which corresponds to the typical acquisition frequencies of usual image sensors.
Usual acoustic-optical modulators may typically shift the frequency of a light beam by a value in the order from a few MHz to a few tens of MHz, for example, by a value in the range from 5 to 100 MHz. The marking ultrasound wave applied to sample 105 causes a frequency shift of the light crossing the excited portion of the object by a value for example in the range from 1 to 15 MHz. Modulators MAO1 and MAO2 are selected to obtain a beat frequency fB compatible with the use of a low-frequency sensor, for example, such that frequency FC=K*fB is in the range from 20 Hz to 20 kHz.
As a numerical illustration, and without this being limiting, in the system of FIG. 6, sensor 609 may have an acquisition frequency FC of 8 kHz, and the frequency shifts fMAO1, fMAO2 and FUS introduced by modulators MAO1 and MAO2 and by the ultrasound wave of excitation of sample 105 may respectively be 77.001 MHz, 79.999 MHz, and 3 MHz, to obtain a beat frequency fB=fMAO1+FUS−fMAO2=2 kHz=FC/4 (considering case K=4).
As in the embodiment of FIG. 5, it is provided to reconstruct an absorption profile of the marking column, having an axial resolution L, based on a series of M measurements, L and M being integers greater than or equal to 2.
During each measurement, a first random or pseudo-random sequence Φ of phase jumps is applied to the acoustic marking signal, identically or similarly to what has been described in the example of FIG. 5.
During the acquisition phase, having a duration Tacq=TB=1/fB, K images are successively acquired by means of sensor 609. For each of the M measurements, a set of K values corresponding to the four phases of beat period TB is thus obtained for each pixel. For simplification, case K=4 will be considered hereafter and references I0, I1, I2, and I3 will be used to designate the four samples acquired by a pixel during a same acquisition phase Tacq (corresponding to the values of the pixel in the 4 images successively acquired during phase Tacq).
As in the embodiment of FIG. 5, it is provided, for each measurement, to apply to the measured signal a demodulation sequence Ω, in the form of a random or pseudo-random phase jump sequence which is not correlated to sequence Φ.
The embodiment of FIG. 6 differs from the embodiment of FIG. 5 in that, in the embodiment of FIG. 6, instead of being applied in the electronic domain by a demodulation device integrated to the sensor, demodulation sequence Ω is applied in the optical domain, via the acoustic-optical modulator MAO2 placed on reference arm R. In other words, in this example, demodulation sequence Ω is a random or pseudo-random sequence directly applied to reference beam R, before the interference of beam R with object beam O.
The signal s(t) measured at the level of a pixel of sensor 609 at the end of the interference between the object arm and the reference arm is given by the following equation:
s(t)=|Er(t)=Es(t)|2 [Eq. 47]
with:
Er(t)=ARei(ωRt=n(t−δ)) [Eq. 48]
and:
Es(t)=ADeiωOt+∫zaM(z)ei(ωOt+ωUS(t−z/υUS)+Φ((t−z/υUS))dz+∫zaM(z)ei(ωOt−ωUS(t−z/υUS)−Φ((t−z/υUS))dz [Eq. 49]
that is:
Es(t)=ADeiωOt+∫zaM+(z)ei((ωO+ωUS)t+Φ((t−z/υUS))dz+∫zaM−(z)ei((ωO−ωUS)t−Φ((t−z/υUS))dz [Eq. 50]
where:
aM+(z)=aM(z)e−iωUSz/υUS [Eq. 51]
aM−(z)=aM(z)eiωUSz/υUS [Eq. 52]
and where:
ωR=2π(fL+fMAO2) [Eq. 53]
ωO=2π(fL+fMAO1) [Eq. 54]
ωR−ωO=ωUS−ωCAM [Eq. 55]
(ωCAM=2πFCAM [Eq. 56]
where FCAM designates the acquisition frequency of the system, such that FCAM=1/Tacq=FC/4.
Equation 50 can then be rewritten as follows:
Es(t)=ADeiwOt+∫zaM+(Z)ei((ωR+ωCAM)t+Φ(t−Z/υUS))dz+∫zaM−(z)ei((ωR−2ωUS+ωCAM)t−Φ(t−Z/υUS))dz [Eq. 57]
whereby:
At the level of sensor 609, K=4 images are successively acquired with an acquisition duration TCAM/4 per image (with TCAM=1/FCAM). Paths I and Q can be expressed as follows:
I=∫0TCAM/4s(t)dt−∫TCAM/23TCAM/4s(t)dt [Eq. 60]
Q=∫TCAM/4TCAM/2s(t)dt−∫3TCAM/4TCAMs(t)dt [Eq. 61]
Functions F and G are defined, such that:
F(z,t)=ei(ωCAMt−ωUSz/υUS+Φ(t−z/υUS)−Ω(t−δ)) [Eq. 62]
G(z,t)=ei((ωCAM−2ωUS)t+ωUSz/υUS−Φ((t−z/υUS)−Ω(t−δ)) [Eq. 63]
One then has:
As in the embodiment of FIG. 5, the terms which do not depend on AR are neglected, the energy carried by the reference arm widely dominating over the other terms. Further, the terms depending on AD are here neglected. Due to their high-frequency component independent from ωCAM, these terms will indeed be averaged over the integration time of each image, in the order of TCAM/4, and the residual value will be taken to zero by the four-phase demodulation (difference I0−I2 and I1−I3). One then obtains:
whereby:
Terms I and Q are then rewritten in the form described in relation with FIG. 5:
For each measurement of rank j, these equations can be rewritten as follows:
Qj=∫z Re(aM(z))Cj(z)dz−∫z Im(aM(z))Dj(z)dz [Eq. 72]
Ij=Re(aM(z))Ej(z)dz−∫z Im(aM(z))Fj(z)dz [Eq. 73]
with:
and:
Fj(t,z)=ei(ωCAMt−ωUSz/υUS+Φj(t−z/υUS)−Ωj(t−δ)) [Eq. 78]
Gj(t,z)=ei((ωCAM−2ωUS)t+ωUSz/υUS−Φj(t−z/υUS)−Ωj(t−δ)) [Eq. 79]
Each of the M obtained measurements can be rewritten as Yj=Ij+iQj, with:
This provides the form of I and Q defined in the example of FIG. 5 (equations 25 and 26). The same method of construction of transition matrices as in the previous example can thus be used (equations 27 to 38), by replacing term G(t,z) with F(z,t)+G(z,t), followed by the same method of reconstruction of absorption profile aM (equation 42).
The system described in relation with FIG. 6 enables, for each of the M measurements of the phase of acquisition of the absorption profile, to perform the four-phase demodulation by means of 4 images acquired via a standard camera (with an acquisition period of TCAM/4 for each image). Such a technique further enables to use high acoustic modulation frequencies FUS.
FIG. 7 schematically shows an example of an acoustic-optical imaging system according to a third embodiment.
In this embodiment, it is provided to extend the method described in relation with FIG. 6 to applications of depth measurement in a scene.
In the system of FIG. 7, sample 105 is replaced with a reflective object 705, corresponding to a scene where depth information is desired to be measured. In this example, no local marking of a portion of the object beam is provided. In other words, no ultrasound transducer arranged to apply a local mechanical wave to a portion of the scene is provided. In FIG. 7, for simplification, the means for splitting the beam emitted by the source into an object beam and a reference beam have not been shown.
Modulation sequence Φ is applied to the object beam upstream of scene 705, via acoustic-optical modulator MAO1. Demodulation sequence Ω is applied to the reference beam via acoustic-optical modulator MAO2.
Modulators MAO1 and MAO2 are selected to obtain a beat frequency fB (here equal to the absolute value of difference fMAO1−fMAO2) compatible with the use of a low-frequency sensor, for example, such that frequency FC=K*fB is in the range from 20 Hz to 20 kHz.
As a numerical illustration, and without this being limiting, in the system of FIG. 7, sensor 609 may have an acquisition frequency FC of 8 kHz, and the frequency shifts fMAO1 and fMAO2 introduced by modulators MAO1 and MAO2 may respectively be 80.001 MHz, 79.999 MHz, to obtain a beat frequency fB=fMAO1−fMAO2=2 kHz=FC/4 (considering case K=4).
In the embodiment of FIG. 7, the modulus of the measured vector aM corresponds to the absorption profile of the source beam along the optical path between the output of modulator MAO1 and sensor 609, mainly corresponding to the absorption of the reflective elements of scene 705 along the optical propagation path (of ballistic nature in this case).
The signal s(t) measured at the level of a pixel of sensor 609 at the end of the interference between the object arm and the reference arm is given by the following equation:
S(t)=|Er(t)+Es(t)|2 [Eq. 82]
with:
Er(t)=ARei(ωRt+Ω(t−δ)) [Eq. 83]
and:
Es(t)=∫zaM(z)ei(ωO(t−z/υL)+Φ((t−z/υL))dz [Eq. 84]
υL being the propagation speed of light in the medium, and:
ωR=2π(fL+fMAO2) [Eq. 85]
ωO=2π(fL+fMAO1) [Eq. 86]
ωR−ωO=ωCAM [Eq. 87]
ωCAM=2λFCAM [Eq. 88]
One then has:
s(t)=(|ei(ωRt+Ω(t−δ))|·|AR+∫zaMei(ωO(t−z/υL)+Φ(t−Z/υL)−ωRt−Ω(t−δ))dz|)2 [Eq. 89]
S(t)=|AR+∫zaMei(ωCAMt−ωOz/υL+Φ(t−z/υL)−Ω(t−δ))dZ|2 [Eq. 90]
At the level of sensor 609, K=4 images are successively acquired with an acquisition duration TCAM/4 per image. Paths I and Q may as previously be expressed as follows:
I=∫0TCAM/4s(t)dt−∫TCAM/23TCAM/4s(t)dt [Eq. 91]
Q=∫TCAM/4TCAM/2s(t)dt−∫3TCAM/4TCAMs(t)dt [Eq. 92]
A function F is defined, such that:
F(z,t)=ei(ωCAMt−ωOz/υL+Φ(t−z/υL)−Ω(t−δ)) [Eq. 93]
One then has:
whereby:
For each measurement of rank j, these equations may be rewritten as follows:
Qj=∫z Re(aM(z))Cj(z)dz−∫z Im(aM(z))Dj(z)dz [Eq. 100]
Ij=∫z Re(aM(z))Ej(z)dz−∫z Im(aM(z))Fj(z)dz [Eq. 101]
with:
Cj(z)=∫TCAM/4TCAM/2(2 Re(AR)Re(Fj(z,t))+2 Im(AR)Im(Fj(z,t)))dt−∫3TCAM/4TCAM(2 Re(AR)Re(Fj(z,t))+2 Im(AR)Im(Fj(z,t)))dt [Eq. 102]
Dj(z)=∫TCAM/4TCAM/2(2 Re(AR)Im(Fj(z,t))−2 Im(AR)Re(Fj(z,t)))dt−∫3TCAM/4TCAM(2 Re(AR)Im(Fj(z,t))−2 Im(AR)Re(Fj(z,t)))dt [Eq. 103]
Ej(z)=∫0TCAM/4(2 Re(AR)Re(Fj(z,t))+2 Im(AR)Im(Fj(z,t)))dt−∫TCAM/23TCAM/4(2 Re(AR)Re(Fj(z,t))+2 Im(AR)Im(Fj(z,t)))dt [Eq. 104]
Fj(z)=∫0TCAM/4(2 Re(AR)Im(Fj(z,t))−2 Im(AR)Re(Fj(z,t)))dt−∫TCAM/23TCAM/4(2 Re(AR)Im(Fj(z,t))−2 Im(AR)Re(Fj(z,t)))dt [Eq. 105]
Fj(z,t)=ei(ωCAMt−ωOz/υL+Φj(t−Z/υL)−Ωj(t−δ)) [Eq. 106]
One can also rewrite:
Cj(z)=2 Re(AR)(∫TCAM/4TCAM/2 Re(Fj(z,t))dt−∫3TCAM/4TCAM Re(Fj(z,t))dt)+2 Im(AR)(∫TCAM/4TCAM/2 Im(Fj(z,t))dt−∫3TCAM/4TCAM Im(Fj(z,t))dt [Eq. 107]
Dj(z)=2 Re(AR)(∫TCAM/4TCAM/2 Im(Fj(z,t))dt−∫3TCAM/4TCAM Im(Fj(z,t))dt)−2 Im(AR)(∫TCAM/4TCAM/2 Re(Fj(z,t))dt−∫3TCAM/4TCAM Re(Fj(z,t))dt [Eq. 108]
Ej(z)=2 Re(AR)(∫0TCAM/4 Re(Fj(z,t))dt−∫TCAM/23TCAM/4 Re(Fj(z,t))dt)+2 Im(AR)(∫0TCAM/4 Im(Fj(z,t))dt−∫TCAM/23TCAM/4 Im(Fj(z,t))dt [Eq. 109]
Fj(z)=2 Re(AR)(∫0TCAM/4 Im(Fj(z,t))dt−∫TCAM/23TCAM/4 Im(Fj(z,t))dt)−2 Im(AR)(∫0TCAM/4 Re(Fj(z,t))dt−∫TCAM/23TCAM/4 Re(Fj(z,t))dt [Eq. 110]
The following notations are used:
H13 Re j(z)=∫0TCAM/4 Re(Fj(z,t))dt−∫TCAM/23TCAM/4 Re(Fj(z,t))dt [Eq. 111]
H13 Im j(z)=∫0TCAM/4 Im(Fj(z,t))dt−∫TCAM/23TCAM/4 Im(Fj(z,t))dt [Eq. 112]
H24 Re j(z)=∫TCAM/4TCAM/2 Re(Fj(z,t))dt−∫3TCAM/4TCAM Re(Fj(z,t))dt [Eq. 113]
H24 Im j(z)=∫TCAM/4TCAM/2 Im(Fj(z,t))dt−∫3TCAM/4TCAM Im(Fj(z,t))dt [Eq. 114]
One can then rewrite:
Cj(z)=2 Re(AR)H24 Re j(z)+2 Im(AR)H24 Im j(z) [Eq. 115]
Dj(z)=2 Re(AR)H24 Im j(z)−2 Im(AR)H24 Re j(z) [Eq. 116]
Ej(z)=2 Re(AR)H13 Re j(z)+2 Im(AR)H13 Im j(z) [Eq. 117]
Fj(z)=2 Re(AR)H13 Im j(z)−2 Im(AR)H13 Re j(z) [Eq. 118]
Each of the M obtained measurements may, as previously, be rewritten in form Yj=Ij+iQj, with:
In this example, pitch Δz will preferably be selected to be much smaller than wavelength λL with λL (3*10{circumflex over ( )}8)*2π/ω0.
Considering a reference absorption profile aMref along the light propagation path, one has:
that is:
Having identified the relation between the measurements and the absorption profile according to equal to equation 128, the following matrix relation can be established:
Yj=Re j,k Re(aM)−Im j,k Im(aM) [Eq. 129]
where ARe and AIm are two matrices each having M rows and L columns, and aM is a vector of L rows and one column.
With:
Re j,k=2(Re(AR)∫(k−1)·Δzk·Δz(H13 Re j(z)+iH24 Re j(z))dz+Im(AR)∫(k−1)·Δzk·Δz(H13 Im j(z)+iH24 Im j(z))dz) [Eq. 130]
Im j,k=2(Re(AR)∫(k−1)·Δzk·Δz(H13 Im j(z)+iH24 Im j(z))dz−Im(AR)∫(k−1)·Δzk·Δz(H13 Re j(z)+iH24 Re j(z))dz) [Eq. 131]
The same method of reconstruction of absorption profile aM as in the previous examples (equation 42) can then be used.
It should be noted that the embodiment of FIG. 5 may also be adapted for applications of depth measurement in a scene. In this case, the assembly will be similar to that of FIG. 7, but with a single acoustic modulator MAO1, placed on the object arm, enabling to apply modulation sequence Φ, and by replacing the low-frequency sensor 609 of FIG. 6 with a synchronous sensor 109 of the type described in relation with FIG. 5, enabling to apply demodulation sequence Q in the electronic field.
FIG. 8 illustrates the operation of an alternative embodiment of the acoustic-optical imaging system of FIG. 5.
In the example of FIG. 5, a system of acquisition of a one-dimension absorption profile aM has been provided. For this purpose, a single ultrasound transducer (or a single group of simultaneously controlled ultrasound transducers) 120 (FIG. 5) applies a marking sine wave ballistically propagating in a local area of the sample, for example, a column-shaped area. The propagation direction of the marking acoustic wave corresponds to the direction of reconstruction of absorption profile aM (direction z in the above-detailed formulas).
In the variant of FIG. 8, it is provided to extend the measurement principle detailed in relation with FIG. 5 to the reconstruction of a two-dimensional absorption profile, x and z, for example, orthogonal.
For this purpose, a reference acoustic excitation wave exhibiting phase jumps defined according to a random or pseudo-random pattern Φ is used as previously. This wave is emitted not longer by means of a single transducer according to a single trajectory, but by means of a linear array 820 of W elementary transducers 820q spatially distributed along axis x, W being an integer greater than or equal to 2 and q being an integer in the range from 1 to W. In this example, the W transducers 820q are aligned along axis x, and each transducer 820q is arranged to emit along a main propagation direction parallel to axis z.
In the variant of FIG. 8, the acoustic excitation wave is emitted in fractions by the different transducers in random or pseudo-random fashion, according to a function γ. Each wave fraction emitted by a transducer preferably corresponds to a single period TUS or to an integral number of periods TUS of the acoustic excitation wave. A same fraction of the acoustic excitation wave may be simultaneously emitted by a plurality of transducers 820q (that is, simultaneously from a plurality of positions x).
The measurement principle can then be described by means of the following equations, by using the previously-used formalism. The pitch between transducers is designated hereafter by Δx and γx designates a random or pseudo-random binary sequence enabling to model the activation of the transducer located at position x over time. Term ax(z) is a column vector of L values representing the absorption profile along z at position x.
The wave of the reference beam and the wave of the object beam can be written as:
The interference sj(t) measured at the level of the sensor corresponding to a measurement of rank j can then be written as:
Four phases are measured by demodulation at the sensor level:
whereby, for example, for path Q:
The terms which do not comprise the component AR of the reference arm are neglected and the following term is defined:
G(t,z)=cos(ωUS(t−z/υUS)+πΦj(t−z/υUS)) [Eq. 142]
Equation 141 is then simplified as follows:
One then has:
Similarly, it can be shown that:
For each of the M measurements of the phase of acquisition of absorption profile aM, the following can be written:
with:
Each measurement can be rewritten in form Yj=Ij+iQj, with:
Term Δz here represents a fraction of the wavelength λUS=TUS*υUS of the ultrasound excitation signal of pulse ωUS, selected to be able to consider the absorption as being constant over an interval Δz. According to the notation used, the considered medium has a thickness L*Δz.
To construct the transition matrix of the system, terms I and Q are rewritten as follows, considering a contribution of reference arefx(z) of each position x of the acoustic marking area, for example, enabling to model the acoustic marking efficiency at each position of the marking area:
whereby:
with:
that is:
Having identified the relation between the measurements and the absorption profile according to equal to equation 161, the following matrix relation can be established:
Yj=Re Re(axz)+Im Im(axz) [Eq. 162]
where axz is a vector of L*W rows and one column corresponding to the concatenation of the W absorption profiles of L points each, respectively corresponding to the W measurement positions on axis x, such that:
axz=(aΔx(Δz), . . . ,aΔx(LΔz),a2Δx(Δz), . . . ,a2Δx(LΔz), . . . ,aWΔx(Δz), . . . ,aWΔx(LΔz))T [Eq. 163]
and where ARe and AIm are two matrices of M rows and L*W columns, each, such that:
Re=(Re j,1,1, . . . ,Re j,1,L,Re j,2,1, . . . ,Re j,2,L,Re j,W,1, . . . ,Re j,W,L) [Eq. 164]
Im=(Im j,1,1, . . . ,Im j,1,L,Im j,2,1, . . . ,Im j,2,L,Im j,W,1, . . . ,Im j,W,L [Eq. 165]
Re j,Δ,k=4 Re(AR)Re(aRef qΔx(kΔz))∫(k−1)ΔzkΔz(s13jq(z)+is24jq(z))dz [Eq. 166]
Re j,Δ,k=4 Im(AR)Im(aRef qΔx(kΔz))∫(k−1)ΔzkΔz(s13jq(z)+is24jq(z))dz [Eq. 167]
FIG. 9 schematically shows the matrix representation of the operation of reconstruction of two-dimensional absorption profile axz.
The cost function associated with the reconstruction of vector axz may have the same form as in the example of FIG. 5 (equation 42), with a fidelity term of same nature and a regularization term enabling to constrain the 2D reconstruction of the absorption of the medium.
Such a two-dimensional measurement principle has been disclosed for a synchronous detection mode with a demodulation in the electronic domain. It may however be easily adapted to a low-frequency acquisition mode with a demodulation in the optical domain, as described in relation with FIG. 6.
Further, this principle may be extended to three-dimensional measurements by adding an axis y, for example, orthogonal to axes x and z, and by using a two-dimensional array of ultrasound transducers and an activation function γ depending on x and y.
To increase the lateral resolution (in x and/or in y) of the measurement, an array, linear or not, of transducers may further be used and a simultaneous activation of overlapping sliding groups of a plurality of neighboring transducers may be provided.
It may further be provided to use different modulation functions Φ according to the activated transducers rather than a modulation function common to all transducers.
More generally, sub-groups of transducers according to the desired acoustic excitation form may be sequentially activated (according to variable directions, for example, in a given plane or volume), to increase the lateral resolution.
An acquisition model enabling to reconstruct the amplitude of the absorption profile aM of the object beam has been provided hereabove. This vector, which is a function of position z (aM(z)), has now been considered as having a constant phase according to the index and the measurement position (that is, as if a single pixel measuring a static medium was considered).
Now noting:
aM(z)=ρ(z)eiθ*(z) [Eq. 168]
it is desired to reconstruct the absorption profile ρ which is a vector formed of real positive values in the range from 0 to 1.
Now considering an array of nv by nh pixels sampling speckle grains (that is, speckles of the interference pattern), nv*nh different variants of coefficient Yj are simultaneously measured. Assuming that the signal received by each pixel has its own random phase θ uniformly distributed between 0 and 2π, this provides a profile aM depending on index p of the pixel, that is:
aM(p,z)=ρ(z)eiθ{circumflex over ( )}(p,z) [Eq. 169]
The measurement of coefficient Yj can thus be considered as a set of nv*nh samples modeled by the scalar product of the transition matrices previously-defined with profile aM such that:
Yj[p]=(Re)j,Re(aM[p])+(Im)j,Im(aM[p]) [Eq. 170]
Yj[p]=(Re)j,ρ⊙ cos(θ[p])+(Im)j,ρ⊙ sin(θ[p]) [Eq. 171]
where symbol [Eq. 172]⊙ designates the point-to-point multiplication.
One thus has:
Yj[p]=(Re)j⊙ cos(θ[p]),ρ+(Im)j⊙ sin(θ[p]),ρ [Eq. 173]
Yj[p]=(Re)j⊙ cos(θ[p])+(Im)j⊙ sin(θ[p]),ρ [Eq. 174]
Yj[p]=Re((Re)j⊙ cos(θ[p])+(Im)j⊙ sin(θ[p])),ρ+iIm((Re)j⊙ cos(θ[p])+(Im)j⊙ sin(θ[p])),ρ [Eq. 175]
Yj[p]=βRe[p],ρ+iβIm[p],ρ [Eq. 176]
βRe j[p]=Re((Re)j)⊙ cos(θ[p])+Re((Im)j)⊙ sin(θ[p]) [Eq. 177]
βIm j[p]=Im((Re)j)⊙ cos(θ[p])+Im((Im)j)⊙ sin(θ[p]) [Eq. 178]
Considering that phase θ follows a uniform law between 0 and 2π, the variance according to the pixel axis, noted Varp, can be expressed as follows:
The following notations are then used:
A measurement model specifying that the variance of the signals measured by the p pixels for each acquisition j is a function of the previously-identified transition matrices is thus available. This model may be used to reconstruct the modulus ρ of the absorption profile.
Various embodiments and variants have been described. Those skilled in the art will understand that certain features of these various embodiments and variants may be combined, and other variants will occur to those skilled in the art. In particular, the described embodiments are not limited to the mathematical formulas used in the present disclosure. More generally, the provided method is applicable for other forms of equations modeling the optical process. As an example, in the acquisition model discussed hereabove, an amplitude aM common to order ωL+ωUS and to order ωL−ωUS has been considered. As a variant, the model may be refined by considering a specific vector for each of the two orders.
Further, in the above-described examples, a single phase jump per acoustic modulation period TUS is provided. As a variation, a number of phase jumps per period TUS greater than 1 may be provided to improve the spatial selectivity during the reconstruction of the absorption profile.
Further, the above-described examples may be adapted by replacing the phase jumps applied by the modulation and demodulation devices with frequency jumps or amplitude jumps.
Further, in the above-described examples, term aMref may be used to take into account, in the transition matrix, a possible lack of uniformity of the acoustic marking beam of the sample in the marking area.
The above-described absorption profile measurement methods may further be combined with a phase conjugation method to obtain a better signal-to-noise ratio. Each measurement is then performed in three steps:
1) Measurement of vector Yj by any adapted method, possibly a known method of the state of the art;
2) Extraction of the complex measurement phase Yj and then refocusing of the incident beam on the sample by phase conjugation, for example, by means of a device of the type described in application FR1903526 (B18214/DD19189) previously filed by the applicant; and
3) New measurement of vector Yj+1 according to the above-described method.
In another variant, a series of measurements may be performed with a same pair of patterns Ω and Φ for each measurement, an analysis of the phase variation of the measurements being implemented to extract information relative to the observed medium. In particular, in free field, the phase variations may enable to detect motions along the beam axis, or also density variations of a diffusion medium (for example, fog). In a diffusion medium, phase variations may enable to detect possible motions of the diffusers in the medium (variations of the motions of platelets enabling to trace back heart rate information for example). The cost function will possibly be adapted accordingly (the regularization term weighted by coefficient λ may apply to the modulus or the phase of the measurements).
Further, a learning of absorption patterns with patterns Ω and Φ may be provided to allow a direct detection of the absorption profiles from the measurements, without passing through the phase of reconstruction by minimization of the cost function.
It should further be noted that although, in the above-described embodiments, the modulation of the object light beam and, possibly, the modulation of the reference light beam, are performed by acoustic-optical modulation, the described embodiments are not limited to this specific case. More generally, any other light beam modulation method may be used, for example, Pockels-effect modulation methods. In particular, in the examples of FIGS. 6 and 7, acoustic-optical modulators MAO1 and MAO2 may be replaced with Pockels effect modulators. In this case, the modulation signal is an electric field and no longer an acoustic wave.
Such alterations, modifications, and improvements are intended to be part of this disclosure, and are intended to be within the spirit and the scope of the present invention. Accordingly, the foregoing description is by way of example only and is not intended to be limiting. The present invention is limited only as defined in the following claims and the equivalents thereto.