METHOD FOR AUTOMATICALLY IDENTIFYING AN ACOUSTIC SOURCE FROM A PRODUCED ACOUSTIC SIGNAL

Information

  • Patent Application
  • 20240003854
  • Publication Number
    20240003854
  • Date Filed
    June 29, 2023
    11 months ago
  • Date Published
    January 04, 2024
    4 months ago
Abstract
This method comprises: a step of reconstructing the acoustic signal produced during the occurrence or the evolution of the defect, this reconstruction step comprising constructing an estimate Se(w) of the signal produced at the position of the defect based on an ultrasonic signal Fj(t) measured by a sensor and using an estimate or a measurement of a product Rj(w)Gj(w) that relates, in the frequency domain, the measured ultrasonic signal Fj(w) to the produced acoustic signal S(w), anda step of automatically classifying the acoustic source, carried out based on the estimate of the reconstructed acoustic signal.
Description

The invention relates to a method and a device for automatically identifying an acoustic source from an acoustic signal produced by the occurrence or the evolution of a defect in a structure. The invention also relates to an information recording medium for implementing this method.


Such identification methods are used to identify the type of defect that has occurred in a structure. Indeed, in an industrial structure, acoustic sources may be varied (falling of an object, cracking, corrosion, breakup of material, friction, etc.). Depending on the purpose of the method or of the identification device, this will be calibrated so as to be sensitive only to certain events. The temporal distribution of these events makes it possible to monitor the state of health of the structure. However, this calibration may be rough, or multiple different events may be of interest, for example various types of damage such as fibre breakage or matrix cracking in the case of a composite. This is why individually monitoring events of each type may provide significantly finer information for supervision (for example, triggering an alert when fibre breakage occurs, but not for matrix cracking, or eliminating interfering signals).


For example, the structure to be inspected is a fibre-reinforced plastic sheet or the like. In this case, the acoustic source to be identified is the type of defect that has occurred or that is evolving. In this text, the expression “type of defect” denotes the physical nature of the defect. For example, in the case of structures consisting of a fibre-reinforced plastic matrix, there are primarily four types of defects, namely:

    • fibre breakages,
    • delamination,
    • breakup of the fibre matrix, and
    • cracking of the matrix.


These methods are therefore used in the field of non-destructive testing or SHM (“structural health monitoring”).


Known methods for identifying an acoustic source from the acoustic signal produced by the occurrence or the evolution of a defect in the structure comprise the following steps:

    • fitting out the structure with one or more sensors capable of measuring ultrasonic signals,
    • measuring ultrasonic signals caused by the occurrence or the evolution of the defect, and then
    • extracting certain physical characteristics of the measured ultrasonic signals, and then
    • automatically classifying the acoustic source into a class from among multiple possible classes based on the extracted physical characteristics.


Examples of such known methods are described in the following articles:

  • 1) Pashmforoush, F. et al.: “Damage characterization of glass/epoxy composite under three-point bending test using acoustic emission technique”, Journal of materials engineering and performance, vol. 21(7), pp. 1380-1390, 2012,
  • 2) Morizet, N. et al.: “Classification of acoustic emission signals using wavelets and Random Forests: Application to localized corrosion”, Mechanical Systems and Signal Processing, 70, 1026-1037, 2016
  • 3) Hisham A Sawan et al.: “Unsupervised learning for classification of acoustic emission events from tensile and bending experiments with open-hole carbon fiber composite samples”, Composites Science and Technology, Volume 107, 2015, Pages 89-97,
  • 4) Farova Zuzana et al.: “Experimental Signal Deconvolution in Acoustic Emission Identification Setup”, 6th International Workshop NDT in Progress 201, 12 Oct. 2011, pages 33-40.


The articles numbered 1) to 3) above are respectively denoted in this text by the following references: Pashmforoush2012, Morizet2016 and Sawan2015.


The article numbered 4) requires the time reversal of signals measured by sensors fastened to the structure before reinjecting them into the structure so as to ultimately measure the reinjected signals using an additional sensor situated at the location where the defect that generated the measured signals has arisen. Using this additional sensor situated at the location where the defect has arisen makes this identification method difficult or even impossible to implement. For example, when the defect occurs inside the structure, it is not possible to place the additional sensor at the location of this defect, or if the emission areas are expansive, this method would have to be fitted out in a highly challenging manner.


One such known method is also described in application RU2737235C1. It is emphasized that the method described in application RU2737235C1 additionally comprises a step of locating the position of the acoustic source before carrying out the classification. In that method, the position of the acoustic source is used to determine the distance separating each sensor from the acoustic source. Next, the type of defect is identified from a ratio between the amplitude of certain frequency components of the ultrasonic signal measured by a sensor and the distance separating this sensor from the acoustic source.


These known identification methods in particular have the advantage that it is not necessary to equip the structure with an electronic emitter emitting waves such as ultrasonic waves. Indeed, it is the ultrasonic wave generated by the defect itself that is used to carry out this identification.


The known identification methods work correctly for small structures. On the contrary, these methods do not work as well with large structures in which the distance between the acoustic source and the sensors may be far greater. Indeed, the identification is carried out by way of parameters of the signals, but these parameters are greatly impacted by the propagation of the wave in the structure, this effect possibly being predominant over variations resulting from the physical nature of the source. Thus, the classification process in practice risks discriminating the propagation distance rather than the physical nature, which distance may vary due to the use of multiple sensors and sources located randomly on the structure in practice. In this regard, it will be recalled that the location of the defect is not known a priori.


The invention aims to propose a method for identifying an acoustic source that exhibits the same advantages as the known methods while at the same time working just as well with small structures as it does with large structures.


One subject of the invention is therefore an identification method.


Another subject of the invention is an information recording medium, able to be read by a microprocessor, wherein this medium comprises non-transitory instructions for the execution of steps b) to d) of the claimed identification method when these instructions are executed by the microprocessor.


Another subject of the invention is a device for automatically identifying an acoustic source that implements the above identification method.





The invention will be better understood on reading the following description, which is given merely by way of non-limiting example, with reference to the drawings, in which:



FIG. 1 is a schematic illustration of the architecture of a device for automatically identifying an acoustic source;



FIG. 2 is a flowchart of a method for automatically identifying an acoustic source using the device of FIG. 1;



FIG. 3 is a graph illustrating the experimental results obtained by implementing the method of FIG. 2, and



FIG. 4 is a flowchart of another identification method able to be implemented by the device of FIG. 1.





In these figures, the same references have been used to designate elements that are the same. In the remainder of this description, features and functions that are well known to those skilled in the art are not described in detail.


Section I: Exemplary Embodiments


FIG. 1 shows a device 2 for automatically identifying an acoustic source from an acoustic signal produced by the occurrence or the evolution of a defect in a structure 4. In this case, the acoustic source is the defect that has occurred or that is evolving.


In this first embodiment, the structure 4 may be any structure in which, when a defect occurs or evolves, an acoustic signal S(t) is generated by this defect. The majority of the power of the signal S(t) is generally situated in the ultrasonic frequency band, that is to say in a frequency band ranging from 16 kHz to 10 MHz. In general, the power spectrum of the signal S(t) covers only part of the ultrasonic frequency band.


The structure 4 is made of a material that allows this signal S(t) to propagate inside this structure over a distance sufficient to be able to be measured by a sensor and distinguished from background noise.


By way of illustration, here, the structure 4 is a tube made of composite material formed of a fibre-reinforced plastic matrix. This tube extends along an axis 6 of revolution over a distance of 2 metres. Its cross section is constant and circular over its entire length.


A reference frame R is attached, without any degree of freedom, to the structure 4. The position of each point of the structure 4 is referenced by coordinates in the reference frame R. Here, the reference frame R comprises an axis X coincident with the axis 6 of revolution and an axis Y perpendicular to the axis X. For example, the axis Y is vertical and the axis X is horizontal.


Hereinafter, given that, in the specific case described here, the structure 4 is invariant to any rotation about the axis 6, the position of each point P of the structure 4 is defined by cylindrical coordinates (xp, θp), where:

    • xp represents the abscissa of this point along the axis X of the reference frame R, and
    • θp is the angle between a vector OkP and a direction parallel to the axis Y, where the vector OkP is the vector perpendicular to the axis 6 that extends from the point Ok situated on the axis 6 to the point P of the structure 4.


The device 2 comprises:

    • an electronic computer 10,
    • a human/machine interface 12 connected to the computer 10, and
    • N sensors C1 to CN connected to the computer 10.


The computer 10 comprises a programmable microprocessor 14 and a memory 16. The memory 16 comprises the instructions and the data needed to execute the method of FIG. 2 or FIG. 4 when these instructions are executed by the microprocessor 14.


The human/machine interface 12 is capable of communicating, in a manner directly intelligible to a human being, the results of the implementation of the identification method of FIG. 2 or 4. For example, the interface 12 comprises a screen.


Each sensor Cj measures the ultrasonic signal produced by the occurrence or the evolution of a defect in the structure 4. In this text, the index j is an integer between 1 and N that identifies the sensor Cj. The computer 10 acquires the measurements from each sensor Cj in order to process them.


The device 2 comprises at least three sensors Cj and, more often than not, at least four or six or eight sensors Cj. Here, by way of illustration, the total number N of sensors Cj is equal to eight. Only the sensors C1, C2, Cj and CN are shown in FIG. 1. The depiction of the other sensors has been replaced with dashed lines to simplify FIG. 1.


Here, the sensors Cj are all structurally identical to one another and differ from one another only in terms of their positions in the reference frame R. For example, the sensors Cj are piezoelectric sensors whose bandwidth covers at least the [100 kHz; 500 kHz] band, and preferably the [50 kHz; 0.5 MHz] or [20 kHz; 0.5 MHz] band.


Each sensor Cj is fastened to the structure 4 at a respective location Pj where it is capable of measuring an ultrasonic signal propagating in the structure 4. The coordinates (xj; θj) of each location Pj are known and stored in the memory 16.


The sampling frequency fe of the measurements performed by the sensors Cj is high, that is to say typically greater than 1 MHz.


Here, by way of illustration, the locations Pj are distributed uniformly along an axis parallel to the axis 6. However, other distributions of the locations Pj on the surface of the structure 4 are possible. For example, a plurality of the locations Pj may also be distributed along the circular circumference of the structure 4. Preferably, the distance between any one of the locations Pj and the closest location Pj+1 is computed so as to ensure good coverage of the structure, specifically that at least two sensors are able to measure a significant signal for a source at any point of the structure.


The operation of the device 2 will now be described with reference to the method of FIG. 2.


In a fitting-out step 50, the sensors Cj are each fastened to a respective location P1 on the structure 4. The coordinates (xj; θj) of each position Pj are recorded in the memory 16 in association with the identifier j of the sensor Cj.


A phase 52 then starts of using the device 2 to detect and identify an acoustic source.


In a step 54, each sensor Cj measures the ultrasonic signal Fj(t) propagating in the structure 4. Typically, the recording and the processing of the signal by each sensor is triggered by the detection of a wave by this sensor, for example by the ultrasonic signal Fj(t) passing above a threshold. The recording stops when the wave is no longer detected, for example when the ultrasonic signal is below the threshold for a significant duration that is determined beforehand. The threshold and the significant duration recorded after the first detection of the wave are therefore chosen so as not to lose information. The recorded duration is typically of the order of a millisecond.


By contrast, the significant duration after the last detection is short enough and the threshold is high enough for, in the vast majority of cases, a single defect to occur or evolve during the recording and noise not to trigger the recording. For example, for this purpose, the significant duration after the last detection is generally of the order of a few hundred μs.


At the same time, in step 54, the measurements from the sensors Cj are acquired by the computer 10 at the sampling frequency fe. The frequency fe is high enough to make it possible to acquire the measured signals Fj(t) while avoiding or eliminating aliasing phenomena. For example, here, the frequency fe is greater than 1 MHz. Advantageously, the frequency fe is twice as great as the upper bound of the bandwidth of the sensors Cj.


The signal Fj(t) measured by each sensor Cj is related, in the frequency domain, to the signal S(t) generated by the defect by the following relationship: Fj(w)=Rj(w)Gj(w)S(w), where:

    • j is an index identifying the sensor Cj,
    • w is the angular frequency in radians,
    • Fj(w) is the Fourier transform of the ultrasonic signal Fj(t) measured in the time domain by the sensor Cj,
    • S(w) is the Fourier transform of the acoustic signal S(t) produced by the occurrence or the evolution of the defect in the structure,
    • Rj(w) is the response, in the frequency domain, of the sensor Cj,
    • Gj(W) is the propagation function, in the frequency domain, of the acoustic signal in the structure between the position Ps where the defect occurs and the location Pj of the sensor Cj.


Once the signals Fj(t) have been acquired, in a step 56, the computer 10 locates the position Ps of the defect that generated the ultrasonic signals acquired by the sensors Cj. This locating consists in determining the coordinates (xs; θs) of the position Ps in the reference frame R from the measured and acquired signals Fj(t) and the known coordinates of the locations Pj. In addition, here, the speed c at which the signal S(t) propagates in the structure 4 is also determined in step 56. For example, here, to determine the coordinates (xs; θs) and the speed c, the triangulation method described in the following article is implemented: Zhang, F. et al.: “Evaluation of acoustic emission source localization accuracy in concrete structures”, Structural Health Monitoring, vol. 19(6), pages 2063-2074, 2020.


More specifically, the times of arrival tj of the signal S(t) at each location Pj are derived from the acquired ultrasonic signals Fj(t). For example, the time t1 corresponds to the first time at which the amplitude of the signal Fj(t) exceeds a predetermined threshold.


Next, the position Ps and the speed c are taken to be equal to the position and the speed that minimizes the following cost function J(Ps; c):







J

(


P
s

,
c

)

=




i
=
1


N
-
1







j
=

i
+
1


N




(





P
s

-

P
i




-




P
s

-

P
j




-

c

(


t
i

-

t
j


)


)

2







For this purpose, an algorithm for minimizing the cost function J(Ps; c) is implemented. For example, this may be an algorithm such as Newton's algorithm or the simplex algorithm. Genetic algorithms may also be implemented.


Once the locating of the acoustic source is complete, in a reconstruction step 58, the computer 10 constructs an estimate Se(w) of the signal S(w) produced at the position Ps by the defect. This step aims to compensate for the propagation of the signal S(t) between the position Ps and the locations Pj. To this end, in this first embodiment, the computer 10 uses the signals Fj(t), the position Ps and the speed c that are determined in step 56.


More specifically, in this first embodiment, what is called a “blind” reconstruction method is implemented. This method is said to be “blind” because it does not require each propagation function Gj(w) of the signal S(t) from the position Ps to the position Pj to be learned beforehand in a learning step. These methods are better known by the term “blind source deconvolution”.


Here, the method implemented by the computer 10 is described in the following article: Sabra, K. G. et al.: “Blind deconvolution in ocean waveguides using artificial time reversal”, The Journal of the Acoustical Society of America, vol. 116(1), pages 262-271, 2004. Below, this article is designated by the reference Sabra2004.


According to this method, in an operation 60, the computer 10 constructs an estimate Ge,j(w) of each propagation function Gj(w) of the signal S(t) from the position Ps to the location Pj. Here, each propagation function Ge,j(w) is a Green function. To this end, the estimate Ge,j(w) of the function Gj(w) is taken to be equal to the signal Fj(t) normalized by the L2 norm of all of the signals Fj(t) and multiplied by an estimated phase Γ. More specifically, the estimate Ge,j(w) is constructed using the following relationships:








G

e
,
j


(
w
)

=




F
j

(
w
)






a
=
1

N





F
a

(
w
)

2






e


-
i


Γ







where:

    • i is the imaginary number such that i2=−1,
    • Γ is a phase defined by the following relationship:






Γ
=

arg





j
=
1

N




W
j




F
j

(
w
)








where:

    • “arg” is the function that returns the argument of a complex number, and
    • Wj are predefined weights.


In addition, here, to make the method of FIG. 2 independent of the characteristics and the physical properties of the structure 4, the weights Wj are computed using the following relationship:







W
j

=

e



-




P
s

-

P
i





c


w






where:

    • i is the imaginary number such that i2=−1,
    • ∥Ps−Pj∥ is the distance separating the positions Ps and Pj,
    • c is the speed determined in step 56, and
    • w is the angular frequency in radians.


Next, in an operation 62, the estimate Se(w) is constructed by backpropagating the signals Fj(t) to the position Ps. To this end, here, each signal Fj(t) is backpropagated to the position Ps. Backpropagating a signal Fj(t) consists in multiplying the signal Fj(w) by the conjugate of the complex propagation function. In other words, the signal Fj(w) backpropagated to the position Ps is given by the following relationship: (Rj(w)Ge,j(w))*Fj(w), where the symbol “( . . . )*” denotes the conjugate of the complex function Rj(w)Ge,j(w) between parentheses.


The estimate Se(w) is therefore constructed from the terms (Rj(w)Ge,j(w))*Fj(w) and each term (Rj(w)Ge,j(w))*Fj(w) is the product of the conjugate of the function Rj(w)Ge,j(w) and the function Fj(w). Here, the estimate Se(w) is taken to be equal to the average of the N terms (Rj(w)Ge,j(w))*Fj(w). The estimate Se(w) is thus constructed using the following relationship:








S
e

(
w
)

=


1
N






j
=
1

N





(



R
j

(
w
)




G

e
,
j


(
w
)


)

*




F
j

(
w
)








In addition, in this embodiment, for simplification, each function Rj(w) is equal to one regardless of the value of the angular frequency w. The estimate Se(w) is thus simply constructed using the following relationship:








S
e

(
w
)

=


1
N






j
=
1

N





(


G

e
,
j


(
w
)

)

*




F
j

(
w
)








Once the estimate Se(w) has been constructed, in this embodiment, in a step 70, the estimate Se(t) of the signal S(t) in the time domain is obtained, for example, by applying an inverse Fourier transformation to the estimate Se(w).


Next, in this embodiment, the computer 10 executes a dimensionality reduction step 72. Such a step is also known by the term “dimension reduction”. Specifically, the estimates Se(w) and Se(t) are generally large, that is to say often composed of several tens of thousands or hundreds of thousands of samples. Step 72 is aimed at reducing the amount of information to be processed while still retaining the essential information that will allow reliable and robust identification of the defect.


In this embodiment, step 72 consists in extracting physical characteristics inherent to the signal S(t) from the estimates Se(w) and Se(t) constructed beforehand. For example, here, the procedure is as described in section 2.2 of the article Morizet2016, except that no energy in a band greater than 200 kHz is extracted.


Finally, a classification step 76 is executed by the computer 10. In this step 76, the computer 10 classifies the signal S(t) into a class chosen from among multiple possible classes based on the characteristics extracted in step 72. To this end, the computer 10 executes an automatic classification method. For example, here, the computer 10 executes a known unsupervised classification method. This known unsupervised classification method is for example an automatic classification method using a Gaussian mixture model, or the method known as the k-means method and described for example in the article Pashmforoush2012. Here, in order to implement this classification method, the number Nk of classes has been determined beforehand. For example, for the structure 4, the number Nk of classes was chosen to be equal to 3. The most appropriate number Nk of classes was determined here by trialling multiple possible values for the number Nk and then by computing, for each of these values of the number Nk, the value of the DB (Davies-Bouldin) index. The value of the number Nk for which the value of the DB index is minimum was selected. This computing of the value of the DB index is described for example in section 3.2 of the article Pashmforoush2012. Each class is related to a physical nature of the defect that produced the classified acoustic signal. The identification method thus makes it possible not only to detect the occurrence or the evolution of a defect, but also to identify the physical nature of this defect.


That which has been described above is repeated for each new event measured by the sensors Cj.



FIG. 3 illustrates, in the form of a graph, the results obtained by deforming the structure 4 repeatedly so as to make defects occur. In this figure, each point corresponds to an acoustic source and therefore to the occurrence or to the evolution of a defect in the structure 4. The ordinate axis is scaled in an arbitrary unit proportional to the amplitude of the signal Se(t), and the abscissa axis is in hertz. For each identified acoustic source, the amplitude of the estimate Se(t) and the frequency centroid of the estimate Se(w) were computed, and then this acoustic source was placed on the graph of FIG. 3 using the value of its frequency centroid as abscissa and the computed amplitude as ordinate. The frequency centroid is computed as described in section 6.1.1 of the following article: Geoffroy Peeters: “A large set of audio features for sound description (similarity and classification) in the CUIDADO project”, 2004. The frequency centroid is expressed in hertz. Each acoustic source was classified into one of the three possible classes by implementing the method of FIG. 2. The graph of FIG. 3 comprises three areas 31, 32 and 33 that each surround the points classified into a respective class. The overlaps between these three areas 31, 32 and 33 are very small, thereby illustrating the fact that the classification that is carried out is robust.



FIG. 4 shows another identification method able to be implemented instead of the method of FIG. 2 by the computer 10. The method of FIG. 4 is identical to the method of FIG. 2 except that:

    • a step 80 of learning each propagation function Gj(w) is introduced between step 50 and the use phase, and
    • the use phase 52 is replaced with a use phase 82.


In step 80, a broad-spectrum known acoustic signal Sc(t) is applied to various locations Pk of the structure 4. The signal Sc(t) has a spectrum that covers the bandwidth of the structure 4, that is to say that includes all frequencies able to propagate and to be measured in the structure 4 without being excessively attenuated. For example, here, the spectrum of the signal Sc(t) covers the ultrasonic frequency band. For example, in this embodiment, the signal Sc(t) applied to each location Pk is a Hsu-Nielsen source, that is to say an acoustic source produced by the snapping of pencil lead. Since this source is known, it is not described in any more detail.


The locations Pk are uniformly distributed here over the surface of the structure 4. For example, the locations Pk correspond to the vertices of a mesh covering the surface 4. Typically, the tiles of this mesh are identical or similar. The shortest distance between two contiguous locations Pk is for example between 1 mm and 100 mm or between 3 mm and 20 mm. Here, the shortest distance between two contiguous locations Pk is between 5 mm and 10 mm. The number Npk of locations Pk is typically proportional to the surface area of the structure 4. For example, the number Npk is taken to be equal to the ratio S4/Sref, where:

    • S4 is the surface area of the structure 4, and
    • Sref is a reference surface area, for example between 10−2 m2 and 1 m2.


The coordinates in the reference frame R of each location Pk are known and stored in the memory 16.


Each time the signal Sc(t) is applied to a location Pk, each sensor Cj measures the ultrasonic signal Fj,k(w) generated by the signal Sc(t) propagating in the structure 4. The ultrasonic signal Fj,k(w) is related to the acoustic signal Sc(t) by the following relationship: Fj,k(w)=Rj(w)Gj,k(w)Sc(w), where:

    • k is an index identifying the location Pk whose coordinates are known,
    • Fj,k(w) is the Fourier transform of the ultrasonic signal Fj,k(t) measured in the time domain by the sensor Cj when the known acoustic signal Sc(t) is applied to the location Pk,
    • Sc(w) is the Fourier transform of the known acoustic signal Sc(t) applied to the location Pk,
    • Rj(w) is the response, in the frequency domain, of the sensor Cj,
    • Gj,k(w) is the propagation function, in the frequency domain, of the known acoustic signal Sc(t) in the structure 4 between the location Pk and the location of the sensor Cj.


Next, for each location Pk where the signal Sc(t) was applied, the product Rj(W)Gj,k(W) of the functions Rj(w) and Gj,k(w) is recorded in the memory 16 in association with the identifier j of the sensor Cj and the location Pk where the signal Sc(t) was applied. The product Rj(w)Gj,k(w) is equal to the ratio Fj,k(w)/Sc(w) of the ultrasonic signal Fj,k(w) measured by the sensor Cj and the known acoustic signal Sc(w).


The phase 82 is identical to the use phase 52 except that the reconstruction step 58 is replaced with a reconstruction step 88. The reconstruction step 88 comprises:

    • an operation 90 of selecting, from the memory 16, the stored products Rj(w)Gj,k(w), and
    • an operation 92 of determining the estimate Se(w) of the signal S(w).


In the operation 90, the computer 10 selects the products Rj(w)Gj,k(w) associated with the location Pk closest to the position Ps obtained at the end of the locating step 56.


In the operation 92, the estimate Se(w) is constructed only from the products Rj(w)Gj,k(w) selected in the operation 90. More specifically, here, the estimate Se(w) is constructed from the terms (Rj(w)Gj,k(w))*Fj(w) in a manner similar to what was described in the case of the operation 62. For example, the estimate Se(w) is constructed using the following relationship:








S
e

(
w
)

=


1
N






j
=
1

N





(



R
j

(
w
)




G

e
,
j


(
w
)


)

*




F
j

(
w
)








where the products Rj(w)Gj,k(w) are those selected in the operation 90.


Section II: Variants

Variants of the “Blind” Reconstruction:


The values of the weights Wj may be different. In particular, in one simplified embodiment, the values of the weights Wj do not depend on the position Ps of the defect. For example, the weights Wj are constants all equal to +1. In this case, the locating step 56 may be omitted. In addition, when the locating of the defect is omitted, the number N of sensors Cj may be fewer than three. For example, the number N of sensors Cj is equal to two.


The values of the weights Wj may also be determined on the basis of the physical characteristics of the structure 4, such as for example the theoretical dispersion of guided waves in the structure. In the latter case, the weights Wj depend on the physical characteristics of the structure, such that the identification method is configured specifically to identify an acoustic signal in this structure.


As a variant, rather than using backpropagation of the signals measured by the sensors to obtain the estimate of the signal Se(w), it is possible to use inverse filtering, as described in the article Sabra2004. In the latter case, it is the terms Fj(w)/(Rj(w)Ge,j(w)) that are used to construct the estimate Se(w) rather than using the terms (Rj(w)Ge,j(w))*Fj(w). In this case, the estimate Se(w) is obtained using the following relationship:








S
e

(
w
)

=


1
N






j
=
1

N





F
j

(
w
)

/

(



R
j

(
w
)




G

e
,
j


(
w
)


)








In the latter case, the bandwidth of the estimate Se(w) is preferably limited so as to exclude highly attenuated frequencies, for example high frequencies attenuated by the structure 4. In other words, beyond a predetermined cutoff frequency fc, the estimate Se(w) is zero. This makes it possible to limit the sensitivity of the estimate Se(w) to noise. Specifically, for frequencies greater than fc, the product Rj(w)Ge,j(w) is close to zero, such that noise is highly amplified beyond the frequency fc if the bandwidth of the estimate Se(w) is not limited.


The reconstruction step may be carried out using “blind” reconstruction methods other than the one described. Numerous examples of other “blind” reconstruction methods may be found in the field of blind source deconvolution.


The estimate Se(w) may also be constructed from a weighted sum of the terms (Rj(w)Ge,j(w))*Fj(w) or Fj(w)/(Rj(w)Ge,j(w)). For example, the estimate Se(w) is constructed using the following relationship:








S
e

(
w
)

=




j
=
1

N






q
j

(



R
j

(
w
)




G

e
,
j


(
w
)


)

*




F
j

(
w
)







where qj is a weighting coefficient. The sum of the coefficients qj is equal to one. For example, the closer the sensor Cj is to the position Ps, the larger the coefficient qj, so as to give more weight to the measurements from the sensors closest to the defect. In a simplified case, only the coefficients qj of the M closest sensors Cj are non-zero and the coefficients qj of the (N−M) remaining sensors Cj are zero, where M is an integer less than N. For example, M is equal to one or two or three.


As a variant, the response Rj(w) is not a constant equal to one, and the estimate Se(w) is constructed using the following relationship:








S
e

(
w
)

=


1
N






j
=
1

N





(



R
j

(
w
)




G

e
,
j


(
w
)


)

*




F
j

(
w
)








To this end, if the impulse response of each sensor Cj is known, then Rj(w) may be constructed from this known impulse response. The response Rj(w) of each sensor Cj may thus be measured. In the latter cases, the response Rj(w) is generally not a constant.


Learning-Based Reconstruction Variants:


The applied known acoustic signal may be generated by an acoustic source other than a Hsu-Nielsen source. For example, the known acoustic signal may also be generated by an electronic acoustic emitter.


In the learning step, as a variant, it is not the same known acoustic signal that is applied to each location Pk. In this case, the product stored in association with each location Pk is obtained using the following relationship: Rj(w)Gj,k(w)=Fj,k(w)/Sc,k(w), where Sc,k(w) is the acoustic signal applied to the location Pk.


Like in the case of the “blind” reconstruction, the estimate Se(w) may also be constructed by inverse filtering. It may also be constructed from a weighted or unweighted sum of the terms (Rj(w)Gj,k(w))*Fj(w) or Fj(w)/(Rj(w)Gj,k(w)).


The data used in the learning step 80 may result from numerical simulations.


Variants Common to all Embodiments:


A step of preprocessing the ultrasonic signals Fj(t) measured by the sensors Cj may be carried out before the locating step and/or before the reconstruction step in order to carry out these steps on preprocessed ultrasonic signals and not on the raw ultrasonic signals delivered directly by the sensors. Examples of preprocessing operations that may be applied to the raw ultrasonic signals Fj(t) are described for example in section 2.1 of the article Morizet2016.


Other embodiments of the locating step 56 are possible. For example, as a variant, the location of the position Ps of the defect is obtained by implementing the method described in the following article: Pearson M. R. et al.: “Improved acoustic emission source location during fatigue and impact events in metallic and composite structures”, Structural Health Monitoring, vol. 16(4), pages 382-399, 2017. In addition, that article also describes another method for deriving the times of arrival tj.


If the speed c at which the signal S(t) propagates within the structure 4 is known, then it is not necessary to determine it in the locating step 56. For example, the known speed c is recorded beforehand in the memory 16.


When the estimate Se(w) is constructed from the terms Fj(w)/(Rj(w)Ge,j(w)) or Fj(w)/(Rj(w)Gj,k(w)), if the denominator of these fractions is close to zero within a frequency range, then this frequency range is excluded and is not taken into consideration in the classification step. Such a situation may be encountered for example when the structure greatly attenuates the propagation of waves at frequencies greater than fmax. For example, if the product Rj(w)Ge,j(w) or Rj(w)Gj,k(w) is less than 0.1 beyond the frequency fmax, then the estimate Se(w) is limited to the frequency range [0; fmax]. This makes it possible to work within the frequency range in which the signal-to-noise ratio is good and to exclude the frequency range where this signal-to-noise ratio is poor. This ultimately improves the robustness of the identification method.


Other embodiments of the dimensionality reduction step 72 are possible. For example, characteristics other than those cited in the article Morizet2016 may be used in addition to or instead of the characteristics cited in this article. It is also possible to use a smaller or a larger number of physical characteristics. It is also possible to reduce the amount of information to be processed in the classification step by methods other than the extraction of physical characteristics. For example, mathematical methods such as the principal component analysis method may be used to carry out the dimensionality reduction step. A brief description of this method may be found in section 3.1 of the article Pashmforoush2012.


The dimensionality reduction step may also be omitted. In this case, the number of samples of the signal Se(w) or Se(t) is not reduced before the classification step.


Classification methods other than those described above may be implemented to identify the acoustic source. For example, it is also possible to use the KGA (“K-means Genetic Algorithm”) method, or the WPT (“Wavelet Packet Transform”) method, both described in the article Pashmfouroush2012. The expectation maximization method, known by the acronym EM (“expectation maximization algorithm”), or else the unsupervised classification methods described in the article Sawan2015 may also be used.


Supervised classification methods may also be implemented. For example, the “Random Forest” method described in the article Morizet2016 may be implemented. The following supervised classification methods may also be used: k-NN (“k-Nearest Neighbours”), SVM (“Support Vector Machine”), an artificial neural network and the like.


The classification may be carried out based on the estimate Se(w) alone. In this case, the step 70 of constructing the estimate Se(t) may be omitted. On the contrary, the classification may also be carried out just based on the signal Se(t).


Other Variants:


The structure is not necessarily a tube made of composite material. That which has been described here is applicable to other structures, such as for example:

    • sheets, for example made of metal or composite material,
    • concrete structures, such as bridge piers.


The microprocessor 14 may be a generic processor, a specific processor, an application-specific integrated circuit (ASIC) or a field-programmable gate array (FPGA).


A plurality of the variants described here may be combined in one and the same embodiment.


Section III: Advantages of the Described Embodiments

Directly using the estimate of the acoustic signal S(t) produced at the position Ps for the classification instead of the measured ultrasonic signals makes it possible to make the classification more robust and therefore to increase the reliability of the method for identifying the acoustic signal. In particular, this makes it possible to reduce the influence of distortions of the acoustic signal that occur when it propagates in the structure so as to reach the locations Pj where it is measured by the sensors Cj. By virtue of this, the identification methods described here may be applied to any structure and, in particular, to large structures in which the acoustic signal travels a significant distance before reaching the sensors Cj.


Using the estimates Ge,j(w) of the propagation functions Gj(w) makes it possible to improve the robustness of the identification of the acoustic source without this requiring the execution of a learning step in which various functions Gj,k(w) are measured for various possible positions Pk of the defect and for a known acoustic signal. In addition, since the estimates Ge,j(w) are constructed each time the signals Fj(t) are measured, these estimates Ge,j(w) follow and adapt automatically to the variations of the functions Gj(w). Indeed, the functions Gj(w) may vary over time, in particular depending on the state of the structure and the environmental conditions in which the structure is placed.


Using weights Wj whose values depend on the distance separating the location Pj and the position Ps makes it possible to improve the precision of the estimate Se(w) and therefore to improve the reliability of the described method. In addition, in this case, the values of the weights Wj are independent of the characteristics of the fitted-out structure. Thus, in this case, the identification method works with any type of structure, and not only with pipes or flat structures such as sheets.


The fact that the response, in the frequency domain, of each sensor is assimilated to a constant equal to one simplifies the implementation of the identification method.


The step of learning the products Rj(w)Gj,k(w) makes it possible to construct a precise estimate of the signal Se(w) and therefore to obtain a robust identification method.

Claims
  • 1. A method for automatically identifying an acoustic source from an acoustic signal S(t) produced by an occurrence or an evolution of a defect in a structure, comprising: a) fitting out the structure by fastening at least one sensor to the structure, and thenb) using the sensor to measure an ultrasonic signal Fj(t) caused by the acoustic signal produced by the occurrence or the evolution of the defect in the structure, the measured ultrasonic signal being related, in the frequency domain, to the produced acoustic signal by the following relationship: Fj(w)=Rj(w)Gj(w)S(w), where: j is an index identifying the sensor,w is an angular frequency in radians,Fj(w) is a Fourier transform of the ultrasonic signal Fj(t) measured in the time domain by the sensor j,S(w) is the Fourier transform of the acoustic signal S(t) produced by the occurrence or the evolution of the defect in the structure,Rj(w) is a response, in the frequency domain, of the sensor j,Gj(w) is a propagation function, in the frequency domain, of the acoustic signal in the structure between a position of the defect and a location of the sensor j,and thenc) automatically classifying the acoustic source into a class of acoustic sources chosen from among multiple possible classes of acoustic sources,wherein: the method also comprises, between b) and c), d) reconstructing the acoustic signal produced during the occurrence or the evolution of the defect, the reconstructing comprising constructing an estimate Se(w) of the signal produced at the position of the defect based on the ultrasonic signal Fj(t) measured by the sensor and using an estimate or a measurement of a product Rj(w)Gj(w) that relates, in the frequency domain, the measured ultrasonic signal Fj(w) to the produced acoustic signal S(w), andthe automatic classification is carried out based on the estimate of the acoustic signal as obtained at an end of step d).
  • 2. The method according to claim 1, wherein: the fitting-out comprises fastening at least three sensors to the structure at locations whose coordinates are known in a reference frame fixed with respect to the structure,b) comprises using each of the sensors to measure a respective ultrasonic signal Fj(t) caused by the acoustic signal produced by the occurrence or the evolution of the defect in the structure, each of the measured ultrasonic signals being related, in the frequency domain, to the produced acoustic signal by the following relationship: Fj(w)=Rj(w)Gj(w)S(w),the method comprises, between b) and d), locating, in the fixed reference frame, the position of the defect that produced the acoustic signal based on the measurements carried out by the sensors in b) and on known coordinates of the sensors in this fixed reference frame, andd) comprises obtaining the estimate or the measurement of the product Rj(w)Gj(w) using the position of the defect as obtained at an end of the locating.
  • 3. The method according to claim 1, wherein d) comprises: for each sensor, an operation of constructing an estimate Ge,j(w) of the propagation function Gj(w) of the acoustic signal in the structure between the position where the defect occurs and the location of this sensor, using the following relationship:
  • 4. The method according to claim 2, wherein d) comprises:for each sensor, an operation of constructing an estimate Ge,j(w) of the propagation function Gj(w) of the acoustic signal in the structure between the position where the defect occurs and the location of this sensor, using the following relationship:
  • 5. The method according to claim 2, wherein each function Rj(w) is equal to one regardless of a value of the angular frequency w.
  • 6. The method according to claim 3, wherein the propagation function Ge,j(w) is a Green function.
  • 7. The method according to claim 2, wherein: d), the method comprises learning the product Rj(w)Gj(w) of the functions Rj(w) and Gj(w) for various possible locations of the defect, the learning comprising:i) for each sensor, measuring or estimating, using numerical simulation, an ultrasonic signal Fj,k(t) measured by the sensor when a known acoustic signal Sc,k(t) is applied to the structure at a location Pk whose coordinates are known in the fixed reference frame, each of the ultrasonic signals Fj,k(t) being related, in the frequency domain, to the known acoustic signal Sc,k(t) by the following relationship: Fj,k(w)=Rj(w)Gj,k(w)Sc,k(w), where:k is an index identifying the location Pk whose coordinates are known,Fj,k(w) is a Fourier transform of the ultrasonic signal Fj,k(t) measured or estimated in the time domain by the sensor j when the known acoustic signal is applied to the location Pk,Sc,k(w) is the Fourier transform of the known acoustic signal Sc,k(t) applied to the location Pk,Rj(w) is the response, in the frequency domain, of the sensor j,Gj,k(w) is the propagation function, in the frequency domain, of the known acoustic signal Sc,k(t) in the structure between the location Pk and the location of the sensor j, and thenii) for each sensor, storing, in association with the sensor j and the coordinates of the location Pk, the product Rj(w)Gj,k(w) of the functions Rj(w) and Gj,k(w), the product Rj(w)Gj,k(w) being obtained using the following relationship: Rj(w)Gj,k(w)=Fj,k(w)/Sc,k(w),iii) repeating i) and ii) for multiple possible locations Pk,d) comprises:iv) selecting the products Rj(w)Gj,k(w) stored in association with the location Pk closest to the position of the defect as obtained at the end of the locating, and thenv) determining the estimate Se(w) of the signal S(w) from the terms (Rj(w)Gj,k(w))*Fj(w) or the terms Fj(w)/(Rj(w)Gj,k(w)), where:the products Rj(w)Gj,k(w) are those selected in operation iv), andthe symbol “( . . . )*” denotes a conjugate of the complex function between parentheses.
  • 8. The method according to claim 1, wherein: the method comprises constructing the estimate Se(t), in the time domain, of the acoustic signal S(t) by applying an inverse Fourier transformation to the estimate Se(w), and thenthe automatic classification is carried out based on the estimate Se(t) in the time domain.
  • 9. The method according to claim 1, wherein the method comprises a dimensionality reduction step in which physical characteristics inherent to the signal produced at the position of the defect are extracted from the estimate Se(w) constructed in d), and then, in c), the automatic classification is carried out based on the characteristics extracted in the dimensionality reduction step.
  • 10. A non-transitory information recording medium, able to be read by a microprocessor, wherein the medium comprises non-transitory instructions for execution of b) to d) of an identification method according to claim 1 when these instructions are executed by the microprocessor.
  • 11. A device for automatically identifying an acoustic source from an acoustic signal S(t) produced by an occurrence or an evolution of a defect in a structure, comprising: at least one sensor fastened to the structure, the sensor being configured to measure an ultrasonic signal Fj(t) caused by the acoustic signal produced by the occurrence or the evolution of the defect in the structure, the measured ultrasonic signal being related, in the frequency domain, to the produced acoustic signal by the following relationship: Fj(w)=Rj(w)Gj(w)S(w), where:j is an index identifying the sensor,w is an angular frequency in radians,Fj(w) is a Fourier transform of the ultrasonic signal Fj(t) measured in the time domain by the sensor j,S(w) is the Fourier transform of the acoustic signal S(t) produced by the occurrence or the evolution of the defect in the structure,Rj(w) is a response, in the frequency domain, of the sensor j,Gj(w) is a propagation function, in the frequency domain, of the acoustic signal in the structure between the position where the defect occurs and a location of the sensor j,an electronic computer configured to:acquire the measurements from the sensor, and thenautomatically classify the acoustic source into a class of acoustic sources chosen from among multiple possible classes of acoustic sources,wherein the electronic computer is also configured to:before the automatic classification, reconstruct the acoustic signal produced during the occurrence or the evolution of the defect, the reconstruction comprising constructing an estimate of the signal produced at the position of the defect based on the ultrasonic signal Fj(t) measured by the sensor and using an estimate or a measurement of the product Rj(w)Gj(w) that relates, in the frequency domain, the measured ultrasonic signal Fj(w) to the produced acoustic signal S(w), andcarry out the automatic classification based on the estimate of the acoustic signal as obtained at an end of the reconstruction step.
Priority Claims (1)
Number Date Country Kind
22 06583 Jun 2022 FR national