This application is based upon and claims priority to Chinese Patent Application No. 202210215447.5, filed on Mar. 7, 2022, the entire contents of which are incorporated herein by reference.
The present invention relates to the field of bridge detection, in particular to a bridge damage identification method considering uncertainty, and the method is more accurate in damage identification and may be used for evaluating the safety state of a bridge structure.
A structural health monitoring (SHM) system, important for the safety of a large structure, has a core and most critical part of structural damage identification. Structural damage identification is intended to identify structural damage and use various types of available information to evaluate the condition of the structure. After years of development, there are many effective methods for structural damage identification, and “model-based” methods are commonly used. That is, the entire structure is modeled by using a finite element model and then serves as a benchmark to be compared with an actual structure that may be damaged. The finite element model can also simulate damage by modifying an integrating model, to further understand the impact of damage.
However, the requirements of model computation also become the main bottleneck of its engineering application, and the gap between the finite element model and the actual structure is still large even through a model updating. Therefore, there is a need to consider uncertainty errors in practice. For example, in practical applications, modeling errors and measurement noise are inevitable. These uncertainties may lead to large variations in the damage indicator proposed in the above method.
In order to overcome the defects in the prior art, the present invention provides a bridge damage identification method considering uncertainty, and aims to solve the problem of difference between a bridge finite element model and a real bridge structure due to environment or modeling errors and obtain a damage identification model with good identification precision. Therefore, the correctness of the model-based damage identification method applied to a real structure is improved.
In order to achieve the objective of the present invention, the following technical solution is used:
(Omitting Duplication)
Compared with the prior art, the present invention has the following beneficial effects:
1. The objective of the present invention is to make the domain adversarial network find common features between the bridge finite element model and the real bridge structure, and the features should be some basic properties of the real bridge structure, such that the actual damage identification performance for the real bridge structure is desirable.
2. The real bridge structure does not need to have damage labels to participate in network training to obtain a classifier. Only a large amount of well-labeled data needs to be obtained from the bridge finite element model, and a satisfactory classifier is trained based on the data, such that the problem that the real bridge structure lacks damage labels is solved.
3. According to the method, not all uncertainty factors in the real bridge structure need to be known, the difference between the real bridge structure and the finite element model may be implicitly related by constructing a partial uncertainty dataset of the bridge finite element model and the real bridge structure, and therefore, modeling of all possible uncertainties in the finite element model may be avoided, and the complexity problem of the real world may be solved.
4. According to the method, only an original acceleration response signal in time domain is needed as input, features related to bridge structure damage may be extracted from the signal without an advanced signal processing tool, so as to achieve efficient bridge damage identification.
A structure of a domain adversarial transfer network in the present invention is shown in
step 1: construct a bridge acceleration response dataset;
step 1.1: determine a geometric material parameter of a bridge, establish a bridge finite element model, divide the bridge finite element model into units and number the units as [1, 2, 3, . . . , 10] in sequence, number nodes between the elements as [1, 2, 3, . . . , 9] in sequence, and arrange accelerometers at the nodes between the elements;
step 1.2: construct a bridge lossless dataset in a bridge lossless state:
firstly, apply a Gaussian distributed random load once to the node 1, where the random load obeys Gaussian distribution with a mean value of 0 and a standard deviation of 200, and obtain an acceleration response signal matrix [a1, . . . , ac, . . . , aC]T of each node between the elements through a Newmark−β method, where ac is an acceleration response signal with the length of w at the c-th node between the elements, C is the number of the accelerometers, C=n−1, an acceleration signal in this example has a length w=5120, and C=9;
step 1.3: normalize the acceleration response signals of all the nodes, to obtain a normalized acceleration response signal matrix [ā1, . . . , āc, . . . , ā9]T; where āc represents the normalized acceleration response signal at the c-th node;
step 1.4: randomly intercept the normalized acceleration response signal matrix [ā1, . . . , āc, . . . , āC]T according to the length {tilde over (w)}=1024, to enhance data diversity, so as to obtain an acceleration response matrix segment A=[ã1, . . . , ãc, . . . , ã9]T of the bridge when the random load is applied to the node 1, and take the segment as a sample; where ãc represents the normalized acceleration response signal with the length of {tilde over (w)} at the c-th node;
step 1.5, repeatedly apply m times of Gaussian distributed random load to the node 1 between the elements, and performing steps 1.2-1.4 for processing, to obtain m samples, where in order to balance the samples in this example, computation is repeated for m=80 times at each random load position in a lossless state; and
step 1.6: process the nodes between the elements in sequence according to steps 1.2-1.5, so as to obtain the acceleration response matrix segment of the bridge when the random load is applied to each node between the elements, where a total of j1=m×(n−1)=80×9=720 samples form the bridge lossless dataset;
step 2: construct a bridge damage dataset:
simulate elemental damage in a mode of reducing stiffness of one element, and set t different reduction coefficients to simulate damage grades, where in this embodiment, t=5 damage grades are set, that is, the stiffness of the element is reduced by 10%, 20%, 30%, 40% and 50%;
step 2.1: select a damage grade at the element 1 for damage simulation, and process each node between the elements according to steps 1.2-1.6 when the element 1 is at the current damage grade, where computation is repeated for m=16 times at each random load position in a lossless state, so as to obtain an acceleration response matrix segment of the bridge under the current damage grade of the element 1;
step 2.2: obtain, according to step 2.1, the acceleration response matrix segments of the bridge under all loss grades of the element 1, with a total of t×(n−1)×m=5×9×16=720 damage samples; and
step 2.3: process all the elements in sequence according to steps 2.1-2.2, so as to obtain acceleration response matrix segment samples of the bridge under different loss grades of each element, with a total of j2=t×(n−1)×m×n=720×10=7200 damage samples to form the bridge damage dataset;
step 3: set labels Y=[y1, . . . , yl, . . . , yL] for the bridge lossless dataset and the bridge damage dataset, where the yl represents the label corresponding to the l-th sample, L is the total number of the samples, and L=j1+j2=720+7200=7920; and yl∈[0, 1, 2, . . . , 10], when yl=0, it represents the bridge lossless state corresponding to the l-th sample, and when yl=1, 2, . . . , n, it represents a serial number of a damaged unit of the bridge corresponding to the l-th sample;
step 4: combine the bridge lossless dataset and the label thereof with the bridge damage dataset and the label thereof to obtain a source domain dataset S(X,Y) with a label, where X represents a union set of the bridge lossless dataset and the bridge damage dataset, X=[A1, . . . , Al, . . . , AL], and Al represents an l-th combined sample;
step 5: simulate the real bridge structure by adding uncertainty to the bridge finite element model;
step 5.1: take the uncertainty of stiffness of the bridge as a working condition (1), and simulate the working condition (1) by multiplying stiffness of each element of the bridge finite element model by a random factor δ, where the random factor δ obeys Gaussian distribution with a mean value of 1 and a standard deviation of 0.02, then a target domain dataset T1(X,Y) is obtained according to steps 1 and 2, the target domain is free of labels, and for later verification of an effect of the method, a label is also set here, and Y is the same as the source domain;
step 5.2: take a geometric error and a material error of the bridge as a working condition (2), simulate the material error by changing density and elastic modulus parameters of the bridge finite element model, where the material has density increased by 2% and elastic modulus reduced by 2% in this example, and simulate the geometric error by changing a bridge length, a cross section width and a height of the finite element model, where the bridge has a length set to be 1.98 m (with a relative error of 1%), a cross section width of 0.29 m (with a relative error of 3.33%), and a height of 0.098 m (with a relative error of 2%) in this example, so as to obtain a target domain dataset T2(X,Y) according to steps 1 and 2 under the condition of the working condition (1) and the working condition (2), where the target domain is free of labels, and for later verification of an effect of the method, a label is also set here, and Y is the same as the source domain; and
step 5.3: take an influence of environmental noise in actual measurement as a working condition (3), simulate the working condition (3) by adding noise Dnoise−N(0,σ2) obeying Gaussian distribution with a mean value of 0 and a variance of σ2 into the bridge finite element model, where in this example, Dnoise−N(0, 0.12), that is, add noise Dnoise−N(0,σ2) into the normalized acceleration response signal matrix [ā1, . . . , āc, . . . , āC]T, so as to obtain a target domain dataset T3(X,Y) according to steps 1 and 2 under the conditions of the working condition (1), the working condition (2) and the working condition (3), where the target domain dataset is free of labels, and for later verification of an effect of the method, a label is also set here, and Y is the same as the source domain;
step 6: construct a domain adversarial transfer learning neural network, where the domain adversarial transfer learning neural network includes: a feature extractor Gf, a label predictor Gy and a domain classifier Gq;
the feature extractor Gf includes e1 convolution layers, a LeakRelu layer is added to each convolution layer, a normalization layer and a maximum pooling layer are added between every two convolution layers, convolution kernels of the convolution layers have a size of k1, the number of h1, and a step length of s1, and convolution kernels of the maximum pooling layers have a size of k2 and a step length of s2, where in this embodiment, e1 is 6, the convolution kernels have the size k1=16, the number h1 of the convolution kernels is 32, 32, 64, 64, 128 and 128 in sequence, the step length s1=1, and the convolution kernels of the maximum pooling layers have the size of k2=4 and the step length of s2=4;
the label predictor Gy is composed of e2 full connection layers, and a LeakRelu layer is added between each full connection layer, where in this example, e2 is 3, a vector dimension flattened by the above feature extractor is 2048, the input of the full connection layers is 2048, 256 and 128 in sequence, and the final output is 11;
the domain classifier Gq is composed of e3 full connection layers, and a ReLU layer and a normalization layer are added between each full connection layer, where in this example, a3 is 3, a vector dimension flattened by the above feature extractor is 2048, the input of the full connection layers is 2048, 1024 and 256 in sequence, and the final output is 2;
step 7, preprocess data, divide the source domain dataset S(X,Y) and the target domain dataset Ti(X), i=1, 2, 3 into a source domain training set Dstra, verification set Dsval, and test set Dstest and a target domain training set Dttra, verification set Dtval and test set Dttest separately according to a proportion of 7:1:2, where the divided test set is used for an application stage, to check the effectiveness of the method, such that training set sample size: verification set sample size: test set sample size in each of the two dataset is =0.7:0.1:0.2=5544:792:1584;
step 8, perform a training and verification stage, where a training flow diagram is shown in
step 8.1: randomly extract P=64 source domain samples Xs=(As1, . . . , Asp, . . . , As64), Ys=(y1, . . . , yp, . . . , y64) and target domain samples Xt=(At1, . . . , Atp, . . . , At64) from the source domain training set Dstra and the target domain training set Dttra each time as a small batch to be input a domain adversarial network and then trained until all the samples of the source domain training set Dstra and the target domain training set Dttra are extracted; where Asp represents the p-th source domain sample of the source domain samples in a small batch; and yp is a label corresponding to the p-th sample of the source domain samples, and Atp represents the p-th target domain sample of the target domain samples in a small batch;
step 8.2: map the p-th source domain sample Asp and the target domain sample Atp in a small batch into a source domain feature vector fsp and a target domain feature vector ftp by means of the feature extractor Gf(A,θf) respectively; where the A is the source domain sample and target domain sample, and θf represents parameter vectors of all layers in mapping;
step 8.3: map the source domain feature vector fsp by means of the label predictor Gy to obtain a prediction label ŷp=Gy(fsp,θy), where θy represents a mapping parameter of the label predictor Gy;
step 8.4: compute a loss Ly(ŷp,yp) of the label predictor by using equation (1):
Ly(ŷp,yp)=−yp log(ŷp) (1)
step 8.5: set a domain label QS=(qs1, . . . , qsp, . . . , qsP) for the source domain sample Xs=(As1, . . . , Asp, . . . , AsP); and set a domain label Qt=(qt1, . . . , qtp, . . . , qtP) for the target domain sample Xt=(At1, . . . , Atp, . . . , AtP), qsp and qtp represent the domain label corresponding to the p-th source domain sample Asp and the domain label corresponding to the p-th target domain sample Atp respectively, qsp=0, and qtp=1;
step 8.6: input the source domain feature vector fsp and the target domain feature vector ftp into the domain classifier Gq(f,θq) for mapping, to obtain a prediction domain label {circumflex over (q)}p, where f represents the source domain feature vector and the target domain feature vector, and θq represents a mapping parameter of the domain classifier Gq;
step 8.7: compute a loss Lq({circumflex over (q)}p,qp) of the domain classifier by using equation (2):
Lq({circumflex over (q)}p,qp)=−qp log({circumflex over (q)}p) (2)
where in equation (2), qp is a domain label corresponding to the p-th sample in a small batch;
step 8.8: establishing a global objective function E(θf,θy,θq) by using equation (3):
where in equation (3), Lyp and Lqp represent a label classifier loss function and a domain classifier loss function calculated by means of the p-th sample in a small batch respectively, λ represents measurement of a weight between the two targets; and λ represents measurement of a weight between two targets, the domain classifier needs to identify a source domain and a target domain as far as possible, to form an adversarial relation with the feature extractor, such that the greater loss is better, which is not facilitates the solution of the entire loss, and a negative sign is added before the domain adversarial loss;
step 8.9: use a standard stochastic gradient descent (SGD) solver to carry out gradient descent search on saddle point parameters of equations (4) and (5), where SGD optimizer parameters are set as follows: a learning rate is set to be 0.01, and momentum is set to be 0.9, so as to obtain a saddle point parameter
where in equations (4) and (5), {circumflex over (θ)}f,{circumflex over (θ)}y,{circumflex over (θ)}q represent mapping parameters of the feature extractor, the label predictor and the domain classifier respectively when the global objective function E(θf,θy,θq) is converged to an optimal state, at the saddle point, the parameters θq of the definition domain classifier minimize a definition domain classification loss (because they come in with a negative sign), while the parameters θy of the label predictor minimize a label prediction loss, and the feature mapping parameters θf minimize a label prediction loss (that is, the features are discriminative) while maximizing a domain classification loss (that is, the features have domain invariance);
step 8.10: repeat steps 8.2-8.9 to complete training of the domain adversarial transfer network by all small batches, and finally obtain a saddle point parameter {circumflex over (θ)}f,{circumflex over (θ)}y,{circumflex over (θ)}q when the global objective function E(θf,θy,θq) is converged to be optimal and an optimal network model; and
step 8.11: perform a validation stage:
verify a label prediction correct rate index of the optimal network model based on the source domain verification set Dsval and the target domain verification set Dtval, indicate, under the condition that the label prediction correct rate index reaches a threshold, that training of the domain adversarial transfer learning network is completed, save the mapping parameter {circumflex over (θ)}f,{circumflex over (θ)}y,{circumflex over (θ)}q when the global objective function E(θf,θy,θq) is converged to the optimal state, and otherwise, return to steps 8.1-8.10 for retraining, and repeat steps 8.1-8.10 for 64 times, such that the verification set correct rate of each group of working conditions basically tends to a threshold; and
step 9: an application stage is performed:
take the Ti(X,Y) test set Dttest as input in an application process as shown in
Number | Name | Date | Kind |
---|---|---|---|
20160034621 | Huang | Feb 2016 | A1 |
Number | Date | Country |
---|---|---|
106556498 | Apr 2017 | CN |
109839440 | Jun 2019 | CN |
113392986 | Sep 2021 | CN |
113567130 | Oct 2021 | CN |
Entry |
---|
M. Y. Choi et al., “Damage Detection System of a Real Steel Truss Bridge by Neural Networks,” Smart Systems for Bridges, Structures, and Highways, S. C. Liu. Editor, Proceedings of SPIE vol. 3988 (Year: 2000). |
Jiangpeng Shu et al., “The application of a damage detection method using Artificial Neural Network and train-induced vibrations on a simplified railway bridge model,” Engineering Structures 52 (2013) 408-421 (Year: 2013). |
A. C. Neves et al., “Structural health monitoring of bridges: a model-free ANN-based approach to damage detection,” J Civil Struct Health Monit (2017) 7:689-702 (Year: 2017). |
He Wen-Yu, et al., Bridge Damage Detection Based on the Moving-vehicle-induced Response and L1 Regularization, China J. Highw. Transp., 2021, pp. 61-70, vol. 34, No. 4. |