This application claims priority to Chinese Patent Application No. 201710186738.5, filed on Mar. 27, 2017, entitled “an fault identifying method for sludge bulking based on a recurrent RBF neural network,” which is hereby incorporated by reference in its entirety.
In this present disclosure, an identifying method is designed for detecting the sludge bulking and find the fault reasons of sludge bulking in the urban wastewater treatment process (WWTP) by a recurrent RBF neural network. The sludge volume index-SVI is an important parameter of characterization of sludge bulking. The basic link of proposed identifying method is to predict SVI on the basis of the relationships between variables. The technology of this present disclosure is part of advanced manufacturing technology, belongs to both the field of control engineer and environment engineer. Therefore, the identifying method for detecting the sludge bulking and find the fault reasons of sludge bulking in WWTP is of great significance.
The urban WWTP also happen the sludge bulking in the activated sludge process. However, the various influencing factors for SVI are various and complex. Therefore, it is difficult to identify the cause variables for sludge bulking, which seriously affected the stable operation of the urban WWTP. The identifying method for sludge bulking, based on recurrent RBF neural network, is helpful to detect the fault phenomenon of sludge bulking and identify the cause variables that resulted in the sludge bulking, which strengthen delicacy management, ensure water quality effluent standards of urban WWTP. It has better economic benefit as well as significant environmental and social benefits. Thus, the research achievements have wide application prospect in this present disclosure.
At present, the activated sludge process has widely used in the urban wastewater treatment process of papermaking, printing and dyeing, chemical industry, and many other industrial wastewater. However, sludge bulking fault problem has always been a thorny problems existing in the activated sludge system. Due to the sludge and water can't normal separate, cause the failure of wastewater treatment process. Sludge bulking occurs frequently and it has basically different degrees of sludge bulking in the wastewater treatment process. High sludge bulking coverage, in many countries such as Germany, Britain, South Africa's wastewater treatment plants according to the survey, more than half of the wastewater treatment plants exist the situation of the excessive growth of filamentous bacteria. Thus, sludge bulking is a common problem in wastewater treatment plants at home and abroad. Many scholars of most countries have studied the prevention and control method of the sludge bulking. Although some progress has been made, but so far, there is no effective control measures of sludge bulking; Moreover, in the event of the sludge bulking, the reason is not easy to explore, and need longer time to treat the failure of wastewater treatment. To sum up, once in the event of sludge bulking, consequences cannot be ignored. For this failure phenomenon of sludge bulking, therefore, early diagnosis and prevention is the most effective methods to solve the problem of sludge bulking, so it has high practical significance.
In this present disclosure, an identifying method for sludge bulking, is presented by building a soft-computing model based on recurrent RBF neural network. The neural network uses fast gradient descent algorithm to ensure the accuracy of recurrent RBF neural network. Once the sludge bulking is detected, an identifying cause variables (CVI) algorithm will be exploited to implement the identification of fault variables. This method can effectively prevent the happen of sludge bulking and reduce the loss of the wastewater treatment plant.
A fault identification method is designed for the sludge bulking based on a recurrent RBF neural network. Its characteristic and steps include following steps:
(1) Determine the Input and Output Variables of SVI:
For sewage treatment process of activated sludge system, by analyzing the detailed mechanism of sludge bulking, five process variables are analyzed and select the input variables of SVI soft-computing model: dissolved oxygen concentration-DO, mixed liquor suspended solids concentration-MLSS, temperature-T, chemical oxygen demand-COD and total nitrogen-TN. The output value of soft-computing model is detected SVI concentration.
(2) Initial Recurrent RBF Neural Network:
The structure of recurrent RBF neural network comprise three layers: input layer, hidden layer and output layer. The network is 5-J−1, named the number of input layer is 5 and hidden neurons is J. Connection weights between input layer and hidden layer are assigned 1, the connection weights between hidden layer and output layer randomly assign values, the assignment interal is [1, 1]. The number of the training sample is N and the input of recurrent RBF neural network is x(t)=[x1(t), x2(t), x3(t), x4(t), x5(t)] at time t. The expectations output of neural network output is expressed as yd(t) and the actual output is expressed as y(t). Soft-computing method of SVI can be described:
{circle around (1)} The input Layer: There are 5 neurons which represent the input variables in this layer. The output values of each neuron are as follows:
ui(t)=xi(t) (1)
wherein ui(t) is the ith output value at time t, i=1, 2, . . . , 5, and the input vector is x(t)=[x1(t), x2(t), . . . , x5(t)].
{circle around (2)} The Hidden Layer: There are J neurons of hidden layer. The outputs of hidden neurons are:
cj(t) denotes the center vector of the jth hidden neuron and cj(t)=[cj1(t), cj2(t), . . . , cjn+1(t)]T at time t, ∥hj(t)−cj(t)∥ is the Euclidean distance between hj(t) and cj(t), and σj(t) is the radius or width of the jth hidden neuron at time t, hj(t) is input vector of the jth hidden neuron at time t described as
hj(t)=[u1(t),u2(t), . . . u5(t),vj(t)×y(t−1)]T (3)
y(t−1) is the output value of the output layer at time t−1, 1), vj(t) denotes the connection weight from output layer to the jth hidden neuron at time t, and v(t)=[v1(t), v2(t), . . . , vJ(t)]T, T represents transpose.
{circle around (3)} The Output Layer: There is only one node in this layer, the output is:
wherein w(t)=[w1(t), w2(t), . . . , wJ(t)]T is the connection weights between the hidden neurons and output neuron at time t, θ(t)=[θ1(t), θ2(t), . . . , θJ(t)]T is the output vector of the hidden layer, y(t) represents the output of recurrent RBF neural network at time t.
The error of recurrent RBF neural network is:
yd(t) is the expectation output of neural network and the actual output is expressed as y(t).
(3) Train Recurrent RBF Neural Network:
{circle around (1)} Given the recurrent RBF neural network, the initial number of hidden layer neurons is J, J>2 is a positive integer. The input of recurrent RBF neural network is x(1), x(2), . . . , x(t), . . . , x(N), correspondingly, the expectation output is yd(1), yd(2), . . . , yd(t), . . . , yd(N), expected error value is set to Ed, Ed∈(0, 0.01). The every variable of initial centre value c1(1)∈(−2, 2), width value σj(1)∈(0, 1), initial feedback weight vj(1)∈(0, 1), j=1, 2, . . . , J. Initial weight w(1)∈(0, 1).
{circle around (2)} Set the learning step s=1;
{circle around (3)} t=s; According to Eqs. (1)-(4), calculate the output of recurrent RBF neural network, exploiting fast gradient descent algorithm:
ηc, ησ, ηv, ηw are the learning rate of centre, width, feedback connection weight from output layer to hidden layer and the connection weight between hidden layer and output layer, respectively. In addition, ηc∈(0, 0.01], ησ∈(0, 0.01], ηv ∈(0, 0.02], ηw∈(0, 0.01]. cj(t+1)=[c1j(t+1), c2j (t+1), . . . , c5j(t+1)] denotes the center vector of the jth hidden neuron at time t+1. σj(t+1) is the radius or width of the jth hidden neuron at time t+1. vj(t+1) denotes the connection weight from output layer to the jth hidden neuron at time t+1. wj(t+1) is the connection weights between the hidden neurons and output neuron at time t+1.
{circle around (4)} increase 1 learning step s, if s<N, then turn to step {circle around (3)}. If s=N, turn to step {circle around (5)}.
{circle around (5)} according to Eq. (5), calculate the performance of recurrent RBF neural network. If E(t)≥Ed, then turn to step {circle around (3)}; If E(t)<Ed, stop the training process.
(4) SVI Concentration Prediction:
The testing samples used as the input of recurrent RBF neural network, the output of neural network is the soft-computing values of SVI.
(5) CVI Algorithm for Sludge Bulking:
{circle around (1)} Calculate the residual of the expectation output and the output of recurrent RBF neural network, if
then turn to step {circle around (2)}, otherwise, stop the process of fault identification for sludge bulking.
{circle around (2)} Define two formula:
wherein IC1(t) is the Mahalanobis distance of input variables at time t and IC2(t) is the squared prediction error at time t, M is the numbers of principal component of input data, θM(t) is the output vector of the hidden layer for M principal components training data at time t, K is the number of the front samples. δ(t)=[δ1(t), . . . , δm(t), . . . , δM(y)]T is the projection of the training data, and δm(t) is
wherein
wherein ΛM(t) is the diagonal matrix of eigenvalues at time t and it satisfies
{tilde over (C)}(t)=Z(t)ΛM(t)ZT(t)+l′(t)(I(t)−Z(t)ZT(t)) (14)
wherein l′(t) is a constant value, I(t) is a unit matrix, {tilde over (C)}(t) is the regularized covariance matrix of C(t) at time t:
wherein Ω(t) is the energy of each variable. θk(t) is the hidden output vector of the kth principal component at time t.
The constant vector a(t)=[a1(t), . . . ak(t), . . . aK(t)]T is given as
G(t) is the Gaussian matrix and λ(t) denotes the eigenvalue
G(t)={θi(t)·θj(t)}K×K (19)
λ(t)P(t)=C(t)p(t) (20)
p(t) denotes eigenvector of the covariance matrix C(t) at time t.
{circle around (3)} For the ith input variable, the contribution degree index satisfy:
wherein Gi(t) is the contribution degree index of ith variable at time t, a hidden layer neuron corresponds to an input variable, κi(t) is the contribution degree which is calculate by the mutual information between this testing samples and difference sets for testing samples xi(t) of ith variable at time t, which can be expressed as
κi(t)=I(xi(t),VΔ(t)) (22)
I(xi(t), VΔ(t)) is the mutual information of xi(t) and VΔ(t) at time t, VΔ(t) is the difference matrix sets of data of training set and testing set at time t, which is expressed as
VΔ(t)=Vtr(t)−Vte(t) (23)
wherein Vtr(t) and Vte(t) is independent data sets of training set and testing set at time t, respectively.
Vtr(t)=D−1(t)G(t) (24)
Vte(t)=Dte−1(t)Gte(t) (25)
D(t) is the covariance matrix of Φ(t) at time t
D(t)=E{Φ(t)ΦT(t)} (26)
Φ(t)=[θ(t−K+1), . . . ,θ(t−1),θ(t−1),θ(t)]T (27)
Φ(t) is output matrix of hidden layer at time t, θ(t−K+1) is the hidden output vector at time t−K+1, K is the number of the front samples.
{circle around (4)} For the ith input variable, a hidden layer neuron corresponds to an input variable, if the Gi(t) at time t satisfies:
G1(t)+ . . . Gi(t)≥0.8 (28)
the variables 1, . . . , i is the cause variables resulted in sludge bulking.
The Novelties of this Present Disclosure Contain:
(1) To detect the sludge bulking and identify the cause variables that resulted in the sludge bulking, an identifying method for sludge bulking is developed in this present disclosure. The results demonstrate that the SVI trends in WWTP can be predicted with acceptable accuracy using the DO, MLSS, T, TN and COD as input variables. This method can not only solve the problem of measured online for SVI concentration with acceptable accuracy but also detect the happen of sludge bulking.
(2) This identifying fault variables method is based on the CVI algorithm. And it identify the fault variables of sludge bulking in the WWTP with high identifying precision. Thus, it can realize the effective regulation of the sludge bulking control in advance.
Attention: this present disclosure utilizes five input variables in this identifying method to predict the SVI. In fact, it is in the scope of this present disclosure that any of the variables: DO, T, MLSS, COD and TN, are used to predict the SVI concentration. Moreover, this identifying method is also able to predict the others variables in urban WWTP.
This invention takes MLSS, DO, T, COD and TN as characteristic variables for SVI, the above unit is mg/L;
The experimental data comes from water quality analysis statement of a wastewater treatment plant in 2014; choosing data of MLSS, DO, T, COD and TN as experimental samples, after eliminating abnormal sample, 100 groups of data are available, and the group of 60 used as training samples, the remaining 40 groups as test samples.
This present disclosure adopts the following technical scheme and implementation steps:
A fault identification method is designed for the sludge bulking based on a recurrent RBF neural network. Its characteristic and steps include following steps:
(1) Determine the Input and Output Variables of SVI:
For sewage treatment process of activated sludge system, by analyzing the detailed mechanism of sludge bulking, five process variables are analyzed and select the input variables of SVI soft-computing model: dissolved oxygen concentration-DO, mixed liquor suspended solids concentration-MLSS, temperature-T, chemical oxygen demand-COD and total nitrogen-TN. The output value of soft-computing model is detected SVI concentration.
(2) Initial Recurrent RBF Neural Network:
The structure of recurrent RBF neural network comprise three layers: input layer, hidden layer and output layer in
{circle around (1)} The input Layer: There are 5 neurons which represent the input variables in this layer. The output values of each neuron are as follows:
ui(1)=xi(t) (29)
wherein ui(t) is the ith output value at time t, i=1, 2, . . . , 5, and the input vector is x(t)=[x1(t), x2(t), . . . , x5(t)].
{circle around (2)} The Hidden Layer: There are J neurons of hidden layer. The outputs of hidden neurons are:
cj (t) denotes the center vector of the jth hidden neuron and cj(t)=[cj1(t), cj2(t), . . . , cjn+1(t)]T at time t, ∥hj(t)−cj(t)∥ is the Euclidean distance between hj(t) and cj(t), and σj(t) is the radius or width of the jth hidden neuron at time t, hj(t) is input vector of the jth hidden neuron at time t described as
hj(t)=[u1(t),u2(t), . . . u5(t),vj(t)×y(t−1)]T (31)
y(t−1) is the output value of the output layer at time t−1, vj(t) denotes the connection weight from output layer to the jth hidden neuron at time t, and v(t)=[v1(t), v2(t), . . . , vJ(t)]T, T represents transpose.
{circle around (3)} The Output Layer There is only one node in this layer, the output is:
wherein w(t)=[w1(t), w2(t), . . . , wJ(t)]T is the connection weights between the hidden neurons and output neuron at time t, θ(t)=[θ1(t), θ2(t), . . . , θJ(t)]T is the output vector of the hidden layer, y(t) represents the output of recurrent RBF neural network at time t.
The error of recurrent RBF neural network is:
yd(t) is the expectation output of neural network and the actual output is expressed as y(t).
(3) Train Recurrent RBF Neural Network:
{circle around (1)} Given the recurrent RBF neural network, the initial number of hidden layer neurons is J, J>2 is a positive integer. The input of recurrent RBF neural network is x(1), x(2), . . . , x(t), . . . , x(N), correspondingly, the expectation output is yd(1), yd(2), . . . , yd(t), . . . , yd(N), expected error value is set to Ed, Ed∈(0, 0.01). The every variable of initial centre value cj(1)∈(−2, 2), width value σj(1)∈(0, 1), initial feedback weight vj(1)∈(0, 1), j=1, 2, . . . , J. Initial weight w(1)∈(0, 1).
{circle around (2)} Set the learning step s=1;
{circle around (3)} t=s; According to Eqs. (1)-(4), calculate the output of recurrent RBF neural network, exploiting fast gradient descent algorithm:
ηc, ησ, ηv, ηw are the learning rate of centre, width, feedback connection weight from output layer to hidden layer and the connection weight between hidden layer and output layer, respectively. In addition, ηc∈(0, 0.01], ησ∈(0, 0.01], ηv∈(0, 0.02], ηw∈(0, 0.01]. cj(t+1)=[c1j(t+1), c2j(t+1), . . . , c5j(t+1)] denotes the center vector of the jth hidden neuron at time t+1. σj(t+1) is the radius or width of the jth hidden neuron at time t+1. vj(t+1) denotes the connection weight from output layer to the jth hidden neuron at time t+1. wj(t+1) is the connection weights between the hidden neurons and output neuron at time t+1.
{circle around (4)} increase 1 learning step s, if s<N, then turn to step {circle around (3)}. If s=N, turn to step {circle around (5)}.
{circle around (5)} according to Eq. (5), calculate the performance of recurrent RBF neural network. If E(t)≥Ed, then turn to step {circle around (3)}; If E(t)<Ed, stop the training process.
(4) SVI Concentration Prediction:
The testing samples used as the input of recurrent RBF neural network, the output of neural network is the soft-computing values of SVI. The testing result is shown in
(5) CVI Algorithm for Sludge Bulking:
{circle around (1)} Calculate the residual of the expectation output and the output of recurrent RBF neural network, if
then turn to step {circle around (2)}, otherwise, stop the process of fault identification for sludge bulking.
{circle around (2)} Define two formula:
wherein IC1(t) is the Mahalanobis distance of input variables at time t and IC2(t) is the squared prediction error at time t, M is the numbers of principal component of input data. K is the number of the front samples, θM(t) is the output vector of the hidden layer for M principal components training data at time t. δ(t)=[δ1(t), . . . , δm(t), . . . , δM(t)]T is the projection of the training data, and δm(t) is
wherein
wherein ΛM(t) is the diagonal matrix of eigenvalues at time t and it satisfies
{tilde over (C)}(t)=Z(t)ΛM(t)ZT(t)+l′(t)(I(t)−Z(t)Zt(t)) (42)
wherein l′(t) is a constant value, I(t) is a unit matrix, {tilde over (C)}(t) is the regularized covariance matrix of C(t) at time t:
wherein Ω(t) is the energy of each variable. θk(t) is the hidden output vector of the kth principal component at time t.
The constant vector a(t)=[a1(t), . . . aj(t), . . . aK(t)]T is given as
G(t) is the Gaussian matrix and λ(t) denotes the eigenvalue
G(t)={θi(t)·θj(t)}K×K (47)
λ(t)p(t)=C(t)p(t) (48)
p(t) denotes eigenvector of the covariance matrix C(t) at time t.
{circle around (3)} For the ith input variable, the contribution degree index satisfy:
wherein Gi(t) is the contribution degree index of ith variable at time t, a hidden layer neuron corresponds to an input variable, κi(t) is the contribution degree which is calculate by the mutual information between this testing samples and difference sets for testing samples xi(t) of ith variable at time t, which can be expressed as
κi(t)=I(xi(t),VΔ(t)) (50)
I(xi(t), VΔ(t)) is the mutual information of xi(t) and VΔ(t) at time t, VΔ(t) is the difference matrix sets of data of training set and testing set at time t, which is expressed as
VΔ(t)=Vtr(t)−Vte(t) (51)
wherein Vtr(t) and Vte(t) is independent data sets of training set and testing set at time t, respectively.
Vtr(t)=D−1(t)G(t) (52)
Vte(t)=Dte−1(t)Gte(t) (53)
D(t) is the covariance matrix of Φ(t) at time t
D(t)=E{Φ(t)ΦT(t)} (54)
Φ(t)=[(t−K+1), . . . ,θ(t−1),θ(t)]T (55)
Φ(t) is output matrix of hidden layer at time t, θ(t−K+1) is the hidden output vector at time t−K+1, K is the number of the front samples.
{circle around (4)} For the ith input variable, a hidden layer neuron corresponds to an input variable, if the Gi(t) at time t satisfies:
G1(t)+ . . . Gi(t)≥0.8 (56)
the variables 1, . . . , i is the cause variables resulted in sludge bulking.
The IC1(t) outputs is shown in
Tables 1-12 show the experimental data in this present disclosure. Tables 1-5 show the training samples of COD, DO, T, MLSS and TN. Tables 6-10 show the testing samples of COD, DO, T, MLSS and TN. Table 11 shows real output values of SVI. Table 12 shows the outputs of the recurrent RBF neural network in the predicting process. Moreover, the samples are imported as the sequence from the tables. The first data is in the first row and the first column. Then, the second data is in the first row and the second column. Until all of data is imported from the first row, the data in the second row and following rows are inputted as the same way.
Training samples are provided as follows:
Number | Date | Country | Kind |
---|---|---|---|
201710186738.5 | Mar 2017 | CN | national |
Entry |
---|
Hong-Gui Han and Jun-Fei Qiao, “Prediction of activated sludge bulking based on a self-organizing RBF neural network”, Available online Apr. 27, 2012, Journal of Process Control 22 (2012), pp. 1103-1112. (Year: 2012). |
Hong-Gui Han, Ying Li, Ya-Nan Guo, and Jun-Fei Qiao, “A soft computing method to predict sludge volume index based on a recurrent self-organizing neural network”, Available online Oct. 19, 2015, Applied Soft Computing 38 (2016), pp. 477-486. (Year: 2015). |
Number | Date | Country | |
---|---|---|---|
20180276531 A1 | Sep 2018 | US |