This application claims priority to Chinese Patent Application No. 201610606146.X, filed on Jul. 28, 2016, entitled “a method for effluent total nitrogen based on a recurrent self-organizing RBF neural network,” which is hereby incorporated by reference in its entirety.
In this present disclosure, a computing implemented method is designed for predicting the effluent total nitrogen TN concentration (TN) in the urban wastewater treatment process (WWTP) by a recurrent self-organizing RBF neural network (RSORBFNN). To improve the measurement efficiency, the RSORBFNN can adjust the structure and parameters concurrently: a growing and pruning algorithm is proposed to design the structure, and an adaptive second-order algorithm is utilized to train the parameters. The technology of this present disclosure is part of advanced manufacturing technology and belongs to both the field of control engineer and environment engineer.
The urban WWTP not only guarantees the reliability and stability of the wastewater treatment system but also meets the water quality national discharge standard. However, the influence factors are various for effluent TN of wastewater treatment process and the relationship between different influencing factors are complex. Therefore, it is hard to make real-time detecting for effluent TN, which seriously affected the stable operation of the urban WWTP. The computing implemented a method for effluent TN, based on RSORBFNN, is helpful to improve the efficiency, strengthen delicacy management and ensure water quality effluent standards of urban WWTP. It has better economic benefit as well as significant environmental and social benefits. Thus, the research achievements have wide application prospect in this present disclosure.
The control target of urban WWTP is to make the water quality meet the national discharge standards, mainly related to the parameters of effluent TN, chemical oxygen demand (COD), effluent suspended solids (SS), ammonia nitrogen (NH4-N), biochemical oxygen demand (BOD) and effluent total phosphorus (TP). Effluent TN refers to the sum of all the nitrogen pollution of the water after dealing with the sewage treatment plant process facilities, mainly for the ammonia nitrogen, nitrate nitrogen, inorganic nitrogen, protein, amino acid and organic amine organic nitrogen combined. According to statistics, nitrogen fixation rate of about 150 million tons per year in nature and chemical nitrogen fertilizer production rate of about 5000˜6000 tons a year. If nature denitrification reaction failed to complete the nitrogen cycle, too much nitrogen compounds and the ammonia nitrogen nutrient caused a significant number of algae in the water, the plants breeding, appearance of eutrophication status. To curb the trend of worsening of water environment, many sewage treatment facilities have spent a large sum of money to build and put into operation in the country, the cities, and towns. The general method for determination is the alkaline potassium persulfate UV spectrophotometry and molecular absorption spectrometry. However, the determination of total nitrogen TN is often offline and can't realize the effluent TN real-time measurement, which led directly to the sewage treatment process is hard to achieve closed loop control. Moreover, it is a big challenge for detection due to a significant amount of pollutants in wastewater and different content. Developing new hardware measuring instrument, although directly solving various wastewater treatment process variables and the detection problem of water quality parameters, due to the very complex organic matter in sewage, research and development of the new sensor will be a significant cost and a time-consuming project. Hence, the new method presented to solve the problem of the real-time measurement of the process parameters of WWTP has become an important topic to research in the field of wastewater control engineering and has important practical significance.
To obtain more reliable information on effluent TN in urban WWTP, we have investigated a computing implemented method based on the RSORBFNN. The neural network uses competitiveness of the hidden neuron to determine whether to add or delete the hidden neurons and to use an adaptive second order algorithm to ensure the accuracy of RSORBFNN. The objective of this present disclosure is to develop a computing implemented method for estimating the effluent TN online and with high precision.
A computing implemented method is designed for the effluent TN based on an RSORBFNN in this present disclosure. For this computing implemented method, the inputs are those variables that are easy to measure and the outputs are estimates of the effluent TN. By constructing the RSORBFNN, it realizes the mapping between auxiliary variables and effluent TN. Also, the method can obtain a real-time measurement of effluent TN, solve the problems of long measurement cycle for effluent TN.
A computing implemented method for the effluent TN based on an RSORBFNN, its characteristic and steps include the following steps:
(1) Determine the input and output variables of effluent TN:
For sewage treatment process of activated sludge system, the variables of sewage treatment process are analyzed and select the input variables of effluent TN soft-computing model: ammonia nitrogen —NH4—N, nitrate nitrogen —NO3—N, effluent suspended solids—SS, biochemical oxygen demand—BOD, total phosphorus—TP, The output value of soft-computing model is detected effluent TN.
(2) Initialize RSORBFNN
The initial structure of RSORBFNN consists of three layers: an input layer, hidden layer, and an output layer. There are 5 neurons in the input layer, J neurons in the hidden layer and 1 neuron in the output layer; J>2 is a positive integer. Connection weights between input layer and hidden layer are assigned 1, the feedback weights between hidden layer and output layer randomly assign values, the assignment internal is [1, 1]; the number of the training sample is P, and the input vector of RSORBFNN is x(t)=[x1(t), x2(t), x3(t), x4(t), x5(t)] at time t; y(t) is the output of RSORBFNN, and yd(t) is the real value of effluent TN at time t, respectively; The output of RSORBFNN can be described:
wherein w2 j(t) is the output weight between the jth hidden neuron and the output neuron, w2(t)=[w2 1(t), w2 2(t), . . . , w2 J(t)]T is the output weight vector between hidden neurons and output neuron, j=1, 2, . . . , J, J is the number of hidden neurons, and θj(t) is the output value of the jth hidden neuron which is usually defined by a normalized Gaussian function:
wherein ||hj-cj|| represents the Euclidean distance between hj and cj, cj(t)=[c1j(t), c2j(t), . . . , c5j(t)]T and σj represent the center vector and radius of the jth hidden neuron, respectively; cij(t) is ith element of jth hidden neuron, and hj is the input vector of jth hidden neuron
h
j(t)=[hj1(t), hj2(t), . . . , hj1(t)], (3)
h
ij(t)=uj(t), (4)
wherein w1j(t) is the feedback weight connecting the jth hidden neuron with the output neuron, w1(t)=[w1 1(t), w1 2(t), . . . , w1 J(t)]T is the feedback weight vector connecting the jth hidden neuron with the output neuron and y(t-1) is the output value of RSORBFNN at time t-1.
The training error function of RSORBFNN is defined
wherein P is the number of the training samples.
(3) Train RSORBFNN
1) Given RSORBFNN, the initial number of hidden layer neurons is J; J>2 is a positive integer. The input of RSORBFNN is x(1), x(2), . . . , x(t), . . . , x(P), the desired output is yd(1), yd(2), . . . , yd(t), . . . , yd(P); the desired error value is set to Ed, Ed∈(0, 0.01), the initial center is cj(1)∈(−2, 2), the initial width value σj(1)∈(0, 1), the initial feedback weight is w1 j(1)∈(0, 1), and the initial weight is w2 j(1)∈(0, 1), j=1, 2, . . . , J;
2) Set the learning step s=1;
3) t=s, calculate the output y(t) of RSORBFNN, update the weight, width, and center of RSORBFNN using the rule:
Θ(t+1)=Θ(t)+(Ψ(t)+η(t)×I)−1×Ω(t), (6)
where Θ(t)=[w1(t), w2(t), C(t), σ(t)] is the variable vector at time t, Ψ(t) is quasi
Hessian matrix at time t, I is the identity matrix, η(t) is the adaptive learning rate defined as:
wherein μ(t) is the adapting factor at time t, and the initial value of μ(t) is μ(1)=1, βmax(t) and βmin(t) are the maximum and minimum eigenvalues of Ψ(t), respectively; 0<βmin(t)<βmax(t), 0<η(t)<1 and η(1)=1. Θ(t) contains four kinds of variables: the feedback connection weight vector w1(t) at time t, the connection weight vector w2(t) at time t, the centre matrix C(t)=[c1(t), c2(t), . . . , cj(t)]T and width vector σ(t)=[σ1(t), σ2(t), . . . , σj(t)]T at time t.
Θ(1)=[w1(1), w2(1), C(1), σ(1)], (9)
the quasi Hessian matrix Ψ(t) and the gradient vector Ω(t) are accumulated as the sum of related submatrices and vectors:
Ψ(t)=jT(t)j(t), (10)
Ω(t)=jTe(t), (11)
e(t)=yd(t)−y(t), (12)
e(t) is the approximating error at time t, yd(t) is the desired output and y(t) is the network output at time t, and the Jacobian-vector j(t) is calculated as:
4) t>3, calculate competitiveness of the jth hidden neuron:
wherein cpj(t) is the competitiveness of the jth hidden neuron, ρ denotes the correlation coefficient between the hidden layer output and network output, ρ∈(0, 1), fj(t) is the active state of the jth hidden neuron, σj(t) is the width of the jth hidden neuron; the active state fj(t) is defined as
fj(t)=χ−|j(t)−cj(t)|, (15)
wherein χ∈(1,2), and f(t)=[f1(t), f2(t), . . . , fj(t)], the correlation coefficient ρj(t) at time t is calculated as
wherein the correlation coefficient of hidden neurons Aj(t)=w2 j(t)θj (t), the correlation coefficient of output layer B(t)=y(t), Ā(t) is the average value of correlation coefficient of hidden neurons at time t,
5) Adjust the structure of RSORBFNN:
If the competitiveness of the jth hidden neuron and training error at time t and t+τ satisfy
denotes the value of j when cpj(t) owns the maximum value. E(t) and E(t+τ) are the training errors at times t and t+τ, respectively, τ is a time interval, τ>2, and ε is the preset threshold, ε∈(0,0.01). Add one hidden neuron, and the number of hidden neurons is M1=J+1. Otherwise, the structure of RSORBFNN will be not adjusted, M1=J.
When the competitiveness of the jth hidden neuron satisfies
cp
j(t)<ξ, (19)
wherein ξ is the preset pruning threshold, ξ∈(0, Ed), Ed is the preset error, Ed∈(0,0.01]. The jth hidden neuron will be pruned, the number of hidden neurons will be updated M2=M1−1. Otherwise, the structure of RSORBFNN will be not adjusted, M2=M1.
6) Increase 1 learning step for s, if s<P, go to step 3); if s=N, proceed to step 7).
7) Per Eq. (5), calculate the performance of RSORBFNN. If E(t)≧Ed, proceed to step 3); if E(t)<Ed, stop the training process.
(4) Effluent TN concentration prediction;
The testing samples are used as the input of RSORBFNN, the output of RSORBFNN is the soft-computing values of effluent TN.
The Novelties of this Present Disclosure Contain:
(1) To detect the effluent TN online and with acceptable accuracy, a computing implemented method is developed in this present disclosure. The results demonstrate that the effluent TN trends in WWTP can be predicted with acceptable accuracy using the NH4—N, NO3—N, effluent SS, BOD, TP as input variables. This computing implemented method can predict the effluent TN with acceptable accuracy and solve the problem that the effluent TN ‘s hard to be measured online.
(2) Since wastewater treatment process has the features of a complicated mechanism, and many influential factors, it was difficult to build a precise mathematical model to predict the effluent TN. Hence, the computing implemented method is based on the RSORBFNN in this present disclosure, which is proposed to predict it. The advantages of the proposed RSORBFNN are that it can simplify and accelerate the structure optimization process of the recurrent neural network, and can predict the effluent TN accurately. Moreover, the predicting performance shows that the RSORBFNN-based computing implemented method can adapt well to environment change. Therefore, this computing implemented method performs well in the whole operating space.
Attention: this present disclosure utilizes five input variables in this RSORBFNN method to predict the effluent TN. In fact, it is in the scope of this present disclosure that any of the variables: the NH4—N, NO3—N, effluent SS, BOD, TP are used to predict the effluent TN. Moreover, this RSORBFNN method is also able to predict the others variables in urban WWTP.
Various A computing implemented method is developed to predict the effluent TN based on an RSORBFNN in this present disclosure. For this computing implemented method, the inputs are those variables that are easy to measure and the outputs are estimates of the effluent TN. In general, the procedure of computing implemented method consists of three parts: data acquisition, data pre-processing and model design. For this present disclosure, an experimental hardware is set up as shown in
This present disclosure adopts the following technical scheme and implementation steps:
A computing implemented method for the effluent TN based on an RSORBFNN, its characteristic and steps include the following steps:
(1) Determine the input and output variables of effluent TN:
For sewage treatment process of activated sludge system, the variables of sewage treatment process are analyzed and select the input variables of effluent TN soft-computing model: ammonia nitrogen—NH4—N, nitrate nitrogen—NO3—N, effluent suspended solids—SS, biochemical oxygen demand—BOD, total phosphorus—TP, The output value of soft-computing model is detected effluent TN.
(2) Initialize RSORBFNN
The initial structure of RSORBFNN consists of three layers: input layer, hidden layer, and output layer. There are 5 neurons in the input layer, J neurons in the hidden layer and 1 neuron in the output layer; J>2 is a positive integer. Connection weights between input layer and hidden layer are assigned 1, the feedback weights between hidden layer and output layer randomly assign values, the assignment internal is [1, 1]; the number of the training sample is P, and the input vector of RSORBFNN is x(t)=[x1(t), x2(t), x3(t), x4(t), x5(t)] at time t; y(t) is the output of RSORBFNN, and yd(t) is the real value of effluent TN at time t, respectively; The output of RSORBFNN can be described:
wherein w2 j(t) is the output weight between the jth hidden neuron and the output neuron, w2(t)=[w2 1(t), w2 2(t), . . . , w2 J(t)]T is the output weight vector between hidden neurons and output neuron, j=1, 2, . . . , J, J is the number of hidden neurons, and θj(t) is the output value of the jth hidden neuron which is usually defined by a normalized Gaussian function:
wherein ||hj-cj|| represents the Euclidean distance between hj and cj, cj(t)=[c1j(t), c2j(t), . . . , c5j(t)]T and σj represent the center vector and radius of the jth hidden neuron, respectively; cij(t) is ith element of jth hidden neuron, and hj is the input vector of jth hidden neuron
h
j(t)=[hj1(t), hj2(t), . . . , hj1(t)], (22)
h
ij(t)=ui(t), (23)
wherein w1 j(t) is the feedback weight connecting the jth hidden neuron with the output neuron, w1(t)=[w1 1(t), w1 2(t), . . . , w1 J(t)]T is the feedback weight vector connecting the jth hidden neuron with the output neuron and y(t-1) is the output value of RSORBFNN at time t-1.
The training error function of RSORBFNN is defined
wherein P is the number of the training samples.
(3) Train RSORBFNN
1) Given RSORBFNN, the initial number of hidden layer neurons isJ;J>2 is a positive integer. The input of RSORBFNN is x(1), x(2), . . . , x(t), . . . , x(P), the desired output is yd(1), yd(2), . . . , yd(t), . . . , yd(P); the desired error value is set to Ed, Ed∈(0, 0.01), the initial center is CJ(1)∈(−2, 2), the initial width value σj(1) ∈(0, 1), the initial feedback weight is w1 j(1)∈(0, 1), and the initial weight is w2 j(1)∈(0, 1), j=1, 2, . . . , J;
2) Set the learning step s=1;
3) t=s, calculate the output y(t) of RSORBFNN, update the weight, width, and center of RSORBFNN using the rule:
Θ(t+1)=Θ(t)+(Ψ(t)+η(t)×I)−1×Ω(t), (25)
where Θ(t+1)=[w1(t), w2(t), C(t), σ(t)[ is the variable vector at time t, Ψ(t) is quasi
Hessian matrix at time t, I is the identity matrix, η(t) is the adaptive learning rate defined as:
wherein μ(t) is the adapting factor at time t, and the initial value of μ(t) is μ(1)=1, βmax(t) and βmin(t) are the maximum and minimum eigenvalues of Ψ(t), respectively; 0<βmin(t)<βmax(t), 0<η(t)<1 and η(1)=1. Θ(t) contains four kinds of variables: the feedback connection weight vector w1(t) at time t, the connection weight vector w2(t) at time t, the centre matrix C(t)=[c1(t), c2(t), . . . , cj(t)]T and width vector σ(t)=[σ1(t), σ2(t), . . . , σj(t)]T at time t.
Θ(1)=[w1(1), w2(1), C(1), σ(1)], (28)
the quasi Hessian matrix Ψ(t) and the gradient vector Ω(t) are accumulated as the sum of related submatrices and vectors:
Ψ(t)=jT(t)j(t), (29)
Ω(t)=jTe(t), (30)
e(t)=yd(t)−y(t), (31)
e(t) is the approximating error at time t, yd(t) is the desired output and y(t) is the network output at time t, and the Jacobian-vector j(t) is calculated as:
4) t>3, calculate competitiveness of the jth hidden neuron:
cp
j(t)=ρfj(t)σj(t), j=1, 2, . . . ,J, (33)
wherein cpj(t) is the competitiveness of the jth hidden neuron, ρ denotes the correlation coefficient between the hidden layer output and network output, ρ∈(0, 1),fi(t) is the active state of the jth hidden neuron, σj(t) is the width of the jth hidden neuron; the active state fj(t) is defined as
wherein χ∈(1,2), and f(t)=[f1(t), f2(t), . . . , fj(t)], the correlation coefficient AM at time t is calculated as
wherein the correlation coefficient of hidden neurons Aj(t)=w2 j(t) θj(t), the correlation coefficient of output layer B(t)=y(t), Ā(t) is the average value of correlation coefficient of hidden neurons at time t,
5) Adjust the structure of RSORBFNN:
If the competitiveness of the jth hidden neuron and training error at time t and t+τ satisfy
denotes the value of jwhen cpj(t) obtain the maximum value. E(t) and E(t+τ) are the training errors at times t and t+τ, respectively, τ is a time interval, τ=5, and ε is the preset threshold, ε=0.001. Add one hidden neuron, and the number of hidden neurons is M1=J+1. Otherwise, the structure of RSORBFNN will be not adjusted, M1=J.
When the competitiveness of the jth hidden neuron satisfies
cp
j(t)<ξ, (38)
wherein ξ is the preset pruning threshold, ξ∈(0, Ed), Ed is the preset error, Ed=0.002. The jth hidden neuron will be pruned, the number of hidden neurons will be updated M2=M1−1. Otherwise, the structure of RSORBFNN will be not adjusted, M2=M1.
6) Increase 1 learning step for s, if s<P, go to step 3); if s=N, proceed to step 7).
7) According to Eq. (24), calculate the performance of RSORBFNN. If E(t)≧Ed, proceed to step 3); if E(t)<Ed, stop the training process.
The training result of the computing implemented method for effluent TN is shown in
(4) Effluent TN concentration prediction;
The testing samples are used as the input of RSORBFNN, and the output of RSORBFNN is the soft-computing values of effluent TN. The predicting result is shown in
Tables 1-14 show the experimental data in this present disclosure. Tables 1-6 show the training samples of biochemical oxygen demand - BOD, ammonia nitrogen—NH4—N, nitrate nitrogen—NO3—N, effluent suspended solids —SS, total phosphorus—TP real effluent TN. Table 7 shows the outputs of the RSORBFNN in the training process. Tables 8-14 show the testing samples of biochemical oxygen demand—BOD, ammonia nitrogen—NH4—N, nitrate nitrogen —NO3—N, effluent suspended solids—SS, total phosphorus—TP and real effluent TN. Table 14 shows the outputs of the RSORBFNN in the predicting process
Training samples are provided as follow.
Number | Date | Country | Kind |
---|---|---|---|
201610606146.X | Jul 2016 | CN | national |