CN111027776A - Sewage treatment water quality prediction method based on improved long-short term memory LSTM neural network - Google Patents

Sewage treatment water quality prediction method based on improved long-short term memory LSTM neural network Download PDF

Info

Publication number
CN111027776A
CN111027776A CN201911279211.2A CN201911279211A CN111027776A CN 111027776 A CN111027776 A CN 111027776A CN 201911279211 A CN201911279211 A CN 201911279211A CN 111027776 A CN111027776 A CN 111027776A
Authority
CN
China
Prior art keywords
data
network
lstm
layer
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911279211.2A
Other languages
Chinese (zh)
Inventor
刘晔晖
冯骁
夏文泽
曲晓川
王喆
钱志明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Huazhan Huiyuan Information Technology Co Ltd
Original Assignee
Beijing Huazhan Huiyuan Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Huazhan Huiyuan Information Technology Co Ltd filed Critical Beijing Huazhan Huiyuan Information Technology Co Ltd
Priority to CN201911279211.2A priority Critical patent/CN111027776A/en
Publication of CN111027776A publication Critical patent/CN111027776A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2135Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/26Government or public services
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A20/00Water conservation; Efficient water supply; Efficient water use
    • Y02A20/152Water filtration

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Resources & Organizations (AREA)
  • Tourism & Hospitality (AREA)
  • Molecular Biology (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Computing Systems (AREA)
  • Strategic Management (AREA)
  • Biomedical Technology (AREA)
  • Economics (AREA)
  • Biophysics (AREA)
  • General Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Development Economics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Primary Health Care (AREA)
  • Educational Administration (AREA)
  • Evolutionary Biology (AREA)
  • Game Theory and Decision Science (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Feedback Control In General (AREA)

Abstract

The invention discloses a sewage treatment water quality prediction method based on an improved long-short term memory (LSTM) neural network, which comprises the following steps of: 1) processing data; 2) establishing and training an improved long-short term memory (LSTM) neural network prediction model; 3) and importing the data to be predicted into an overall improved long-short term memory (LSTM) neural network prediction model, and outputting to obtain a predicted water quality value. The invention firstly effectively cleans and combs the data source, removes interference factors and noise points and clarifies the relation among data, and makes early work for data analysis. During data analysis, the improved LSTM network can well analyze large-time-lag and strong-coupling data, can accurately map a numerical relation of a long time span, and provides great convenience for sewage quality prediction.

Description

Sewage treatment water quality prediction method based on improved long-short term memory LSTM neural network
Technical Field
The invention belongs to the field of sewage treatment, and particularly relates to a sewage treatment water quality prediction method based on an improved long-short term memory (LSTM) neural network.
Background
A Recurrent Neural Network (RNN) is a Neural Network used to process sequence data. Compared with a general neural network, the neural network can process data with sequence variation. For example, if the meaning of a word is different because of the difference in the above-mentioned contents, the RNN can solve such a problem well. The RNN has the characteristics that the RNN can not only operate the space dimension, but also operate the time sequence dimension, the structure characteristic of the RNN is that one-step feedback operation is added on a hidden layer, and the structure can enable the input content of the RNN to not only contain the data of the input layer, but also contain the data fed back by the hidden layer at the last moment, and the RNN is also a main reason for the time sequence characteristic. However, the structure has a fatal problem, namely the problem of abnormal gradient of the structure is accompanied; since the application of the self-feedback loop accelerates the occurrence of the gradient vanishing phenomenon, and in order to avoid this phenomenon, a Long Short Term Memory Network (LSTM) was developed.
The LSTM network can effectively solve the problem that the RNN network has gradient disappearance caused by network defects, namely, long-time content cannot be recorded, and the LSTM network realizes the characteristic of recording content for a long time by adding a forgetting gate and improved structures such as an input gate, an updating gate and an output gate. However, even if the characteristics exist, the method cannot be applied to sewage treatment analysis because the sewage treatment process is a large-scale flow industrial process and has the characteristics of large time variation, strong coupling, large time lag, serious interference and the like; meanwhile, biochemical reaction in the sewage treatment process is complex, and a lot of variables are involved in the operation process, so that the large-time-span data can be greatly memorized when the data is analyzed, and at the moment, the LSTM can not meet the application requirement, so that the LSTM neural network needs to be improved and adjusted to meet the requirement of data analysis in the sewage treatment process.
Disclosure of Invention
Aiming at the defects of the existing sewage treatment process characteristics and the existing sewage treatment water quality analysis method, the invention aims to provide a high-precision analysis method based on neural network analysis, namely a sewage treatment water quality prediction method based on an improved long-short term memory (LSTM) neural network.
In order to achieve the purpose, the invention adopts the technical scheme that:
a sewage treatment water quality prediction method based on an improved long-short term memory (LSTM) neural network comprises the following steps:
1) data processing:
preprocessing the sewage treatment reported data, performing PCA (principal component analysis) dimensionality reduction on the preprocessed data to obtain result data, selecting a main parameter corresponding relation of the result data by utilizing a decision tree and pruning operation, and establishing a parameter mapping relation table;
2) establishing and training an improved long-short term memory (LSTM) neural network prediction model:
determining input and output parameters of the improved long-short term memory LSTM neural network according to the parameter mapping relation table, establishing an improved long-short term memory LSTM neural network prediction model, and training to obtain an overall improved long-short term memory LSTM neural network prediction model;
the improved long-short term memory LSTM neural network comprises an input layer, a front-end network parallel layer, an LSTM network layer, a rear-end network serial layer and an output layer;
3) and importing the data to be predicted into an overall improved long-short term memory (LSTM) neural network prediction model, and outputting to obtain a predicted water quality value.
Preferably, the preprocessing is to remove data outside the unreasonable limit value in the sewage treatment reported data, then perform 3-time standard deviation processing on the reported data, remove data with the difference between the maximum value or the minimum value of the reported data and the mean value exceeding 3-time standard deviation, perform integral dimensionality reduction operation on the sorted data, and remove redundant information in the data to obtain the result data.
Preferably, an input layer in the LSTM improved neural network model is a docking interface of a data parametric input network, and a data set X is divided into n parts according to a data inclusion time length, that is:
X=[x(1),x(2),...,x(n)]
i.e. a data of a long time span is divided into n data of a relatively short time span, and it is noted that the value of the fraction n that the network divides the data set is larger as the time span of the data set is longer.
Preferably, the front-end network parallel layer includes a parallel form of multiple groups of fully-connected neural networks and dropout layers, and the weight value between a network node i and a network node j in the multiple groups of fully-connected neural networks is wijThe threshold value of the node j is bjThe output value of each node is xjThe output value of each node is determined by the activation function f according to the output values of all nodes on the upper layer, the weights and biases of the current node and all nodes on the upper layer, and the effect is as follows:
Figure BDA0002316277160000021
xj=f(Sj)
Figure BDA0002316277160000031
Figure BDA0002316277160000032
after the output of the fully-connected network is finished, the method enters a dropout network hierarchical structure for processing, and the method specifically comprises the following steps:
Figure BDA0002316277160000033
Figure BDA0002316277160000034
the Bernoulli function in the above formula randomly generates a vector of 0 and 1 with probability p.
Preferably, the LSTM network layer is a main data processing layer in the whole improved LSTM neural network and is composed of a plurality of LSTM networks connected in parallel, wherein the current input xtAnd c passed down from the previous statet-1、ht-1The common determination is as follows:
Figure BDA0002316277160000035
Figure BDA0002316277160000036
Figure BDA0002316277160000037
Figure BDA0002316277160000038
wherein z isf,zi,z0After the splicing vector is multiplied by the weight matrix, the splicing vector is converted into a value between 0 and 1 through a sigmoid activation function to serve as a gating state, and z is the value between-1 and 1 obtained by converting the result through a tanh activation function:
Figure BDA0002316277160000039
Figure BDA00023162771600000310
yt=δ(W'ht)
yseq=yt
Figure BDA00023162771600000311
is the multiplication of the corresponding elements in the matrix,
Figure BDA00023162771600000312
matrix addition is performed.
Preferably, the back-end network serial layer includes multiple sets of fully-connected neural networks and dropout layers, and after receiving the LSTM network layer result, the back-end network serial layer needs to accumulate and sum the parallel results to obtain a unique solution, and then sends the unique solution to the back fully-connected and dropout network structure layer for processing, and the specific effects are as follows:
Figure BDA00023162771600000313
the method firstly carries out numerical value limit processing on the reported data of the sewage treatment, and removes the data outside the unreasonable limit. It should be noted that, since the sewage data are reported in a time logic sequence, the data under the same timestamp represent the water quality condition at that time, that is, when an index in the water quality is removed due to a failure, other data at that time should also be removed, so as to ensure the accuracy of the time logic of the data. After the limit value processing, 3 times of standard deviation processing is needed to be carried out on the data, because the distribution condition of the reported data of the sewage treatment is just too distributed, the probability of existence of the numerical value exceeding 3 times of standard deviation is about 0.3%, and when the difference between the maximum value (or the minimum value) and the mean value of the reported data exceeds 3 times of standard deviation, the extreme values with problems in the maximum value (or the minimum value) and the mean value are likely to have problems, so that the extreme values with problems in the maximum value (or the minimum value) and the mean value are removed to.
The sewage treatment process has the characteristics of large time variation, strong coupling, large time lag, serious interference, complex biochemical reaction and the like, so that the reported data quality is poor, the water quality prediction difficulty is high, and the problems can be well overcome by the method. The method firstly effectively cleans and combs the data source, removes interference factors and noise points, and makes clear the relation between data, so as to make early work for data analysis. During data analysis, the improved LSTM network can well analyze large-time-lag and strong-coupling data, can accurately map a numerical relation of a long time span, and provides great convenience for sewage quality prediction.
The invention has the beneficial effects that:
1) the improved long-short term memory LSTM network can well analyze large-time-lag and strong-coupling data, can accurately map the numerical relationship of a long time span, and provides great convenience for sewage quality prediction;
2) the invention has high analysis and prediction precision.
Drawings
FIG. 1 is a diagram of a sewage prediction method architecture;
FIG. 2 is a decision tree relationship resolution graph, which is a tree form graph after pruning, a decision tree overall graph and a corresponding graph of error effect after pruning;
FIG. 3 is a diagram of an improved Long Short Term Memory (LSTM) neural network;
FIG. 4 wastewater quality NH4And a COD prediction effect graph.
Detailed Description
In order to further illustrate the technical effects of the present invention, the present invention is specifically described below by way of examples.
Example 1
The invention discloses a sewage treatment water quality prediction method based on an improved long-short term memory (LSTM) neural network, which is operated as shown in figure 1, and has two major aspects and five parts in total. Taking the CAST sewage treatment process as an example, the CAST sewage treatment process reports a great deal of data, such as inflow rate, inflow PH value, inflow COD, inflow ammonia nitrogen, inflow temperature, CAST pool ss, CAST pool do value, CAST pool total air volume, CAST pool fan current total amount, output water quality ammonia nitrogen, output water quality ss, output water quality COD, output water quality total phosphorus, output water quality total nitrogen and the like. The overall pretreatment is carried out on the data reported by the sewage treatment systems, and the specific operation steps are as follows.
Step 1: firstly, numerical limit value processing is carried out on reported data of sewage treatment, and data outside unreasonable limit values are removed. It should be noted that, since the sewage data are reported in a time logic sequence, the data under the same timestamp represent the water quality condition at that time, that is, when an index in the water quality is removed due to a failure, other data at that time should also be removed, so as to ensure the accuracy of the time logic of the data.
Step 2: after the limit value processing, 3 times of standard deviation processing is needed to be carried out on the data, because the distribution condition of the reported data of the sewage treatment is just too distributed, the probability of existence of the numerical value exceeding 3 times of standard deviation is about 0.3%, and when the difference between the maximum value (or the minimum value) and the mean value of the reported data exceeds 3 times of standard deviation, the extreme values with problems in the maximum value (or the minimum value) and the mean value are likely to have problems, so that the extreme values with problems in the maximum value (or the minimum value) and the mean value are removed to.
And step 3: performing PCA (principal component analysis) dimension reduction on the processed data, and removing redundant information in the data, wherein the dimension reduction method is to establish a data matrix X firstly as follows:
Figure BDA0002316277160000051
firstly, carrying out standardization processing on an X matrix:
Figure RE-GDA0002388583210000052
wherein the content of the first and second substances,
Figure BDA0002316277160000053
then, a correlation coefficient matrix is calculated, and the method is as follows:
Figure BDA0002316277160000054
Figure BDA0002316277160000061
computingEigenvalues (λ) of the correlation coefficient matrix R12,., λ p) and the corresponding feature vector:
ai=(ai1,ai2,Λ,aip),i=1,2,Λ,p
selecting important principal components, selecting the first k principal components according to 95% of the actual total information amount, removing 5% of small probability event components, selecting a proportion probability that a certain characteristic value accounts for the total of all the characteristic values, and obtaining a matrix after PCA dimension reduction by using the remaining k principal components.
And 4, step 4: after data processing is finished, the causal relationship among the data also needs to be combed, and due to the sewage process, certain result water quality indexes are the factors of other result parameters, so that the clear causal relationship of the data is obtained. And selecting the main parameter corresponding relation of the result data by utilizing the decision tree and the pruning operation, and establishing a parameter mapping relation table.
Before the decision tree is operated, the result parameters serving as the classification need to be clustered, that is, the result parameters are clustered into 5 classes according to the value domain distribution condition.
And classifying the data by using the information gain rate, and solving the probability of the result which can be covered by classification, wherein the information gain rate formula is solved by the information gain and the splitting information quantity, and the following steps are shown:
Figure BDA0002316277160000062
wherein the information gain G (S, a) is defined as:
Figure BDA0002316277160000063
Figure BDA0002316277160000064
e (S) is the entropy of the data set S, PiIs the proportion of the number of samples of the ith attribute value in the subset, SvThe feature A in the set S takes the v-th value to obtain a sample subset.
The split information quantity is defined as:
Figure BDA0002316277160000071
Siis a subset of samples in S that belong to the i-th class.
Finally, the post pruning operation of the decision tree is utilized to delete the node branches and replace the node branches with leaf sub-nodes, the leaf nodes are generally marked by the maximum probability category replacement in the subset, so that the main decision factors of the category are found out, a parameter mapping relation table is compiled, and the result form is shown in fig. 2.
And 5: and confirming that the input and output parameters of the improved LSTM neural network are put into the network according to the established parameter mapping relation table. The improved LSTM neural network consists of a plurality of network levels, namely an input layer, a front-end network parallel layer, an LSTM network layer, a rear-end network serial layer and an output layer. Wherein the input layer X represents the input data parameter data set, and Y represents the output result parameter, the specific form is shown in fig. 3.
1) The input layer is a butt joint interface of a data parameter input network, firstly, a data set X is divided into n parts according to the time length of data inclusion, namely:
X=[x(1),x(2),...,x(n)]
one data of a long time span is divided into n data of a relatively short time span, and it is noted that the score n of the data set divided by the network is larger as the data set time span is longer. During the actual operation, according to the size of the reported data set and the water quality analysis, the reported data of the water plant is one year, and the change of four seasons in one year can cause the water quality parameters to have larger jump, so the integral input value is divided into 4 parts, namely the n value is set as 4.
2) The front-end network parallel layer mainly comprises a parallel form consisting of a plurality of groups of fully-connected neural networks and a dropout layer, wherein the weight value between a network node i and a network node j is set as wijThe threshold value of the node j is bjThe output value of each node is xjAnd the output value of each node is based on the aboveThe output values of all nodes in the layer, the weights and the biases of the current node and all nodes in the previous layer are jointly determined by an activation function f (Relu type activation function), and the effects are as follows:
Figure BDA0002316277160000072
xj=f(Sj)
wherein the activation function ReLU is as follows:
Figure BDA0002316277160000073
after the output of the fully connected network is completed, the method also needs to enter a dropout network hierarchical structure for processing, which specifically includes the following steps:
Figure BDA0002316277160000081
Figure BDA0002316277160000082
Figure BDA0002316277160000083
Figure BDA0002316277160000084
the Bernoulli function in the above formula randomly generates a vector of 0 and 1 with probability p.
3) The LSTM network layer is the main data processing layer in the whole network and is composed of a plurality of LSTM networks connected in parallel, wherein the current input xtAnd c passed down from the previous statet-1、ht-1The common determination is as follows:
Figure BDA0002316277160000085
Figure BDA0002316277160000086
Figure BDA0002316277160000087
Figure BDA0002316277160000088
wherein z isf,zi,z0After the splicing vector is multiplied by the weight matrix, the value is converted into a value between 0 and 1 through a sigmoid activation function to serve as a gating state. And z is the value between-1 and 1 that the result will be converted by a tanh activation function (tanh is used here because it is used as input data, not as a gating signal).
Figure BDA0002316277160000089
Figure BDA00023162771600000810
yt=δ(W'ht)
yseq=yt
Figure BDA00023162771600000811
Is the multiplication of the corresponding elements in the matrix,
Figure BDA00023162771600000812
matrix addition is performed.
4) The content of the serial layer of the back-end network is very similar to that of the parallel layer of the front-end network in structure, except that after receiving the results of the LSTM network layer, the parallel results need to be accumulated and summed to obtain a unique solution, and then the unique solution is sent to the full-connection and dropout network structure layer for processing, and the specific effects are as follows:
Figure BDA00023162771600000813
5) it should be further noted that some hyper-parametric designs in the network result, in the network structure, the initial learning rate is defined as 0.01, the Mini batch size is defined as 400, the optimizer selects adam, the gradient threshold is 1, and the front and rear fully-connected layers are defined as 3 layers.
6) Finally, the network can obtain an overall prediction network after training, then the water quality value at the next moment is predicted, known data are copied and divided into n parts to be put into the network for data prediction, and finally the predicted water quality value is obtained, and the effect is shown in fig. 4.
Finally, it should be noted that the above embodiments are only used for illustrating the technical solutions of the present invention and not for limiting, and although the technical solutions of the present invention are described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications or equivalent substitutions can be made on the technical solutions of the present invention without departing from the spirit and scope of the present invention, which should be covered by the protection scope of the present invention.

Claims (6)

1. A sewage treatment water quality prediction method based on an improved long-short term memory (LSTM) neural network is characterized by comprising the following steps:
1) data processing:
preprocessing the sewage treatment reported data, performing PCA (principal component analysis) dimensionality reduction on the preprocessed data to obtain result data, selecting a main parameter corresponding relation of the result data by utilizing a decision tree and pruning operation, and establishing a parameter mapping relation table;
2) establishing and training an improved long-short term memory (LSTM) neural network prediction model:
determining input and output parameters of the improved long-short term memory LSTM neural network according to the parameter mapping relation table, establishing an improved long-short term memory LSTM neural network prediction model, and training to obtain an overall improved long-short term memory LSTM neural network prediction model;
the improved long-short term memory LSTM neural network comprises an input layer, a front-end network parallel layer, an LSTM network layer, a rear-end network serial layer and an output layer;
3) and importing the data to be predicted into an overall improved long-short term memory (LSTM) neural network prediction model, and outputting to obtain a predicted water quality value.
2. The method for predicting the quality of the sewage treatment water according to claim 1, wherein the pretreatment is to remove the data outside the unreasonable limit value in the reported data of the sewage treatment, then perform 3 times standard deviation treatment on the reported data, and remove the data of which the difference between the maximum value or the minimum value of the reported data and the mean value exceeds 3 times standard deviation.
3. The method according to claim 1, wherein the input layer is a docking interface of a data parameter input network, and the data set X is divided into n parts according to the data-containing time span, that is, one data of a long time span is divided into n data of a relatively short time span:
X=[x(1),x(2),...,x(n)]。
4. the wastewater treatment water quality prediction method according to claim 1, wherein the front-end network parallel layer comprises a parallel connection form of a plurality of groups of fully-connected neural networks and a dropout layer, and a weight value between a network node i and a network node j in the plurality of groups of fully-connected neural networks is wijThe threshold value of the node j is bjThe output value of each node is xjThe output value of each node is determined by the activation function f according to the output values of all nodes on the upper layer, the weights and biases of the current node and all nodes on the upper layer, and the effect is as follows:
Figure FDA0002316277150000011
xj=f(Sj)
after the output of the fully-connected network is finished, the method enters a dropout network hierarchical structure for processing, and the method specifically comprises the following steps:
Figure FDA0002316277150000021
Figure FDA0002316277150000022
Figure FDA0002316277150000023
Figure FDA0002316277150000024
the Bernoulli function in the above formula randomly generates a vector of 0 and 1 with probability p.
5. The wastewater treatment water quality prediction method of claim 1, wherein the LSTM network layer is a main data processing layer in the whole modified LSTM neural network and is composed of a plurality of LSTM networks connected in parallel, wherein the current input x istAnd c passed down from the previous statet-1、ht-1The common determination is as follows:
Figure FDA0002316277150000025
Figure FDA0002316277150000026
Figure FDA0002316277150000027
Figure FDA0002316277150000028
wherein z isf,zi,z0After the splicing vector is multiplied by the weight matrix, the splicing vector is converted into a value between 0 and 1 through a sigmoid activation function to serve as a gating state, and z is the value between-1 and 1 obtained by converting the result through a tanh activation function:
Figure FDA0002316277150000029
Figure FDA00023162771500000210
yt=δ(W'ht)
yseq=yt
Figure FDA00023162771500000211
is the multiplication of the corresponding elements in the matrix,
Figure FDA00023162771500000212
matrix addition is performed.
6. The sewage treatment water quality prediction method according to claim 1, wherein the back-end network serial layer comprises a plurality of sets of fully-connected neural networks and dropouts, and after receiving the LSTM network layer results, the back-end network serial layer needs to accumulate and sum the parallel results to obtain a unique solution, and then sends the unique solution to the back fully-connected and dropouts network structure layer for processing, and the specific effects are as follows:
Figure FDA00023162771500000213
CN201911279211.2A 2019-12-13 2019-12-13 Sewage treatment water quality prediction method based on improved long-short term memory LSTM neural network Pending CN111027776A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911279211.2A CN111027776A (en) 2019-12-13 2019-12-13 Sewage treatment water quality prediction method based on improved long-short term memory LSTM neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911279211.2A CN111027776A (en) 2019-12-13 2019-12-13 Sewage treatment water quality prediction method based on improved long-short term memory LSTM neural network

Publications (1)

Publication Number Publication Date
CN111027776A true CN111027776A (en) 2020-04-17

Family

ID=70206622

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911279211.2A Pending CN111027776A (en) 2019-12-13 2019-12-13 Sewage treatment water quality prediction method based on improved long-short term memory LSTM neural network

Country Status (1)

Country Link
CN (1) CN111027776A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112598151A (en) * 2020-11-16 2021-04-02 华南理工大学 High-voltage cable surface temperature prediction method in cable trench laying mode
CN113033917A (en) * 2021-04-19 2021-06-25 重庆工商大学 Sewage treatment plant prediction planning operation management method based on peripheral data
CN113554321A (en) * 2021-07-28 2021-10-26 陕西科技大学 Dairy product cold-chain logistics quality safety early warning method
CN113886118A (en) * 2021-09-16 2022-01-04 杭州安恒信息技术股份有限公司 Abnormal resource processing method, device, system, electronic device and storage medium
WO2023236601A1 (en) * 2022-06-08 2023-12-14 腾讯科技(深圳)有限公司 Parameter prediction method, prediction server, prediction system and electronic device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109508811A (en) * 2018-09-30 2019-03-22 中冶华天工程技术有限公司 Parameter prediction method is discharged based on principal component analysis and the sewage treatment of shot and long term memory network
CN109598451A (en) * 2018-12-27 2019-04-09 东北大学 A kind of non-intrusion type load discrimination method based on PCA Yu LSTM neural network
CN109754113A (en) * 2018-11-29 2019-05-14 南京邮电大学 Load forecasting method based on dynamic time warping Yu length time memory
CN109859469A (en) * 2019-02-15 2019-06-07 重庆邮电大学 A kind of vehicle flowrate prediction technique based on integrated LSTM neural network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109508811A (en) * 2018-09-30 2019-03-22 中冶华天工程技术有限公司 Parameter prediction method is discharged based on principal component analysis and the sewage treatment of shot and long term memory network
CN109754113A (en) * 2018-11-29 2019-05-14 南京邮电大学 Load forecasting method based on dynamic time warping Yu length time memory
CN109598451A (en) * 2018-12-27 2019-04-09 东北大学 A kind of non-intrusion type load discrimination method based on PCA Yu LSTM neural network
CN109859469A (en) * 2019-02-15 2019-06-07 重庆邮电大学 A kind of vehicle flowrate prediction technique based on integrated LSTM neural network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
陈亮 等: "基于改进的LSTM的药品温湿度预测方法", 《电子测量与仪器学报》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112598151A (en) * 2020-11-16 2021-04-02 华南理工大学 High-voltage cable surface temperature prediction method in cable trench laying mode
CN113033917A (en) * 2021-04-19 2021-06-25 重庆工商大学 Sewage treatment plant prediction planning operation management method based on peripheral data
CN113033917B (en) * 2021-04-19 2022-04-12 重庆工商大学 Sewage treatment plant prediction planning operation management method based on peripheral data
CN113554321A (en) * 2021-07-28 2021-10-26 陕西科技大学 Dairy product cold-chain logistics quality safety early warning method
CN113886118A (en) * 2021-09-16 2022-01-04 杭州安恒信息技术股份有限公司 Abnormal resource processing method, device, system, electronic device and storage medium
WO2023236601A1 (en) * 2022-06-08 2023-12-14 腾讯科技(深圳)有限公司 Parameter prediction method, prediction server, prediction system and electronic device

Similar Documents

Publication Publication Date Title
CN111027776A (en) Sewage treatment water quality prediction method based on improved long-short term memory LSTM neural network
CN106022954B (en) Multiple BP neural network load prediction method based on grey correlation degree
CN110782658B (en) Traffic prediction method based on LightGBM algorithm
CN111105332A (en) Highway intelligent pre-maintenance method and system based on artificial neural network
CN110188936B (en) Short-term traffic flow prediction method based on multi-factor space selection deep learning algorithm
CN109558893B (en) Rapid integrated sewage treatment fault diagnosis method based on resampling pool
CN111738520A (en) System load prediction method fusing isolated forest and long-short term memory network
KR102009284B1 (en) Training apparatus for training dynamic recurrent neural networks to predict performance time of last activity in business process
CN114757432A (en) Future execution activity and time prediction method and system based on flow log and multi-task learning
CN109214444B (en) Game anti-addiction determination system and method based on twin neural network and GMM
CN110956309A (en) Flow activity prediction method based on CRF and LSTM
CN110830291A (en) Node classification method of heterogeneous information network based on meta-path
CN112765894B (en) K-LSTM-based aluminum electrolysis cell state prediction method
CN113077271A (en) Enterprise credit rating method and device based on BP neural network
CN114091794A (en) Patent value evaluation model training method, evaluation method, device and equipment
CN116542701A (en) Carbon price prediction method and system based on CNN-LSTM combination model
CN111310974A (en) Short-term water demand prediction method based on GA-ELM
CN114943328A (en) SARIMA-GRU time sequence prediction model based on BP neural network nonlinear combination
CN114820074A (en) Target user group prediction model construction method based on machine learning
CN113111588A (en) NO of gas turbineXEmission concentration prediction method and device
Jiang et al. A new test for common breaks in heterogeneous panel data models
CN112650770B (en) MySQL parameter recommendation method based on query work load analysis
CN114943304B (en) Bayesian-based 3C electronic product assembly error transfer method
Önay Koçoğlu et al. Machine Learning Approach and Model Performance Evaluation for Tele-Marketing Success Classification
Ali Qasimi Examining and Evaluating Classification Algorithms Based on Decision Trees

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200417