CN112485394A - Water quality soft measurement method based on sparse self-coding and extreme learning machine - Google Patents

Water quality soft measurement method based on sparse self-coding and extreme learning machine Download PDF

Info

Publication number
CN112485394A
CN112485394A CN202011249258.7A CN202011249258A CN112485394A CN 112485394 A CN112485394 A CN 112485394A CN 202011249258 A CN202011249258 A CN 202011249258A CN 112485394 A CN112485394 A CN 112485394A
Authority
CN
China
Prior art keywords
sample data
layer
hidden layer
data
output
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011249258.7A
Other languages
Chinese (zh)
Inventor
杨秦敏
曹伟伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN202011249258.7A priority Critical patent/CN112485394A/en
Publication of CN112485394A publication Critical patent/CN112485394A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N33/00Investigating or analysing materials by specific methods not covered by groups G01N1/00 - G01N31/00
    • G01N33/18Water
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2136Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on sparsity criteria, e.g. with an overcomplete basis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Biology (AREA)
  • Molecular Biology (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computing Systems (AREA)
  • Chemical & Material Sciences (AREA)
  • Food Science & Technology (AREA)
  • Medicinal Chemistry (AREA)
  • Analytical Chemistry (AREA)
  • Biochemistry (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Measuring Or Testing Involving Enzymes Or Micro-Organisms (AREA)

Abstract

The invention discloses a water quality soft measurement method based on sparse self-coding and an extreme learning machine, which combines the sparse self-coding method with the extreme learning machine method in the application of sewage treatment, takes the characteristics of extremely high learning speed, high model estimation precision, sparse self-coding, nonlinear data dimension reduction extraction characteristic value and the like of the extreme learning machine into consideration, can realize the rapid and effective estimation of the concentration of key water quality in sewage treatment, and effectively reduces the calculation complexity by combining a method of repeated resampling and averaging on the premise of ensuring the performance. The method is applied to the soft measurement of the ammonia nitrogen ion concentration in the sewage treatment, and can realize the rapid and accurate estimation of the ammonia nitrogen ion concentration, thereby realizing the soft measurement effect aiming at key components in the sewage treatment, reducing the restriction and limitation of the cost of the sensor on the sewage treatment process, and further providing support for the improvement of the sewage treatment process and the improvement of the effluent quality.

Description

Water quality soft measurement method based on sparse self-coding and extreme learning machine
Technical Field
The invention relates to the field of control science and engineering and environmental science and engineering, in particular to a water quality soft measurement method based on sparse self-coding and an extreme learning machine.
Background
The water resource reserves all over the world are abundant, but the fresh water resources account for only 2.53 percent, mainly deep water and glaciers, and the fresh water of lakes and rivers only accounts for 0.3 percent of the fresh water resources, so the fresh water resources available for human beings are very limited. Meanwhile, the development and the utilization of water resources by human beings are unreasonable, so that the water resources are greatly wasted and polluted, and the space of the available water resources is further compressed. Meanwhile, polluted water resources can cause damage to the environment, vegetation decline, animal death and the like, and can also harm human society and endanger human health and life safety. Therefore, water resource remediation is a research subject with great significance, and sewage treatment is a very effective means, and various kinds of sewage generated by human society are purified and discharged into rivers and lakes after the water quality reaches the standard or harmlessly utilized, so that not only can damage to the environment and the human society be avoided, but also the problem of water resource shortage caused by excessive development of the human society can be relieved.
At present, most sewage treatment in the world adopts anaerobic and aerobic biochemical reactions to realize sewage treatment, the main reason for adopting the method is that the sewage contains a large amount of organic matters which come from a garbage landfill, resident domestic wastewater, pharmaceutical factories, food factories and other scenes, the most main content of the organic matters is nitrogen element, ammonia nitrogen comes from coking plants, chemical fertilizer plants, petrochemical plants and the like, the wastewater containing a large amount of ammonia nitrogen ions is discharged into the nature to enrich and blacken and odorize water, and toxic action is generated on human beings and organisms. Therefore, the detection of the concentration of the ammonia nitrogen sample in the effluent quality of the sewage treatment plant is important. However, the existing sensor specially used for detection has high unreliability and high cost, and a plurality of soft measurement methods are provided for realizing accurate and rapid detection, but the effect of the existing soft measurement method can be further restrained due to the complexity of the sewage treatment process and the mutual influence among highly coupled components, and in order to further improve the soft measurement effect, the invention provides a water quality soft measurement method based on sparse self-coding and an extreme learning machine to overcome the existing difficulty.
Disclosure of Invention
In order to realize the rapid estimation of some components which are difficult to measure in the sewage treatment water quality and facilitate workers to adjust a control strategy in time, a plurality of scientific research works try to realize soft measurement methods for the components which are difficult to measure by using a machine learning method, but as the sewage treatment is a very complex system with strong coupling and the variable types are complex and diverse, different types of components are difficult to separate manually. Aiming at the problem, the invention provides a water quality soft measurement method based on sparse self-coding and an extreme learning machine, which can quickly separate a plurality of components and quickly and accurately estimate key components which are difficult to measure in sewage treatment reaction, thereby providing reasonable guidance for technicians to timely adjust control strategies.
The purpose of the invention is realized by the following technical scheme: a water quality soft measurement method based on sparse self-coding and extreme learning machine comprises the following steps:
(1) acquiring sample data: obtaining N from a wastewater treatment process0Group sample data
Figure BDA0002771055280000021
Each set of input vectors XiCharacterizing a plurality of wastewater quality components, corresponding expected output TiAnd characterizing the concentration of ammonia nitrogen ions in the effluent quality.
(2) Compressing the sample data by adopting a sampling mode, which specifically comprises the following steps: in [1,10 ]]Randomly selecting an integer initial value a, and acquiring data which is ten times compressed in a batch by acquiring every 10 points
Figure BDA0002771055280000022
And repeatedly sampling, and resetting the initial value a every time to obtain p batches of sample data.
(3) Sample data normalization: and respectively carrying out descaler dimensionalization on each batch of sample data, and normalizing the data with different dimensions to the range of [ -1,1] to obtain normalized sample data x.
(4) Performing dimensionality reduction on data according to a sparse self-encoder, specifically:
from the input layer to the hidden layer:
h=f(W1x+b1)
from the hidden layer to the output layer:
Figure BDA0002771055280000023
where h is the output of the hidden layer,
Figure BDA0002771055280000024
for the output of the output layer, i.e., the reconstructed vector, f (-) is the non-linear mapping, and W and b are the neural network weights and bias parameters.
The decoding function is a linear function or a Sigmoid function, so that the reconstruction error is minimum, and the reconstruction error is as follows:
Figure BDA0002771055280000025
adding sparseness limitation in encoder to control number of hidden layer neuron activation, supposing aj(x) Representing the activation function of the jth neuron in the hidden layer, the average activation amount of the jth neuron
Figure BDA0002771055280000026
Can be expressed as:
Figure BDA0002771055280000027
in order to render most of the hidden neurons inactive, let
Figure BDA0002771055280000028
Equal to a constant p, called the sparseness constant, close to 0. Selecting KL divergence as an expression of a penalty term PN:
Figure BDA0002771055280000031
wherein M is the number of neurons in the hidden layer,
Figure BDA0002771055280000032
is the KL divergence. The KL divergence expression is:
Figure BDA0002771055280000033
for an auto-encoder, the cost function is:
Figure BDA0002771055280000034
wherein λ is weight decay constant, nlNumber of layers of neural network, slThe number of the layer I neurons is shown as,
Figure BDA0002771055280000035
is the ji weight value of the l-th layer neural network. The total cost function containing the sparse penalty term is then:
Jsparse(W,b)=J(W,b)+βPN
where β is the sparse penalty term coefficient.
Updating the weight W and the bias b, the update equation can be obtained as:
Figure BDA0002771055280000036
Figure BDA0002771055280000037
wherein b isi lThe optimal W and b are obtained for the ith bias value of the l-th layer neural network and alpha is the learning rate, and better hidden layer output h belongs to RN×MThe method is used for representing the characteristics of the detection sample data, so that the dimension reduction of the detection sample data to M dimension is realized; let Y be h.
(5) The method comprises the following steps of constructing an extreme learning machine to realize water quality key component soft measurement, wherein a neural network of the extreme learning machine is formed by an input layer, a hidden layer and an output layer together, setting the input layer of the neural network to have M nodes according to the characteristics of sample data, setting the hidden layer to have L nodes, and setting the output layer to have M nodes, and the method comprises the following steps:
step 1: according to the reduced sample data set
Figure BDA0002771055280000038
Determines the type M and the data length N of the input data.
Figure BDA0002771055280000039
Wherein G (-) is the excitation function of the neural network, al,bl(L ═ 1, 2., L) are weights and offset values from the input layer to the hidden layer, L represents the number of hidden layer nodes of the neural network, Y represents a total of N groups of neural network input data, each group has M eigenvalues, i.e. the number of nodes corresponding to the input layer of the neural network, and H is the output of the hidden layer of the neural network;
step 2: taking the effluent quality of the sewage as target historical data T:
Figure BDA0002771055280000041
wherein t isj(j 1, 2.... N) is an output vector of the j-th group of target history data;
step 3: constructing a network from the hidden layer to the output layer has
Figure BDA0002771055280000042
Writing this formula as a matrix form
T=βH
Wherein wlmIs a weight vector from the hidden layer to the output layer, and the matrix is beta epsilon Rm×L,G(al,blY) is the hidden layer output and also the output layer input, in matrix formThen is H ∈ RL×N
Step 4: obtaining the weight value from the hidden layer to the output layer by adopting a Moore-Penrose method:
Figure BDA0002771055280000043
wherein ILIs an identity matrix of dimension L, and C is a normal value.
(6) Completing the calculation of sample data of one batch, and then calculating the sample data of the next batch until the calculation of the sample data of p batches is completed; and finally, respectively carrying out soft measurement by using the training results of the p batches of sample data, and averaging the soft measurement results to obtain a final soft measurement result, namely the concentration of the ammonia nitrogen ions in the effluent quality. By the method of dividing the soft measurement model at intervals, the performance of the soft measurement model is guaranteed, single calculated amount can be effectively reduced, the requirement on hardware equipment is reduced, and the calculation complexity can be reduced, so that the calculation complexity is reduced by multiple times.
Further, in the step (1), N is obtained from the sewage treatment process0Group sample data
Figure BDA0002771055280000044
Wherein each set of input vectors is of specific form Xi=[SI,i,SS,i,XI,i,XS,i,XBH,i,XBA,i,XP,i,SNO,i,SO,i,SND,i,XND,i]TRespectively representing 11 components of soluble inert organic matters, easily biodegradable substrates, insoluble inert organic matters, slowly biodegradable substrates, active heterotrophic organisms, active autotrophic organisms, biomass decay insoluble products, nitrate and nitrite, ammonium ions, soluble degradable organic nitrogen and insoluble degradable organic nitrogen in the sewage.
Further, in the step (3), the data of different dimensions are normalized to [ -1,1] by a method of minimum and maximum normalization, respectively, for each batch of sample data, and the formula is as follows:
Figure BDA0002771055280000051
wherein X is sample data compressed in sewage treatment, and X isminIs the minimum value of X, and XmaxThen it is the maximum value in X, which is the normalized sample data.
The invention has the beneficial effects that: the method combines the sparse self-coding with the extreme learning machine, on one hand, the sparse self-coding is utilized to carry out dimensionality reduction feature extraction on sample data, on the other hand, the extreme learning machine is utilized to quickly and effectively realize quick estimation on the key components of sewage treatment, meanwhile, the calculated amount is considered to be large, and the calculation complexity and the calculation conditions are effectively reduced by adopting repeated resampling and mean value calculation, so that the effect of soft measurement is effectively realized, and the constraint and limitation of the sensor cost on the sewage treatment process are reduced.
Drawings
FIG. 1 is a schematic view of the structure of the water quality soft measurement of the present invention;
FIG. 2 is a flow chart of a water quality soft measurement method based on sparse self-coding and extreme learning machine.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be further described in detail with reference to the accompanying drawings and examples. It should be understood that the detailed description and specific examples, while indicating the scope of the invention, are intended for purposes of illustration only and are not intended to limit the scope of the invention.
The invention provides a water quality soft measurement method based on sparse self-coding and extreme learning machine by combining sparse self-coding with extreme learning machine and utilizing a method of repeated resampling and averaging, which can quickly and accurately estimate key components in sewage treatment. As shown in fig. 1 and 2, the specific implementation steps are as follows:
(1) obtaining sample data
Obtaining N from a wastewater treatment process0Group sample data
Figure BDA0002771055280000052
Wherein each set of input vectors is of specific form Xi=[SI,i,SS,i,XI,i,XS,i,XBH,i,XBA,i,XP,i,SNO,i,SO,i,SND,i,XND,i]TRespectively representing 11 components in the sewage, such as soluble inert organic matters, easily biodegradable substrates, insoluble inert organic matters, slowly biodegradable substrates, active heterotrophic organisms, active autotrophic organisms, biomass decay insoluble products, nitrate and nitrite, ammonium ions, soluble degradable organic nitrogen, insoluble degradable organic nitrogen and the like, and correspondingly expecting output Ti=SNH,iAnd characterizing the component of ammonia nitrogen ion concentration in the effluent quality.
(2) Compressing sample data
In order to reduce the computational complexity without affecting the performance of the method, the invention adopts a method of multi-sampling computation. Knowing the original sample data
Figure BDA0002771055280000061
Sampling method is adopted to obtain compressed sample data in [1,10 ]]Randomly selecting an integer initial value a, and acquiring data which is ten times compressed in a batch by acquiring every 10 points
Figure BDA0002771055280000062
Sampling is repeatedly carried out, the initial value a is reset every time, p batches of sample data are obtained, and the following processing and training are respectively carried out on the p batches of sample data.
(3) Sample data normalization
Because the dimensions of different components in sewage treatment are different and the scale difference between numerical values is huge, in order to eliminate the influence caused by the dimensions, the data of different dimensions are normalized to between [ -1,1] by a minimum maximum value normalization method aiming at each batch of sample data respectively, so that the influence of the dimensions on soft measurement is eliminated. The concrete form is as follows:
Figure BDA0002771055280000063
wherein X is sample data compressed in sewage treatment, and X isminIs the minimum value of X, and XmaxThen it is the maximum value in X, where X is the normalized sample data, specifically X ═ SIn,SSn,XIn,XSn,XBHn,XBAn,XPn,SNOn,SOn,SNDn,XNDn]T
The following calculation steps are performed for each batch of normalized sample data.
(4) Dimensionality reduction of data from sparse self-encoder
The self-encoder is a three-layer symmetric deep learning neural network which extracts the hierarchical characteristics of high-dimensional complex input data from a non-tag data center by using unsupervised learning and optimization system parameters and obtains the distribution characteristic representation of original data. The neural network consists of an input layer, an implicit layer and an output layer, wherein the implicit layer encodes original input data, and the output layer decodes the implicit expression to reconstruct the original data, so that the reconstruction error is minimum to obtain the optimal implicit expression.
From the input layer to the hidden layer there are:
h=f(W1x+b1)
from the hidden layer to the output layer there are:
Figure BDA0002771055280000064
where h is the output of the hidden layer,
Figure BDA0002771055280000065
for the output of the output layer, i.e., the reconstructed vector, f (-) is a non-linear mapping, and W and b are the neural network weights and bias parameters.
The decoding function is typically a linear function or Sigmoid function, so that the reconstruction error is minimal, and is:
Figure BDA0002771055280000066
in order to enable the self-encoder to learn useful features, the encoder is added with sparse limitation to control the number of hidden layer neuron activations, and the self-encoder with hidden layer added with sparse limitation is called a sparse self-encoder. Suppose aj(x) Representing the activation function of the jth neuron in the hidden layer, the average activation amount of the jth neuron
Figure BDA0002771055280000071
Can be expressed as:
Figure BDA0002771055280000072
in order to render most of the hidden neurons inactive, let
Figure BDA0002771055280000073
Equal to a constant p, called the sparseness constant, close to 0. Selecting KL divergence as an expression of a penalty term PN:
Figure BDA0002771055280000074
wherein M is the number of hidden layer neurons (M < 11 in this embodiment),
Figure BDA0002771055280000075
is the KL divergence. The KL divergence expression is:
Figure BDA0002771055280000076
for an auto-encoder, the general cost function can be written as:
Figure BDA0002771055280000077
wherein λ is weight decay constant, nlNumber of layers of neural network, slThe number of the layer I neurons is shown as,
Figure BDA0002771055280000078
is the ji weight value of the l-th layer neural network. The total cost function containing the sparse penalty term can be written as:
Jsparse(W,b)=J(W,b)+βPN
where β is the sparse penalty term coefficient.
Updating the weight W and the bias b, the update equation can be obtained as:
Figure BDA0002771055280000079
Figure BDA00027710552800000710
wherein
Figure BDA00027710552800000711
The optimal W and b are obtained for the ith bias value of the l-th layer neural network and alpha is the learning rate, and then better hidden layer output h belongs to RN×MThe method is used for representing the characteristics of the detection sample data, thereby realizing the dimension reduction of the detection sample data to M dimension.
For convenience of subsequent description, Y ═ h is introduced here;
(5) water quality key component soft measuring-extreme learning machine
The neural network of the general extreme learning machine is composed of an input layer, a hidden layer and an output layer, wherein according to the characteristics of sample data, the input layer of the neural network is set to have M nodes (M equals to 3 in the embodiment), the hidden layer has L nodes (L equals to 100 in the embodiment), and the output layer has M nodes (M equals to 1 in the embodiment). Comprises the following steps:
step 1: according to the reduced sample data set
Figure BDA0002771055280000081
Determines the type M and the data length N of the input data.
Figure BDA0002771055280000082
Wherein G (-) is the excitation function of the neural network, al,bl(L ═ 1, 2., L) are weights and offset values from the input layer to the hidden layer, L represents the number of hidden layer nodes of the neural network, Y represents a total of N groups of neural network input data, each group has M eigenvalues, i.e. the number of nodes corresponding to the input layer of the neural network, and H is the output of the hidden layer of the neural network;
step 2: taking the effluent quality of the sewage as target historical data T:
Figure BDA0002771055280000083
wherein t isj(j 1, 2.... N) is an output vector of the j-th group of target history data;
step 3: constructing a network from a hidden layer to an output layer, and selecting a Purelin function according to the output layer, wherein the Purelin function has the following characteristics
Figure BDA0002771055280000084
Writing this formula as a matrix form
T=βH
Wherein wlmIs a weight vector from the hidden layer to the output layer, and the matrix is beta epsilon Rm×L,G(al,blY) is the hidden layer output and the output layer input, and the matrix form is H e RL×N
Step 4: under the premise of obtaining step1 and step2, step3 is processed by a Moore-Penrose method to obtain a weight value from a hidden layer to an output layer:
Figure BDA0002771055280000085
wherein ILIs an identity matrix of dimension L, and C is a normal value.
(6) And finishing the calculation of one batch of sample data, and then calculating the next batch of sample data until the last p batches of sample data are calculated. And finally, respectively carrying out soft measurement by using the training results of the p batches of sample data, and averaging the soft measurement results to obtain a final soft measurement result, namely the concentration of the ammonia nitrogen ions in the effluent quality.
The foregoing is only a preferred embodiment of the present invention, and although the present invention has been disclosed in the preferred embodiments, it is not intended to limit the present invention. Those skilled in the art can make numerous possible variations and modifications to the present teachings, or modify equivalent embodiments to equivalent variations, without departing from the scope of the present teachings, using the methods and techniques disclosed above. Therefore, any simple modification, equivalent change and modification made to the above embodiments according to the technical essence of the present invention are still within the scope of the protection of the technical solution of the present invention, unless the contents of the technical solution of the present invention are departed.

Claims (3)

1. A water quality soft measurement method based on sparse self-coding and extreme learning machine is characterized by comprising the following steps:
(1) acquiring sample data: obtaining N from a wastewater treatment process0Group sample data
Figure FDA0002771055270000011
Each set of input vectors XiCharacterizing a plurality of wastewater quality components, corresponding expected output TiAnd characterizing the concentration of ammonia nitrogen ions in the effluent quality.
(2) Compressing the sample data by adopting a sampling mode, which specifically comprises the following steps: in [1,10 ]]Randomly selecting an integer initial value a, and acquiring a batch of data at intervals of 10 pointsTen times as much data is compressed
Figure FDA0002771055270000012
And repeatedly sampling, and resetting the initial value a every time to obtain p batches of sample data.
(3) Sample data normalization: and respectively carrying out descaler dimensionalization on each batch of sample data, and normalizing the data with different dimensions to the range of [ -1,1] to obtain normalized sample data x.
(4) Performing dimensionality reduction on data according to a sparse self-encoder, specifically:
from the input layer to the hidden layer:
h=f(W1x+b1)
from the hidden layer to the output layer:
Figure FDA0002771055270000013
where h is the output of the hidden layer,
Figure FDA0002771055270000014
for the output of the output layer, i.e., the reconstructed vector, f (-) is the non-linear mapping, and W and b are the neural network weights and bias parameters.
The decoding function is a linear function or a Sigmoid function, so that the reconstruction error is minimum, and the reconstruction error is as follows:
Figure FDA0002771055270000015
adding sparseness limitation in encoder to control number of hidden layer neuron activation, supposing aj(x) Representing the activation function of the jth neuron in the hidden layer, the average activation amount of the jth neuron
Figure FDA0002771055270000016
Can be expressed as:
Figure FDA0002771055270000017
in order to render most of the hidden neurons inactive, let
Figure FDA0002771055270000018
Equal to a constant p, called the sparseness constant, close to 0. Selecting KL divergence as an expression of a penalty term PN:
Figure FDA0002771055270000019
wherein M is the number of neurons in the hidden layer,
Figure FDA0002771055270000021
is the KL divergence. The KL divergence expression is:
Figure FDA0002771055270000022
for an auto-encoder, the cost function is:
Figure FDA0002771055270000023
wherein λ is weight decay constant, nlNumber of layers of neural network, slThe number of the layer I neurons is shown as,
Figure FDA0002771055270000024
is the ji weight value of the l-th layer neural network. The total cost function containing the sparse penalty term is then:
Jsparse(W,b)=J(W,b)+βPN
where β is the sparse penalty term coefficient.
Updating the weight W and the bias b, the update equation can be obtained as:
Figure FDA0002771055270000025
Figure FDA0002771055270000026
wherein
Figure FDA0002771055270000027
The optimal W and b are obtained for the ith bias value of the l-th layer neural network and alpha is the learning rate, and better hidden layer output h belongs to RN×MThe method is used for representing the characteristics of the detection sample data, so that the dimension reduction of the detection sample data to M dimension is realized; let Y be h.
(5) The method comprises the following steps of constructing an extreme learning machine to realize water quality key component soft measurement, wherein a neural network of the extreme learning machine is formed by an input layer, a hidden layer and an output layer together, setting the input layer of the neural network to have M nodes according to the characteristics of sample data, setting the hidden layer to have L nodes, and setting the output layer to have M nodes, and the method comprises the following steps:
step 1: according to the reduced sample data set
Figure FDA0002771055270000028
Determines the type M and the data length N of the input data.
Figure FDA0002771055270000029
Wherein G (-) is the excitation function of the neural network, al,bl(L ═ 1, 2., L) are weights and offset values from the input layer to the hidden layer, L represents the number of hidden layer nodes of the neural network, Y represents a total of N groups of neural network input data, each group has M eigenvalues, i.e. the number of nodes corresponding to the input layer of the neural network, and H is the output of the hidden layer of the neural network;
step 2: taking the effluent quality of the sewage as target historical data T:
Figure FDA0002771055270000031
wherein t isj(j 1, 2.... N) is an output vector of the j-th group of target history data;
step 3: constructing a network from the hidden layer to the output layer has
Figure FDA0002771055270000032
Writing this formula as a matrix form
T=βH
Wherein wlmIs a weight vector from the hidden layer to the output layer, and the matrix is beta epsilon Rm×L,G(al,blY) is the hidden layer output and the output layer input, and the matrix form is H e RL×N
Step 4: obtaining the weight value from the hidden layer to the output layer by adopting a Moore-Penrose method:
Figure FDA0002771055270000033
wherein ILIs an identity matrix of dimension L, and C is a normal value.
(6) Completing the calculation of sample data of one batch, and then calculating the sample data of the next batch until the calculation of the sample data of p batches is completed; and finally, respectively carrying out soft measurement by using the training results of the p batches of sample data, and averaging the soft measurement results to obtain a final soft measurement result, namely the concentration of the ammonia nitrogen ions in the effluent quality.
2. The sparse self-coding and extreme learning machine-based water quality soft measurement method according to claim 1, wherein in the step (1), N is obtained from a sewage treatment process0Group sample data
Figure FDA0002771055270000034
Wherein each set of input vectors is of specific form Xi=[SI,i,SS,i,XI,i,XS,i,XBH,i,XBA,i,XP,i,SNO,i,SO,i,SND,i,XND,i]TRespectively representing 11 components of soluble inert organic matters, easily biodegradable substrates, insoluble inert organic matters, slowly biodegradable substrates, active heterotrophic organisms, active autotrophic organisms, biomass decay insoluble products, nitrate and nitrite, ammonium ions, soluble degradable organic nitrogen and insoluble degradable organic nitrogen in the sewage.
3. The sparse self-coding and extreme learning machine-based water quality soft measurement method according to claim 1, wherein in the step (3), the data of different dimensions are normalized to [ -1,1] by a minimum maximum normalization method by respectively performing de-dimensioning on each batch of sample data, and the formula is as follows:
Figure FDA0002771055270000035
wherein X is sample data compressed in sewage treatment, and X isminIs the minimum value of X, and XmaxThen it is the maximum value in X, which is the normalized sample data.
CN202011249258.7A 2020-11-10 2020-11-10 Water quality soft measurement method based on sparse self-coding and extreme learning machine Pending CN112485394A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011249258.7A CN112485394A (en) 2020-11-10 2020-11-10 Water quality soft measurement method based on sparse self-coding and extreme learning machine

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011249258.7A CN112485394A (en) 2020-11-10 2020-11-10 Water quality soft measurement method based on sparse self-coding and extreme learning machine

Publications (1)

Publication Number Publication Date
CN112485394A true CN112485394A (en) 2021-03-12

Family

ID=74929454

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011249258.7A Pending CN112485394A (en) 2020-11-10 2020-11-10 Water quality soft measurement method based on sparse self-coding and extreme learning machine

Country Status (1)

Country Link
CN (1) CN112485394A (en)

Citations (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103728431A (en) * 2014-01-09 2014-04-16 重庆科技学院 Industrial sewage COD (chemical oxygen demand) online soft measurement method based on ELM (extreme learning machine)
CN106485289A (en) * 2016-11-21 2017-03-08 东北大学 A kind of sorting technique of the grade of magnesite ore and equipment
CN107480777A (en) * 2017-08-28 2017-12-15 北京师范大学 Sparse self-encoding encoder Fast Training method based on pseudo- reversal learning
CN107563567A (en) * 2017-09-18 2018-01-09 河海大学 Core extreme learning machine Flood Forecasting Method based on sparse own coding
CN107679543A (en) * 2017-02-22 2018-02-09 天津大学 Sparse autocoder and extreme learning machine stereo image quality evaluation method
CN107886161A (en) * 2017-11-03 2018-04-06 南京航空航天大学 A kind of global sensitivity analysis method for improving Complex Information System efficiency
CN108021947A (en) * 2017-12-25 2018-05-11 北京航空航天大学 A kind of layering extreme learning machine target identification method of view-based access control model
CN108154260A (en) * 2017-12-15 2018-06-12 南京信息工程大学 A kind of short-term wind power forecast method
CN108469507A (en) * 2018-03-13 2018-08-31 北京工业大学 A kind of water outlet BOD flexible measurement methods based on Self organizing RBF Neural Network
CN108562709A (en) * 2018-04-25 2018-09-21 重庆工商大学 A kind of sewage disposal system water quality monitoring method for early warning based on convolution self-encoding encoder extreme learning machine
CN108665100A (en) * 2018-05-09 2018-10-16 中国农业大学 A kind of water quality prediction technique, system and device
CN108710738A (en) * 2018-05-11 2018-10-26 哈尔滨理工大学 A kind of extreme learning machine response phase method calculating leaf dish vibration reliability
CN109102012A (en) * 2018-07-30 2018-12-28 上海交通大学 A kind of defect identification method and system of local discharge signal
CN109187898A (en) * 2018-09-03 2019-01-11 中国农业大学 The flexible measurement method and device of Water quality ammonia nitrogen content in culture environment of aquatic products
CN109829627A (en) * 2019-01-04 2019-05-31 三峡大学 A kind of safe confidence appraisal procedure of Electrical Power System Dynamic based on integrated study scheme
CN109948583A (en) * 2019-03-28 2019-06-28 合肥京东方显示技术有限公司 Extreme learning machine, face identification method, readable storage medium storing program for executing and computer equipment
CN109948194A (en) * 2019-02-27 2019-06-28 北京航空航天大学 A kind of high-voltage circuitbreaker mechanical defect integrated study diagnostic method
CN110096985A (en) * 2019-04-23 2019-08-06 东北电力大学 A kind of City Building recognition methods based on characteristics of image and GPS positioning
CN110245781A (en) * 2019-05-14 2019-09-17 贵州科学院 The modelling application predicted based on the extreme learning machine of self-encoding encoder in industrial production
CN110320335A (en) * 2019-07-19 2019-10-11 东北大学 A kind of polynary robust flexible measurement method about wastewater treatment effluent quality index
CN110470477A (en) * 2019-09-19 2019-11-19 福州大学 A kind of Fault Diagnosis of Roller Bearings based on SSAE and BA-ELM
CN110473140A (en) * 2019-07-18 2019-11-19 清华大学 A kind of image dimension reduction method of the extreme learning machine based on figure insertion
CN110849626A (en) * 2019-11-18 2020-02-28 东南大学 Self-adaptive sparse compression self-coding rolling bearing fault diagnosis system
WO2020077232A1 (en) * 2018-10-12 2020-04-16 Cambridge Cancer Genomics Limited Methods and systems for nucleic acid variant detection and analysis
CN111476301A (en) * 2019-12-26 2020-07-31 山东中科先进技术研究院有限公司 Medical image classification method and system based on machine learning
CN111650834A (en) * 2020-06-16 2020-09-11 湖南工业大学 Sewage treatment process prediction control method based on Extreme Learning Machine (ELM)
CN111783959A (en) * 2020-07-08 2020-10-16 湖南工业大学 Electronic skin touch pattern recognition method based on classification of hierarchical extreme learning machine

Patent Citations (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103728431A (en) * 2014-01-09 2014-04-16 重庆科技学院 Industrial sewage COD (chemical oxygen demand) online soft measurement method based on ELM (extreme learning machine)
CN106485289A (en) * 2016-11-21 2017-03-08 东北大学 A kind of sorting technique of the grade of magnesite ore and equipment
CN107704883A (en) * 2016-11-21 2018-02-16 东北大学 A kind of sorting technique and system of the grade of magnesite ore
CN107679543A (en) * 2017-02-22 2018-02-09 天津大学 Sparse autocoder and extreme learning machine stereo image quality evaluation method
CN107480777A (en) * 2017-08-28 2017-12-15 北京师范大学 Sparse self-encoding encoder Fast Training method based on pseudo- reversal learning
CN107563567A (en) * 2017-09-18 2018-01-09 河海大学 Core extreme learning machine Flood Forecasting Method based on sparse own coding
CN107886161A (en) * 2017-11-03 2018-04-06 南京航空航天大学 A kind of global sensitivity analysis method for improving Complex Information System efficiency
CN108154260A (en) * 2017-12-15 2018-06-12 南京信息工程大学 A kind of short-term wind power forecast method
CN108021947A (en) * 2017-12-25 2018-05-11 北京航空航天大学 A kind of layering extreme learning machine target identification method of view-based access control model
CN108469507A (en) * 2018-03-13 2018-08-31 北京工业大学 A kind of water outlet BOD flexible measurement methods based on Self organizing RBF Neural Network
CN108562709A (en) * 2018-04-25 2018-09-21 重庆工商大学 A kind of sewage disposal system water quality monitoring method for early warning based on convolution self-encoding encoder extreme learning machine
CN108665100A (en) * 2018-05-09 2018-10-16 中国农业大学 A kind of water quality prediction technique, system and device
CN108710738A (en) * 2018-05-11 2018-10-26 哈尔滨理工大学 A kind of extreme learning machine response phase method calculating leaf dish vibration reliability
CN109102012A (en) * 2018-07-30 2018-12-28 上海交通大学 A kind of defect identification method and system of local discharge signal
CN109187898A (en) * 2018-09-03 2019-01-11 中国农业大学 The flexible measurement method and device of Water quality ammonia nitrogen content in culture environment of aquatic products
WO2020077232A1 (en) * 2018-10-12 2020-04-16 Cambridge Cancer Genomics Limited Methods and systems for nucleic acid variant detection and analysis
CN109829627A (en) * 2019-01-04 2019-05-31 三峡大学 A kind of safe confidence appraisal procedure of Electrical Power System Dynamic based on integrated study scheme
CN109948194A (en) * 2019-02-27 2019-06-28 北京航空航天大学 A kind of high-voltage circuitbreaker mechanical defect integrated study diagnostic method
CN109948583A (en) * 2019-03-28 2019-06-28 合肥京东方显示技术有限公司 Extreme learning machine, face identification method, readable storage medium storing program for executing and computer equipment
CN110096985A (en) * 2019-04-23 2019-08-06 东北电力大学 A kind of City Building recognition methods based on characteristics of image and GPS positioning
CN110245781A (en) * 2019-05-14 2019-09-17 贵州科学院 The modelling application predicted based on the extreme learning machine of self-encoding encoder in industrial production
CN110473140A (en) * 2019-07-18 2019-11-19 清华大学 A kind of image dimension reduction method of the extreme learning machine based on figure insertion
CN110320335A (en) * 2019-07-19 2019-10-11 东北大学 A kind of polynary robust flexible measurement method about wastewater treatment effluent quality index
CN110470477A (en) * 2019-09-19 2019-11-19 福州大学 A kind of Fault Diagnosis of Roller Bearings based on SSAE and BA-ELM
CN110849626A (en) * 2019-11-18 2020-02-28 东南大学 Self-adaptive sparse compression self-coding rolling bearing fault diagnosis system
CN111476301A (en) * 2019-12-26 2020-07-31 山东中科先进技术研究院有限公司 Medical image classification method and system based on machine learning
CN111650834A (en) * 2020-06-16 2020-09-11 湖南工业大学 Sewage treatment process prediction control method based on Extreme Learning Machine (ELM)
CN111783959A (en) * 2020-07-08 2020-10-16 湖南工业大学 Electronic skin touch pattern recognition method based on classification of hierarchical extreme learning machine

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
朱超岩等: "堆栈稀疏降噪自编码网络在变压器故障诊断中的应用", 《中国科技论文》 *
林雨: "极限学习机与自动编码器的融合算法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
汤芳等: "稀疏自编码深度神经网络及其在滚动轴承故障诊断中的应用", 《机械科学与技术》 *

Similar Documents

Publication Publication Date Title
CN108469507B (en) Effluent BOD soft measurement method based on self-organizing RBF neural network
Karul et al. Case studies on the use of neural networks in eutrophication modeling
CN102854296B (en) Sewage-disposal soft measurement method on basis of integrated neural network
Zaghloul et al. Application of machine learning techniques to model a full-scale wastewater treatment plant with biological nutrient removal
CN110647037B (en) Cooperative control method for sewage treatment process based on two-type fuzzy neural network
CN111367181B (en) Hybrid drive intelligent judgment control method for sewage treatment system
CN115345344A (en) Sewage treatment plant effluent quality prediction method based on attention mechanism and long-short term memory network
CN108764588A (en) A kind of temperature influence power prediction method based on deep learning
Yang et al. Fuzzy modeling approach to predictions of chemical oxygen demand in activated sludge processes
Wang et al. A full-view management method based on artificial neural networks for energy and material-savings in wastewater treatment plants
Liu et al. Effluent quality prediction of papermaking wastewater treatment processes using stacking ensemble learning
Mahmod et al. Dynamic modelling of aerobic granular sludge artificial neural networks
CN113918873B (en) Method for estimating dissolved oxygen concentration in sewage, storage medium, electronic device, and system
CN113111576B (en) Mixed coding particle swarm-long-short-term memory neural network-based effluent ammonia nitrogen soft measurement method
Peng et al. Monitoring of wastewater treatment process based on multi-stage variational autoencoder
Wang et al. Monitoring of wastewater treatment process based on slow feature analysis variational autoencoder
CN110929809B (en) Sewage key water quality index soft measurement method of characteristic self-enhanced cyclic neural network
CN106769748B (en) Intelligent detection system for water permeability of membrane bioreactor-MBR (Membrane bioreactor)
CN112485394A (en) Water quality soft measurement method based on sparse self-coding and extreme learning machine
Zhang et al. An echo state network based adaptive dynamic programming approach for time-varying parameters optimization with application in algal bloom prediction
CN110837886A (en) Effluent NH4-N soft measurement method based on ELM-SL0 neural network
CN115905821A (en) Urban sewage treatment process state monitoring method based on multi-stage dynamic fuzzy width learning
CN112183676A (en) Water quality soft measurement method based on mixed dimensionality reduction and kernel function extreme learning machine
CN112862190B (en) Water quality time sequence prediction method based on improved Seq2Seq frame
CN114864015A (en) Water eutrophication detection method, device, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210312

RJ01 Rejection of invention patent application after publication