CN112862173A - Lake and reservoir cyanobacterial bloom prediction method based on self-organizing deep confidence echo state network - Google Patents

Lake and reservoir cyanobacterial bloom prediction method based on self-organizing deep confidence echo state network Download PDF

Info

Publication number
CN112862173A
CN112862173A CN202110126626.7A CN202110126626A CN112862173A CN 112862173 A CN112862173 A CN 112862173A CN 202110126626 A CN202110126626 A CN 202110126626A CN 112862173 A CN112862173 A CN 112862173A
Authority
CN
China
Prior art keywords
echo state
sub
state network
self
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110126626.7A
Other languages
Chinese (zh)
Other versions
CN112862173B (en
Inventor
张慧妍
胡博
王小艺
王立
孙茜
王昭洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Technology and Business University
Original Assignee
Beijing Technology and Business University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Technology and Business University filed Critical Beijing Technology and Business University
Priority to CN202110126626.7A priority Critical patent/CN112862173B/en
Publication of CN112862173A publication Critical patent/CN112862173A/en
Application granted granted Critical
Publication of CN112862173B publication Critical patent/CN112862173B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • G06Q50/26Government or public services
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A20/00Water conservation; Efficient water supply; Efficient water use
    • Y02A20/152Water filtration

Landscapes

  • Business, Economics & Management (AREA)
  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Economics (AREA)
  • Human Resources & Organizations (AREA)
  • Strategic Management (AREA)
  • General Physics & Mathematics (AREA)
  • Tourism & Hospitality (AREA)
  • Health & Medical Sciences (AREA)
  • Marketing (AREA)
  • General Health & Medical Sciences (AREA)
  • Development Economics (AREA)
  • General Business, Economics & Management (AREA)
  • Educational Administration (AREA)
  • Biophysics (AREA)
  • Operations Research (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Game Theory and Decision Science (AREA)
  • Primary Health Care (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Quality & Reliability (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a lake and reservoir cyanobacteria bloom prediction method based on a self-organizing deep confidence echo state network, and belongs to the technical field of cyanobacteria bloom prediction and information science cross fusion. The lake and reservoir cyanobacterial bloom forecasting method adopts a mutual information method to screen input variables and output variables, then constructs a structure of a deep confidence echo state network, designs self-organization mechanisms aiming at the deep confidence network and the echo state network respectively, and obtains a self-organization deep confidence echo state network model after optimizing the structure self-organization mechanisms so as to effectively forecast the lake and reservoir cyanobacterial bloom and facilitate subsequent lake and reservoir cyanobacterial bloom treatment. The method fully learns the deep characteristics of the training data, designs a self-organization mechanism for the deep confidence echo state network, realizes the dynamic adjustment of the number of the hidden layer neurons and the sub-reservoirs, is suitable for the lake and reservoir cyanobacterial bloom data containing abnormal values such as detection noise and the like, and can improve the precision and the robustness of a prediction result.

Description

Lake and reservoir cyanobacterial bloom prediction method based on self-organizing deep confidence echo state network
Technical Field
The invention belongs to the technical field of blue algae bloom prediction and information science cross fusion, and particularly relates to a lake and reservoir blue algae bloom prediction method based on a self-organizing deep confidence echo state network.
Background
The lake-reservoir cyanobacterial bloom refers to the pollution phenomenon that algae and plankton in eutrophic lakes and reservoirs are abnormally and rapidly propagated, and a large amount of blue-green algae layers visible to naked eyes are gathered on the surface layer of a water body and thickly cover the water surface. As urban and industrial wastewater is continuously discharged into lakes and reservoirs, the content of nutrient substances such as nitrogen, phosphorus and the like in the water body is higher and higher, which provides an environmental foundation for the outbreak of cyanobacterial bloom. Generally, factors such as water temperature, wind speed and nutrient substances influence the outbreak of the cyanobacterial bloom in lakes and reservoirs. Therefore, the indexes can provide basis for targeted prediction, early warning and treatment of the cyanobacterial bloom. The lake and reservoir cyanobacteria bloom generation process has a chaotic attribute, and the time series prediction is carried out by taking the chlorophyll a concentration as a representation output variable and taking the water temperature, nutrient substances and the like as modeling input variables. The scholars in the environmental and biological fields carry out extensive research on the formation mechanism of the lake and reservoir cyanobacteria bloom, including modeling on environmental factors and plankton dynamics, and have better embodiment on the basic law of the lake and reservoir cyanobacteria bloom generation. Although the mechanism model has good interpretability, the evolution of the lake and reservoir cyanobacterial bloom is a complex nonlinear dynamic process and has certain sensitivity, and the establishment of the mechanism model with ideal quantitative prediction precision based on the existing research accumulation is difficult. With the development of the technology, the accessibility of data is continuously improved, and the application of a data driving method mainly based on a machine learning algorithm in the field of blue algae bloom prediction is more and more concerned. But the existing lake and reservoir cyanobacterial bloom prediction method has great defects in the aspects of prediction precision and robustness.
Disclosure of Invention
The invention provides a blue algae water bloom prediction method based on a self-organization deep confidence echo state network, which aims to effectively solve the problems of insufficient precision and poor robustness of the existing lake and reservoir blue algae water bloom prediction method. And after determining the input variable and the output variable, constructing a structure of a deep confidence echo state network, respectively designing a self-organization mechanism aiming at the deep confidence network and the echo state network, and obtaining a self-organization deep confidence echo state network model after optimizing the structure self-organization mechanism so as to effectively predict the lake and reservoir cyanobacterial bloom and facilitate the subsequent lake and reservoir cyanobacterial bloom treatment.
The invention provides a lake and reservoir cyanobacterial bloom prediction method based on a self-organizing deep confidence echo state network, which comprises the following four steps:
determining an input variable and an output variable of a deep confidence echo state network model;
determining the characterization variables of the lake and reservoir cyanobacterial bloom as output variables according to the domain knowledge, and screening out the influence variables of the lake and reservoir cyanobacterial bloom from the candidate water quality variables as input variables based on a mutual information method.
Step two, establishing a structure of a deep confidence echo state network;
the method comprises the steps of constructing a structure of a deep confidence echo state network, wherein the structure comprises the deep confidence network and the echo state network, and particularly, the echo state network adopts a modularized subreserve pool structure and adopts a robust loss function to solve an output weight matrix.
Designing a self-organization mechanism of the deep confidence echo state network and optimizing the deep confidence echo state network;
after the structure of the deep confidence echo state network is constructed, firstly, the importance index of a neuron is defined, then respective self-organization mechanisms of the deep confidence network and the echo state network are respectively designed, and the deep confidence echo state network is trained and optimized to obtain a self-organization deep confidence echo state network model.
Predicting based on the self-organizing deep confidence echo state network model;
and predicting the cyanobacterial bloom by using the self-organizing deep confidence echo state network model.
Compared with other methods in the prior art, the method provided by the invention has the advantages of feasibility and effectiveness.
The invention has the advantages that:
1. the invention constructs a self-organizing deep confidence echo state network model for forecasting the lake and reservoir cyanobacterial bloom, and can fully learn the deep characteristics of training data, thereby realizing the effective forecasting of the lake and reservoir cyanobacterial bloom.
2. The invention provides a neuron importance index for measuring the importance degree of neurons, and the neuron importance index is used as the basis of self-organizing mechanism design and is beneficial to training and optimizing the deep confidence echo state network.
3. The invention designs a self-organization mechanism for the deep belief network and the echo state network respectively, so that the deep belief echo state network model can automatically determine the network structure in the training process, and the dynamic adjustment of the number of hidden layer neurons and sub reserve pools is realized.
4. In the invention, the echo state network part utilizes a robust loss function to solve the output weight matrix. Therefore, the proposed self-organization deep confidence echo state network model is suitable for lake and reservoir cyanobacterial bloom data containing abnormal values such as detection noise and the like, and can improve the accuracy and robustness of a prediction result.
Drawings
FIG. 1 is a flow chart of a lake and reservoir cyanobacterial bloom prediction method based on a self-organizing deep confidence echo state network provided by the invention;
FIG. 2 is a flow chart of establishing a deep confidence echo state network structure according to the present invention;
FIG. 3 is a flow chart of the establishment and training optimization of the self-organizing mechanism of the deep confidence echo state mesh structure in the present invention;
FIG. 4A is a diagram illustrating the mutual information between the chlorophyll-a concentration and the lag input variable outputted in the embodiment of the present invention;
FIG. 4B is a diagram illustrating the mutual information between the input variables and the chlorophyll-a concentration according to an embodiment of the present invention;
fig. 5A, 5B, and 5C are a convergence graph of the number of neurons in the hidden layer of the deep belief network in the self-organizing deep belief echo state network structure, a convergence graph of the size of the reservoir of the echo state network, and a training error RMSE convergence graph of the deep belief echo state network model in the training process of the embodiment, respectively;
FIG. 6 is a schematic diagram showing the comparison between the predicted result of cyanobacterial bloom in lakes and reservoirs and the result of other conventional prediction methods in the embodiment of the present invention;
FIG. 7 is a comparison diagram of the predicted result of cyanobacterial bloom in lakes and reservoirs obtained by adding abnormal values with different proportions to the training data in the embodiment of the present invention.
Detailed Description
The present invention will be described in detail below with reference to the accompanying drawings and examples.
The invention provides a lake and reservoir cyanobacterial bloom prediction method based on a self-organizing deep confidence echo state network, wherein the self-organizing deep confidence echo state network comprises a deep confidence network and an echo state network, and in order to better predict lake and reservoir cyanobacterial bloom, the structure of the echo state network is required to be effectively optimized, and the characteristics of input variables are required to be subjected to targeted refining treatment. The deep belief network is a deep neural network model based on an energy function, can overcome the defect of local minimum and has good performance in the time series prediction problem. According to the method, the deep characteristics of the time sequence data in the input variables are extracted by using the unsupervised learning process of the deep confidence network, then the echo state network is adopted to model the deep characteristics and predict the chlorophyll a concentration at the next moment, so that the processing capacity of the self-organized deep confidence echo state network model on the time sequence information can be improved, and the blue-green algae bloom can be conveniently predicted.
In order to solve the optimization design problem of the neural network structure, the invention defines a neuron importance index by adopting a mutual information method, further defines the importance index of the hidden layer neuron and the importance index of the sub reserve pool respectively, and realizes the dynamic adjustment of the number of the hidden layer neuron and the sub reserve pool by designing a self-organization mechanism. In addition, the invention also adds a robust loss function for solving the output weight matrix of the echo state network so as to improve the robustness of the echo state network. Therefore, the prediction method provided by the invention has good prediction performance and good robustness on time sequence data containing abnormal values such as detection noise and the like, is suitable for modeling and prediction of practical lake and reservoir cyanobacterial bloom, and can provide prediction and early warning support for outbreak of the lake and reservoir cyanobacterial bloom.
The invention provides a lake and reservoir cyanobacteria bloom prediction method based on a self-organizing deep confidence echo state network, the flow of which is shown in figure 1, and the method mainly comprises the following four steps:
determining an input variable and an output variable of a deep confidence echo state network model;
the self-organizing deep confidence echo state network model is constructed by respectively determining an input variable and an output variable of the deep confidence echo state network model. In this embodiment, the output variable is determined as the concentration of chlorophyll a, and the input variable needs to be screened from a plurality of water quality variables affecting the generation of blue-green algae in lakes and reservoirs. The invention takes a mutual information method as a judgment criterion for screening input variables. Mutual information is used as a method for measuring the degree of interdependence between two variables, and can describe the nonlinear correlation of the two variables. When the mutual information value between the variables is larger, the correlation between the variables is higher. By respectively calculating mutual information values of the candidate water quality variables and the output variables, proper water quality variables can be screened as input variables according to the conditions of prediction precision requirements, speed and the like. Here, when the mutual information value of the candidate water quality variable and the output variable is greater than a set threshold (e.g., 0.2), the candidate water quality variable is selected as the input variable, otherwise, the candidate water quality variable is eliminated. And the screened input variables and the screened output variables participate in the training and prediction of the deep confidence echo state network model together.
Step two, establishing a structure of a deep confidence echo state network;
the self-organizing deep confidence echo state network model is composed of a deep confidence network based on limited Boltzmann machine stacking and a modular echo state network based on a sub-reserve pool. The deep confidence echo state network model firstly extracts deep features of input variables through a conventional deep confidence network. The limited Boltzmann machine is a basic unit forming a deep belief network, and comprises two layers of neurons, wherein one layer is a visual layer and is used for inputting variables; the other layer is a hidden layer for extracting deep features of the input variables. In particular, the deep confidence network part of the deep confidence echo state network is formed by stacking two limited Boltzmann machines. Specifically, as shown in fig. 2, the structure for establishing the deep confidence echo state includes the following steps:
inputting the input variable into a deep belief network, carrying out unsupervised learning through a contrast divergence method, and training the deep belief network to extract deep features of the input variable.
Inputting the deep features output by the hidden layer of the deep belief network into an echo state network, initializing the weight matrix of the deep features and the weight matrix of a sub-reserve pool by the echo state network, and collecting an internal state matrix.
The echo state network in the deep confidence echo state network is an echo state network based on a sub-reserve pool. The echo state network not only can meet the echo state characteristics, but also can reduce the complexity of parameter setting. The reserve pool in the echo state network without output feedback in the deep confidence echo state network comprises a plurality of sub reserve pools, and each sub reserve pool is mutually independent, so that the decoupling of partial neurons in the reserve pool is ensured.
Setting the number of the neutron reserve tanks in the original reserve tank as NtotalN in each sub-reserve poolsubEach neuron then consists of NtotalWeight matrix W of reserve pool formed by individual sub reserve pools* resIn a block diagonal matrix, i.e.,
Figure BDA0002924297260000041
wherein, Wi(1≤i≤Ntotal) And the weight matrix is corresponding to the ith sub-reserve pool. WiGenerated by singular value decomposition, i.e. Wi=UiSiVi. Wherein, the diagonal matrix
Figure BDA0002924297260000042
Randomly generated by a given singular value distribution, and the interior of the sub-pool matrix is fullAnd (4) connecting. n issubIs the size of the ith sub-reserve pool, i.e. all the sub-reserve pool weight matrixes in the invention are nsub×nsubA dimension matrix.
Figure BDA0002924297260000043
Is two random orthogonal matrices generated simultaneously, where upk,vpk∈(-1,1),p=1,2,…,nsub,k=1,2,…,nsub
The mathematical expression of the echo state network based on the sub-reserve pool is as follows:
Figure BDA0002924297260000044
Figure BDA0002924297260000045
wherein u (n) is an input vector at the time of n with K x 1 dimension, namely the deep-layer feature extracted by the deep belief network at the time of n, and K is the number of neurons of the last hidden layer of the deep belief network;
Figure BDA0002924297260000051
xi(n) is 1 XnsubState vector of ith sub reserve pool n moment of dimension; and y (n) is the output value of the echo state network at the moment n.
Figure BDA0002924297260000052
In order to input the weight matrix, the weight matrix is input,
Figure BDA0002924297260000053
is nsubThe input weight matrix of the ith sub-pool in xK dimension,
Figure BDA0002924297260000054
is 1 × (N)total×nsub) The output weight matrix of the dimension. f. ofresIs the activation function of the reserve pool neurons, and takes the sigmoid function.
Here, to overcome the effect of the initial transient, assume that n is the number nminFrom time +1 to LtrainTime-of-day collection internal state matrix H ═ x (n)min+1),…,x(Ltrain)]TThe corresponding desired output vector is T ═ T (n)min+1),…,t(Ltrain)]T,t(nmin+1) is nminThe desired output value at time + 1.
In addition, in order to overcome the ill-conditioned solution problem possibly caused by abnormal values including detection noise and the like and improve the robustness of prediction, the output weight matrix is iteratively solved by adopting a robust loss function including L2 regularization
Figure BDA0002924297260000055
Initializing the solving iteration times k of the output weight matrix to be 1, initializing the robust weight matrix to be a unit matrix, and calculating a robust loss function and residual robust scale estimation; and updating the robust weight matrix according to the robust weight function.
And initializing the solving iteration number k of the output weight matrix to be 1, initializing the robust weight matrix to be a unit matrix, calculating a robust loss function and residual robust scale estimation in the iteration process, updating the robust weight matrix according to the robust weight function, and calculating the output weight matrix. Combining the robust loss function E (k) of the regularization term and the output weight matrix from iteration to k
Figure BDA0002924297260000056
The solving results of (1) are respectively:
Figure BDA0002924297260000057
Figure BDA0002924297260000058
where C is the regularization coefficient and I is (N)total×nsub)×(Ntotal×nsub) Identity matrix of dimension,
Figure BDA0002924297260000059
Is a 2-norm, ρ (-) is a robust objective function, ξ[k](n)=T(n)-y[k](n) is the training error at the nth time instant of iteration to step k,
Figure BDA00029242972600000510
for the residual robust scale estimation from iteration to k steps, MAR is the median absolute deviation.
Figure BDA00029242972600000511
Is represented by (L)train-nmin)×(Ltrain-nmin) The robust weight matrix of dimension, w (-) is the robust weight function. In the invention, a Welsch function is taken as a robust weight function, and a robust target function rho (-) and a robust weight function w (-) thereof are respectively as follows:
Figure BDA00029242972600000512
Figure BDA00029242972600000513
wherein z is a function variable, kset=μkdefMu is a robust coefficient, the robust weight function is selected according to experience, the Welsch function is selected as the robust weight function, and then the coefficient kdef=2.985。
Designing a self-organization mechanism of the deep confidence network and the echo state network and training the deep confidence network and the echo state network;
the invention respectively designs a self-organization mechanism and a corresponding training process aiming at a deep confidence network and an echo state network. Namely, in the step two, on the basis of the step two, the adjustment of the hidden layer neuron of the deep confidence network and the neutron reserve pool of the echo state network is respectively realized in each iteration of the respective training process.
As shown in fig. 3, for deep belief networksFor each hidden layer neuron, first initialize the iterative training times k 11, training a weight matrix of the deep belief network according to a contrast divergence method, and calculating the importance index of each layer of neuron, wherein the process is iterated to the k-th arbitrary neuron1Importance index of neurons in step
Figure BDA00029242972600000616
Is defined as:
Figure BDA0002924297260000061
wherein,
Figure BDA0002924297260000062
and
Figure BDA0002924297260000063
respectively, the input and output of the jth neuron of the l layer.
Figure BDA0002924297260000064
Is shown as
Figure BDA0002924297260000065
And
Figure BDA0002924297260000066
the mutual information value between the two information blocks,
Figure BDA0002924297260000067
is shown as
Figure BDA0002924297260000068
And the desired output vector T. For the deep belief network part, the self-organization process of hidden layer neurons includes splitting and deleting, and a specific self-organization mechanism based on neuron importance is shown as follows.
(1) The mechanism of neuronal cleavage in the hidden layer: when iterating to the k1When the step is carried out,
Figure BDA00029242972600000615
the higher the neuron is, the more active it is processing information. The present invention therefore chooses to split the most active neurons in the hidden layer. That is, when the jth neuron of the l-th layer satisfies the following condition:
Figure BDA0002924297260000069
the jth neuron splits into two neurons,
Figure BDA00029242972600000610
is iterated to the k < th >1Total number of layer I neurons in step (ii).
(2) Mechanism of pruning of hidden layer neurons: when in use
Figure BDA00029242972600000611
At lower, the neuron processes the information less strongly and should be considered to delete it. Thus, the present invention defines iterating through the kth1The adaptive pruning threshold at step time is as follows:
Figure BDA00029242972600000612
wherein beta is (0,1)]. Then, according to the above formula, when the jth neuron satisfies the condition
Figure BDA00029242972600000613
Then the jth neuron is deleted.
After the number of neurons in the hidden layer of the deep belief network and the weight matrix are subjected to iterative training, the number of the sub-reserve pools of the echo state network and the output weight matrix can be subjected to iterative training. Taking the output vector of the last hidden layer of the trained deep belief network as the input of the echo state network, and initializing the iteration times k of the echo state network2Defining the control parameter vector by user, randomly generating a temporary reserve pool weight and a temporary input weight which are consistent with the original reserve pool in size, and for 1The specific screening and growing mechanism of the sub-reserve pool of the echo state network is shown as follows:
(1) the screening mechanism of the child reserve pool: the invention defines the importance index of the ith sub-reserve pool in the reserve pool
Figure BDA00029242972600000614
Comprises the following steps:
Figure BDA0002924297260000071
wherein
Figure BDA0002924297260000072
Is the input vector of the p-th neuron of the ith sub-reservoir,
Figure BDA0002924297260000073
is the output vector of the p neuron of the ith sub-reservoir. Thus, training iteratively to k2Randomly generating temporary sub-reserve pool with structure consistent with that of original reserve pool in step time
Figure BDA00029242972600000716
Sorting according to the size of the importance indexes:
Figure BDA0002924297260000074
the invention defines the adaptive screening threshold as follows:
Sth(k2)=NS′sub(INT(αimax(k2))) (12)
wherein INT (-) is an integer function.
Figure BDA0002924297260000075
The sorted child pool importance vectors are used. And alpha epsilon (0,1) is a customized control parameter, namely different screening degrees of the control sub-reserve pool in each circulation. The parameter may take a plurality of values
Figure BDA0002924297260000076
Together forming a control parameter vector
Figure BDA0002924297260000077
But need to satisfy alpha1<α2<…<α,NαIs the dimension of the control parameter vector.
The training goal of the echo state network is to minimize the robust loss function of equation (4). In order to ensure that the performance of the reserve pool after screening can be kept or better than the sub-reserve pool set before screening, when the kth iteration2The ith sub-reserve pool meets the following requirements:
Figure BDA0002924297260000078
and the robust loss function E (k) of all sub-pools that satisfy the condition2) And when the value is less than or equal to the minimum value of the historical robust loss function, the sub-reserve pools are reserved as new reserve pools, and the rest sub-reserve pools are deleted. And taking the screened sub reserve pools as temporary reserve pools and calculating training errors.
(2) Growth mechanism of child reserve pool: after screening, the increase of the sub reserve pool is realized, the temporary reserve pool is used as a new reserve pool, and a new randomly generated sub reserve pool is merged, so that the output weight matrix of the merged echo state network is as follows:
Figure BDA0002924297260000079
wherein HoFor the state matrix corresponding to the reserve pool after the screening mechanism is completed, HgFor the state matrix corresponding to the growing reserve pool,
Figure BDA00029242972600000710
is the state matrix corresponding to the growing reserve pool.
Figure BDA00029242972600000711
Is composed of
Figure BDA00029242972600000712
An identity matrix of dimensions, wherein,
Figure BDA00029242972600000713
is the total number of growing child pools. Further, a combined output weight matrix may be derived based on equation (14)
Figure BDA00029242972600000714
The updated mathematical expression is:
Figure BDA00029242972600000715
wherein, IoIs (N)o×nsub)×(No×nsub) Identity matrix of dimension, NoThe number of the child reserve pools after the screening mechanism is completed. I isgIs nsub×nsubAn identity matrix of dimensions. I isLIs (L)train-nmin)×(Ltrain-nmin) An identity matrix of dimensions.
And obtaining a self-organizing deep confidence echo state network model.
Predicting based on the self-organizing deep confidence echo state network model;
through the design of a self-organization mechanism, the self-organization deep confidence echo state network model can automatically learn and optimally design the proper number of neurons in the hidden layer of the deep confidence network and the number of the sub reserve pools of the echo state network in the training process, and meanwhile, the weight matrix solution corresponding to each neural network is realized. And inputting the input variable into the trained self-organizing depth confidence echo state network model, so that the characterization index of the lake and reservoir cyanobacterial bloom, namely the prediction of the chlorophyll a concentration, can be realized.
The technical solution of the present invention is further illustrated by the following examples.
The first embodiment is as follows:
the embodiment provides a lake and reservoir cyanobacterial bloom prediction method based on a self-organizing deep confidence echo state network, which comprises the following specific implementation steps:
step one, determining an input variable and an output variable of a prediction model;
the data in the examples were derived from the water quality data set of west fal-thao harbor, usa. The data set contains 6 water quality variables, and table 1 specifically shows the abbreviations, units and meanings of the individual variables in the data set.
TABLE 1 Water quality variables information
Figure BDA0002924297260000081
The sampling frequency of the data is 20 minutes, the acquisition time starts from 18 o 'clock 01 at 6 h in 2017 to 13 o' clock 21 at 31 h in 2017, 8 h in 2017, and 2491 groups of data are shared. In order to overcome the influence of the redundant indexes on the modeling effect, the experiment measures the correlation between the water quality variable and the output variable chlorophyll a concentration by using mutual information values. In the experiment, not only the correlation of the water quality variables but also the autoregressive characteristic of the time series of the chlorophyll a concentration are considered. As can be seen from fig. 4A, the mutual information value of the lag variable of chlorophyll a gradually decreases as the lag time increases. Fig. 4B shows the mutual information values of 5 water quality variables for the chlorophyll a concentration at the next time. The experiment selects a water quality variable with a mutual information value greater than 0.2. Therefore, the input variables of the self-organizing depth confidence echo state network are the water temperature, the salinity, the oxygen saturation, the specific conductivity, the chlorophyll a concentration at the same moment and the chlorophyll a concentration at the lag three moments, and the output variable is the chlorophyll a concentration at the next moment. That is, the number of input variables is 8, and the number of output variables is 1.
Step two, establishing a structure of a deep confidence echo state network;
in the experiment for predicting the cyanobacterial bloom in the lake and reservoir, the self-organizing deep confidence echo state network starts to collect data after running 200 data to form a state matrix, the length of training data is 1600, and the length of testing data is 691. Wherein, the neuron of the hidden layer of the deep belief network part is initialized to 3, the iterative training time is 50, and the learning batchThe size is 50, the learning rate is 0.1, and β is 0.98. The initialization range of the input weight matrix element of the echo state network is [ -1,1 [ -1 [ ]]Taking the singular value of the diagonal matrix in SVD to be [0.1,0.99 ]]The size of the sub-reserve pool is 5, the regularization coefficient C is 1e-7, the robust coefficient mu is 1, the iteration frequency of output weight matrix solving is 15, the iteration frequency of the self-organization process of the reserve pool is 50, and the control parameter vector is uniformly distributed
Figure BDA0002924297260000091
Is (0.5,0.6,0.7,0.8, 0.9).
Designing a self-organizing mechanism and a training process of the optimized deep confidence echo state network;
the self-organizing process of hidden layers and reserve pools in the self-organizing deep confidence echo state network is shown in fig. 5A and 5B. In FIG. 5A, the neurons of the first hidden layer H1 and the second hidden layer H2 are finally stable at 7 and 6, respectively, and thus the final hidden layer structure is 7-6. During the process of reservoir size learning, the iterative training time is set to 100. As shown in fig. 5B, the pool size iteratively converges to 120, containing a total of 24 child pools, based on the self-organizing mechanism. Therefore, the structure of the self-organizing deep confidence echo state network in the experiment is 8-7-6-120-1. Fig. 5C shows the convergence curve of Root Mean Square Error (RMSE) during training. The training error of the self-organizing deep confidence echo state network finally converges to be near the minimum value of 0.383.
Predicting based on the self-organizing deep confidence echo state network model;
FIG. 6 shows the comparison of the predicted results of cyanobacterial bloom in lakes and reservoirs by using the self-organizing deep confidence echo state network and the predicted results by using other echo state network methods. It can be seen that the self-organizing deep confidence echo state network (SDBMESN) provided by the embodiment of the invention can effectively learn the evolution rule of the cyanobacterial bloom in the lake and reservoir relative to other echo state network models. Table 2 shows the comprehensive performance of the basic echo state network (OESN), the Regularized Echo State Network (RESN), the Growing Echo State Network (GESN), the adaptive regularized echo state network (DRESN), and the deep confidence echo state network (DBESN) in training and testing, including neural network structure and RMSE indices. Therefore, the lake and reservoir cyanobacterial bloom prediction method based on the self-organizing deep confidence echo state network has high prediction precision and good generalization capability. Meanwhile, the size of a reserve pool of the self-organizing deep confidence echo state network is smaller than that of other echo state networks, and the self-organizing deep confidence echo state network has the simplest neural network structure. Here, the DBESN for each set of experiments employs the same neural network structure as the self-organizing deep confidence echo state network. But under the condition of consistent neural network structure, the predicted performance of DBESN is lower than that of the self-organizing deep confidence echo state network provided by the invention. The self-organizing mechanism of the self-organizing deep confidence echo state network not only realizes structural simplification, but also reserves neurons and sub reserve pools with better relative performance in the existing neurons in the self-organizing process, so that the neurons and the reserve pools in the self-organizing deep confidence echo state network have better prediction effect, and the capability of the neurons and the reserve pools for processing dynamic information is further improved. Therefore, the self-organizing deep confidence echo state network is suitable for the prediction application of the lake and reservoir cyanobacterial bloom.
TABLE 2 blue algae bloom prediction experiment results and different method comparison
Figure BDA0002924297260000101
In the self-organizing deep confidence echo state network in the embodiment, the robustness loss function is used as the target function, so that the robustness of time series data prediction of abnormal values such as monitoring noise can be improved. To verify this feature, a 10% to 40% proportion of the pulse function was added to each training sample of the example dataset. The test results are shown in FIG. 7. It can be seen that the robustness of the self-organizing deep confidence echo state network of the embodiment of the invention is obviously superior to that of other echo state networks, and the robustness is better.

Claims (9)

1. A lake and reservoir cyanobacterial bloom prediction method based on a self-organizing deep confidence echo state network is characterized by comprising the following steps: comprises the following steps of (a) carrying out,
determining an input variable and an output variable of a deep confidence echo state network;
the output variable is the chlorophyll a concentration of the characterization variable of the lake and reservoir cyanobacterial bloom, and the input variable is obtained by screening the water quality variables influencing the lake and reservoir cyanobacterial bloom by taking a mutual information method as a judgment criterion;
step two, establishing a structure of a deep confidence echo state network;
the structure of the deep confidence echo state network comprises a deep confidence network and an echo state network, wherein the echo state network adopts a modular reserve pool structure; the method comprises the following specific steps:
2.1, adopting a limited Boltzmann mechanism to form a basic unit of a deep confidence network, and extracting deep features of input variables;
2.2, learning the deep features and predicting the chlorophyll a concentration at the next moment by an echo state network;
designing a self-organization mechanism of the deep confidence echo state network and training the deep confidence echo state network;
firstly, defining importance indexes of neurons in the deep confidence network, then respectively designing respective self-organization mechanisms of the deep confidence network and the echo state network, and optimizing the structure of the deep confidence echo state network to obtain a self-organization deep confidence echo state network model;
predicting based on the self-organizing deep confidence echo state network model;
and predicting the cyanobacterial bloom by using the self-organizing deep confidence echo state network model.
2. The lake and reservoir cyanobacterial bloom prediction method based on the self-organizing deep confidence echo state network as claimed in claim 1, wherein: the mutual information value of the input variable and the chlorophyll a concentration at the same moment is more than 0.2; the input variables also include the chlorophyll-a concentration at three moments later.
3. The lake and reservoir cyanobacterial bloom prediction method based on the self-organizing deep confidence echo state network as claimed in claim 1, wherein: the deep belief network is formed by stacking two limited Boltzmann machines, each limited Boltzmann machine comprises two layers of neurons, and one layer is a visual layer and is used for being used as input of input variables of training data; the other layer is a hidden layer for extracting deep features as input variables of the training data.
4. The lake and reservoir cyanobacterial bloom prediction method based on the self-organizing deep confidence echo state network as claimed in claim 3, wherein: the reserve pool in the echo state network comprises a plurality of sub reserve pools, and each sub reserve pool is independent; setting the number of the sub-reserve pools as NtotalEach is then composed of NtotalReserve pool weight matrix W formed by individual reserve pools* resIs a block diagonal matrix, namely:
Figure FDA0002924297250000021
wherein each weight matrix element WiI is more than or equal to 1 and less than or equal to N for the weight matrix corresponding to the ith sub-reserve pooltotal;WiGenerated by singular value decomposition, i.e. Wi=UiSiViWherein a diagonal matrix
Figure FDA0002924297250000022
Randomly generated by a given singular value distribution, and the weight matrix inside the sub-reserve pool is fully connected,
Figure FDA0002924297250000023
nsubis the size of the ith sub-pool,
Figure FDA0002924297250000024
is two random orthogonal matrices generated simultaneously, where upk,vpk∈(-1,1),p=1,2,…,nsub,k=1,2,…,nsubThen, the mathematical expression of the echo state network is:
x(n)=fres(W* resx(n-1)+Winu(n)) (2)
y(n)=W* outx(n) (3)
wherein u (n) is an input vector of K multiplied by 1 dimension at n time, K is the number of neurons of the last hidden layer of the deep belief network, namely the deep features of the deep belief network at n time;
Figure FDA0002924297250000025
xi(n) is 1 XnsubState vector of ith sub reserve pool n moment of dimension; y (n) is the output value of the echo state network at the moment n,
Figure FDA0002924297250000026
as an input weight matrix, Wi inIs nsubInput weight matrix, W, of ith sub-pool of dimension xK* outIs 1 × (N)total×nsub) Output weight matrix of dimension, fresIs the activation function of the reserve pool neurons, and takes the sigmoid function.
5. The lake and reservoir cyanobacterial bloom prediction method based on the self-organizing deep confidence echo state network as claimed in claim 4, wherein: the output weight matrix W* outAdopting a robust loss function containing L2 regularization to iteratively solve output, combining the robust loss function E (k) of the regularization item from iteration to k step and outputting a weight matrix W* out[k]The solving results of (1) are respectively:
Figure FDA0002924297250000027
W* out[k]=(HTWN [k]H+CI)-1HTWN [k]T (5)
where C is the regularization coefficient and I is (N)total×nsub)×(Ntotal×nsub) The identity matrix of the dimension(s),
Figure FDA0002924297250000028
2-norm, ρ (-) is a robust objective function,
Figure FDA00029242972500000211
for the training error iterated to the nth time of step k,
Figure FDA0002924297250000029
for residual robust scale estimation from iteration to k steps, MAR is the median absolute deviation,
Figure FDA00029242972500000210
to represent
Figure FDA00029242972500000212
A robust weight matrix of dimension, w (-) is a robust weight function, and the robust objective function ρ (-) and the robust weight function w (-) are respectively:
Figure FDA0002924297250000031
Figure FDA0002924297250000032
wherein z is a variable, kset=μkdefMu is a robust coefficient, kdef=2.985。
6. The lake and reservoir cyanobacterial bloom prediction method based on the self-organizing deep confidence echo state network as claimed in claim 5, wherein: the robust weight function is a Welsch function.
7. The lake and reservoir cyanobacterial bloom prediction method based on the self-organizing deep confidence echo state network as claimed in claim 4, wherein: the echo state network further comprises the step of collecting an internal state matrix H, in particular, assumed from nminFrom time +1 to LtrainTime-of-day collection internal state matrix H ═ x (n)min+1),…,x(Ltrain)]TThe corresponding desired output vector is T ═ T (n)min+1),…,t(Ltrain)]T,t(nmin+1) is nminThe desired output value at time + 1.
8. The lake and reservoir cyanobacterial bloom prediction method based on the self-organizing deep confidence echo state network as claimed in claim 5, wherein: the self-organization mechanism of the deep belief network in step three comprises a splitting mechanism of hidden layer neurons and a pruning mechanism of hidden layer neurons, and specifically,
for each hidden layer neuron of the deep belief network, iterate to kth1Importance index of neurons in step
Figure FDA0002924297250000033
Is defined as:
Figure FDA0002924297250000034
wherein,
Figure FDA0002924297250000035
and
Figure FDA0002924297250000036
respectively the input and output of the jth neuron of the l-th layer,
Figure FDA0002924297250000037
is shown as
Figure FDA0002924297250000038
And
Figure FDA0002924297250000039
the mutual information value between the two information blocks,
Figure FDA00029242972500000310
is shown as
Figure FDA00029242972500000311
And the mutual information value between the desired output vector T,
(3.1) mechanism of cleavage of hidden layer neurons: when the jth neuron of the l-th layer meets the following condition:
Figure FDA00029242972500000312
the jth neuron splits into two neurons,
Figure FDA00029242972500000313
is iterated to the k < th >1Total number of layer I neurons in step (ii);
(3.2) mechanism of pruning of hidden layer neurons: define iteration to kth1The adaptive pruning threshold at step time is as follows:
Figure FDA00029242972500000314
wherein beta is (0,1)]When the jth neuron satisfies the condition
Figure FDA00029242972500000315
Then the jth neuron is deleted.
9. The lake and reservoir cyanobacterial bloom prediction method based on the self-organizing deep confidence echo state network as claimed in claim 8, wherein: the self-organization mechanism of the echo state network in step three comprises a sub-pool screening and growing mechanism, specifically,
(3.3) screening mechanism of child pools: defining importance index of ith sub-reserve pool in reserve pool
Figure FDA00029242972500000316
Comprises the following steps:
Figure FDA0002924297250000041
wherein
Figure FDA0002924297250000042
Is the input vector of the p-th neuron of the ith sub-reservoir,
Figure FDA0002924297250000043
for the output vector of the p-th neuron of the ith sub-reservoir, iteratively training to k2Randomly generating i consistent with the structure of the original reserve pool in stepmax(k2) A temporary sub-reserve pool
Figure FDA00029242972500000414
Sorting according to the size of the importance indexes:
Figure FDA0002924297250000044
the adaptive screening threshold is defined as follows:
Sth(k2)=NS′sub(INT(αimax(k2))) (12)
wherein INT (-) is an integer function,
Figure FDA0002924297250000045
for the sorted sub-reserve pool importance vectors, alpha is an element (0,1) which is a self-defined control parameter;
when iterating the k2When the step is carried out, the ith sub-reserve pool meets the following requirements:
Figure FDA0002924297250000046
and robust loss function E (k) for all sub-pools2) When the value is less than or equal to the minimum value of the historical robust loss function, the sub reserve pools are reserved, and the rest sub reserve pools are deleted;
(3.4) growth mechanism of child reserve pool: combining each screened sub reserve pool with a new randomly generated sub reserve pool, wherein the weight matrix of the output vector of the echo state network after combination is as follows:
Figure FDA0002924297250000047
wherein HoFor the state matrix corresponding to the reserve pool after the screening mechanism is completed, HgFor the state matrix corresponding to the growing reserve pool,
Figure FDA0002924297250000048
to merge the state matrices corresponding to the grown pools,
Figure FDA0002924297250000049
is composed of
Figure FDA00029242972500000410
An identity matrix of dimensions, wherein,
Figure FDA00029242972500000411
the total number of the merged and increased child reserve pools;
further, an output weight matrix is obtained based on the formula (14)
Figure FDA00029242972500000412
The updated mathematical expression is:
Figure FDA00029242972500000413
wherein, IoIs (N)o×nsub)×(No×nsub) Identity matrix of dimension, NoFor the number of child pools after completion of the screening mechanism, IgIs nsub×nsubIdentity matrix of dimension, ILIs (L)train-nmin)×(Ltrain-nmin) An identity matrix of dimensions.
CN202110126626.7A 2021-01-29 2021-01-29 Lake and reservoir cyanobacterial bloom prediction method based on self-organizing deep confidence echo state network Active CN112862173B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110126626.7A CN112862173B (en) 2021-01-29 2021-01-29 Lake and reservoir cyanobacterial bloom prediction method based on self-organizing deep confidence echo state network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110126626.7A CN112862173B (en) 2021-01-29 2021-01-29 Lake and reservoir cyanobacterial bloom prediction method based on self-organizing deep confidence echo state network

Publications (2)

Publication Number Publication Date
CN112862173A true CN112862173A (en) 2021-05-28
CN112862173B CN112862173B (en) 2022-10-11

Family

ID=75986842

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110126626.7A Active CN112862173B (en) 2021-01-29 2021-01-29 Lake and reservoir cyanobacterial bloom prediction method based on self-organizing deep confidence echo state network

Country Status (1)

Country Link
CN (1) CN112862173B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114282639A (en) * 2021-12-24 2022-04-05 上海应用技术大学 Water bloom early warning method based on chaos theory and BP neural network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170185892A1 (en) * 2015-12-27 2017-06-29 Beijing University Of Technology Intelligent detection method for Biochemical Oxygen Demand based on a Self-organizing Recurrent RBF Neural Network
CN107506857A (en) * 2017-08-14 2017-12-22 北京工商大学 Urban lake storehouse blue-green alga bloom multi variant based on fuzzy support vector machine
CN108416460A (en) * 2018-01-19 2018-08-17 北京工商大学 Cyanobacterial bloom prediction technique based on the random depth confidence network model of multifactor sequential-
CN109886454A (en) * 2019-01-10 2019-06-14 北京工业大学 A kind of fresh water environment wawter bloom prediction technique based on self-organizing deepness belief network and Method Using Relevance Vector Machine
CN111860306A (en) * 2020-07-19 2020-10-30 陕西师范大学 Electroencephalogram signal denoising method based on width depth echo state network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170185892A1 (en) * 2015-12-27 2017-06-29 Beijing University Of Technology Intelligent detection method for Biochemical Oxygen Demand based on a Self-organizing Recurrent RBF Neural Network
CN107506857A (en) * 2017-08-14 2017-12-22 北京工商大学 Urban lake storehouse blue-green alga bloom multi variant based on fuzzy support vector machine
CN108416460A (en) * 2018-01-19 2018-08-17 北京工商大学 Cyanobacterial bloom prediction technique based on the random depth confidence network model of multifactor sequential-
CN109886454A (en) * 2019-01-10 2019-06-14 北京工业大学 A kind of fresh water environment wawter bloom prediction technique based on self-organizing deepness belief network and Method Using Relevance Vector Machine
CN111860306A (en) * 2020-07-19 2020-10-30 陕西师范大学 Electroencephalogram signal denoising method based on width depth echo state network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
李丁园: "回声状态网络结构设计及应用研究", 《中国优秀博士论文全文数据库》 *
王磊: "回声状态网络优化设计及应用研究", 《中国优秀博士论文数据库》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114282639A (en) * 2021-12-24 2022-04-05 上海应用技术大学 Water bloom early warning method based on chaos theory and BP neural network
CN114282639B (en) * 2021-12-24 2024-02-02 上海应用技术大学 Water bloom early warning method based on chaos theory and BP neural network

Also Published As

Publication number Publication date
CN112862173B (en) 2022-10-11

Similar Documents

Publication Publication Date Title
CN107688850B (en) Deep neural network compression method
CN102622418B (en) Prediction device and equipment based on BP (Back Propagation) nerve network
CN108416755A (en) A kind of image de-noising method and system based on deep learning
CN111324990A (en) Porosity prediction method based on multilayer long-short term memory neural network model
CN108764540B (en) Water supply network pressure prediction method based on parallel LSTM series DNN
CN106022954B (en) Multiple BP neural network load prediction method based on grey correlation degree
CN109948029A (en) Based on the adaptive depth hashing image searching method of neural network
CN108416460B (en) Blue algae bloom prediction method based on multi-factor time sequence-random depth confidence network model
CN102622515B (en) A kind of weather prediction method
CN109214579B (en) BP neural network-based saline-alkali soil stability prediction method and system
CN111242380A (en) Lake (reservoir) eutrophication prediction method based on artificial intelligence algorithm
CN113761777B (en) HP-OVMD-based ultra-short-term photovoltaic power prediction method
CN113408799A (en) River total nitrogen concentration prediction method based on hybrid neural network
CN106971241A (en) The method that sewage quality data are predicted based on fuzzy neural network
CN112862173B (en) Lake and reservoir cyanobacterial bloom prediction method based on self-organizing deep confidence echo state network
Fan et al. Daily suspended sediment concentration forecast in the upper reach of Yellow River using a comprehensive integrated deep learning model
CN115640901A (en) Small sample load prediction method based on hybrid neural network and generation countermeasure
CN109408896B (en) Multi-element intelligent real-time monitoring method for anaerobic sewage treatment gas production
CN109978024B (en) Effluent BOD prediction method based on interconnected modular neural network
CN107729988A (en) Blue-green alga bloom Forecasting Methodology based on dynamic depth confidence network
Goswami et al. Automatic object recognition from satellite images using artificial neural network
CN114357877A (en) Fishpond water quality evaluation prediction system and method based on fuzzy evaluation and improved support vector machine
Akinwale Adio et al. Translated Nigeria stock market price using artificial neural network for effective prediction
CN109781951B (en) Fishpond water quality monitoring system and monitoring method
CN116681159A (en) Short-term power load prediction method based on whale optimization algorithm and DRESN

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant