CN112765894A - K-LSTM-based aluminum electrolysis cell state prediction method - Google Patents
K-LSTM-based aluminum electrolysis cell state prediction method Download PDFInfo
- Publication number
- CN112765894A CN112765894A CN202110111679.1A CN202110111679A CN112765894A CN 112765894 A CN112765894 A CN 112765894A CN 202110111679 A CN202110111679 A CN 202110111679A CN 112765894 A CN112765894 A CN 112765894A
- Authority
- CN
- China
- Prior art keywords
- lstm
- information
- gate
- model
- algorithm
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 54
- 229910052782 aluminium Inorganic materials 0.000 title claims abstract description 18
- XAGFODPZIPBFFR-UHFFFAOYSA-N aluminium Chemical compound [Al] XAGFODPZIPBFFR-UHFFFAOYSA-N 0.000 title claims abstract description 13
- 238000005868 electrolysis reaction Methods 0.000 title claims description 11
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 36
- 238000012549 training Methods 0.000 claims abstract description 19
- 238000012545 processing Methods 0.000 claims abstract description 9
- 238000012360 testing method Methods 0.000 claims abstract description 9
- 238000011478 gradient descent method Methods 0.000 claims abstract description 4
- 238000010606 normalization Methods 0.000 claims abstract description 4
- 210000004027 cell Anatomy 0.000 claims description 36
- 230000006870 function Effects 0.000 claims description 36
- 230000008569 process Effects 0.000 claims description 27
- 238000004364 calculation method Methods 0.000 claims description 15
- 210000002569 neuron Anatomy 0.000 claims description 7
- 230000008859 change Effects 0.000 claims description 6
- 230000001537 neural effect Effects 0.000 claims description 4
- 230000009471 action Effects 0.000 claims description 3
- 230000000644 propagated effect Effects 0.000 claims description 3
- 230000009467 reduction Effects 0.000 claims description 3
- 230000000717 retained effect Effects 0.000 claims description 3
- AZDRQVAHHNSJOQ-UHFFFAOYSA-N alumane Chemical compound [AlH3] AZDRQVAHHNSJOQ-UHFFFAOYSA-N 0.000 abstract description 5
- 230000000694 effects Effects 0.000 description 6
- 238000013528 artificial neural network Methods 0.000 description 5
- 238000010801 machine learning Methods 0.000 description 4
- 238000004519 manufacturing process Methods 0.000 description 4
- 238000005070 sampling Methods 0.000 description 4
- 238000002474 experimental method Methods 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 230000007774 longterm Effects 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 1
- PNEYBMLMFCGWSK-UHFFFAOYSA-N aluminium oxide Inorganic materials [O-2].[O-2].[O-2].[Al+3].[Al+3] PNEYBMLMFCGWSK-UHFFFAOYSA-N 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000010835 comparative analysis Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000007405 data analysis Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000003066 decision tree Methods 0.000 description 1
- 230000006866 deterioration Effects 0.000 description 1
- 230000008034 disappearance Effects 0.000 description 1
- 239000003792 electrolyte Substances 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000007636 ensemble learning method Methods 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 238000009440 infrastructure construction Methods 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/20—Design optimisation, verification or simulation
- G06F30/27—Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
-
- C—CHEMISTRY; METALLURGY
- C25—ELECTROLYTIC OR ELECTROPHORETIC PROCESSES; APPARATUS THEREFOR
- C25C—PROCESSES FOR THE ELECTROLYTIC PRODUCTION, RECOVERY OR REFINING OF METALS; APPARATUS THEREFOR
- C25C3/00—Electrolytic production, recovery or refining of metals by electrolysis of melts
- C25C3/06—Electrolytic production, recovery or refining of metals by electrolysis of melts of aluminium
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/30—Computing systems specially adapted for manufacturing
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Chemical & Material Sciences (AREA)
- Chemical Kinetics & Catalysis (AREA)
- Electrochemistry (AREA)
- Materials Engineering (AREA)
- Metallurgy (AREA)
- Organic Chemistry (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Computer Hardware Design (AREA)
- Geometry (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
The invention discloses a K-LSTM-based aluminum electrolytic cell state prediction method, which comprises the following steps: step 1: carrying out normalization processing on the data; step 2: constructing a training set and a testing set according to a set sliding window size m; and step 3: constructing an improved LSTM model, and initializing parameters of the model; and 4, step 4: training the prediction model by using a training set, updating parameters by adopting a gradient descent method, and iterating for several times until the precision requirement is met; and 5: and feeding the test set into a trained model, and predicting the predicted value at the t +1 moment by using historical data. Based on the improved K-LSTM algorithm, the invention eliminates the sample unbalance by setting the weight aiming at the problem of sample unbalance in the LSTML forgetting gate unit, and can effectively predict the aluminum cell state.
Description
Technical Field
The invention relates to the technical field of aluminum electrolysis industry, in particular to a K-LSTM-based aluminum electrolysis cell state prediction method.
Background
The production data of the aluminum electrolysis cell is a time series and has the characteristic of high data dimension. There are various models of existing prediction algorithms for time series data, such as artificial neural networks, autoregressive moving average, wavelet neural networks, etc. The time series prediction study begins with a regression equation that predicts the number of sunblack seeds in a year from data analysis. Autoregressive moving average models (ARMA) and autoregressive integrated moving average models (ARIMA) indicate that time series prediction models based on regression methods are becoming more and more popular.
Therefore, these models are also the simplest and most important models for time series prediction. However, due to the complexity, irregularity, randomness and nonlinearity of the actual data, it is difficult to realize high-precision prediction by a complex model. By adopting a machine learning method, a nonlinear prediction model based on a large amount of historical data can be established. In fact, through iterative training iterations and learning approximations, machine learning models can obtain more accurate predictions than traditional statistical-based models. Typical methods are support vector regression or kernel-based classification, artificial neural multi-order (ANN) and strong nonlinear function approximation and tree-based ensemble learning methods such as gradient enhanced regression or decision trees (GBRT, GBDT). However, the above method has limited effect in the time series prediction task because it lacks effective processing of the sequence dependency between the input variables.
With continuous and deep research on a deep learning algorithm, the problem that the deep learning algorithm can be suitable for predicting time series data is found, the algorithm gradually analyzes input data information, then extracts effective characteristics and extracts implicit relations from a data sequence. In order to enable RNN networks to process time series data more efficiently, a time series concept is introduced in the neural network architecture of RNN. An improved algorithm of the RNN is a long-short term memory neural network, the problems of gradient explosion, gradient disappearance, long-term sequence data memory and the like existing in the RNN structure are solved, and long-term sequence information can be effectively processed. The LSTM model is applied to many fields such as voice recognition, stock price prediction, rainfall prediction, traffic flow prediction, image and character recognition, and achieves good application effects.
Because aluminum electrolysis is an industrial process system with large time lag and the state is not changed frequently, the problem of unbalanced samples can be encountered when the existing data are used for training a neural network. The first gate of the LSTM is the forgetting gate, which is used to determine if some information is lost from the memory cell, and this process is what the forgetting gate needs to do and is handled by a sigmoid function. However, in the working process of the forgetting gate, if the state of the input data changes infrequently, the forgetting gate can be in the state 1 for a long time, namely, the last state is used, the state does not need to be updated, and the forgetting gate has the problem of unbalanced sample. I.e., for the ft model, there is a sample imbalance problem.
The sample imbalance problem mainly exists in a supervised machine learning task, and is mainly shown in that when sample imbalance data is encountered, a model tends to pay more attention to classes with a large number of classes in a classification prediction task, so that the sample prediction effect of a few classes is poor, and most common machine learning cannot effectively work on existing unbalanced data sets. There are generally two approaches to solving the sample imbalance problem, namely undersampling and oversampling.
(1) Undersampling: undersampling is the reduction of the total number of classes, and is usually chosen when the amount of data is sufficiently supported. The samples are balanced by retaining the number of samples contained in a small number of classes, then reducing the number of samples contained in a large number of classes, and then modeling further.
(2) Oversampling: when the amount of data is not sufficient to support method (1), an over-sampling method is selected that takes the addition of a small number of classes of data sets to balance the data sets, rather than removing the number of majority classes, by using a repeat, bootstrap, or composite minority over-sampling method to add the small number of classes.
However, neither over-sampling nor under-sampling methods have absolute advantages in comparison, and the application of these two methods depends on the data set to be applied.
Disclosure of Invention
The invention aims to provide a K-LSTM-based aluminum cell state prediction method to support the communication infrastructure construction mentioned in the background art and achieve the purposes of cost reduction and efficiency improvement.
In order to achieve the purpose, the invention aims at the problem of unbalanced samples in a forgotten door, adopts two methods in the background technology, judges the conditions and then respectively sets a weight, and reduces the weight when the ft model is in the same state for a long time so as to balance the sample problem.
Therefore, the K-LSTM-based aluminum electrolysis cell state prediction method specifically comprises the following steps:
step 1: carrying out normalization processing on the data;
step 2: constructing a training set and a testing set according to a set sliding window size m;
and step 3: constructing an improved LSTM model, and initializing parameters of the model;
and 4, step 4: training the prediction model by using a training set, updating parameters by adopting a gradient descent method, and iterating for several times until the precision requirement is met;
and 5: and feeding the test set into a trained model, and predicting the predicted value at the t +1 moment by using historical data.
The improved LSTM model is constructed based on the LSTM algorithm and the improved K-LSTM algorithm.
The K-LSTM algorithm is implemented as follows:
the three gate structures of the LSTM comprise an input gate, an output gate and a forgetting gate; the calculation process is as follows:
(1) forget the door: the information used for judging the discarding of the previous memory information is obtained by outputting the value h at the moment t-1t-1Input value x at the current time ttLinear combination, compressing the values to [0,1 ] by sigmoid function]In the range, the closer the value is to 1, the more f in the current cell state is representedtThe more information to be retained, the closer to 0 the value is, the more f is representedtThe more the information is selected to be abandoned, the calculation process of the forgetting gate is as follows:
ft=sigmoid(Wf·[ht-1,xt]+bf) (1)
(2) an input gate: for processing input value x at current time ttAnd the output h of the previous momentt-1The method is also realized by a sigmoid function as total input information; then use xtAnd ht-1Obtaining new candidate cell information through tanh layerThe total amount of the input information at the current moment and the memory information at the previous moment is calculated, and the calculation process is as follows:
it=sigmoid(Wi·[ht-1,xt]+bi) (2)
then, the old cell information C is updatedt-1Updated to new cell information Ct(ii) a The updating method is that the old forgotten cell information is selected by the forgotten gate and the newly added information at the current moment is added by the input gate, and the two determine the finally updated cell information CtThe calculation process is as follows:
(3) an output gate: for determining the information h to be transmitted when outputting to the next momentt(ii) a The output information also needs to obtain a judgment condition through a sigmoid function, and then the cell state is obtained through a tanh function to obtain a range of [ -1,1]A value in between; the value is a vector, and the vector is multiplied by the previously obtained judgment condition to obtain the output at the current moment. The calculation process is as follows:
ot=sigmoid(Wo[ht-1,xt]+bo) (5)
ht=ot*tanh(Ct) (6)
when the network is learned, the offset derivatives of the loss functions on all the parameters are calculated, then the parameters are updated, the iteration is carried out in sequence until the loss functions are converged, and the offset derivatives of the loss functions on ft are intervened on the basis of the original LSTM gate control unit to avoid the problem of sample imbalance; according to whether the slot state is changed, if the slot state is changed, a larger weight is selected, and if the slot state is not changed, a smaller weight is selected;
f=f*k+tf.stop_gradient(f-f*k) (7)
using tf.stop _ gradient () function, according to the characteristics of tf.stop _ gradient () function, tf.stop _ gradient () does not work in the forward process, so + (f × k) and- (f × k) cancel out, leaving only f forward pass; in the reverse process, the gradient of f-f × k becomes 0 due to the action of tf.
The improved K-LSTM algorithm comprises the following steps:
step (1): calculating input and output in the forward propagation of the K-LSTM and output values of each neuron of the hidden layer;
step (2): calculating an output error through a cross entropy function, and reversely transmitting the error to each layer of neural units through a back propagation algorithm;
and (3): in the backward propagation process of the forgetting gate, selecting a weight to change an original propagation function according to a judgment condition;
and (4): updating the parameters of each layer of neurons according to a gradient descent algorithm and the propagated error;
and (5): and (4) repeating the step (2), the step (3) and the step (4) according to the set iteration number until convergence, and finishing the model training.
Compared with the prior art, the method has the beneficial effects that: based on the improved K-LSTM algorithm, the invention eliminates the sample unbalance by setting the weight aiming at the problem of sample unbalance in the LSTML forgetting gate unit, and can effectively predict the aluminum cell state.
Drawings
FIG. 1 is a flow chart of the improved K-LSTM algorithm.
FIG. 2 is a flow chart of K-LSTM-based slot state prediction.
Fig. 3 is a schematic diagram of production data storage.
FIG. 4 is a graph of the predicted results of K-LSTM.
FIG. 5 is a graph of LSTM prediction results.
FIG. 6 is a graph of EA-LSTM prediction results.
Detailed Description
The technical solution of the present patent will be described in further detail with reference to the following embodiments.
The K-LSTM algorithm is implemented as follows:
the three gate structures of the LSTM comprise an input gate, an output gate and a forgetting gate; the calculation process is as follows:
(1) forget the door: the information used for judging the discarding of the previous memory information is obtained by outputting the value h at the moment t-1t-1Input value x at the current time ttLinear combination, compressing the values to [0,1 ] by sigmoid function]In the range, the closer the value is to 1, the more f in the current cell state is representedtThe more information to be retained, the closer to 0 the value is, the more f is representedtThe more the information is selected to be abandoned, the calculation process of the forgetting gate is as follows:
ft=sigmoid(Wf·[ht-1,xt]+bf) (1)
(2) an input gate: for processing input value x at current time ttAnd the output h of the previous momentt-1The method is also realized by a sigmoid function as total input information; then use xtAnd ht-1Obtaining new candidate cell information through tanh layerThe total amount of the input information at the current moment and the memory information at the previous moment is calculated, and the calculation process is as follows:
it=sigmoid(Wi·[ht-1,xt]+bi) (2)
then, the old cell information C is updatedt-1Updated to new cell information Ct(ii) a The updating method is that the old forgotten cell information is selected by the forgotten gate and the newly added information at the current moment is added by the input gate, and the two determine the finally updated cell information CtThe calculation process is as follows:
(3) an output gate: for determining the information h to be transmitted when outputting to the next momentt(ii) a The output information also needs to obtain a judgment condition through a sigmoid function, and then the cell state is obtained through a tanh function to obtain a range of [ -1,1]A value in between; the value is a vector, and the vector is multiplied by the previously obtained judgment condition to obtain the output at the current moment. The calculation process is as follows:
ot=sigmoid(Wo[ht-1,xt]+bo) (5)
ht=ot*tanh(Ct) (6)
when the network is learned, the offset derivatives of the loss functions on all the parameters are calculated, then the parameters are updated, the iteration is carried out in sequence until the loss functions are converged, and the offset derivatives of the loss functions on ft are intervened on the basis of the original LSTM gate control unit to avoid the problem of sample imbalance; according to whether the slot state is changed, if the slot state is changed, a larger weight is selected, and if the slot state is not changed, a smaller weight is selected;
f=f*k+tf.stop_gradient(f-f*k) (7)
using tf.stop _ gradient () function, according to the characteristics of tf.stop _ gradient () function, tf.stop _ gradient () does not work in the forward process, so + (f × k) and- (f × k) cancel out, leaving only f forward pass; in the reverse process, the gradient of f-f × k becomes 0 due to the action of tf.
The improved algorithm flow chart is shown in fig. 1, and the improved K-LSTM algorithm comprises the following steps:
step (1): calculating input and output in the forward propagation of the K-LSTM and output values of each neuron of the hidden layer;
step (2): calculating an output error through a cross entropy function, and reversely transmitting the error to each layer of neural units through a back propagation algorithm;
and (3): in the backward propagation process of the forgetting gate, selecting a weight to change an original propagation function according to a judgment condition;
and (4): updating the parameters of each layer of neurons according to a gradient descent algorithm and the propagated error;
and (5): and (4) repeating the step (2), the step (3) and the step (4) according to the set iteration number until convergence, and finishing the model training.
The modified K-LSTM algorithm is shown in Table 1:
TABLE 1 improved K-LSTM Algorithm
In this embodiment, an LSTM slot state prediction model is constructed using a Keras framework of tensrflow, all programs are written using Python language, a prediction experiment is performed on a computer of a cpu2.50ghz, memory 8GB, Windows7 operating system, and data is data with cluster attributes.
Setting a sliding window m by using an improved K-LSTM algorithm, constructing a training set and a testing set according to the sliding window, training a model, selecting cross entropy as a loss function to reflect the deviation between predicted data and real data, and finally predicting the state of the aluminum electrolysis cell by using the trained model.
To further verify the validity of the algorithm, the improved algorithm was compared with the accuracy of the conventional LSTM and attention-based LSTM algorithms, respectively.
As shown in FIG. 2, the flow of predicting the aluminum cell state using the modified K-LSTM is as follows:
step 1: carrying out normalization processing on the data;
step 2: constructing a training set and a testing set according to a set sliding window size m;
and step 3: constructing an improved LSTM model, and initializing parameters of the model;
and 4, step 4: training the prediction model by using a training set, updating parameters by adopting a gradient descent method, and iterating for several times until the precision requirement is met;
and 5: and feeding the test set into a trained model, and predicting the predicted value at the t +1 moment by using historical data.
In this example, the data is derived from real aluminum cell production data (as shown in fig. 3), and the data is collected once a day, wherein each cell contains 13 characteristics of Fe content, aluminum level, molecular ratio, Si content, alumina concentration, electrolyte level, electrolysis temperature, and the like; before analyzing and mining the information hidden in the data, the data with many problems needs to be preprocessed; because the data used in this embodiment has a null value, the data is subjected to null value and noise processing, and then normalized according to the characteristic of high data dimensionality.
The results of the prediction of the state of the aluminum electrolysis cell based on K-LSTM are shown in FIG. 4. For clearer observation effect, only the first 200 pieces of data are shown; in fig. 4, the abscissa is a time series, and the ordinate is a slot state, there being a total of two states 0 and 1; the two dotted lines are the true bin state and the model predicted bin state, where the cross entropy LOSS is 0.0418, and the accuracy reaches 99.6%.
Fig. 5 is a diagram of a prediction result of a conventional LSTM model, and it can be seen that in the prediction of the tank state by the model, the overall fitting effect of the predicted value and the true value is good when the tank state does not change, but when the tank state changes suddenly, accurate prediction cannot be made in time, and the prediction is successful after the tank state changes for a period of time.
Fig. 6 shows that the model prediction based on EA-LSTM construction of the attention mechanism is more accurate than the conventional LSTM model, but when the groove state changes, the change of the groove state is not predicted accurately, and the groove state at the previous time is generally continued.
In order to verify the effectiveness of the algorithm, the K-LSTM algorithm is compared and analyzed with the traditional LSTM algorithm and the EA-LSTM algorithm based on the attention mechanism respectively, and the same iteration number, sliding window size and neuron number are adopted.
From Table 2, it can be seen that the improved K-LSTM improves the accuracy of the prediction of the slot state to some extent, regardless of whether the traditional LSTM-constructed model or the EA-LATM model is inferior to the improved K-LSTM model in terms of error and accuracy.
TABLE 2 comparative analysis of the models
By comparison, it can be seen that the improved K-LSTM of the present invention provides a significant improvement in this problem, enabling a quicker prediction of the change in the state of the cell. The method is convenient for operators to find the abnormal state of the tank more quickly and make a decision in time to prevent the further deterioration of the tank state.
The embodiment performs experiments to verify the feasibility and the effectiveness of the algorithm, and simultaneously performs comparison experiments with the traditional LSTM model and the EA-LSTM model based on the attention mechanism, finally proves that the improved K-LSTM prediction effect is obviously superior to the other two models, and the accuracy of the groove state prediction is improved.
Although the preferred embodiments of the present patent have been described in detail, the present patent is not limited to the above embodiments, and various changes can be made without departing from the spirit of the present patent within the knowledge of those skilled in the art.
Claims (4)
1. A K-LSTM-based aluminum electrolysis cell state prediction method is characterized by comprising the following steps:
step 1: carrying out normalization processing on the data;
step 2: constructing a training set and a testing set according to a set sliding window size m;
and step 3: constructing an improved LSTM model, and initializing parameters of the model;
and 4, step 4: training the prediction model by using a training set, updating parameters by adopting a gradient descent method, and iterating for several times until the precision requirement is met;
and 5: and feeding the test set into a trained model, and predicting the predicted value at the t +1 moment by using historical data.
2. The method of claim 1, wherein the improved LSTM model is constructed based on the LSTM algorithm modified K-LSTM algorithm.
3. The method of claim 2, wherein the improved K-LSTM algorithm comprises the steps of:
step (1): calculating input and output in the forward propagation of the K-LSTM and output values of each neuron of the hidden layer;
step (2): calculating an output error through a cross entropy function, and reversely transmitting the error to each layer of neural units through a back propagation algorithm;
and (3): in the backward propagation process of the forgetting gate, selecting a weight to change an original propagation function according to a judgment condition;
and (4): updating the parameters of each layer of neurons according to a gradient descent algorithm and the propagated error;
and (5): and (4) repeating the step (2), the step (3) and the step (4) according to the set iteration number until convergence, and finishing the model training.
4. The K-LSTM based aluminum reduction cell condition prediction method of claim 1, wherein the K-LSTM algorithm is implemented as follows:
the three gate structures of the LSTM comprise an input gate, an output gate and a forgetting gate; the calculation process is as follows:
(1) forget the door: the information used for judging the discarding of the previous memory information is obtained by outputting the value h at the moment t-1t-1Input value x at the current time ttLinear combination, compressing the values to [0,1 ] by sigmoid function]In the range, the closer the value is to 1, the more f in the current cell state is representedtThe more information to be retained, the closer to 0 the value is, the more f is representedtThe more the information is selected to be abandoned, the calculation process of the forgetting gate is as follows:
ft=sigmoid(Wf·[ht-1,xt]+bf) (1)
(2) an input gate: for processing input value x at current time ttAnd the output h of the previous momentt-1The method is also realized by a sigmoid function as total input information; then use xtAnd ht-1Obtaining new candidate cell information through tanh layerThe total amount of the input information at the current moment and the memory information at the previous moment is calculated, and the calculation process is as follows:
it=sigmoid(Wi·[ht-1,xt]+bi) (2)
then, the old cell information C is updatedt-1Updated to new cell information Ct(ii) a The updating method is that the old forgotten cell information is selected by the forgotten gate and the newly added information at the current moment is added by the input gate, and the two determine the finally updated cell information CtThe calculation process is as follows:
(3) an output gate: for determining the information h to be transmitted when outputting to the next momentt(ii) a The output information also needs to obtain a judgment condition through a sigmoid function, and then the cell state is obtained through a tanh function to obtain a range of [ -1,1]A value in between; the value is a vector, then the vector is multiplied by the previously obtained judgment condition to obtain the output of the current moment, and the calculation process is as follows:
ot=sigmoid(Wo[ht-1,xt]+bo) (5)
ht=ot*tanh(Ct) (6)
when the network is learned, the offset derivatives of the loss functions on all the parameters are calculated, then the parameters are updated, the iteration is carried out in sequence until the loss functions are converged, and the offset derivatives of the loss functions on ft are intervened on the basis of the original LSTM gate control unit to avoid the problem of sample imbalance; according to whether the slot state is changed, if the slot state is changed, a larger weight is selected, and if the slot state is not changed, a smaller weight is selected;
f=f*k+tf.stop_gradient(f-f*k) (7)
using tf.stop _ gradient () function, according to the characteristics of tf.stop _ gradient () function, tf.stop _ gradient () does not work in the forward process, so + (f × k) and- (f × k) cancel out, leaving only f forward pass; in the reverse process, the gradient of f-f × k becomes 0 due to the action of tf.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2020113346304 | 2020-11-25 | ||
CN202011334630 | 2020-11-25 |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112765894A true CN112765894A (en) | 2021-05-07 |
CN112765894B CN112765894B (en) | 2023-05-05 |
Family
ID=75706135
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110111679.1A Active CN112765894B (en) | 2020-11-25 | 2021-01-27 | K-LSTM-based aluminum electrolysis cell state prediction method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112765894B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114959797A (en) * | 2022-07-04 | 2022-08-30 | 广东技术师范大学 | Aluminum electrolysis cell condition diagnosis method based on data amplification and SSKELM |
CN115081744A (en) * | 2022-07-22 | 2022-09-20 | 重庆师范大学 | Method for predicting unit energy consumption of aluminum electrolysis manufacturing system |
CN116288532A (en) * | 2023-03-13 | 2023-06-23 | 赛富能科技(深圳)有限公司 | Method and equipment for monitoring electrolytic tank |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CA2230882A1 (en) * | 1997-03-14 | 1998-09-14 | Dubai Aluminium Company Limited | Intelligent control of aluminium reduction cells using predictive and pattern recognition techniques |
CN1471627A (en) * | 2000-10-26 | 2004-01-28 | �Ʒ� | A fault tolerant liquid measurement system using multiple-model state estimators |
CN201334531Y (en) * | 2008-12-02 | 2009-10-28 | 北方工业大学 | Novel potline stop-start shunting device and system |
WO2017026010A1 (en) * | 2015-08-07 | 2017-02-16 | 三菱電機株式会社 | Device for predicting amount of photovoltaic power generation, and method for predicting amount of photovoltaic power generation |
CN109543699A (en) * | 2018-11-28 | 2019-03-29 | 北方工业大学 | Image abstract generation method based on target detection |
CN109614885A (en) * | 2018-11-21 | 2019-04-12 | 齐鲁工业大学 | A kind of EEG signals Fast Classification recognition methods based on LSTM |
CN110770760A (en) * | 2017-05-19 | 2020-02-07 | 渊慧科技有限公司 | Object-level prediction of future states of a physical system |
WO2020075767A1 (en) * | 2018-10-10 | 2020-04-16 | 旭化成株式会社 | Planning device, planning method, and planning program |
CN111563706A (en) * | 2020-03-05 | 2020-08-21 | 河海大学 | Multivariable logistics freight volume prediction method based on LSTM network |
US20200348662A1 (en) * | 2016-05-09 | 2020-11-05 | Strong Force Iot Portfolio 2016, Llc | Platform for facilitating development of intelligence in an industrial internet of things system |
-
2021
- 2021-01-27 CN CN202110111679.1A patent/CN112765894B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CA2230882A1 (en) * | 1997-03-14 | 1998-09-14 | Dubai Aluminium Company Limited | Intelligent control of aluminium reduction cells using predictive and pattern recognition techniques |
CN1471627A (en) * | 2000-10-26 | 2004-01-28 | �Ʒ� | A fault tolerant liquid measurement system using multiple-model state estimators |
CN201334531Y (en) * | 2008-12-02 | 2009-10-28 | 北方工业大学 | Novel potline stop-start shunting device and system |
WO2017026010A1 (en) * | 2015-08-07 | 2017-02-16 | 三菱電機株式会社 | Device for predicting amount of photovoltaic power generation, and method for predicting amount of photovoltaic power generation |
US20200348662A1 (en) * | 2016-05-09 | 2020-11-05 | Strong Force Iot Portfolio 2016, Llc | Platform for facilitating development of intelligence in an industrial internet of things system |
CN110770760A (en) * | 2017-05-19 | 2020-02-07 | 渊慧科技有限公司 | Object-level prediction of future states of a physical system |
WO2020075767A1 (en) * | 2018-10-10 | 2020-04-16 | 旭化成株式会社 | Planning device, planning method, and planning program |
CN109614885A (en) * | 2018-11-21 | 2019-04-12 | 齐鲁工业大学 | A kind of EEG signals Fast Classification recognition methods based on LSTM |
CN109543699A (en) * | 2018-11-28 | 2019-03-29 | 北方工业大学 | Image abstract generation method based on target detection |
CN111563706A (en) * | 2020-03-05 | 2020-08-21 | 河海大学 | Multivariable logistics freight volume prediction method based on LSTM network |
Non-Patent Citations (2)
Title |
---|
侯婕等: "基于LSTM的铝电解槽况预测" * |
孔淑麒: "铝电解槽状态预测算法研究" * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114959797A (en) * | 2022-07-04 | 2022-08-30 | 广东技术师范大学 | Aluminum electrolysis cell condition diagnosis method based on data amplification and SSKELM |
CN115081744A (en) * | 2022-07-22 | 2022-09-20 | 重庆师范大学 | Method for predicting unit energy consumption of aluminum electrolysis manufacturing system |
CN115081744B (en) * | 2022-07-22 | 2024-05-07 | 重庆师范大学 | Unit energy consumption prediction method for aluminum electrolysis manufacturing system |
CN116288532A (en) * | 2023-03-13 | 2023-06-23 | 赛富能科技(深圳)有限公司 | Method and equipment for monitoring electrolytic tank |
Also Published As
Publication number | Publication date |
---|---|
CN112765894B (en) | 2023-05-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112765894B (en) | K-LSTM-based aluminum electrolysis cell state prediction method | |
CN111860785A (en) | Time sequence prediction method and system based on attention mechanism cyclic neural network | |
CN112990556A (en) | User power consumption prediction method based on Prophet-LSTM model | |
CN110084610B (en) | Network transaction fraud detection system based on twin neural network | |
CN110909926A (en) | TCN-LSTM-based solar photovoltaic power generation prediction method | |
CN110782658B (en) | Traffic prediction method based on LightGBM algorithm | |
CN111079931A (en) | State space probabilistic multi-time-series prediction method based on graph neural network | |
CN111277434A (en) | Network flow multi-step prediction method based on VMD and LSTM | |
CN112053560B (en) | Short-time traffic flow prediction method, system and storage medium based on neural network | |
CN113642225A (en) | CNN-LSTM short-term wind power prediction method based on attention mechanism | |
CN111860989B (en) | LSTM neural network short-time traffic flow prediction method based on ant colony optimization | |
CN113487855B (en) | Traffic flow prediction method based on EMD-GAN neural network structure | |
EP3792841A1 (en) | Automated feature generation for machine learning application | |
CN112232604B (en) | Prediction method for extracting network traffic based on Prophet model | |
CN113449919B (en) | Power consumption prediction method and system based on feature and trend perception | |
CN112766603A (en) | Traffic flow prediction method, system, computer device and storage medium | |
CN116542701A (en) | Carbon price prediction method and system based on CNN-LSTM combination model | |
CN113052373A (en) | Monthly runoff change trend prediction method based on improved ELM model | |
CN114500004A (en) | Anomaly detection method based on conditional diffusion probability generation model | |
CN116665483A (en) | Novel method for predicting residual parking space | |
CN114219531A (en) | Waste mobile phone dynamic pricing method based on M-WU concept drift detection | |
CN116303786B (en) | Block chain financial big data management system based on multidimensional data fusion algorithm | |
CN107979606A (en) | It is a kind of that there is adaptive distributed intelligence decision-making technique | |
CN111524348A (en) | Long-short term traffic flow prediction model and method | |
CN109033413B (en) | Neural network-based demand document and service document matching method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |