CN112765894A - K-LSTM-based aluminum electrolysis cell state prediction method - Google Patents

K-LSTM-based aluminum electrolysis cell state prediction method Download PDF

Info

Publication number
CN112765894A
CN112765894A CN202110111679.1A CN202110111679A CN112765894A CN 112765894 A CN112765894 A CN 112765894A CN 202110111679 A CN202110111679 A CN 202110111679A CN 112765894 A CN112765894 A CN 112765894A
Authority
CN
China
Prior art keywords
lstm
information
gate
model
algorithm
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110111679.1A
Other languages
Chinese (zh)
Other versions
CN112765894B (en
Inventor
曹丹阳
田学法
陈云博
孔淑麒
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
North China University of Technology
Original Assignee
North China University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by North China University of Technology filed Critical North China University of Technology
Publication of CN112765894A publication Critical patent/CN112765894A/en
Application granted granted Critical
Publication of CN112765894B publication Critical patent/CN112765894B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • G06F30/27Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
    • CCHEMISTRY; METALLURGY
    • C25ELECTROLYTIC OR ELECTROPHORETIC PROCESSES; APPARATUS THEREFOR
    • C25CPROCESSES FOR THE ELECTROLYTIC PRODUCTION, RECOVERY OR REFINING OF METALS; APPARATUS THEREFOR
    • C25C3/00Electrolytic production, recovery or refining of metals by electrolysis of melts
    • C25C3/06Electrolytic production, recovery or refining of metals by electrolysis of melts of aluminium
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/30Computing systems specially adapted for manufacturing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Software Systems (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Chemical & Material Sciences (AREA)
  • Chemical Kinetics & Catalysis (AREA)
  • Electrochemistry (AREA)
  • Materials Engineering (AREA)
  • Metallurgy (AREA)
  • Organic Chemistry (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Computer Hardware Design (AREA)
  • Geometry (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a K-LSTM-based aluminum electrolytic cell state prediction method, which comprises the following steps: step 1: carrying out normalization processing on the data; step 2: constructing a training set and a testing set according to a set sliding window size m; and step 3: constructing an improved LSTM model, and initializing parameters of the model; and 4, step 4: training the prediction model by using a training set, updating parameters by adopting a gradient descent method, and iterating for several times until the precision requirement is met; and 5: and feeding the test set into a trained model, and predicting the predicted value at the t +1 moment by using historical data. Based on the improved K-LSTM algorithm, the invention eliminates the sample unbalance by setting the weight aiming at the problem of sample unbalance in the LSTML forgetting gate unit, and can effectively predict the aluminum cell state.

Description

K-LSTM-based aluminum electrolysis cell state prediction method
Technical Field
The invention relates to the technical field of aluminum electrolysis industry, in particular to a K-LSTM-based aluminum electrolysis cell state prediction method.
Background
The production data of the aluminum electrolysis cell is a time series and has the characteristic of high data dimension. There are various models of existing prediction algorithms for time series data, such as artificial neural networks, autoregressive moving average, wavelet neural networks, etc. The time series prediction study begins with a regression equation that predicts the number of sunblack seeds in a year from data analysis. Autoregressive moving average models (ARMA) and autoregressive integrated moving average models (ARIMA) indicate that time series prediction models based on regression methods are becoming more and more popular.
Therefore, these models are also the simplest and most important models for time series prediction. However, due to the complexity, irregularity, randomness and nonlinearity of the actual data, it is difficult to realize high-precision prediction by a complex model. By adopting a machine learning method, a nonlinear prediction model based on a large amount of historical data can be established. In fact, through iterative training iterations and learning approximations, machine learning models can obtain more accurate predictions than traditional statistical-based models. Typical methods are support vector regression or kernel-based classification, artificial neural multi-order (ANN) and strong nonlinear function approximation and tree-based ensemble learning methods such as gradient enhanced regression or decision trees (GBRT, GBDT). However, the above method has limited effect in the time series prediction task because it lacks effective processing of the sequence dependency between the input variables.
With continuous and deep research on a deep learning algorithm, the problem that the deep learning algorithm can be suitable for predicting time series data is found, the algorithm gradually analyzes input data information, then extracts effective characteristics and extracts implicit relations from a data sequence. In order to enable RNN networks to process time series data more efficiently, a time series concept is introduced in the neural network architecture of RNN. An improved algorithm of the RNN is a long-short term memory neural network, the problems of gradient explosion, gradient disappearance, long-term sequence data memory and the like existing in the RNN structure are solved, and long-term sequence information can be effectively processed. The LSTM model is applied to many fields such as voice recognition, stock price prediction, rainfall prediction, traffic flow prediction, image and character recognition, and achieves good application effects.
Because aluminum electrolysis is an industrial process system with large time lag and the state is not changed frequently, the problem of unbalanced samples can be encountered when the existing data are used for training a neural network. The first gate of the LSTM is the forgetting gate, which is used to determine if some information is lost from the memory cell, and this process is what the forgetting gate needs to do and is handled by a sigmoid function. However, in the working process of the forgetting gate, if the state of the input data changes infrequently, the forgetting gate can be in the state 1 for a long time, namely, the last state is used, the state does not need to be updated, and the forgetting gate has the problem of unbalanced sample. I.e., for the ft model, there is a sample imbalance problem.
The sample imbalance problem mainly exists in a supervised machine learning task, and is mainly shown in that when sample imbalance data is encountered, a model tends to pay more attention to classes with a large number of classes in a classification prediction task, so that the sample prediction effect of a few classes is poor, and most common machine learning cannot effectively work on existing unbalanced data sets. There are generally two approaches to solving the sample imbalance problem, namely undersampling and oversampling.
(1) Undersampling: undersampling is the reduction of the total number of classes, and is usually chosen when the amount of data is sufficiently supported. The samples are balanced by retaining the number of samples contained in a small number of classes, then reducing the number of samples contained in a large number of classes, and then modeling further.
(2) Oversampling: when the amount of data is not sufficient to support method (1), an over-sampling method is selected that takes the addition of a small number of classes of data sets to balance the data sets, rather than removing the number of majority classes, by using a repeat, bootstrap, or composite minority over-sampling method to add the small number of classes.
However, neither over-sampling nor under-sampling methods have absolute advantages in comparison, and the application of these two methods depends on the data set to be applied.
Disclosure of Invention
The invention aims to provide a K-LSTM-based aluminum cell state prediction method to support the communication infrastructure construction mentioned in the background art and achieve the purposes of cost reduction and efficiency improvement.
In order to achieve the purpose, the invention aims at the problem of unbalanced samples in a forgotten door, adopts two methods in the background technology, judges the conditions and then respectively sets a weight, and reduces the weight when the ft model is in the same state for a long time so as to balance the sample problem.
Therefore, the K-LSTM-based aluminum electrolysis cell state prediction method specifically comprises the following steps:
step 1: carrying out normalization processing on the data;
step 2: constructing a training set and a testing set according to a set sliding window size m;
and step 3: constructing an improved LSTM model, and initializing parameters of the model;
and 4, step 4: training the prediction model by using a training set, updating parameters by adopting a gradient descent method, and iterating for several times until the precision requirement is met;
and 5: and feeding the test set into a trained model, and predicting the predicted value at the t +1 moment by using historical data.
The improved LSTM model is constructed based on the LSTM algorithm and the improved K-LSTM algorithm.
The K-LSTM algorithm is implemented as follows:
the three gate structures of the LSTM comprise an input gate, an output gate and a forgetting gate; the calculation process is as follows:
(1) forget the door: the information used for judging the discarding of the previous memory information is obtained by outputting the value h at the moment t-1t-1Input value x at the current time ttLinear combination, compressing the values to [0,1 ] by sigmoid function]In the range, the closer the value is to 1, the more f in the current cell state is representedtThe more information to be retained, the closer to 0 the value is, the more f is representedtThe more the information is selected to be abandoned, the calculation process of the forgetting gate is as follows:
ft=sigmoid(Wf·[ht-1,xt]+bf) (1)
(2) an input gate: for processing input value x at current time ttAnd the output h of the previous momentt-1The method is also realized by a sigmoid function as total input information; then use xtAnd ht-1Obtaining new candidate cell information through tanh layer
Figure BDA0002919510980000031
The total amount of the input information at the current moment and the memory information at the previous moment is calculated, and the calculation process is as follows:
it=sigmoid(Wi·[ht-1,xt]+bi) (2)
Figure BDA0002919510980000032
then, the old cell information C is updatedt-1Updated to new cell information Ct(ii) a The updating method is that the old forgotten cell information is selected by the forgotten gate and the newly added information at the current moment is added by the input gate, and the two determine the finally updated cell information CtThe calculation process is as follows:
Figure BDA0002919510980000033
(3) an output gate: for determining the information h to be transmitted when outputting to the next momentt(ii) a The output information also needs to obtain a judgment condition through a sigmoid function, and then the cell state is obtained through a tanh function to obtain a range of [ -1,1]A value in between; the value is a vector, and the vector is multiplied by the previously obtained judgment condition to obtain the output at the current moment. The calculation process is as follows:
ot=sigmoid(Wo[ht-1,xt]+bo) (5)
ht=ot*tanh(Ct) (6)
when the network is learned, the offset derivatives of the loss functions on all the parameters are calculated, then the parameters are updated, the iteration is carried out in sequence until the loss functions are converged, and the offset derivatives of the loss functions on ft are intervened on the basis of the original LSTM gate control unit to avoid the problem of sample imbalance; according to whether the slot state is changed, if the slot state is changed, a larger weight is selected, and if the slot state is not changed, a smaller weight is selected;
f=f*k+tf.stop_gradient(f-f*k) (7)
using tf.stop _ gradient () function, according to the characteristics of tf.stop _ gradient () function, tf.stop _ gradient () does not work in the forward process, so + (f × k) and- (f × k) cancel out, leaving only f forward pass; in the reverse process, the gradient of f-f × k becomes 0 due to the action of tf.
The improved K-LSTM algorithm comprises the following steps:
step (1): calculating input and output in the forward propagation of the K-LSTM and output values of each neuron of the hidden layer;
step (2): calculating an output error through a cross entropy function, and reversely transmitting the error to each layer of neural units through a back propagation algorithm;
and (3): in the backward propagation process of the forgetting gate, selecting a weight to change an original propagation function according to a judgment condition;
and (4): updating the parameters of each layer of neurons according to a gradient descent algorithm and the propagated error;
and (5): and (4) repeating the step (2), the step (3) and the step (4) according to the set iteration number until convergence, and finishing the model training.
Compared with the prior art, the method has the beneficial effects that: based on the improved K-LSTM algorithm, the invention eliminates the sample unbalance by setting the weight aiming at the problem of sample unbalance in the LSTML forgetting gate unit, and can effectively predict the aluminum cell state.
Drawings
FIG. 1 is a flow chart of the improved K-LSTM algorithm.
FIG. 2 is a flow chart of K-LSTM-based slot state prediction.
Fig. 3 is a schematic diagram of production data storage.
FIG. 4 is a graph of the predicted results of K-LSTM.
FIG. 5 is a graph of LSTM prediction results.
FIG. 6 is a graph of EA-LSTM prediction results.
Detailed Description
The technical solution of the present patent will be described in further detail with reference to the following embodiments.
The K-LSTM algorithm is implemented as follows:
the three gate structures of the LSTM comprise an input gate, an output gate and a forgetting gate; the calculation process is as follows:
(1) forget the door: the information used for judging the discarding of the previous memory information is obtained by outputting the value h at the moment t-1t-1Input value x at the current time ttLinear combination, compressing the values to [0,1 ] by sigmoid function]In the range, the closer the value is to 1, the more f in the current cell state is representedtThe more information to be retained, the closer to 0 the value is, the more f is representedtThe more the information is selected to be abandoned, the calculation process of the forgetting gate is as follows:
ft=sigmoid(Wf·[ht-1,xt]+bf) (1)
(2) an input gate: for processing input value x at current time ttAnd the output h of the previous momentt-1The method is also realized by a sigmoid function as total input information; then use xtAnd ht-1Obtaining new candidate cell information through tanh layer
Figure BDA0002919510980000041
The total amount of the input information at the current moment and the memory information at the previous moment is calculated, and the calculation process is as follows:
it=sigmoid(Wi·[ht-1,xt]+bi) (2)
Figure BDA0002919510980000051
then, the old cell information C is updatedt-1Updated to new cell information Ct(ii) a The updating method is that the old forgotten cell information is selected by the forgotten gate and the newly added information at the current moment is added by the input gate, and the two determine the finally updated cell information CtThe calculation process is as follows:
Figure BDA0002919510980000052
(3) an output gate: for determining the information h to be transmitted when outputting to the next momentt(ii) a The output information also needs to obtain a judgment condition through a sigmoid function, and then the cell state is obtained through a tanh function to obtain a range of [ -1,1]A value in between; the value is a vector, and the vector is multiplied by the previously obtained judgment condition to obtain the output at the current moment. The calculation process is as follows:
ot=sigmoid(Wo[ht-1,xt]+bo) (5)
ht=ot*tanh(Ct) (6)
when the network is learned, the offset derivatives of the loss functions on all the parameters are calculated, then the parameters are updated, the iteration is carried out in sequence until the loss functions are converged, and the offset derivatives of the loss functions on ft are intervened on the basis of the original LSTM gate control unit to avoid the problem of sample imbalance; according to whether the slot state is changed, if the slot state is changed, a larger weight is selected, and if the slot state is not changed, a smaller weight is selected;
f=f*k+tf.stop_gradient(f-f*k) (7)
using tf.stop _ gradient () function, according to the characteristics of tf.stop _ gradient () function, tf.stop _ gradient () does not work in the forward process, so + (f × k) and- (f × k) cancel out, leaving only f forward pass; in the reverse process, the gradient of f-f × k becomes 0 due to the action of tf.
The improved algorithm flow chart is shown in fig. 1, and the improved K-LSTM algorithm comprises the following steps:
step (1): calculating input and output in the forward propagation of the K-LSTM and output values of each neuron of the hidden layer;
step (2): calculating an output error through a cross entropy function, and reversely transmitting the error to each layer of neural units through a back propagation algorithm;
and (3): in the backward propagation process of the forgetting gate, selecting a weight to change an original propagation function according to a judgment condition;
and (4): updating the parameters of each layer of neurons according to a gradient descent algorithm and the propagated error;
and (5): and (4) repeating the step (2), the step (3) and the step (4) according to the set iteration number until convergence, and finishing the model training.
The modified K-LSTM algorithm is shown in Table 1:
TABLE 1 improved K-LSTM Algorithm
Figure BDA0002919510980000053
Figure BDA0002919510980000061
In this embodiment, an LSTM slot state prediction model is constructed using a Keras framework of tensrflow, all programs are written using Python language, a prediction experiment is performed on a computer of a cpu2.50ghz, memory 8GB, Windows7 operating system, and data is data with cluster attributes.
Setting a sliding window m by using an improved K-LSTM algorithm, constructing a training set and a testing set according to the sliding window, training a model, selecting cross entropy as a loss function to reflect the deviation between predicted data and real data, and finally predicting the state of the aluminum electrolysis cell by using the trained model.
To further verify the validity of the algorithm, the improved algorithm was compared with the accuracy of the conventional LSTM and attention-based LSTM algorithms, respectively.
As shown in FIG. 2, the flow of predicting the aluminum cell state using the modified K-LSTM is as follows:
step 1: carrying out normalization processing on the data;
step 2: constructing a training set and a testing set according to a set sliding window size m;
and step 3: constructing an improved LSTM model, and initializing parameters of the model;
and 4, step 4: training the prediction model by using a training set, updating parameters by adopting a gradient descent method, and iterating for several times until the precision requirement is met;
and 5: and feeding the test set into a trained model, and predicting the predicted value at the t +1 moment by using historical data.
In this example, the data is derived from real aluminum cell production data (as shown in fig. 3), and the data is collected once a day, wherein each cell contains 13 characteristics of Fe content, aluminum level, molecular ratio, Si content, alumina concentration, electrolyte level, electrolysis temperature, and the like; before analyzing and mining the information hidden in the data, the data with many problems needs to be preprocessed; because the data used in this embodiment has a null value, the data is subjected to null value and noise processing, and then normalized according to the characteristic of high data dimensionality.
The results of the prediction of the state of the aluminum electrolysis cell based on K-LSTM are shown in FIG. 4. For clearer observation effect, only the first 200 pieces of data are shown; in fig. 4, the abscissa is a time series, and the ordinate is a slot state, there being a total of two states 0 and 1; the two dotted lines are the true bin state and the model predicted bin state, where the cross entropy LOSS is 0.0418, and the accuracy reaches 99.6%.
Fig. 5 is a diagram of a prediction result of a conventional LSTM model, and it can be seen that in the prediction of the tank state by the model, the overall fitting effect of the predicted value and the true value is good when the tank state does not change, but when the tank state changes suddenly, accurate prediction cannot be made in time, and the prediction is successful after the tank state changes for a period of time.
Fig. 6 shows that the model prediction based on EA-LSTM construction of the attention mechanism is more accurate than the conventional LSTM model, but when the groove state changes, the change of the groove state is not predicted accurately, and the groove state at the previous time is generally continued.
In order to verify the effectiveness of the algorithm, the K-LSTM algorithm is compared and analyzed with the traditional LSTM algorithm and the EA-LSTM algorithm based on the attention mechanism respectively, and the same iteration number, sliding window size and neuron number are adopted.
From Table 2, it can be seen that the improved K-LSTM improves the accuracy of the prediction of the slot state to some extent, regardless of whether the traditional LSTM-constructed model or the EA-LATM model is inferior to the improved K-LSTM model in terms of error and accuracy.
TABLE 2 comparative analysis of the models
Figure BDA0002919510980000071
By comparison, it can be seen that the improved K-LSTM of the present invention provides a significant improvement in this problem, enabling a quicker prediction of the change in the state of the cell. The method is convenient for operators to find the abnormal state of the tank more quickly and make a decision in time to prevent the further deterioration of the tank state.
The embodiment performs experiments to verify the feasibility and the effectiveness of the algorithm, and simultaneously performs comparison experiments with the traditional LSTM model and the EA-LSTM model based on the attention mechanism, finally proves that the improved K-LSTM prediction effect is obviously superior to the other two models, and the accuracy of the groove state prediction is improved.
Although the preferred embodiments of the present patent have been described in detail, the present patent is not limited to the above embodiments, and various changes can be made without departing from the spirit of the present patent within the knowledge of those skilled in the art.

Claims (4)

1. A K-LSTM-based aluminum electrolysis cell state prediction method is characterized by comprising the following steps:
step 1: carrying out normalization processing on the data;
step 2: constructing a training set and a testing set according to a set sliding window size m;
and step 3: constructing an improved LSTM model, and initializing parameters of the model;
and 4, step 4: training the prediction model by using a training set, updating parameters by adopting a gradient descent method, and iterating for several times until the precision requirement is met;
and 5: and feeding the test set into a trained model, and predicting the predicted value at the t +1 moment by using historical data.
2. The method of claim 1, wherein the improved LSTM model is constructed based on the LSTM algorithm modified K-LSTM algorithm.
3. The method of claim 2, wherein the improved K-LSTM algorithm comprises the steps of:
step (1): calculating input and output in the forward propagation of the K-LSTM and output values of each neuron of the hidden layer;
step (2): calculating an output error through a cross entropy function, and reversely transmitting the error to each layer of neural units through a back propagation algorithm;
and (3): in the backward propagation process of the forgetting gate, selecting a weight to change an original propagation function according to a judgment condition;
and (4): updating the parameters of each layer of neurons according to a gradient descent algorithm and the propagated error;
and (5): and (4) repeating the step (2), the step (3) and the step (4) according to the set iteration number until convergence, and finishing the model training.
4. The K-LSTM based aluminum reduction cell condition prediction method of claim 1, wherein the K-LSTM algorithm is implemented as follows:
the three gate structures of the LSTM comprise an input gate, an output gate and a forgetting gate; the calculation process is as follows:
(1) forget the door: the information used for judging the discarding of the previous memory information is obtained by outputting the value h at the moment t-1t-1Input value x at the current time ttLinear combination, compressing the values to [0,1 ] by sigmoid function]In the range, the closer the value is to 1, the more f in the current cell state is representedtThe more information to be retained, the closer to 0 the value is, the more f is representedtThe more the information is selected to be abandoned, the calculation process of the forgetting gate is as follows:
ft=sigmoid(Wf·[ht-1,xt]+bf) (1)
(2) an input gate: for processing input value x at current time ttAnd the output h of the previous momentt-1The method is also realized by a sigmoid function as total input information; then use xtAnd ht-1Obtaining new candidate cell information through tanh layer
Figure FDA0002919510970000011
The total amount of the input information at the current moment and the memory information at the previous moment is calculated, and the calculation process is as follows:
it=sigmoid(Wi·[ht-1,xt]+bi) (2)
Figure FDA0002919510970000021
then, the old cell information C is updatedt-1Updated to new cell information Ct(ii) a The updating method is that the old forgotten cell information is selected by the forgotten gate and the newly added information at the current moment is added by the input gate, and the two determine the finally updated cell information CtThe calculation process is as follows:
Figure FDA0002919510970000022
(3) an output gate: for determining the information h to be transmitted when outputting to the next momentt(ii) a The output information also needs to obtain a judgment condition through a sigmoid function, and then the cell state is obtained through a tanh function to obtain a range of [ -1,1]A value in between; the value is a vector, then the vector is multiplied by the previously obtained judgment condition to obtain the output of the current moment, and the calculation process is as follows:
ot=sigmoid(Wo[ht-1,xt]+bo) (5)
ht=ot*tanh(Ct) (6)
when the network is learned, the offset derivatives of the loss functions on all the parameters are calculated, then the parameters are updated, the iteration is carried out in sequence until the loss functions are converged, and the offset derivatives of the loss functions on ft are intervened on the basis of the original LSTM gate control unit to avoid the problem of sample imbalance; according to whether the slot state is changed, if the slot state is changed, a larger weight is selected, and if the slot state is not changed, a smaller weight is selected;
f=f*k+tf.stop_gradient(f-f*k) (7)
using tf.stop _ gradient () function, according to the characteristics of tf.stop _ gradient () function, tf.stop _ gradient () does not work in the forward process, so + (f × k) and- (f × k) cancel out, leaving only f forward pass; in the reverse process, the gradient of f-f × k becomes 0 due to the action of tf.
CN202110111679.1A 2020-11-25 2021-01-27 K-LSTM-based aluminum electrolysis cell state prediction method Active CN112765894B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN2020113346304 2020-11-25
CN202011334630 2020-11-25

Publications (2)

Publication Number Publication Date
CN112765894A true CN112765894A (en) 2021-05-07
CN112765894B CN112765894B (en) 2023-05-05

Family

ID=75706135

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110111679.1A Active CN112765894B (en) 2020-11-25 2021-01-27 K-LSTM-based aluminum electrolysis cell state prediction method

Country Status (1)

Country Link
CN (1) CN112765894B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114959797A (en) * 2022-07-04 2022-08-30 广东技术师范大学 Aluminum electrolysis cell condition diagnosis method based on data amplification and SSKELM
CN115081744A (en) * 2022-07-22 2022-09-20 重庆师范大学 Method for predicting unit energy consumption of aluminum electrolysis manufacturing system
CN116288532A (en) * 2023-03-13 2023-06-23 赛富能科技(深圳)有限公司 Method and equipment for monitoring electrolytic tank

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2230882A1 (en) * 1997-03-14 1998-09-14 Dubai Aluminium Company Limited Intelligent control of aluminium reduction cells using predictive and pattern recognition techniques
CN1471627A (en) * 2000-10-26 2004-01-28 �Ʒ� A fault tolerant liquid measurement system using multiple-model state estimators
CN201334531Y (en) * 2008-12-02 2009-10-28 北方工业大学 Novel potline stop-start shunting device and system
WO2017026010A1 (en) * 2015-08-07 2017-02-16 三菱電機株式会社 Device for predicting amount of photovoltaic power generation, and method for predicting amount of photovoltaic power generation
CN109543699A (en) * 2018-11-28 2019-03-29 北方工业大学 Image abstract generation method based on target detection
CN109614885A (en) * 2018-11-21 2019-04-12 齐鲁工业大学 A kind of EEG signals Fast Classification recognition methods based on LSTM
CN110770760A (en) * 2017-05-19 2020-02-07 渊慧科技有限公司 Object-level prediction of future states of a physical system
WO2020075767A1 (en) * 2018-10-10 2020-04-16 旭化成株式会社 Planning device, planning method, and planning program
CN111563706A (en) * 2020-03-05 2020-08-21 河海大学 Multivariable logistics freight volume prediction method based on LSTM network
US20200348662A1 (en) * 2016-05-09 2020-11-05 Strong Force Iot Portfolio 2016, Llc Platform for facilitating development of intelligence in an industrial internet of things system

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2230882A1 (en) * 1997-03-14 1998-09-14 Dubai Aluminium Company Limited Intelligent control of aluminium reduction cells using predictive and pattern recognition techniques
CN1471627A (en) * 2000-10-26 2004-01-28 �Ʒ� A fault tolerant liquid measurement system using multiple-model state estimators
CN201334531Y (en) * 2008-12-02 2009-10-28 北方工业大学 Novel potline stop-start shunting device and system
WO2017026010A1 (en) * 2015-08-07 2017-02-16 三菱電機株式会社 Device for predicting amount of photovoltaic power generation, and method for predicting amount of photovoltaic power generation
US20200348662A1 (en) * 2016-05-09 2020-11-05 Strong Force Iot Portfolio 2016, Llc Platform for facilitating development of intelligence in an industrial internet of things system
CN110770760A (en) * 2017-05-19 2020-02-07 渊慧科技有限公司 Object-level prediction of future states of a physical system
WO2020075767A1 (en) * 2018-10-10 2020-04-16 旭化成株式会社 Planning device, planning method, and planning program
CN109614885A (en) * 2018-11-21 2019-04-12 齐鲁工业大学 A kind of EEG signals Fast Classification recognition methods based on LSTM
CN109543699A (en) * 2018-11-28 2019-03-29 北方工业大学 Image abstract generation method based on target detection
CN111563706A (en) * 2020-03-05 2020-08-21 河海大学 Multivariable logistics freight volume prediction method based on LSTM network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
侯婕等: "基于LSTM的铝电解槽况预测" *
孔淑麒: "铝电解槽状态预测算法研究" *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114959797A (en) * 2022-07-04 2022-08-30 广东技术师范大学 Aluminum electrolysis cell condition diagnosis method based on data amplification and SSKELM
CN115081744A (en) * 2022-07-22 2022-09-20 重庆师范大学 Method for predicting unit energy consumption of aluminum electrolysis manufacturing system
CN115081744B (en) * 2022-07-22 2024-05-07 重庆师范大学 Unit energy consumption prediction method for aluminum electrolysis manufacturing system
CN116288532A (en) * 2023-03-13 2023-06-23 赛富能科技(深圳)有限公司 Method and equipment for monitoring electrolytic tank

Also Published As

Publication number Publication date
CN112765894B (en) 2023-05-05

Similar Documents

Publication Publication Date Title
CN112765894B (en) K-LSTM-based aluminum electrolysis cell state prediction method
CN111860785A (en) Time sequence prediction method and system based on attention mechanism cyclic neural network
CN112990556A (en) User power consumption prediction method based on Prophet-LSTM model
CN110084610B (en) Network transaction fraud detection system based on twin neural network
CN110909926A (en) TCN-LSTM-based solar photovoltaic power generation prediction method
CN110782658B (en) Traffic prediction method based on LightGBM algorithm
CN111079931A (en) State space probabilistic multi-time-series prediction method based on graph neural network
CN111277434A (en) Network flow multi-step prediction method based on VMD and LSTM
CN112053560B (en) Short-time traffic flow prediction method, system and storage medium based on neural network
CN113642225A (en) CNN-LSTM short-term wind power prediction method based on attention mechanism
CN111860989B (en) LSTM neural network short-time traffic flow prediction method based on ant colony optimization
CN113487855B (en) Traffic flow prediction method based on EMD-GAN neural network structure
EP3792841A1 (en) Automated feature generation for machine learning application
CN112232604B (en) Prediction method for extracting network traffic based on Prophet model
CN113449919B (en) Power consumption prediction method and system based on feature and trend perception
CN112766603A (en) Traffic flow prediction method, system, computer device and storage medium
CN116542701A (en) Carbon price prediction method and system based on CNN-LSTM combination model
CN113052373A (en) Monthly runoff change trend prediction method based on improved ELM model
CN114500004A (en) Anomaly detection method based on conditional diffusion probability generation model
CN116665483A (en) Novel method for predicting residual parking space
CN114219531A (en) Waste mobile phone dynamic pricing method based on M-WU concept drift detection
CN116303786B (en) Block chain financial big data management system based on multidimensional data fusion algorithm
CN107979606A (en) It is a kind of that there is adaptive distributed intelligence decision-making technique
CN111524348A (en) Long-short term traffic flow prediction model and method
CN109033413B (en) Neural network-based demand document and service document matching method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant