CN116610919A - Space time sequence prediction method based on cyclic graph algorithm sub-neural network - Google Patents
Space time sequence prediction method based on cyclic graph algorithm sub-neural network Download PDFInfo
- Publication number
- CN116610919A CN116610919A CN202310630846.2A CN202310630846A CN116610919A CN 116610919 A CN116610919 A CN 116610919A CN 202310630846 A CN202310630846 A CN 202310630846A CN 116610919 A CN116610919 A CN 116610919A
- Authority
- CN
- China
- Prior art keywords
- graph
- representing
- neural network
- sub
- operator
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 33
- 238000000034 method Methods 0.000 title claims abstract description 30
- 125000004122 cyclic group Chemical group 0.000 title claims abstract description 17
- 238000004422 calculation algorithm Methods 0.000 title claims description 8
- 238000004364 calculation method Methods 0.000 claims abstract description 12
- 238000012545 processing Methods 0.000 claims abstract description 7
- 230000004931 aggregating effect Effects 0.000 claims abstract description 5
- 239000011159 matrix material Substances 0.000 claims description 16
- 238000009792 diffusion process Methods 0.000 claims description 12
- 239000013598 vector Substances 0.000 claims description 12
- 238000012549 training Methods 0.000 claims description 10
- 230000009466 transformation Effects 0.000 claims description 10
- 230000007246 mechanism Effects 0.000 claims description 7
- 238000005295 random walk Methods 0.000 claims description 6
- 230000007704 transition Effects 0.000 claims description 6
- 230000006870 function Effects 0.000 claims description 5
- 230000003068 static effect Effects 0.000 claims description 3
- 230000004913 activation Effects 0.000 claims description 2
- 238000005520 cutting process Methods 0.000 claims description 2
- 238000010606 normalization Methods 0.000 claims 1
- 238000010586 diagram Methods 0.000 description 10
- 230000007774 longterm Effects 0.000 description 6
- 238000013527 convolutional neural network Methods 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 102100040954 Ephrin-A1 Human genes 0.000 description 2
- 101000965523 Homo sapiens Ephrin-A1 Proteins 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000013439 planning Methods 0.000 description 2
- 101001095088 Homo sapiens Melanoma antigen preferentially expressed in tumors Proteins 0.000 description 1
- 102100037020 Melanoma antigen preferentially expressed in tumors Human genes 0.000 description 1
- 206010039203 Road traffic accident Diseases 0.000 description 1
- 238000012271 agricultural production Methods 0.000 description 1
- 238000013136 deep learning model Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 230000007787 long-term memory Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- YHXISWVBGDMDLQ-UHFFFAOYSA-N moclobemide Chemical compound C1=CC(Cl)=CC=C1C(=O)NCCN1CCOCC1 YHXISWVBGDMDLQ-UHFFFAOYSA-N 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 230000000306 recurrent effect Effects 0.000 description 1
- 238000013468 resource allocation Methods 0.000 description 1
- 230000006403 short-term memory Effects 0.000 description 1
- 238000013068 supply chain management Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/29—Graphical models, e.g. Bayesian networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0499—Feedforward networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computational Linguistics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Evolutionary Biology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
The application discloses a space time sequence prediction method based on a cyclic graph computation sub-neural network, which comprises the following steps: modeling a traffic road network to obtain a graph structure model, and abstracting a sensor in the traffic road network into nodes in the graph structure model; carrying out standardization processing on data acquired by a sensor; constructing various graph calculation sub-networks as feature traps of data, wherein the feature traps respectively use different modes to aggregate information of each node in the graph structure model; constructing GGRU units, and respectively embedding various graph calculation sub-networks into GRU units of the neural network; aggregating the characteristic information captured by different graph sub-networks through an integrator; an encoder-decoder architecture is constructed using a plurality of integrators to enable prediction of spatial multi-temporal sequences. The method has the capability of understanding information in multiple angles, effectively improves the understanding of space multi-element time sequence data, and further improves the prediction accuracy.
Description
Technical Field
The application relates to the technical field of space-time sequence prediction, in particular to a space-time sequence prediction method based on a cyclic graph algorithm sub-neural network.
Background
Space-time series prediction has very wide application requirements in modern times, such as: (1) weather forecast: weather prediction is an important application of spatial time series prediction. By modeling the meteorological data, the change trend of the meteorological variables such as the future temperature, rainfall, wind speed and the like can be predicted, and decision support is provided for weather forecast, disaster early warning, agricultural production and the like. (2) energy demand prediction: it is important for energy suppliers and consumers to understand the trend of future energy demand. By modeling the historical energy demand data, future energy demands can be predicted to adjust production planning and supply chain management. (3) traffic prediction: traffic prediction can help city planners and traffic managers to better manage traffic flow and improve traffic efficiency. By modeling the historical traffic data, future road congestion conditions, traffic accident occurrence rates and the like can be predicted so as to take corresponding measures. (4) population flow prediction: population flow is an important factor in urban planning and social policy formulation. By modeling historical population flow data, future population flow trends can be predicted to rationally plan urban infrastructure and social resource allocation. (5) stock price prediction: stock price prediction is an important application in the financial field. By modeling the historical stock price data, the change trend of the future stock price can be predicted, and decision support is provided for investors.
Currently, a deep learning method is mainly used for multi-element time series prediction. Such neural network-based models can be effectively used for multivariate time series prediction. For example: (1) recurrent neural network RNN: is a neural network that can process sequence data by capturing the timing relationship of the sequence with the output of the previous time as the input of the current time. In multivariate time series prediction, multiple RNNs may be used to process each sequence separately and then their outputs are connected to arrive at the final prediction result. RNNs are very widely used in multivariate time series prediction, for example, in the fields of traffic flow prediction, stock price prediction, etc. to achieve good results. (2) long and short term memory network LSTM: is a special RNN that can better handle long sequences and long-term dependencies. In multivariate time series prediction, multiple LSTMs may be used to process each series and then their outputs are connected to get the final prediction result. LSTM performs better than RNN in handling long sequences and long-term dependencies and may therefore be more suitable in certain applications. (3) convolutional neural network CNN: is a neural network mainly used for image processing, but can also be used for multivariate time series prediction in some cases. In multivariate time series prediction, multiple sequences can be considered as multiple channels, then each channel is convolved using multiple convolution checks, and finally the outputs of all channels are connected to obtain the final prediction result. CNNs have relatively few applications in multivariate time series prediction, but may be more suitable in certain applications. (4) Attention mechanism Attention: the method is a mechanism which can give different weights according to different parts of input data and can be used for multi-element time sequence prediction. In multivariate time series prediction, an attention mechanism is used to automatically learn the importance of different parts of the sequence and an attention weight is applied to the predictions for each sequence. The application of attention mechanisms in multivariate time series prediction is newer, but has achieved good results in some applications.
However, the above-described multivariate time series prediction either completely does not consider the spatial dependency between the sequences, or only considers the spatial dependency between the sequences from a single point of view. In fact, the underlying relationship between these sequences is complex. Thus, there is an urgent need for a new prediction method that ensures that these underlying relationships can be fully mined from a number of different dimensions. Just like human reading text, the metaphors of text tend to be more profound than the superficial meaning.
Disclosure of Invention
The application aims to provide a space time sequence prediction method based on a cyclic graph sub-neural network, which has the capability of understanding information at multiple angles, effectively improves the understanding of space multi-element time sequence data, and further improves the prediction precision.
In order to achieve the above object, the present application provides a space time sequence prediction method based on a cyclic graph algorithm sub-neural network, including:
modeling a traffic road network to obtain a graph structure model, and abstracting a sensor in the traffic road network into nodes in the graph structure model;
carrying out standardization processing on data acquired by a sensor;
constructing various graph calculation sub-networks as feature traps of data, wherein the feature traps respectively use different modes to aggregate information of each node in the graph structure model;
constructing GGRU units, and respectively embedding various graph calculation sub-networks into GRU units of the neural network;
aggregating the characteristic information captured by different graph sub-networks through an integrator;
a sequence-to-sequence encoder-decoder architecture is constructed using a plurality of integrators to implement prediction of spatial multi-element time series.
Further, use ofThe traffic flow data acquired by all the sensors at the moment r is represented, wherein N represents the number of the sensors, and P represents the number of traffic indexes detected by each sensor; use->Representing historical observations within T' time stamps, with +.>Predicted values representing T time stamps in the future; and then determining a function of the neural network learning graph structure model:
wherein ,is the neural network to be fitted.
Further, the specific mode for carrying out standardization processing on the data collected by the sensor is as follows:
wherein X represents training set sample, X * Representing normalized data, E (X) represents training set sample mean and D (X) represents training set sample variance.
Further, constructing various graph calculation sub-networks as feature traps of data, specifically:
constructing a graph diffusion convolution Operator DC Operator, which is a graph calculation sub-network based on a static graph;
a graph gating attention Operator GA Operator is constructed, which is a graph computation sub-network based on dynamic graphs.
Further, the specific way of constructing the graph diffusion convolution Operator DC Operator is as follows:
the graph diffusion convolution operator adopts a random walk strategy:
wherein X represents the input of the model and H represents the output of the model; the diffusion process uses limited S steps for cutting; d (D) O ,D I The out-degree matrix and in-degree matrix in the graph are represented,representing a forward state transition matrix and a reverse state transition matrix; alpha, beta E [0,1 ]]Respectively representing the restarting probability of random walk; />Represents model parameters, and Θ O[q,p,s] =α(1-α) s ,Θ I[q,p,k] =β(1-β) s ;
The critical matrix W is obtained using a distance-based gaussian kernel:
here dist (v) i ,v j ) Representing sensor v i and vj A distance therebetween; σ is the standard deviation of the distance set.
Further, the specific way to construct the graph gating attention Operator GA Operator is as follows:
using the K-head attention mechanism in the graph gating attention operator, there is one K-dimensional gating vector g for each node i i :
wherein ,/>Representing all neighbor node sets of node i; x is x i =X i,: A feature vector representing node i; />Reference vectors of all neighboring nodes of node i, and +.> Max represents taking the maximum value by element; />Representing the final result mapped to K-dimension and scaled to [0,1 ]]Between them;
dynamically acquiring an attention weight matrix among nodes:
wherein ,representing the parameter theta xa Is a linear transformation of (2); />Representing the parameter theta za Is a linear transformation of (2); i.e. < ->Representing a linear layer; θ represents various parameters.
Obtaining an output vector y of the node i i :
Wherein K represents the number of heads of attention, and K representsCurrently executing is the kth attention header;representing the parameter theta 0 Is a linear transformation of (2); />The expression parameter is->Is a linear transformation of (a).
Further, for use in GGRU unitsRepresenting different graphic operators:
H (t) =u (t) ⊙H (t-1) +(1-u (t) )⊙C (t)
wherein ,X(t) ,H (t) Input and output representing a t-th timestamp; r is (r) (t) ,u (t) A gating state of a Reset gate (Reset gate) and an Update gate (Update gate) representing a t-th time stamp; theta (theta) r ,Θ u ,Θ C Is a different filter parameter; Γ -shaped structure g Represented in a specified graph networkMiddle execution->An operator; as indicated by the letter Hadamard product.
As a further step, the integrator aggregates the feature information captured by the different graph sub-networks, specifically:
wherein ,representing GRU units embedded with graphic operators, I representing the number of categories of graphic operators,/-, etc.>Representing an output dimension d 0 Feedforward neural network, τ, means using the tanh activation function in the last layer.
As a further step, the input to the encoder is a historical observationThe output of the decoder is the future prediction value +.>
Compared with the prior art, the technical scheme adopted by the application has the advantages that: according to the application, a deep learning model is adopted, a plurality of graph operator sub-networks are combined to carry out multi-angle modeling on the space dependency relationship among the nodes, and meanwhile, the information captured by a plurality of graph operators can be effectively aggregated through an integrator. The understanding capability of the model on the bottom layer dependency relationship among the multiple time sequences is enhanced, and the prediction accuracy is improved.
Drawings
FIG. 1 is a diagram of a single output architecture using parallel single inputs;
FIG. 2 is a diagram of a multiple output architecture using parallel multiple inputs;
FIG. 3 is a block diagram of a single output and class residual using parallel single inputs;
FIG. 4 is a diagram of a structure using parallel multiple inputs, multiple outputs, and class residuals;
FIG. 5 is a block diagram using a serial single input, single output;
FIG. 6 is a diagram of a multiple output architecture using multiple inputs in series;
FIG. 7 is a block diagram using serial single input, single output, and class residual;
FIG. 8 is a block diagram of a multi-output and residual-like structure using a series of multiple inputs;
FIG. 9 is a block diagram of a GGRU model;
fig. 10 is a diagram of an encoder/decoder architecture.
Detailed Description
The present application will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present application more apparent. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the application, i.e., the embodiments described are merely some, but not all, of the embodiments of the application.
Thus, the following detailed description of the embodiments of the application, as presented in the figures, is not intended to limit the scope of the application, as claimed, but is merely representative of selected embodiments of the application. All other embodiments, which can be made by a person skilled in the art without making any inventive effort, are intended to be within the scope of the present application.
The application provides a space time sequence prediction method based on a cyclic graph algorithm sub-neural network, which specifically comprises the following steps:
step 1: modeling a traffic road network to obtain a graph structure model, and abstracting a sensor in the traffic road network into nodes in the graph structure model;
in particular, byRepresenting traffic flow data collected by all sensors at time t, wherein N represents the number of sensors and P represents the detection of each sensorThe number of traffic indicators measured; use->Representing historical observations within T' time stamps, with +.>Predicted values representing T time stamps in the future; and then determining a function of the neural network learning graph structure model:
wherein ,is the neural network to be fitted.
Step 2: and (3) carrying out standardization processing on the data acquired by the sensor:
wherein X represents training set sample, X * Representing normalized data, E (X) represents training set sample mean and D (X) represents training set sample variance.
Step 3: constructing various graph calculation sub-networks as feature traps of data, wherein the feature traps respectively use different modes to aggregate information of each node in the graph structure model;
specifically, the application includes two types of graph operators that aggregate information of each node in the graph structure model in different ways, respectively, and embed the information into the GRU unit as a substitute for the original linear unit.
Step 3.1: constructing a graph diffusion convolution Operator DC Operator, which is a graph calculation sub-network based on a static graph;
step 3.1.1: the graph diffusion convolution operator adopts a random walk strategy:
here, X represents the input of the model, and H represents the output of the model. The diffusion process uses a limited number of K-step truncations. D (D) O ,D I Representing the out-degree and in-degree matrices in the graph,representing a forward state transition matrix and a reverse state transition matrix. Alpha, beta E [0,1 ]]Respectively, the restart probabilities of the random walks. />Represents model parameters, and Θ O[q,p,s] =α(1-α) s ,Θ I[q,p,s] =β(1-β) s 。/>The probability value representing the S-step diffusion from the i-th node to its neighbor node, the former representing the forward propagation probability and the latter representing the backward propagation probability.
Step 3.1.2: the critical matrix W is obtained using a distance-based gaussian kernel:
here dist (v) i ,v j ) Representing sensor v i and vj A distance therebetween; σ is the standard deviation of the distance set.
Step 3.2: constructing a graph gating attention Operator GA Operator, which is a graph calculation sub-network based on a dynamic graph;
step 3.2.1: using the K-head attention mechanism in the graph gating attention operator, there is one K-dimensional gating vector g for each node i i :
wherein ,representing all neighbor node sets of node i; x is x i =X i,: A feature vector representing node i;reference vectors of all neighboring nodes of node i, and +.> Max represents taking the maximum value by element; />Representing the final result mapped to K-dimension and scaled to [0,1 ]]Between them;
step 3.2.2: dynamically acquiring an attention weight matrix among nodes:
wherein ,representing the parameter theta xa Is a linear transformation of (2); />Representing the parameter theta za Is a linear transformation of (2);
step 3.2.3: obtaining an output vector y of the node i i :
Step 4: constructing GGRU units, embedding multiple graph sub-networks into GRU units of neural network, see FIG. 9, whereRepresenting different graphic operators:
H (t) =u (t) ⊙H (t-1) +(1-u (t) )⊙C (t)
wherein ,X(t) ,H (t) Input and output representing a t-th timestamp; r is (r) (t) ,u (t) A gating state of a Reset gate (Reset gate) and an Update gate (Update gate) representing a t-th time stamp; theta (theta) r ,Θ u ,Θ C Is a different filter parameter;represented in the specified graph network +.>Middle execution->An operator; as indicated by the letter Hadamard product.
Step 5: aggregating the characteristic information captured by different graph sub-networks through an integrator;
wherein ,the GRU element embedded with the graph operator is represented. Fig. 1-8 show eight polymerizer structures.
Step 6: constructing a sequence-to-sequence encoder-decoder architecture using a plurality of integrators to implement prediction of spatial multi-element time series, see fig. 10;
specifically, the input to the encoder is a historical observationThe output of the decoder is the future prediction value +.>
In the embodiment, a Ubuntu system is used as a development environment, python is used as a development language, and Pytorch is used for building a frame. The space time sequence prediction method based on the cyclic graph computation sub-neural network is adopted to predict traffic flow: obtaining a graph structure model representation of a traffic network, and adopting a disclosed METR-LA data set; dividing the data set into a training set, a verification set and a test set, wherein the proportion is 7:1:2. to compare algorithm performance, it is compared with some commonly used multivariate time series prediction methods. Including statistical learning method ARIMA; a machine learning method LSVR; deep learning method FC-LSTM; and a graph-based deep learning method DCRNN, STGCN, gaAN, ASTGCN, GMAN. Meanwhile, various evaluation indexes are used for comprehensively evaluating the performance of the model, including MAE, RMSE, MAPE and the like.
wherein ,xi Representing the actual data of the data set,data representing model predictions.
The comparison results are shown in Table 1: wherein the best evaluation index of all algorithms will be bolded.
Table 1: performance comparison of different traffic speed prediction models on METR-LA datasets
Table 1 compares the prediction results of the icornn model with other baseline models. By comparing the scores of each model over long-term and short-term time series predictions, it can be seen that the graph-based neural network model achieves better prediction accuracy. For the smoother sequence PeMS-BAY, some traditional models achieve good results even in short-term predictions (15 minutes), but perform poorly in long-term predictions (60 minutes). For the unstable sequence Metr-LA, the traditional model performs poorly in both short-term and long-term predictions. The GaAN model, on the other hand, performs well on the Metr-LA dataset. GMAN is more suitable for smoother sequences. All of these illustrate the importance of introducing graph structure in complex time series predictions. Furthermore, the model designed by the present application shows optimal or suboptimal performance in long-term and short-term predictions, especially for more complex time sequences, as it introduces the advantages of multiple graph structures at the same time.
In summary, compared with other advanced methods, the space time sequence prediction method based on the cyclic graph sub-neural network provided by the application has better performance and robustness, and can be suitable for the more complex multi-element time sequence prediction problem.
The foregoing descriptions of specific exemplary embodiments of the present application are presented for purposes of illustration and description. It is not intended to limit the application to the precise form disclosed, and obviously many modifications and variations are possible in light of the above teaching. The exemplary embodiments were chosen and described in order to explain the specific principles of the application and its practical application to thereby enable one skilled in the art to make and utilize the application in various exemplary embodiments and with various modifications as are suited to the particular use contemplated. It is intended that the scope of the application be defined by the claims and their equivalents.
Claims (9)
1. The space time sequence prediction method based on the cyclic graph algorithm sub-neural network is characterized by comprising the following steps of:
modeling a traffic road network to obtain a graph structure model, and abstracting a sensor in the traffic road network into nodes in the graph structure model;
carrying out standardization processing on data acquired by a sensor;
constructing various graph calculation sub-networks as feature traps of data, wherein the feature traps respectively use different modes to aggregate information of each node in the graph structure model;
constructing GGRU units, and respectively embedding various graph calculation sub-networks into GRU units of the neural network;
aggregating the characteristic information captured by different graph sub-networks through an integrator;
a sequence-to-sequence encoder-decoder architecture is constructed using a plurality of integrators to implement prediction of spatial multi-element time series.
2. The method for predicting the space-time sequence based on the cyclic graph sub-neural network according to claim 1, whereinThe traffic flow data acquired by all the sensors at the time t are represented, wherein N represents the number of the sensors, and P represents the number of traffic indexes detected by each sensor; use->Representing historical observations within T' time stamps, with +.>Predicted values representing T time stamps in the future; and then determining a function of the neural network learning graph structure model:
wherein ,is the neural network to be fitted.
3. The space-time sequence prediction method based on the cyclic graph sub-neural network according to claim 1, wherein the specific mode of carrying out normalization processing on the data acquired by the sensor is as follows:
wherein X represents training set sample, X * Representing normalized data, E (X) represents training set sample mean and D (X) represents training set sample variance.
4. The space-time sequence prediction method based on the cyclic graph sub-neural network according to claim 1, wherein a plurality of graph sub-networks are constructed as feature traps of data, specifically:
constructing a graph diffusion convolution Operator DC Operator, which is a graph calculation sub-network based on a static graph;
a graph gating attention Operator GA Operator is constructed, which is a graph computation sub-network based on dynamic graphs.
5. The space-time sequence prediction method based on the cyclic graph Operator neural network according to claim 4, wherein the specific mode of constructing the graph diffusion convolution Operator DC Operator is as follows:
the graph diffusion convolution operator adopts a random walk strategy:
wherein X represents the input of the model and H represents the output of the model; the diffusion process uses limited S steps for cutting; d (D) O ,D I The out-degree matrix and in-degree matrix in the graph are represented,representing a forward state transition matrix and a reverse state transition matrix; alpha, beta E [0,1 ]]Respectively representing the restarting probability of random walk; />Represents model parameters, and Θ O[q,p,s] =α(1-α) s ,Θ I[q,p,k] =β(1-β) s ;
The critical matrix W is obtained using a distance-based gaussian kernel:
here dist (v) i ,v j ) Representing sensor v i and vj A distance therebetween; σ is the standard deviation of the distance set.
6. The space-time sequence prediction method based on the cyclic graph Operator neural network according to claim 4, wherein the specific mode of constructing the graph gating attention Operator GA Operator is as follows:
using the K-head attention mechanism in the graph gating attention operator, there is one K-dimensional gating vector g for each node i i :
wherein ,representing all neighbor node sets of node i; x is x i =X i,: A feature vector representing node i; />Reference vectors of all neighboring nodes of node i, and z i =/>Max represents taking the maximum value by element; />Representing the final result mapped to K-dimension and scaled to [0,1 ]]Between them;
dynamically acquiring an attention weight matrix among nodes:
wherein ,representing the parameter theta xa Is a linear transformation of (2); />Representing the parameter theta za Is a linear transformation of (2);
obtaining an output vector y of the node i i :
Where K represents the number of attention heads and K represents the kth note currently being performedA force-imparting head;representing the parameter theta 0 Is a linear transformation of (2); />The expression parameter is->Is a linear transformation of (a).
7. The method for spatial-temporal sequence prediction based on a cyclic graph sub-neural network of claim 1, wherein the GGRU unit is used inRepresenting different graphic operators:
H (t) =u (t) ⊙H (t-1) +(1-u (t) )⊙C (t)
wherein ,H(t) ,H (t) Input and output representing a t-th timestamp; r is (r) (t) ,u (t) A reset gate and a gate state of the update gate representing a t-th timestamp; theta (theta) r ,Θ u ,Θ C Is a different filter parameter;represented in the specified graph network +.>Middle execution->An operator; as indicated by the letter Hadamard product.
8. The space-time sequence prediction method based on cyclic graph sub-neural network according to claim 1, wherein the integrator is used for aggregating the characteristic information captured by different graph sub-networks, specifically:
wherein ,representing GRU units embedded with graphic operators, I representing the number of categories of graphic operators,/-, etc.>Representing an output dimension d 0 Feedforward neural network, τ, means using the tanh activation function in the last layer.
9. The method of claim 1, wherein the encoder input is a historical observationThe output of the decoder is the future prediction value +.>
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310630846.2A CN116610919A (en) | 2023-05-31 | 2023-05-31 | Space time sequence prediction method based on cyclic graph algorithm sub-neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310630846.2A CN116610919A (en) | 2023-05-31 | 2023-05-31 | Space time sequence prediction method based on cyclic graph algorithm sub-neural network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116610919A true CN116610919A (en) | 2023-08-18 |
Family
ID=87679794
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310630846.2A Pending CN116610919A (en) | 2023-05-31 | 2023-05-31 | Space time sequence prediction method based on cyclic graph algorithm sub-neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116610919A (en) |
-
2023
- 2023-05-31 CN CN202310630846.2A patent/CN116610919A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Wan et al. | CTS-LSTM: LSTM-based neural networks for correlatedtime series prediction | |
Choi et al. | TrajGAIL: Generating urban vehicle trajectories using generative adversarial imitation learning | |
Chen et al. | A novel reinforced dynamic graph convolutional network model with data imputation for network-wide traffic flow prediction | |
Sun et al. | Dual dynamic spatial-temporal graph convolution network for traffic prediction | |
CN109583565B (en) | Flood prediction method based on attention model long-time and short-time memory network | |
CN114220271A (en) | Traffic flow prediction method, equipment and storage medium based on dynamic space-time graph convolution cycle network | |
CN111899510A (en) | Intelligent traffic system flow short-term prediction method and system based on divergent convolution and GAT | |
Liu et al. | Multivariate time-series forecasting with temporal polynomial graph neural networks | |
Jin et al. | A GAN-based short-term link traffic prediction approach for urban road networks under a parallel learning framework | |
Li et al. | Graph neural network for robust public transit demand prediction | |
CN112765896A (en) | LSTM-based water treatment time sequence data anomaly detection method | |
CN110138595A (en) | Time link prediction technique, device, equipment and the medium of dynamic weighting network | |
Gammelli et al. | Predictive and prescriptive performance of bike-sharing demand forecasts for inventory management | |
Liang et al. | Mixed-order relation-aware recurrent neural networks for spatio-temporal forecasting | |
Bai et al. | Deep spatial–temporal sequence modeling for multi-step passenger demand prediction | |
Wu et al. | Msstn: Multi-scale spatial temporal network for air pollution prediction | |
CN116504060A (en) | Diffusion diagram attention network traffic flow prediction method based on Transformer | |
CN117494871A (en) | Ship track prediction method considering ship interaction influence | |
CN116894096A (en) | News event prediction method based on recursive double hypergraph neural network | |
Shuai et al. | Relationship analysis of short-term origin–destination prediction performance and spatiotemporal characteristics in urban rail transit | |
Ikhlasse et al. | Multimodal cloud resources utilization forecasting using a Bidirectional Gated Recurrent Unit predictor based on a power efficient Stacked denoising Autoencoders | |
Kieu et al. | Towards real-time predictions using emulators of agent-based models | |
Hu et al. | Variational expectation maximization attention broad learning systems | |
Chang et al. | Enhanced road information representation in graph recurrent network for traffic speed prediction | |
Yalçın | Weather parameters forecasting with time series using deep hybrid neural networks |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |