CN116596109A - Traffic flow prediction model based on gating time convolution network - Google Patents

Traffic flow prediction model based on gating time convolution network Download PDF

Info

Publication number
CN116596109A
CN116596109A CN202310352647.XA CN202310352647A CN116596109A CN 116596109 A CN116596109 A CN 116596109A CN 202310352647 A CN202310352647 A CN 202310352647A CN 116596109 A CN116596109 A CN 116596109A
Authority
CN
China
Prior art keywords
time
dependence
traffic flow
model
tcn
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310352647.XA
Other languages
Chinese (zh)
Inventor
康明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN202310352647.XA priority Critical patent/CN116596109A/en
Publication of CN116596109A publication Critical patent/CN116596109A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/065Traffic control systems for road vehicles by counting the vehicles in a section of the road or in a parking area, i.e. comparing incoming count with outgoing count
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • G06N3/0442Recurrent networks, e.g. Hopfield networks characterised by memory or gating, e.g. long short-term memory [LSTM] or gated recurrent units [GRU]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/40Business processes related to the transportation industry
    • GPHYSICS
    • G08SIGNALLING
    • G08GTRAFFIC CONTROL SYSTEMS
    • G08G1/00Traffic control systems for road vehicles
    • G08G1/01Detecting movement of traffic to be counted or controlled
    • G08G1/0104Measuring and analyzing of parameters relative to traffic conditions
    • G08G1/0125Traffic data processing
    • G08G1/0133Traffic data processing for classifying traffic situation
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Business, Economics & Management (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Economics (AREA)
  • Strategic Management (AREA)
  • Human Resources & Organizations (AREA)
  • General Business, Economics & Management (AREA)
  • Tourism & Hospitality (AREA)
  • Marketing (AREA)
  • Quality & Reliability (AREA)
  • Operations Research (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Game Theory and Decision Science (AREA)
  • Development Economics (AREA)
  • Analytical Chemistry (AREA)
  • Chemical & Material Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Primary Health Care (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention relates to a traffic flow prediction model based on a gating time convolution network, wherein the traffic flow combination prediction model and a G-TCN model the time dependence, the spatial dependence and the long-time sequence of traffic flow respectively. The time-dependence and spatial-dependence of traffic flows are captured using a Time Convolutional Network (TCN) and a Graph Convolutional Network (GCN), respectively, and the STA-Block module models the time-space dependence of long-time sequences through a time-space attention mechanism and a gating fusion mechanism. An adaptive adjacency matrix is constructed in a G-TCN model, and learning is performed through node embedding, so that the model can accurately capture hidden space-time dependency relations in traffic flow data.

Description

Traffic flow prediction model based on gating time convolution network
Technical Field
The invention relates to the technical field, in particular to a traffic flow prediction model based on a gating time convolution network.
Background
Traffic flow predictions aim at predicting future traffic conditions in a road network based on historical observations. Traffic flow prediction can predict traffic conditions of the road section within a certain time in the future, and great help is provided in the aspects of signal control, traffic guidance, path planning and the like.
ARIMA is a classical statistical model in time series analysis, widely used for traffic flow prediction. Researchers extend the spatial domain into the ARIMA time series model to obtain a spatio-temporal autoregressive integral moving average model. However, since the time series analysis model is a pure inductive method, some ideal prior assumptions are required. And due to the complexity and non-linear nature of traffic data, these methods often perform poorly in practical applications.
The space-time diagram modeling has wide application in solving the problems of complex systems such as traffic speed prediction and the like. For example, in traffic speed prediction, speed sensors on urban roads form a graph in which edge weights are calculated from euclidean distances of two nodes. Since traffic congestion on a road can cause traffic speed on its entrance road to decrease, the network structure of the model floor map is considered as a priori knowledge of the interdependencies between nodes when modeling traffic speed time series data on each road.
At present, the research on time space diagram modeling mainly has two directions: one is to integrate a graph roll-up network (GCN) into a Recurrent Neural Network (RNN). And secondly, the GCN is integrated into a Convolutional Neural Network (CNN). While these studies demonstrate the effectiveness of introducing a graph network structure of data into a model, these approaches still face two major drawbacks. (1) The data graph structure assumed in these studies, when a connection does not contain an interdependence relationship between two nodes, and when an interdependence relationship between two nodes exists, there is a case where the connection is invalid. (2) Current studies of time space graph modeling are ineffective in learning time dependencies. RNN-based methods are prone to gradient extinction when capturing long time sequences.
The invention provides a traffic flow prediction model based on a gate control time convolution network (G-TCN), which consists of a gate control time convolution network module (TCN), a graph convolution network module (GCN) and a space-time attention mechanism module (STA-Block), wherein each STA-Block consists of a space attention mechanism, a time attention mechanism and a gate control fusion mechanism. Meanwhile, a graph roll stacking layer is provided, and an adaptive adjacency matrix can be obtained from data through an end-to-end supervision training mode. The present invention employs a stacked causal convolutional network of dilation to capture time dependence. The G-TCN model is capable of effectively and efficiently processing space-time diagram data with long time sequences, supported by an expansion causal convolutional network.
Disclosure of Invention
In order to solve the technical problems, the technical scheme provided by the invention is as follows: a traffic flow prediction model based on a gating time convolution network, wherein the traffic flow combination prediction model and a G-TCN model respectively model the time dependence, the spatial dependence and the long-time sequence of traffic flow. The time dependence and the space dependence of traffic flows are respectively captured by using TCN and GCN, and the STA-Block module models the time-space dependence of a long-time sequence through a time-space attention mechanism and a gating fusion mechanism. An adaptive adjacency matrix is constructed in a G-TCN model, and learning is performed through node embedding, so that the model can accurately capture hidden space-time dependency relations in traffic flow data.
The invention has the following advantages:
1. the invention constructs an adaptive adjacency matrix that automatically discovers hidden graph structures from data without any prior knowledge guidance.
2. An efficient framework is presented to capture spatio-temporal dependencies simultaneously. The invention combines the proposed GCN with the expansion causal convolution network, so that each graph convolution layer processes the expansion causal convolution layers with different granularity levels to extract the spatial correlation of node information.
3. A spatial attention mechanism and a temporal attention mechanism are proposed to learn the dynamic spatial correlation and nonlinear temporal dependence, respectively, in traffic flow data. In addition, a gating fusion mechanism is designed to adaptively fuse information extracted by a spatio-temporal attention mechanism to reduce error propagation during prediction.
Drawings
FIG. 1 is an overall framework diagram of the G-TCN model.
FIG. 2 is a diagram of an expansion cause and effect convolution network.
Fig. 3 is a graph of a Gated TCN framework.
Fig. 4 is an ST-Conv Block frame diagram.
Fig. 5 is a STA-Block framework diagram.
Detailed Description
The present invention will be described in further detail with reference to examples.
1 model
The G-TCN model captures the time characteristics of traffic flow by using TCN, and simultaneously adopts a coder-decoder structure to capture the space characteristics of traffic flow, wherein the coder and the decoder are composed of a plurality of STA-blocks so as to simulate the influence of space-time factors on traffic conditions. The encoder encodes the input traffic stream characteristics and the decoder predicts the output sequence. Between the encoder and the decoder, a transition attention mechanism is applied to transition the encoded traffic characteristics to generate a future time-step sequence representation as input to the decoder.
2 problem definition
In the present invention, the traffic road network is defined as graph g= (V, E), where V is a set of road nodes, E is a set of edges of the traffic road network, a E R N × N The adjacency matrix of fig. G is shown. If v i ,v j E V and (V) i ,v j ) E, then A ij 1, otherwise 0. In each time step t, the graph G has a dynamic feature matrix X (t) ∈R N×D The traffic flow prediction problem aims at learning a function f that maps the history map signal S to a future T-map signal, given a map G, as follows:
wherein X is (t-S):t ∈R N×D×S And X (t+1):(t+T) ∈R N×D×T
2.1G-TCN model building
The G-TCN model consists of an input layer, a stacked space-time convolutional network layer, ST-Conv Block, STA-Block and an output layer, each network layer of the stack being connected to the ST-Conv Block by hops. The time space layer is composed of a GCN layer and a Gated time convolution layer (Gated TCN), and the Gated TCN is composed of two parallel time convolution layers. ST-Conv Block contains three spatio-temporal convolution blocks, capturing many-to-one effects from three different angles corresponding to space, time and space, respectively. STA-Block combines spatial and temporal attention mechanisms through gating fusion. By stacking multiple time-space layers, the G-TCN is able to handle spatial correlation at different temporal levels. Spatial features in traffic flow history data can be captured by the GCN, wherein the input h of the GCN is a three-dimensional tensor with the size of [ N, C, L ], N is the number of nodes, C is the dimension of a hidden layer, and L is the length of a data sequence.
The present invention chooses to use the Mean Absolute Error (MAE) as a loss function for G-TCN, defined as:
the G-TCN model willAs an overall output, the accepted field size of the G-TCN model is equal to the input sequence length, so that in the last spatio-temporal layer, the output time dimension is equal to 1, and the number of output channels of the last layer is set to a step size T to obtain the dimension required for output.
2.2 Time Convolutional Network (TCN)
The present invention captures the temporal trend of nodes using extended causal convolution as a temporal convolution layer (TCN). The extended causal convolution network realizes larger receptive field by increasing the depth of the layer, can correctly process long-term sequences, and relieves the gradient explosion problem. The extended causal volume operation slides the input by skipping values at specific steps. Mathematically, given a one-dimensional sequence input x εR T And filter f epsilon R K The extended causal convolution operation of x and f at step t is represented as shown in equation (2):
where d is the spreading factor controlling the jump distance.
Gated TCN: the gating mechanism can control information of each layer in the TCN, and the Gated TCN only comprises one output gate. Given input X E N×D×S The form is as follows:
h=g(θ 1 *X+b)·sigmoid(θ 2 *X+c) (4)
wherein θ 1 ,θ 2 B and c are model parameters and g (·) is the activation function of the output.
2.3 graph roll-up network (GCN)
The GCN is a basic operation of extracting node characteristics from the structural information of the nodes, and smoothing the signals of the nodes by aggregating and converting the neighborhood information thereof. Order theRepresenting normalized adjacency matrix, X.epsilon.R N×D Representing the input signal, Z.epsilon.R N×M Representing the output, W.epsilon.R D×M Representing a matrix of model parameters, the GCN is defined as:
the diffusion process of the graph signal was modeled with K finite steps. Generalizing the diffusion convolution network into equation (6), the result is expressed as:
wherein P is k Representing the power series of the transition matrix. In the undirected graph, p=a/rowsum (a). In the directed graph, the diffusion process is divided into forward and backward, forward transfer matrix P f =a/rowsum (a) and backward transfer matrix P b =A T Rowsum (A). Thus, the diffusion graph convolution network can be defined as:
2.4 adaptive adjacency matrix
In the present invention an adaptive adjacency matrix is proposedThe matrix does not require any prior knowledge and is subjected to end-to-end learning by random gradient descent by using a learnable parameter E 1 ,E 2 ∈R N×c Two node embedded dictionaries are randomly initialized to mine traffic flow hidden spatial features. The adaptive adjacency matrix proposed by the present invention is shown in equation (8):
wherein E is 1 Representing source node embedding, E 2 Representing target node embedding, weak connections are eliminated using a ReLU activation function, and the softmax function normalizes the adaptive adjacency matrix. By combining predefined spatial dependencies and self-learning hidden spatial features, the present invention proposes the following graph roll layering:
when the graph network structure is not available, the present invention proposes to use the adaptive adjacency matrix alone to capture hidden spatial dependencies as shown in equation (10):
2.5 space-time convolutional network block
Space-time convolution network block packetA spatio-temporal convolution block containing three kernels captures many-to-one effects from three different angles corresponding to space, time, and space-time, respectively. The temporal kernel captures the temporal dependence of traffic streams at the same location and the spatial kernel captures the spatial dependence of traffic streams at neighboring locations over the same time step. Each spatio-temporal convolution block takes as input the output of the previous spatio-temporal attention block, i.eOutput->Calculated from equation (11):
wherein, the liquid crystal display device comprises a liquid crystal display device,and->Is a spatiotemporal kernel of f×f, f×1 denotes a temporal kernel, and 1×f denotes a spatial kernel. The inakrlu (·) represents inaky modified linear unit function, and the x represents convolution operation. Finally, the outputs of the three convolution kernels are connected and 1X 1 convolution +.>To compress the features while limiting the number of channels.
2.6 space-time attention block
The ST-Attention Block comprises a spatial Attention mechanism, a temporal Attention mechanism and a gating fusion mechanism. The invention represents the input of the first block as H (l-1) Vertex v i At time step t j Is represented as the hidden state of (2)The outputs of the spatial and temporal attentiveness mechanisms in the first block are denoted as +.>And->Vertex v i At time step t j The hidden states of (a) are denoted as +.>And->After the gated fusion, the output of the first block is obtained, denoted as H (l)
The invention designs a spatial attention mechanism to adaptively capture the correlation between sensors in a road network and dynamically allocate different weights to different road segments at different time steps. For time step t j Vertex v of (2) i Calculating a weighted sum of all vertices:
where V is the set of all vertices,is the attention score, the sum of the attention scores is equal to
The invention learns the attention score by considering the traffic characteristics and the graph network structure, and calculates the vertex v by adopting a dot product scaling method i And v:
wherein, the connection operation is represented by the I,<·,·>representing inner product operators, 2D representationIs a dimension of (c). Then, the +.A. is added by the softmax function>Normalization is:
obtaining attention scoreThe hidden state is then updated by equation (14).
In order to stabilize the learning process, the invention expands the spatial attention mechanism to a multi-head attention mechanism. Specifically, K parallel attention mechanisms are connected with different leachable projections:
and->Representing three different nonlinear mappings in the kth head-attention mechanism, yielding an output in d=d/K dimensions.
The present invention designs a time-attention mechanism to adaptively model the nonlinear correlation between different time steps. Calculating attention score by multi-head method by measuring correlation between different time steps in consideration of traffic characteristics and time, taking into consideration vertex v i Time step t j The correlation between t is defined as:
wherein, the liquid crystal display device comprises a liquid crystal display device,representing a time step t j And t, correlation between->Is the attention fraction of the kth head, representing the time step t j Importance to t->And->Representing two different leachable transitions, < >>Representing t j A previous set of time steps. After the attention score is obtained, then at time step t j Vertex v i The hidden state update of (a) is as follows:
representing a non-linear projection, the learnable parameters in equations (19), (20) and (21) are shared between all vertices and time steps by parallel computation.
The invention designs a gating fusion mechanism which adaptively fuses a spatial attention mechanism and a temporal attention mechanism. In the L-th STA-Block, the outputs of the spatial and temporal attention mechanisms are denoted asAndand->Fusion is performed by equation (21):
W z,1 ∈R D×D ,W z,2 ∈R D×D ,b z ∈R D is a learnable parameter, z represents gating.
3. Experiment
3.1 data description
The present invention uses the METR-LA dataset and the PEMS-BAY dataset to verify the performance of the proposed G-TCN model.
The METR-LA recorded four month traffic speed data via 207 sensors. PEMS-BAY contains 325 sensors for 6 months of traffic speed information. In the experiment, the dataset was divided into 7:2:1 as training set, test set and validation set, respectively. For predicting traffic flow rates of 15 minutes, 30 minutes and 60 minutes.
The present invention uses a min-max normalization method to normalize the data, limiting it to [0,1]. The normalization formula is as follows:
wherein x is i Represents the ith raw data, x max And x min Representing the maximum and minimum values of the original data, respectivelyRepresenting the normalized input data.
3.2 Experimental setup
The invention is based on Pytorch deep learning framework, and completes the construction and training of traffic flow prediction model in PyCharm development environment. The invention uses an 8-layer G-TCN network, and the expansion factor sequence is 1,2,1,2,1,2,1,2. Using equation (6) as the GCN layer of the present invention, the diffusion step k=2. The model was trained using Adam optimizer with an initial learning rate of 0.001 and a discard rate p=0.3.
The invention evaluates the error between the actual traffic flow speed and the predicted result based on the following indicators:
average absolute error:
root mean square error:
average absolute percentage error:
wherein y is i Andthe actual traffic speed and the predicted traffic speed n are the number of observations, respectively.
3.3 Baseline model
The present invention compares the G-TCN model with the following models:
(1) HA: historical average model. The average traffic information for the historical period is used as a prediction.
(2) VAR: vector autoregressions.
(3) SVR: and training the model by using a linear support vector machine to obtain the relationship between input and output so as to predict traffic flow.
(4) FNN: feedforward neural network with two hidden layers and L2 regularization.
(5) ARIMA: an autoregressive integrated moving average model with a Kalman filter.
(6) FC-LSTM: recurrent neural networks with fully connected LSTM hidden units.
(7) WaveNet: convolutional network architecture for sequence data.
(8) Graph WaveNet: graph WaveNet, which combines graph convolution with dilated occasional convolution.
(9) STGCN: space-time diagram convolution networks combine GCN and 1D convolutions.
(10) ASTGCN: based on the space-time diagram convolutional network of attention, a space-time attention mechanism is integrated into the STGCN for capturing dynamic space-time patterns.
(11) STSGCN: the spatiotemporal synchronous graph convolution network captures spatiotemporal correlation by stacking multiple local GCN layers with adjacent matrices on the time axis.
3.4 experimental results and analysis
The predicted performance of the G-TCN model and various baseline models on both sets of data was compared for 15 minutes, 30 minutes and 60 minutes. From table 2, it can be observed that the deep learning method is superior to the traditional time series method and machine learning model, demonstrating the ability of the deep neural network to model nonlinear traffic flow data. In deep learning methods, graph WaveNet and STGCN based models (e.g., graph WaveNet and STGCN) generally perform better than the FC-LSTM model, with Graph WaveNet achieving better results on both data sets, which perform far better than the equivalent models of HA, ARIMA and FC-LSTM. The G-TCN model of the invention is superior to the traditional convolution-based method (such as STGCN), and shows that the G-TCN model can better capture the space-time dependence in traffic flow data.
TABLE 1 comparison of Performance of different traffic flow prediction models on two sets of data sets
In contrast to the baseline model, the G-TCN model employs stacked time-space layers, including GCN layers with different parameters. Thus, each GCN layer in the G-TCN model can focus on its own time input range, and the G-TCN model achieves optimal predictive performance. Under different time points, the G-TCN model has better prediction performance, and the effectiveness of the space-time correlation of the G-TCN model in capturing traffic flow data is proved.
While the invention has been described in detail in the foregoing general description and with reference to specific embodiments thereof, it will be apparent to one skilled in the art that modifications and improvements can be made thereto. Accordingly, such modifications or improvements may be made without departing from the spirit of the invention and are intended to be within the scope of the invention as claimed.

Claims (2)

1. A traffic flow prediction model based on a gating time convolution network is characterized in that: the traffic flow combination prediction model (G-TCN) models the time dependence, the space dependence and the long-time sequence of traffic flows respectively. The time-dependence and spatial-dependence of traffic flows are captured using a Time Convolutional Network (TCN) and a Graph Convolutional Network (GCN), respectively, and the STA-Block module models the time-space dependence of long-time sequences through a time-space attention mechanism and a gating fusion mechanism. An adaptive adjacency matrix is constructed in a G-TCN model, and learning is performed through node embedding, so that the model can accurately capture hidden space-time dependency relations in traffic flow data.
2. The traffic flow prediction model based on the gated time convolution network according to claim 1, wherein the traffic flow combination prediction model construction step is as follows: the three modules model the time dependence, the spatial dependence and the long time sequence of the traffic flow respectively. The time-dependence and spatial-dependence of traffic flows are captured using a Time Convolutional Network (TCN) and a Graph Convolutional Network (GCN), respectively, and the STA-Block module models the time-space dependence of long-time sequences through a time-space attention mechanism and a gating fusion mechanism. An adaptive adjacency matrix is constructed in a G-TCN model, and learning is performed through node embedding, so that the model can accurately capture hidden space-time dependency relations in traffic flow data. Finally, the prediction sequence is output through the linear layer to conduct prediction.
CN202310352647.XA 2023-04-04 2023-04-04 Traffic flow prediction model based on gating time convolution network Pending CN116596109A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310352647.XA CN116596109A (en) 2023-04-04 2023-04-04 Traffic flow prediction model based on gating time convolution network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310352647.XA CN116596109A (en) 2023-04-04 2023-04-04 Traffic flow prediction model based on gating time convolution network

Publications (1)

Publication Number Publication Date
CN116596109A true CN116596109A (en) 2023-08-15

Family

ID=87594433

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310352647.XA Pending CN116596109A (en) 2023-04-04 2023-04-04 Traffic flow prediction model based on gating time convolution network

Country Status (1)

Country Link
CN (1) CN116596109A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117290706A (en) * 2023-10-31 2023-12-26 兰州理工大学 Traffic flow prediction method based on space-time convolution fusion probability sparse attention mechanism
CN117579324A (en) * 2023-11-14 2024-02-20 湖北华中电力科技开发有限责任公司 Intrusion detection method based on gating time convolution network and graph
CN118053518A (en) * 2024-04-16 2024-05-17 之江实验室 Chemical process time series data causal relationship graph construction method, device and medium

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117290706A (en) * 2023-10-31 2023-12-26 兰州理工大学 Traffic flow prediction method based on space-time convolution fusion probability sparse attention mechanism
CN117579324A (en) * 2023-11-14 2024-02-20 湖北华中电力科技开发有限责任公司 Intrusion detection method based on gating time convolution network and graph
CN117579324B (en) * 2023-11-14 2024-04-16 湖北华中电力科技开发有限责任公司 Intrusion detection method based on gating time convolution network and graph
CN118053518A (en) * 2024-04-16 2024-05-17 之江实验室 Chemical process time series data causal relationship graph construction method, device and medium

Similar Documents

Publication Publication Date Title
CN108205889B (en) Method for predicting highway traffic flow based on convolutional neural network
CN116596109A (en) Traffic flow prediction model based on gating time convolution network
CN110309732B (en) Behavior identification method based on skeleton video
CN113313947B (en) Road condition evaluation method of short-term traffic prediction graph convolution network
CN110851782A (en) Network flow prediction method based on lightweight spatiotemporal deep learning model
Liu et al. Time series prediction based on temporal convolutional network
CN110570035B (en) People flow prediction system for simultaneously modeling space-time dependency and daily flow dependency
CN110737968A (en) Crowd trajectory prediction method and system based on deep convolutional long and short memory network
CN115578851A (en) Traffic prediction method based on MGCN
Wei et al. Learning motion rules from real data: Neural network for crowd simulation
CN113239897B (en) Human body action evaluation method based on space-time characteristic combination regression
CN115376317B (en) Traffic flow prediction method based on dynamic graph convolution and time sequence convolution network
CN114169647A (en) Traffic prediction method and system of continuous memory self-adaptive heterogeneous space-time diagram convolution
CN111242395A (en) Method and device for constructing prediction model for OD (origin-destination) data
CN113505924A (en) Information propagation prediction method and system based on cascade spatiotemporal features
Xu et al. AGNP: Network-wide short-term probabilistic traffic speed prediction and imputation
CN114495500A (en) Traffic prediction method based on dual dynamic space-time diagram convolution
CN116935649A (en) Urban traffic flow prediction method for multi-view fusion space-time dynamic graph convolution network
Chen et al. Pedestrian behavior prediction model with a convolutional LSTM encoder–decoder
US20220358346A1 (en) Systems, methods, and media for generating and using spiking neural networks with improved efficiency
CN114912169B (en) Industrial building heat supply autonomous optimization regulation and control method based on multisource information fusion
CN115762147A (en) Traffic flow prediction method based on adaptive graph attention neural network
CN115840261A (en) Typhoon precipitation short-term prediction model optimization and prediction method
ABBAS A survey of research into artificial neural networks for crime prediction
CN115456314A (en) Atmospheric pollutant space-time distribution prediction system and method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination