CN115022193B - Local area network flow prediction method based on deep learning model - Google Patents

Local area network flow prediction method based on deep learning model Download PDF

Info

Publication number
CN115022193B
CN115022193B CN202210561413.1A CN202210561413A CN115022193B CN 115022193 B CN115022193 B CN 115022193B CN 202210561413 A CN202210561413 A CN 202210561413A CN 115022193 B CN115022193 B CN 115022193B
Authority
CN
China
Prior art keywords
flow
module
model
convlstm
matrix
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210561413.1A
Other languages
Chinese (zh)
Other versions
CN115022193A (en
Inventor
陈智勇
朱俊宏
马万艺
赖俊宇
余长江
皮昌洪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing Huifa Network Technology Co ltd
University of Electronic Science and Technology of China
Original Assignee
Chongqing Huifa Network Technology Co ltd
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing Huifa Network Technology Co ltd, University of Electronic Science and Technology of China filed Critical Chongqing Huifa Network Technology Co ltd
Priority to CN202210561413.1A priority Critical patent/CN115022193B/en
Publication of CN115022193A publication Critical patent/CN115022193A/en
Application granted granted Critical
Publication of CN115022193B publication Critical patent/CN115022193B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/14Network analysis or design
    • H04L41/147Network analysis or design for predicting network behaviour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/50Reducing energy consumption in communication networks in wire-line communication networks, e.g. low power modes or reduced link rate
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Abstract

The invention discloses a local area network link flow prediction method based on an improved ConvLSTM deep learning model, and relates to the field of network flow prediction. Aiming at the problems that the prediction accuracy is low, the spatial characteristics of the target network flow cannot be predicted at the same time, the flow of all links of the target network cannot be predicted through a single model and the like in the target network flow prediction method based on the traditional linear model and the nonlinear model, the invention provides a local area network flow prediction method based on an improved ConvLSTM deep learning model. The model is added with a similar residual error structure module and a focus mechanism-based Squeeze-and-specification module so as to finally achieve the aims of accurate and rapid simulation and twinning of the target network link flow. The method improves the accuracy of local area network space-time flow matrix prediction, reduces training iteration times and improves the utilization rate of computing resources.

Description

Local area network flow prediction method based on deep learning model
Technical Field
The invention belongs to the technical fields of computer network traffic engineering, network twinning and network simulation, and particularly relates to a local area network traffic prediction method based on an improved ConvLSTM deep learning model.
Background
Network simulation and network twinning are research hotspots in the current computer network and communication arts. Besides the software simulation (simulation) method, the network simulation (simulation) technology can be adopted to realize high fidelity reproduction of key elements such as nodes, links, topology and the like in the target network. For example, a virtual instance in a cloud computing platform (a virtual machine based on a traditional host virtualization technology or a container based on a lightweight virtualization technology) is adopted to realize simulation or twin reproduction of a target network node, and a cloud platform bottom virtual link and a virtual network are adopted to realize simulation or twin reproduction of a target network link and topology. The network simulation technology has the advantages of high simulation fidelity, flexibility, expandability, support of upper layer protocols, direct deployment of application programs and the like.
Network simulation and network twinning are not only required to realize high fidelity reproduction of target network nodes, links and topology, but also required to realize real-time and accurate reproduction of transmission flow in a target network. The main method for realizing the link flow reproduction in the simulation network or the twin network in the target network comprises two methods: the first method is a method based on the target network traffic log. The target network flow can be completely recorded by using a tcpdump and other packet grabbing tools and stored as a TraceFile, and then the flow is reproduced in the simulation network or the twin network according to the file. Although the method can accurately reproduce the flow of the target network within a period of time, the method cannot meet the requirement of real-time reproduction. The second method is based on a target network historical flow data training (statistics, machine learning and deep learning) model, and predicting the future flow of the target network through the obtained model, and based on the model, realizing flow reproduction in a simulation network or a twin network. Although the method sacrifices a certain flow accuracy, the method can better meet the real-time requirements of network flow twinning and synchronization. Currently, the flow prediction methods can be classified into the following two types:
1) The first type of method comprises the following steps: predicting network traffic based on the linear model;
2) The second type of method is as follows: and predicting network traffic based on the nonlinear model.
In one aspect, the first class of methods includes autoregressive models, moving average models, autoregressive synthetic moving average models, and the like. The method is suitable for predicting the traffic generated by the network nodes in a short period, and a comprehensive model capable of accurately predicting the complex service application traffic generated by a plurality of network nodes is difficult to construct by adopting the method one. On the other hand, the second type of method comprises a wavelet model, a regression score integration moving average model, a support vector machine in the machine learning field, an artificial neural network model and the like, can realize the prediction of long-term traffic on the autocorrelation of network traffic, but cannot usually realize the high fidelity requirement of short-term traffic prediction. With the development of deep learning technology in recent years, prediction of target network flow is realized based on a deep neural network model, so that better accuracy can be achieved, and the method is gradually focused on related research fields. In particular, network traffic prediction methods based on Recurrent Neural Network (RNN) models, long and Short Term Memory (LSTM) network models, have enabled accurate prediction of network link traffic. However, the method based on deep learning generally only realizes traffic prediction of the time dimension of the links of the target network, cannot capture the spatial characteristics of multiple links of the target network, and cannot predict all link traffic and spatial association thereof in the target network at the same time.
The local area network is an important component part with the largest quantity and the most complicated flow in a typical target network, so in network simulation and network twinning application, in order to realize accurate and real-time reproduction of the whole network flow, a local area network flow accurate prediction theory method and an implementation technology based on deep learning need to be further researched so as to effectively improve the fidelity of local area network simulation and twinning.
Disclosure of Invention
Aiming at the problems that the prediction accuracy is low, the spatial characteristics of the target network flow cannot be predicted at the same time, the flow of all links of the target network cannot be predicted through a single model and the like in the target network flow prediction method based on the traditional linear model and the nonlinear model, the invention provides a local area network flow prediction method based on an improved ConvLSTM deep learning model. The model is added with a class residual error structure module (RES template) and a focus mechanism-based Squeeze-and-specification module (SE module) to finally achieve the aims of accurate and rapid simulation and twinning of the target network link flow.
The technical scheme of the invention is realized as follows: a local area network link flow prediction method based on an improved ConvLSTM deep learning model firstly sets a target local area network with N three-layer nodes, wherein the N is included c Each client node N s Individual server nodes and N r A plurality of router nodes; the method comprises the following steps:
step 1: collecting a local area network traffic matrix sequence, and performing equidistant averaging on the initial matrix sequence;
step 1.1: collecting the flow of all links of a target network, and counting the access flow reaching N sampling nodes every second to obtain flow data at the moment t:
wherein:
D t ∈R N×N
element(s)The flow of the i node to the j node at the time T is shown, and the T is the total sampling time;
step 1.2: collecting data of K days, and performing interval S on the collected data due to fluctuation of local area network flow data S The internal average is found that the traffic sequence length T= (24×60×60×K)/S S
Step 2: target network traffic matrix conversion, re-weighting N x N point-to-point traffic matrixModeling an M multiplied by N mode traffic matrix of the server node to the user node; where M is the number of uplink and downlink for the user to transmit data with the server, i.e., for N c Each client node N s Individual server nodes and N r Each router node, m= (N) s +N r ) X 2; the post-conversion traffic matrix is as follows:
wherein:
step 3: normalizing the converted flow matrix obtained in the step 2, and accelerating the speed of gradient descent to solve the optimal solution after normalization, thereby improving the prediction precision; the normalization formula is expressed as:
wherein: d, d max And d min Representing D' t The maximum and minimum values at all times,representing normalized values, the normalized flow matrix is represented as D * t
Step 4: to prevent model overfitting, the Loss function Loss is defined as the mean square error of the predicted flow matrix plus epsilon times the mean square error of the actual flow matrix to predict the predicted error for each element, as:
wherein epsilon is an adjustable super parameter,is the traffic matrix D * t True value +.>Is a predicted value of (2);
step 5: and (3) extracting time feature and space feature information by using a SE-ConvLSTM-Res model according to the mode flow matrix sequence obtained in the step (2), wherein the SE-ConvLSTM-Res model comprises the following steps: res module, SE module and ConvLSTM module;
step 5.1: the Res module operation specific flow is as follows:
step 5.1.1: for flow matrix D * t Splitting:
wherein the method comprises the steps of Element->n represents the total number of the split element matrixes;
step 5.1.2: for a pair ofThe elements of (2) are respectively combined with a convolutional neural network with a convolutional kernel of 2x2 after feature extraction, and finally added with the output of an SE module, and the process is as follows:
wherein O is RES For the output of the Res module, ε is an adjustable parameter, ω k Representing the weight of the kth layer of the convolutional neural network, b k Representing a deviation coefficient of the convolutional neural network;
step 5.2: the SE module mainly comprises compression operation and excitation operation, and the specific flow is as follows:
step 5.2.1: the compression operation is realized by global average pooling, which compresses the spatial features into a global feature in a global average way, but retains the information of the channels and the real-time features:
wherein z is E R C C represents a channel, z is the entire time series obtained,is an element of an ith link to a jth node;
step 5.2.2: the excitation operation is to capture the nonlinear relationship between channels; through the realization of an activation function and a full connection layer, the captured channel information is mapped between (0, 1) by using a sigmoid gate mechanism through the dimension reduction and dimension increase of the two full connection layers:
s=σ(ω 2 ⊙ReLU(ω 1 ⊙z))
where s represents the gate activation value, ω, between channels 1 For fully connected FC 1 Weights of layers, omega 2 For fully connected FC 2 The weight of the layer, sigma is a sigmoid activation function, and ReLU is an activation function;
finally, the learned gate activation value between channels is combined with x c Multiplying to obtain the output of the SE module as the time channel weight O SE ,x c Is a matrix element of a channel:
O SE =x c ⊙s;
step 5.3: the ConvLSTM module specifically comprises the following steps:
step 5.3.1: convLSTM improves full connectivity to convolutional neural networks based on LSTM, i.e., input X of model t H from the cell state at the previous time t Merging the signals passing through the convolutional neural network module to be used as the input of the SE module;
wherein ω represents a weight set of the convolutional neural network full-connection layer;
step 5.3.2: the output of the SE module is overlapped with the output of the Res module to be used as the input of the LSTM part, namely the input O RES +O SE
Step 5.3.3: the LSTM portion of the ConvLSTM module performs the following calculations in order:
f t =Sigmoid(Conv(x t ;ω xf )+Conv(h t-1 ;ω hf )+b f ) (4)
i t =Sigmoid(Conv(x t ;ω xi )+Conv(h t-1 ;ω hi )+b i ) (5)
o t =Sigmoid(Conv(x t ;ω xo )+Conv(h t-1 ;ω ho )+b o ) (6)
g t =Tanh(Conv(x t ;ω xg )+Conv(h t-1 ;ω hg )+b g ) (7)
c t =f t ⊙c t-1 +i t ⊙g t (8)
h t =o t ⊙Tanh(c t ) (9)
wherein x is t Is the input element of the corresponding calculation process of the LSTM part in the ConvLSTM module, and x t =O RES +O SE
ω xf 、ω hf 、ω xi 、ω hi 、ω xo 、ω ho 、ω xg 、ω hg Weights corresponding to convolution layers in the convolution calculation process, b f 、b i 、b o 、b g As the deviation coefficient of the neural network, C t-1 And h t-1 The state value of the model at the last moment of the calculation process is calculated;
finally, c t And h t As input of the full connection layer, the full connection layer output is a target matrix; and (3) training the SE-ConvLSTM-Res model by using the data obtained in the step (3), and carrying out flow prediction by using the trained SE-ConvLSTM-Res model.
The invention has the beneficial effects that: the invention provides the following methods for solving the problems of accurate and efficient prediction of the flow of a target local area network in the fields of network simulation and network twinning:
1. the deep learning method is applied to solving the problem of target local area network traffic prediction. Simplifying the original flow matrix data structure into a data structure based on the uplink and downlink of the server based on the unique characteristics of the local area network flow;
2. the variance of the loss value in the flow matrix is added into the loss function, so that the deep learning model is effectively prevented from being over fitted.
3. Based on the concept of deep learning SE module and residual error, a SE-ConvLSTM-Res enhancement model is provided, and the sensitivity of each moment can be better grasped in the time dimension by the enhancement model. In the spatial dimension, the traffic characteristics of the local area network can be learned more quickly.
The method improves the accuracy of local area network space-time flow matrix prediction, reduces training iteration times and improves the utilization rate of computing resources.
Drawings
FIG. 1 is a block diagram of a SE-ConvLSTM-Res deep learning model of the present invention;
fig. 2 is a flowchart illustrating the operation of the lan traffic matrix prediction method according to the present invention.
DETAILED DESCRIPTION OF EMBODIMENT (S) OF INVENTION
For a detailed description of the technical scheme disclosed in the invention, the following is a further detailed description of the invention with reference to the drawings and specific embodiments of the description:
in order to solve the problem of high-precision prediction of the local area network traffic matrix, the invention provides a local area network traffic prediction method based on deep learning. The method provided by the invention simplifies the original data structure into a data structure which is more beneficial to deep learning aiming at the characteristics of the local area network. In order to prevent the excessive difference of the errors of the flow predictions of all the nodes in the flow matrix, the variance of the loss value in the flow matrix is added into a loss function, so that the deep learning model is prevented from being over fitted. The invention adds SE and Res modules based on ConvLSTM, and provides an SE-ConvLSTM-Res improved model which can capture the time and space characteristics of a local area network flow matrix more quickly, improve the training speed of the model and improve certain prediction precision. Finally, the real-time synchronization and the accurate reproduction of the local area network flow in the twin network are facilitated.
The first step: collecting a local area network Traffic Matrix (Traffic Matrix) sequence, and performing equidistant average on the initial Matrix sequence:
(1.1) collecting the flow of all links of a target network, and counting the access flow reaching N sampling nodes every second to obtain flow data at the moment t:
wherein:
D t ∈R N×N
element(s)And the traffic size sent to the j node by the i node at the time t is shown.
(1.2) suppose data are collected for k days. Due to the fluctuation of the local area network flow data, the collected data is processed at small intervals S S Internal average, the traffic sequence length t= (24×60×60×k)/S S
For (1.1) accessing different servers in the local area network, different traffic models are generated, such as accessing a web server, the generated traffic has burstiness, and in order to acquire traffic characteristics of the web server as much as possible, the data sampling time of the node needs to be as short as possible. For (1.2), because the local area network flow data has larger fluctuation, the data with larger fluctuation can reduce the prediction precision of the model, so the sampled data obtained in (1.1) are sampled at equal m intervals, and the data stability is improved.
And a second step of: the local area network is provided with N three layers of nodes, wherein the N three layers of nodes comprise N c Each client node N s Individual server nodes and N r And a router node. The flow from the local area network end to the end can be divided into two parts, namely the flow between users and the flow between the users and the server:
D(t)=D c (t)+D s (t)
the user-to-user traffic can be expressed as:
the flow between the user and the server consists of the flow from the user to the server and the flow from the server to the user, namely the upstream flow of the server and the downstream flow of the server, which can be expressed as follows:
because of D in local area network c (t) is small and D s (t) is relatively large. Spatially D c (t) and D s (t) the large difference in size affects the accuracy of the model convolution extracted features. And D is s D in (t) cs (t) and D sc (t) have partial repetition, structurally transposed with respect to each other. And then, the local area network traffic matrix structure is remodelled, and the N multiplied by N point-to-point traffic matrix is remodelled into an M multiplied by N mode traffic matrix of the server node to the user node. Where M is the number of uplinks and downlinks for the user to transmit data with the server, where m= (N) s +N r ) X 2. The converted traffic matrix is as follows:
wherein:
D′ t ∈R M×N
D′ t elements of (2)Two models representing upstream and downstream traffic of a server, < >>Representing traffic destined for the jth user node by the ith server.
And a third step of: the flow matrix sequence is normalized, the speed of gradient descent to solve the optimal solution is increased after the data normalization, and the model precision is improved; the invention uses maximum and minimum normalization to normalize, and the formula can be expressed as:
fourth step: determining a loss function, a model training process may overstrain a certain space in MXNTraining of points, but neglecting other points in space, results in a decrease in overall prediction accuracy, although the overall traffic matrix error MSE is still in a decreasing trend. In order to prevent the model from being over fitted, the spatial overall prediction precision is synchronously improved, and the training loss function is defined as the mean square error MSE of a predicted flow matrix and an actual flow matrix plus the mean square error of the prediction error of each element of the flow matrix of epsilon times. Namely, the loss function is:
wherein epsilon is an adjustable super-parameter,is the traffic matrix D * t True value +.>Is a predicted value of (a).
Fifth step: and (3) extracting time feature and space feature information by using the SE-ConvLSTM-Res model according to the mode flow matrix sequence obtained in the second step. SE-ConvLSTM-Res model introduction:
(5.1) ConvLSTM model (Convolutional LSTM Network) is an improved model based on long-short-term memory neural network, and ConvLSTM improves the full connection of input and weight into convolution based on the long-short-term memory neural network LSTM model. The original one-dimensional sequence of the LSTM model is improved to input pictures which can be two-dimensional. The original LSTM model can only extract time feature information, and ConvLSTM can extract time and space feature information at the same time due to convolution. The method is suitable for the scene of local area network traffic matrix prediction.
(5.2) because the traffic patterns of different servers in the local area network are very different, such as file servers, web servers, audio and video servers and users, the traffic patterns of the audio and video servers are continuous in time, the traffic patterns of the web servers are bursty in time, the generated traffic is generally smaller than that of the audio and video, the traffic patterns of the file servers are continuous in time, have stronger correlation in space, and the generated traffic is larger. The ConvLSTM is directly used for flow prediction of the local area network, and interference information of other servers is mixed in a predicted flow matrix due to the characteristic of integral convolution when a model is trained, so that the time cost of model training is increased.
The invention is improved based on ConvLSTM model, introduces concepts of a module (Deep residual network) similar to ResNet and a SE module (Squeeze and Excitation Block) and obtains the SE-ConvLSTM-Res model. The SE module is a lightweight module, and utilizes a attention mechanism to screen output channel information of ConvLSTM-Res, so that sensitivity of a model to channel characteristics is improved, characteristics of important channels are strengthened, and characteristics of non-important channels are weakened. The introduced Res-like model will input information X t Information separation of individual servers of (2) feature extraction with a 2x2 size convolution kernel convolution layer. And after the extraction, combining the characteristic information of each server. The final ConvLSTM convolved output is added with the gamma-fold server characteristic information and combined as the in-cell input. The SE-ConvLSTM-Res model structure is shown in FIG. 1, wherein the SE module enhances channel information at sensitive moments and weakens channel information at insensitive moments in time sequence based on the attention mechanism. And the Res module spatially weakens the crossing of the spatial information between different servers and enhances the spatial information of a single server. The training speed of the model is increased, and the predicted loss is reduced.
SE-ConvLSTM-Res model flow introduction:
the specific operation flow of the improved model RES module is as follows:
(1.1) splitting an input traffic matrixSet of server traffic matrix>
Wherein the method comprises the steps of Element->
(1.2) pairThe elements of (2) are respectively combined with a convolutional neural network with a convolutional kernel of 2x2 after feature extraction, and finally added with the output of an SE module, and the process is as follows:
the improved model SE module mainly comprises compression operation and excitation operation, and the specific flow is as follows:
(1.1) the compression operation is implemented by global averaging pooling (global average pooling), which compresses the spatial features into a global feature on the whole average, but retains the information of its channels and the temporal features:
wherein z is E R C C represents a channel and z represents the entire time series.
(1.2) the excitation operation is to capture the nonlinear relationship between channels. By activating functions and fully connected layers. The activation functions include the ReLU and sigmoid functions. Through dimension reduction and dimension increase of two full connection layers, the captured channel information is mapped between (0, 1) by using a sigmoid gate mechanism:
s=σ(ω 2 ⊙ReLU(ω 1 ⊙z))
wherein the method comprises the steps ofFor fully connected FC 1 Weights of layers, ++>For fully connected FC 2 Layer weights. Sigma is a sigmoid activation function.
Finally, the learned gate activation value between channels is combined with x c Multiplication can achieve the effect of adjusting the time channel weight:
O SE =x c ⊙s
the ConvLSTM module specifically executes the following steps:
(1.1) ConvLSTM improves full connectivity to convolutional neural networks based on LSTM, input X t H from the cell state at the previous time t Merging the signals passing through the convolutional neural network module to be used as input of the SE module, wherein the formula is as follows:
I SE =Conv([X t ,h t ];ω)
(1.2) superimposing the output of the SE module and the Res module as the input to the LSTM section, i.e., input O RES +O SE
(1.3) the LSTM portion of ConvLSTM was calculated as follows:
f t =Sigmoid(Conv(x t ;ω xf )+Conv(h t-1 ;ω hf )+b f ) (4)
i t =Sigmoid(Conv(x t ;ω xi )+Conv(h t-1 ;ω hi )+b i ) (5)
o t =Sigmoid(Conv(x t ;ω xo )+Conv(h t-1 ;ω ho )+b o ) (6)
g t =Tanh(Conv(x t ;ω xg )+Conv(h t-1 ;ω hg )+b g ) (7)
c t =f t ⊙c t-1 +i t ⊙g t (8)
h t =o t ⊙Tanh(c t ) (9)
wherein the method comprises the steps ofb is the deviation, C t-1 And H is t-1 Representing the state value of the last moment model.
Sixth step: in order to make the structure of the resulting output consistent with that expected, the output of the SE-ConvLSTM-Res model is finally passed through the full connection layer.
In the experiment, the input dimension is set to be 20, and the output hidden dimension is set to be 16. The convolution kernel size of the ConvLSTM portion in the SE-ConvLSTM-Res model is 3x3. The Res module output weight epsilon is set to be 1, and the learning rate is 10 -3 The attenuation ratio gamma is 0.9.

Claims (1)

1. A local area network link flow prediction method based on an improved ConvLSTM deep learning model firstly sets a target local area network with N three-layer nodes, wherein the N is included c Each client node N s Individual server nodes and N r A plurality of router nodes; the method comprises the following steps:
step 1: collecting a local area network traffic matrix sequence, and performing equidistant averaging on the initial matrix sequence;
step 1.1: collecting the flow of all links of a target network, and counting the access flow reaching N sampling nodes every second to obtain flow data at the moment t:
wherein:
D t ∈R N×N
element(s)The flow of the i node to the j node at the time T is shown, and the T is the total sampling time;
step 1.2: collecting data of K days, and performing interval S on the collected data due to fluctuation of local area network flow data S The internal average is found that the traffic sequence length T= (24×60×60×K)/S S
Step 2: target network traffic momentMatrix conversion, remodelling the N multiplied by N point-to-point traffic matrix into an M multiplied by N mode traffic matrix of the server node to the user node; where M is the number of uplink and downlink for the user to transmit data with the server, i.e., for N c Each client node N s Individual server nodes and N r Each router node, m= (N) s +N r ) X 2; the post-conversion traffic matrix is as follows:
wherein:
step 3: normalizing the converted flow matrix obtained in the step 2, and accelerating the speed of gradient descent to solve the optimal solution after normalization, thereby improving the prediction precision; the normalization formula is expressed as:
wherein: d, d max And d min Representing D' t The maximum and minimum values at all times,representing normalized values, the normalized flow matrix is represented as D * t
Step 4: to prevent model overfitting, the Loss function Loss is defined as the mean square error of the predicted flow matrix plus epsilon times the mean square error of the actual flow matrix to predict the predicted error for each element, as:
wherein epsilon is an adjustable super parameter,is the traffic matrix D * t True value +.>Is a predicted value of (2);
step 5: and (3) extracting time feature and space feature information by using a SE-ConvLSTM-Res model according to the mode flow matrix sequence obtained in the step (2), wherein the SE-ConvLSTM-Res model comprises the following steps: res module, SE module and ConvLSTM module;
step 5.1: the Res module operation specific flow is as follows:
step 5.1.1: for flow matrix D * t Splitting:
wherein the method comprises the steps of Element->n represents the total number of the split element matrixes;
step 5.1.2: for a pair ofThe elements of (2) are respectively combined with a convolutional neural network with a convolutional kernel of 2x2 after feature extraction, and finally added with the output of an SE module, and the process is as follows:
wherein O is RES For the output of the Res module, ε is an adjustable parameter, ω k Representing the weight of the kth layer of the convolutional neural network, b k Representing a deviation coefficient of the convolutional neural network;
step 5.2: the SE module mainly comprises compression operation and excitation operation, and the specific flow is as follows:
step 5.2.1: the compression operation is realized by global average pooling, which compresses the spatial features into a global feature in a global average way, but retains the information of the channels and the real-time features:
wherein z is E R C C represents a channel, z is the entire time series obtained,is an element of an ith link to a jth node;
step 5.2.2: the excitation operation is to capture the nonlinear relationship between channels; through the realization of an activation function and a full connection layer, the captured channel information is mapped between (0, 1) by using a sigmoid gate mechanism through the dimension reduction and dimension increase of the two full connection layers:
s=σ(ω 2 ⊙ReLU(ω 1 ⊙z))
where s represents the gate activation value, ω, between channels 1 For fully connected FC 1 Weights of layers, omega 2 For fully connected FC 2 Weights of layers, sigma is sigmoidAn activation function, reLU is the activation function;
finally, the learned gate activation value between channels is combined with x c Multiplying to obtain the output of the SE module as the time channel weight O SE ,x c Is a matrix element of a channel;
O SE =x c ⊙s;
step 5.3: the ConvLSTM module specifically comprises the following steps:
step 5.3.1: convLSTM improves full connectivity to convolutional neural networks based on LSTM, i.e., input X of model t H from the cell state at the previous time t Merging the signals passing through the convolutional neural network module to be used as the input of the SE module;
wherein ω represents a weight set of the convolutional neural network full-connection layer;
step 5.3.2: the output of the SE module is overlapped with the output of the Res module to be used as the input of the LSTM part, namely the input O RES +O SE
Step 5.3.3: the LSTM portion of the ConvLSTM module performs the following calculations in order:
f t =Sigmoid(Conv(x t ;ω xf )+Conv(h t-1 ;ω hf )+b f ) (4)
i t =Sigmoid(Conv(x t ;ω xi )+Conv(h t-1 ;ω hi )+b i ) (5)
o t =Sigmoid(Conv(x t ;ω xo )+Conv(h t-1 ;ω ho )+b o ) (6)
g t =Tanh(Conv(x t ;ω xg )+Conv(h t-1 ;ω hg )+b g ) (7)
c t =f t ⊙c t-1 +i t ⊙g t (8)
h t =o t ⊙Tanh(c t ) (9)
wherein x is t Is the input element of the corresponding calculation process of the LSTM part in the ConvLSTM module, and x t =O RES +O SE
ω xf 、ω hf 、ω xi 、ω hi 、ω xo 、ω ho 、ω xg 、ω hg Weights corresponding to convolution layers in the convolution calculation process, b f 、b i 、b o 、b g As the deviation coefficient of the neural network, C t-1 And h t-1 The state value of the model at the last moment of the calculation process is calculated;
finally, c t And h t As input of the full connection layer, the full connection layer output is a target matrix; and (3) training the SE-ConvLSTM-Res model by using the data obtained in the step (3), and carrying out flow prediction by using the trained SE-ConvLSTM-Res model.
CN202210561413.1A 2022-05-23 2022-05-23 Local area network flow prediction method based on deep learning model Active CN115022193B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210561413.1A CN115022193B (en) 2022-05-23 2022-05-23 Local area network flow prediction method based on deep learning model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210561413.1A CN115022193B (en) 2022-05-23 2022-05-23 Local area network flow prediction method based on deep learning model

Publications (2)

Publication Number Publication Date
CN115022193A CN115022193A (en) 2022-09-06
CN115022193B true CN115022193B (en) 2024-02-02

Family

ID=83068177

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210561413.1A Active CN115022193B (en) 2022-05-23 2022-05-23 Local area network flow prediction method based on deep learning model

Country Status (1)

Country Link
CN (1) CN115022193B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115187154B (en) * 2022-09-14 2022-12-16 华中科技大学 Neural network-based regional power grid oscillation source risk prediction method and system
CN115426201B (en) * 2022-11-03 2023-01-17 湖南大佳数据科技有限公司 Data acquisition method and system for network shooting range
CN116319378B (en) * 2023-05-16 2023-07-21 合肥工业大学 Network traffic matrix estimation and model training method and system based on deep learning

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110851782A (en) * 2019-11-12 2020-02-28 南京邮电大学 Network flow prediction method based on lightweight spatiotemporal deep learning model
CN112291808A (en) * 2020-11-02 2021-01-29 东南大学 Regional network flow prediction method based on deep learning
CN112508173A (en) * 2020-12-02 2021-03-16 中南大学 Traffic space-time sequence multi-step prediction method, system and storage medium
CN112669606A (en) * 2020-12-24 2021-04-16 西安电子科技大学 Traffic flow prediction method for training convolutional neural network by utilizing dynamic space-time diagram
CN112910711A (en) * 2021-02-03 2021-06-04 山东大学 Wireless service flow prediction method, device and medium based on self-attention convolutional network
WO2021186158A1 (en) * 2020-03-17 2021-09-23 The University Court Of The University Of Edinburgh A distributed network traffic data decomposition method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111130839B (en) * 2019-11-04 2021-07-16 清华大学 Flow demand matrix prediction method and system

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110851782A (en) * 2019-11-12 2020-02-28 南京邮电大学 Network flow prediction method based on lightweight spatiotemporal deep learning model
WO2021186158A1 (en) * 2020-03-17 2021-09-23 The University Court Of The University Of Edinburgh A distributed network traffic data decomposition method
CN112291808A (en) * 2020-11-02 2021-01-29 东南大学 Regional network flow prediction method based on deep learning
CN112508173A (en) * 2020-12-02 2021-03-16 中南大学 Traffic space-time sequence multi-step prediction method, system and storage medium
CN112669606A (en) * 2020-12-24 2021-04-16 西安电子科技大学 Traffic flow prediction method for training convolutional neural network by utilizing dynamic space-time diagram
CN112910711A (en) * 2021-02-03 2021-06-04 山东大学 Wireless service flow prediction method, device and medium based on self-attention convolutional network

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Deep Learning Based Traffic Prediction Method for Digital Twin Network;Junyu Lai;《Springer》;全文 *
基于全注意力机制的多步网络流量预测;郭佳;余永斌;杨晨阳;;信号处理(第05期);全文 *
基于神经网络的网络流量预测研究;吴雨聪;《硕士电子期刊》;全文 *

Also Published As

Publication number Publication date
CN115022193A (en) 2022-09-06

Similar Documents

Publication Publication Date Title
CN115022193B (en) Local area network flow prediction method based on deep learning model
Yu et al. Intelligent edge: Leveraging deep imitation learning for mobile edge computation offloading
CN110851782A (en) Network flow prediction method based on lightweight spatiotemporal deep learning model
CN112491442B (en) Self-interference elimination method and device
CN112152948A (en) Wireless communication processing method and device
CN111245673A (en) SDN time delay sensing method based on graph neural network
WO2021036414A1 (en) Co-channel interference prediction method for satellite-to-ground downlink under low earth orbit satellite constellation
CN111224905B (en) Multi-user detection method based on convolution residual error network in large-scale Internet of things
Jiang et al. Federated learning algorithm based on knowledge distillation
Lu et al. Heterogeneous model fusion federated learning mechanism based on model mapping
Nguyen et al. Hcfl: A high compression approach for communication-efficient federated learning in very large scale iot networks
Wang et al. Deep joint source-channel coding for multi-task network
CN115115510A (en) Image processing method, system, storage medium and terminal equipment
CN114528987A (en) Neural network edge-cloud collaborative computing segmentation deployment method
Dangi et al. 5G network traffic control: a temporal analysis and forecasting of cumulative network activity using machine learning and deep learning technologies
Li et al. A unified federated DNNs framework for heterogeneous mobile devices
CN116362327A (en) Model training method and system and electronic equipment
CN104955059B (en) Cellular network base stations state time-varying model method for building up based on Bayesian network
Nakip et al. Subspace-based emulation of the relationship between forecasting error and network performance in joint forecasting-scheduling for the Internet of Things
KR20220124106A (en) Method and apparatus for graph neural network based virtual network management
CN111541472B (en) Low-complexity machine learning assisted robust precoding method and device
Zhang et al. Network traffic classification method based on subspace triple attention mechanism
CN116996401A (en) Flow prediction method based on space-time enhanced cyclic neural network
Chen et al. FUNOff: Offloading Applications At Function Granularity for Mobile Edge Computing
CN115102767B (en) DDoS active defense system and method based on distributed collaborative learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant