CN111030889B - Network traffic prediction method based on GRU model - Google Patents
Network traffic prediction method based on GRU model Download PDFInfo
- Publication number
- CN111030889B CN111030889B CN201911343425.1A CN201911343425A CN111030889B CN 111030889 B CN111030889 B CN 111030889B CN 201911343425 A CN201911343425 A CN 201911343425A CN 111030889 B CN111030889 B CN 111030889B
- Authority
- CN
- China
- Prior art keywords
- gru
- model
- neural network
- data
- adagrad
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L43/00—Arrangements for monitoring or testing data switching networks
- H04L43/08—Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
- H04L43/0876—Network utilisation, e.g. volume of load or congestion level
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Environmental & Geological Engineering (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Traffic Control Systems (AREA)
Abstract
The invention discloses a network flow prediction method based on a GRU model, relating to the technical field of information communication; inputting a network flow data sequence into a GRU neural network model and completing the prediction of network flow; the accuracy and the effect of network flow prediction are improved by inputting the network flow data sequence into the GRU neural network model and completing the prediction of the network flow and the like.
Description
Technical Field
The invention relates to the technical field of information communication, in particular to a network traffic prediction method based on a GRU model.
Background
The electric power data communication network is a comprehensive wide area network transmission platform and is also an important component of an electric power information infrastructure. With the rapid development of power data networks, the network scale is continuously enlarged, and sufficient and reliable information support is increasingly required to ensure the safe and reliable operation of the power data networks. The prediction of the network flow of the power data network can provide important information for the safe operation of the power data network, and particularly, the method can sense the flow abnormality and the operation state abnormality of the power data network in advance so as to guarantee the operation of the power data network, so that the method has important research value and application prospect. In general, network traffic data is affected by various complex and random factors, but is nonlinear time-series data in nature.
The characteristics of modern internet make network traffic prediction important in improving network efficiency, reliability and adaptability. In recent years, many researchers have made studies on network traffic prediction, and many network traffic prediction methods have been proposed. Currently, prediction models for network traffic prediction include time series models, neural network models, and the like. However, because the network flow data sequence is influenced by various uncertain factors and the data of the influencing factors is difficult to express, the network flow sequence has the complex characteristics of high nonlinearity and non-stationarity, and the traditional time sequence model and the neural network model are difficult to process, so that the accuracy of the network flow prediction by adopting a simple prediction model is low, and the reasonable planning and distribution of the network are influenced.
Therefore, how to improve the accuracy of network traffic prediction to improve the reliability of the network is a problem to be solved by those skilled in the art.
In order to solve the development state of the prior art, the existing patents and documents are searched, compared and analyzed, and the following technical information with high relevance to the invention is screened out:
patent scheme 1:201510793377.1 network traffic prediction method based on traffic trend
The invention provides a method for processing faults of a wireless sensor network, which aims at solving the problems that the wireless sensor network adopting a centralized fault processing mode occupies network resources and the wireless sensor network adopting a distributed fault processing mode consumes sensor resources and energy. The method comprises the following steps: extracting a network traffic trend in a time period before a current time period; predicting the network traffic trend at a future moment according to the extracted network traffic trend; calculating the error between the extracted network flow value and the network flow trend thereof, and predicting the flow error; and predicting the predicted value of the network flow at the future moment according to the predicted network flow trend and the predicted flow error. The invention greatly reduces the number of training samples required by flow error prediction and flow estimation, and saves training time; and the extracted network traffic trend not only highlights the periodic characteristics of the traffic in each time period, but also maintains the local structural characteristics of the traffic.
Patent scheme 2:201611249158.8 neural network-based network traffic prediction system and traffic prediction method thereof
The invention provides a network traffic prediction method based on a BP (back propagation) neural network, which is based on the principle that data is normalized to enable a sample data value to be between 0 and 1, parameters of the BP neural network are initialized, the BP neural network is pre-trained and optimized by using a BP algorithm, and finally the trained BP neural network is used for prediction to obtain a prediction result. The method can not only extract the characteristics of the data, but also optimize the network by using the BP algorithm, thereby solving the problem of complex network structure and difficult training and improving the accuracy of flow prediction to a certain extent. The invention can monitor, detect and analyze various backbone networks, monitor and detect network abnormal events in the backbone networks in real time, and realize early warning of network abnormal conditions.
Patent scheme 3:201810011664.6 traffic prediction method based on neural network
The invention provides a neural network-based traffic prediction method, which is characterized in that computer data are sampled according to a set sampling time period, the window length of a training set is determined, and the abnormal traffic can be prevented and detected by matching the use of data sampling, data set setting, LSTM model training and data judgment. The method comprises the following steps: sampling computer data according to a set sampling time period; dividing a training set and a verification set; substituting into the LSTM model for model training and verification; and sampling the computer flow to be predicted and then bringing the sampled computer flow into a well-trained LSTM model for prediction. The invention can realize the prevention and detection of abnormal flow by matching the data sampling, the data set setting, the LSTM model training and the data judgment, and has the characteristics of high automatic degree, high detection speed and wide application range.
The defects of the above patent scheme 1: the scheme extracts the network traffic trend of a period before the current moment and predicts the network traffic trend of a period in the future according to real-time network traffic data; then, calculating the error of the network flow and the network flow trend in the past period and predicting the future network flow error; finally, predicting a predicted value of future network flow according to the predicted network flow trend and the predicted network flow error; in the scheme, the predefined cycle time has an important influence on the prediction of the future network traffic, so that the accuracy of the predicted value is influenced, and the prediction scheme is more difficult to express a highly complex nonlinear sequence, so that the universality of the scheme is not high.
The defects of the above patent scheme 2: the scheme provides a network flow prediction method based on a BP neural network. The BP neural network is easy to establish and train, has certain expression capacity on complex data sequences, firstly performs data normalization, then performs pre-training on the BP neural network, optimizes the BP neural network by using a BP algorithm, and finally performs prediction by using the trained BP neural network to obtain prediction. In the scheme, the BP neural network is mainly adopted to predict the network traffic data, but the BP neural network has poor memorability to the traffic data, and the improvement of the traffic prediction precision is limited.
The defect of the above patent scheme 3: the scheme provides a flow prediction method based on a neural network, which is used for sampling computer data according to a set sampling time period, training by using an LSTM model and predicting. However, the patent scheme only uses a single LSTM model, and although the LSTM model has better expressiveness in a nonlinear sequence, in practice, due to the defect of the gradient descent method, the learning rate is too fast to skip the optimal point, and there is a certain improvement space in prediction accuracy.
Problems with the prior art and considerations:
how to solve the technical problem of improving the accuracy and the effect of predicting the network flow.
Disclosure of Invention
The technical problem to be solved by the invention is to provide a network traffic prediction method based on a GRU model, which improves the accuracy and effect of network traffic prediction by inputting a network traffic data sequence into the GRU neural network model and completing the prediction of network traffic and the like.
In order to solve the technical problems, the technical scheme adopted by the invention is as follows: a network flow prediction method based on a GRU model inputs a network flow data sequence into a GRU neural network model and completes the prediction of network flow.
The further technical scheme is as follows: the GRU-Adam neural network model is a neural network model using an SGD gradient descent algorithm, the GRU-Adam model is a neural network model using an Adam gradient descent algorithm, the GRU-AdaGrad model is a neural network model using an AdaGrad gradient descent algorithm, the GRU-AdaGrad model is predicted by respectively using the GRU-SGD model, the GRU-Adam model and the GRU-AdaGrad model, and data predicted by each model are added to calculate an average value to obtain predicted network flow data.
The further technical scheme is as follows: specifically comprises the steps S1 to S5,
s1, obtaining historical network flow data;
s2, determining training data and verification data in historical network traffic data;
s3, bringing training data into each GRU neural network model for training;
s4, predicting the verification data through three GRU neural network models;
and S5, adding the predicted data, and averaging to obtain the predicted network flow data.
The further technical scheme is as follows: wherein the step S3 specifically comprises the step S31,
s31, three GRU neural network models are respectively a GRU-SGD model, a GRU-Adam model and a GRU-AdaGrad model, the GRU-SGD model is a neural network model using an SGD gradient descent algorithm, the GRU-Adam model is a neural network model using an Adam gradient descent algorithm, the GRU-AdaGrad model is a neural network model using an AdaGrad gradient descent algorithm, training data are respectively input into each GRU neural network model, and the training data are firstly transmitted in the GRU neural network model in a forward direction.
The further technical scheme is as follows: wherein the step S3 further comprises a step S32,
and S32, calculating a loss function of time.
The further technical scheme is as follows: wherein the step S3 further comprises a step S33,
and S33, sequentially iterating until the loss function is converged by using the reverse chain type derivation.
The further technical scheme is as follows: wherein the step S3 further comprises a step S34,
s34, the GRU-SGD model is updated by using an SGD gradient descent algorithm, the GRU-Adam model is updated by using an Adam gradient descent algorithm, and the GRU-AdaGrad model is updated by using an AdaGrad gradient descent algorithm.
The further technical scheme is as follows: wherein the step S3 further comprises a step S35,
and S35, repeating the steps from S31 to S34, continuously updating, stopping until the loss function is less than 0.2, and finishing the model training.
The further technical scheme is as follows: wherein the step S34 specifically comprises the steps S341 to S343,
s341, calculating the reduction amount of each parameter by using an SGD gradient reduction algorithm through the GRU-SGD model, and updating;
s342, calculating the descending amount of each parameter by using an Adam gradient descending algorithm through the GRU-Adam model, and updating;
and S343, calculating the reduction amount of each parameter by using an AdaGrad gradient reduction algorithm through the GRU-AdaGrad model, and updating.
The further technical scheme is as follows: the method is run on a server basis.
Adopt the produced beneficial effect of above-mentioned technical scheme to lie in:
firstly, the accuracy and the effect of network flow prediction are improved by inputting a network flow data sequence into a GRU neural network model and completing the prediction of network flow and the like.
Secondly, the GRU-SGD model, the GRU-Adam model and the GRU-AdaGrad model are used for prediction respectively, the data predicted by each model are added to obtain the average value to obtain the predicted network traffic data, and the accuracy and the effect of network traffic prediction are further improved.
See detailed description of the preferred embodiments.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a block diagram of a GRU neural network model in the present invention;
FIG. 3 is a graph comparing predicted flow rate data with actual flow rate data in the present invention.
Detailed Description
The technical solutions in the embodiments of the present application will be described clearly and completely with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only some embodiments of the present application, and not all embodiments. The following description of at least one exemplary embodiment is merely illustrative in nature and is in no way intended to limit the application, its application, or uses. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present application, but the present application may be practiced in other ways than those described herein, and it will be appreciated by those skilled in the art that the present application may be practiced without departing from the spirit and scope of the present application, and that the present application is not limited to the specific embodiments disclosed below.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present application. As used herein, the singular forms "a", "an", and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
The relative arrangement of the components and steps, the numerical expressions, and numerical values set forth in these embodiments do not limit the scope of the present application unless specifically stated otherwise. Meanwhile, it should be understood that the sizes of the respective portions shown in the drawings are not drawn in an actual proportional relationship for the convenience of description. Techniques, methods, and apparatus known to those of ordinary skill in the relevant art may not be discussed in detail but are intended to be part of the specification where appropriate. In all examples shown and discussed herein, any particular value should be construed as merely illustrative, and not limiting. Thus, other examples of the exemplary embodiments may have different values. It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, it need not be discussed further in subsequent figures.
In the description of the present application, it is to be understood that the orientation or positional relationship indicated by the directional terms such as "front, rear, upper, lower, left, right", "lateral, vertical, horizontal" and "top, bottom", etc., are generally based on the orientation or positional relationship shown in the drawings, and are used for convenience of description and simplicity of description only, and in the case of not making a reverse description, these directional terms do not indicate and imply that the device or element being referred to must have a particular orientation or be constructed and operated in a particular orientation, and therefore, should not be considered as limiting the scope of the present application; the terms "inner and outer" refer to the inner and outer relative to the profile of the respective component itself.
For ease of description, spatially relative terms such as "over 8230 \ 8230;,"' over 8230;, \8230; upper surface "," above ", etc. may be used herein to describe the spatial relationship of one device or feature to another device or feature as shown in the figures. It will be understood that the spatially relative terms are intended to encompass different orientations of the device in use or operation in addition to the orientation depicted in the figures. For example, if a device in the figures is turned over, devices described as "above" or "on" other devices or configurations would then be oriented "below" or "under" the other devices or configurations. Thus, the exemplary terms "at 8230; \8230; 'above" may include both orientations "at 8230; \8230;' above 8230; 'at 8230;' below 8230;" above ". The device may be otherwise variously oriented (rotated 90 degrees or at other orientations) and the spatially relative descriptors used herein interpreted accordingly.
It should be noted that the terms "first", "second", and the like are used to define the components, and are only used for convenience of distinguishing the corresponding components, and the terms have no special meanings unless otherwise stated, and therefore, the scope of protection of the present application is not to be construed as being limited.
As shown in fig. 1, the present invention discloses a network traffic prediction method based on a GRU model, which includes steps S1 to S5, and the network traffic data sequence is input into a GRU neural network model and the prediction of the network traffic is completed, specifically as follows:
the GRU-Adam neural network model is a neural network model using an Adam gradient descent algorithm, the GRU-AdaGrad model is a neural network model using an AdaGrad gradient descent algorithm, the GRU-AdaGrad model is predicted by respectively using the GRU-SGD model, the GRU-AdaGrad model and the GRU-AdaGrad model, and data predicted by each model are added to calculate an average value to obtain predicted network flow data.
S1, obtaining historical network flow data.
And S2, determining training data and verification data in the historical network flow data.
And S3, bringing the training data into each GRU neural network model for training.
S31, three GRU neural network models are respectively a GRU-SGD model, a GRU-Adam model and a GRU-AdaGrad model, the GRU-SGD model is a neural network model using an SGD gradient descent algorithm, the GRU-Adam model is a neural network model using an Adam gradient descent algorithm, the GRU-AdaGrad model is a neural network model using an AdaGrad gradient descent algorithm, training data are respectively input into each GRU neural network model, and the training data are firstly transmitted in the GRU neural network model in a forward direction.
And S32, calculating a time loss function.
And S33, sequentially iterating until the loss function is converged by using the reverse chain type derivation.
S34, updating the GRU-SGD model by using an SGD gradient descent algorithm, updating the GRU-Adam model by using an Adam gradient descent algorithm, and updating the GRU-AdaGrad model by using an AdaGrad gradient descent algorithm.
S341, calculating the reduction amount of each parameter by using an SGD gradient reduction algorithm through the GRU-SGD model, and updating.
And S342, calculating the reduction amount of each parameter by using an Adam gradient reduction algorithm through the GRU-Adam model, and updating the reduction amount.
And S343, calculating the reduction amount of each parameter by using an AdaGrad gradient reduction algorithm through the GRU-AdaGrad model, and updating.
And S35, repeating the steps from S31 to S34, continuously updating, stopping until the loss function is less than 0.2, and finishing the model training.
And S4, predicting the verification data through three GRU neural network models.
And S5, adding the predicted data, and averaging to obtain the predicted network flow data.
The GRU neural network model, the SGD gradient descent algorithm, the Adam gradient descent algorithm, and the AdaGrad gradient descent algorithm are prior art and are not described herein again.
Description of the drawings:
first, the present invention needs to explain variables used in a GRU-based network traffic prediction method. The variables used were as follows:
zt: an update gate at time t;
rt: a reset gate at time t;
ht: output information of the GRU unit at the time t;
yt: final output information at time t;
Δθt: the value of the gradient decline of the parameter.
The GRU-based network traffic prediction method comprises the steps of inputting a network traffic data sequence into a GRU neural network, training different GRU neural network models by adopting different gradient descent algorithms, and finally adding and restoring data predicted by the models into predicted network traffic data. In accordance with the variables defined above, the solution of the invention is explained in detail below with reference to fig. 1.
As shown in fig. 1, the steps are described as follows:
s1, acquiring historical network flow data;
s2, determining training data and verification data in historical network flow data;
s3, bringing training data into a plurality of GRU models for training;
s4, predicting the verification data by a plurality of GRU neural network models;
and S5, adding the predicted data and averaging to obtain predicted network flow data.
Wherein, step S3 specifically includes:
s31, inputting the training data x (t) into a GRU neural network model, wherein three GRU models are provided, namely GRU-SGD, GRU-Adam and GRU-AdaGrad, and the difference is that the used gradient descent algorithms are different. Training data is first propagated forward in the GRU neural unit.
GRU is composed of update gate and reset gate, and at time step t, update gate ztThe calculation formula of (2) is as follows:
zt=σ(Whz*ht-1+Wxz*xt) (1)
wherein xtAs input vector at the t-th time step, ht-1For information of the previous time step t-1, WhzIs a weight matrix, WhzTo update the matrix, sigma is a sigmoid activation function, and information can be compressed to be between 0 and 1, and the formula and the derivative formula are as follows:
σ′(z)=y(1-y) (3)
as shown in fig. 2, the refresh gate primarily determines how much information of the past time step can be retained until the subsequent time step.
The reset gate mainly determines how much past time step information is forgotten, and the reset gate rtThe calculation formula of (c) is:
rt=σ(Whr*ht-1+Wxr*xt) (4)
wherein, WxrIs a weight matrix, WhrTo update the matrix. Memorizing informationStoring information of past time steps by resetting gates, memorizing informationThe calculation formula of (c) is:
wherein, WxcIs a weight matrix, WhcTo update the matrix, the line is Hadamard product,for the tanh activation function, the tanh activation function formula and its derivative formula are as follows:
GRU output unit ht:
Wherein h ist-1Output cell information for last time step, ztIn order to update the door information,for memorizing information.
Final output value ytPassing again through sigmoid activation function, WoAs a weight matrix:
yt=σ(Wo*ht) (9)
s32, the formula in the forward propagation process shows that the parameter to be learned has Whz、Wxz、Whr、Wxr、Whc、Wxc、WoThe final output of the output layer is ytCalculating a loss function for a certain time:
wherein y isdAre true values. The loss of a single sequence is then:
and S33, gradually solving the derivative of each loss function to each parameter W by using the reverse chain derivation:
wherein each intermediate parameter is:
δy,t=(yd-yt)*σ′ (19)
δh,t=δy,tWo+δz,t+1Whz+δt+1Whc*rt+1+δh,t+1Whr+δh,t+1*(1-zt+1) (20)
after the partial derivatives for each parameter are calculated, the parameters may be updated, and the iterations may be performed until the loss function converges.
And S34, updating parameters by adopting different gradient descent algorithms for the three different GRU network models respectively.
Wherein, S34 step includes:
s341, calculating the reduction quantity delta theta of each parameter by using an SGD gradient reduction algorithm through a GRU-SGD modeltThereby updating the parameter W.
Wherein, gtIs the gradient of the weight(s),for learning rate, Δ θtIs the amount by which the parameter W is decreased.
S342, calculating the quantity delta theta of reduction of each parameter by using an Adam gradient reduction algorithm through a GRU-Adam modeltThereby updating the parameter W.
The Adam algorithm adjusts the learning rate of each parameter by first order moment estimation and second order moment estimation of the gradient. Adam corrects the first moment estimation and the second moment estimation in an offset mode, so that the learning rate of each iteration has a stable range, and parameters are changed stably.
mt=μ*mt-1+(1-μ)*gt (26)
Wherein, gtIs the gradient of the weight, mt,ntFirst and second order moment estimates of the parametric partial derivatives, respectively, mu and v are exponential decay rates, between 0, 1), typically 0.9 and 0.999,is a correction value. Delta thetatIs the amount by which the parameter W is decreased,is the learning rate.
S343, GRU-AdaGrad model calculates the reduction quantity delta theta of each parameter by using an AdaGrad gradient reduction algorithmtThereby updating the parameter W.
The AdaGrad algorithm forms a constraint term by recursion, the early stage gtWhen the gradient is small, the constraint term is large, the gradient can be amplified, and the later period gtWhen the gradient is larger, the constraint term is smaller, and the gradient can be constrained.
Wherein, gtIs the weight gradient, ntIs an estimate of the second moment of the weight gradient, Δ θtIs the amount by which the weight is decreased,to learn the rate, e is used to guarantee that the denominator is not 0.
And S35, repeating the steps S31-S34, continuously updating the W parameter until the loss function E is less than 0.2, and finishing the model training.
Example data for the present invention are illustrated below:
s1, in the present embodiment, 14776 pieces of network traffic sequence data are collected as a data set.
S2, in the present embodiment, the first 12000 pieces of data in the network traffic sequence data set are used as a training set train (t), and the last 2776 pieces of data are used as a verification set val (t).
S3, in the embodiment of the invention, the training set sequence train (t) is input into three GRU neural network models for training, each GRU neural network comprises 32 GRU neural network units, the random batch size is set to be 128, and the training is carried out 100 times, wherein the GRU neural network units areThe SGD model adopts SGD to perform gradient descent, the GRU-Adam model adopts Adam to perform gradient descent, and the GRU-AdaGrad model adopts AdaGrad to perform gradient descent. Obtaining a well-trained GRU-SGD model after the training is finishedsgdGRU-Adam modelAdamGRU-AdaGrad modelAdaGrad. An Adam gradient descent algorithm is used.
S4, inputting the verification set data val (t) into the trained GRU neural network modelsgd、modelAdam、modelAdaGradIn (1), output predicted data presgd(t)、preAdam(t)、preAdaGrad(t)。
S5, adding and summing the predicted sub-sequence data to obtain final predicted network flow data
As shown in fig. 3, the predicted flow value pre (t) is compared with the actual flow value val (t).
The purpose of the invention is as follows:
the network flow prediction aims at accurately predicting the flow change in a future network and providing reliable data for network planning and maintenance. Most of the existing network flow prediction models adopt a method for constructing a linear mathematical model or a neural network model, most of the existing network flow prediction methods are processed by adopting a linear or nonlinear method, and the accuracy and the real-time performance of prediction caused by the planarity are difficult to ensure. In order to solve the above problems, the present patent provides a network traffic prediction method based on GRU. The method comprises the steps of inputting a network flow data sequence into a plurality of built GRU network models, adopting different gradient descent algorithms for each GRU model, respectively predicting by utilizing GRU neural network models, and then adding predicted data to obtain the predicted network flow data by averaging. The invention selects to integrate the GRU neural network in the neural network model into the network flow prediction, and the GRU network has good memorability and expression capability to the time sequence. For network flow data under a common condition, the change of the network flow data is influenced by various factors which are difficult to express, the sequence of the network flow data has the complex characteristics of high nonlinearity and non-stationarity, and different characteristics exist under different network environment conditions. The invention discloses a GRU-based network traffic prediction method, which respectively predicts data by utilizing GRU neural network models adopting different gradient descent algorithms, and finally adds and calculates an average value to obtain predicted network traffic data. The invention aims to overcome the defects in the prior art on the basis of the prior art and further improve the accuracy of network flow data sequence prediction.
The technical contribution of the invention is as follows:
the network flow prediction is widely applied to various fields of networks, and a network flow data sequence of the network flow prediction is a nonlinear time sequence in nature, but has the characteristic of high instability due to the influence of various uncertain factors, so that the network flow data is difficult to express and express, and further planning and maintaining of future networks become difficult. For this reason, network traffic prediction is of paramount importance. The invention provides a network flow prediction method based on GRU. Compared with the prior work, the main contributions of the invention lie in the following aspects:
(1) The invention predicts the network flow sequence by utilizing the GRU neural network algorithm, and the GRU neural network has memorability and can make better prediction on the nonlinear time sequence.
(2) In order to enable the traffic prediction method provided by the invention to carry out network traffic prediction under different situations and keep better prediction accuracy, the network traffic prediction method provided by the invention adopts a method of multi-model prediction addition averaging, three GRU models adopt different gradient descent algorithms to carry out prediction, and finally the prediction data is a value obtained by summing and averaging.
Description of the effects of the invention:
the method utilizes the GRU neural network to predict the network traffic data sequence, simultaneously adopts three GRU models with different gradient descent algorithms, and finally obtains the final predicted network traffic data by summing and averaging the predicted values of the three models.
The invention can memorize the change rule of the past network flow data through the GRU neural network, the GRU neural network is relatively simple, the resource occupation is less, the capability of expressing the nonlinear network flow sequence is strong, and the prediction effect is improved by respectively predicting the subsequence.
The invention adopts three GRU models with different gradient descent methods, and aims to adapt to prediction in different scenes.
Claims (2)
1. A network flow prediction method based on a GRU model is characterized in that: inputting a network flow data sequence into a GRU neural network model and finishing predicting network flow, wherein the number of the GRU neural network model is three, the GRU neural network model is a GRU-SGD model, a GRU-Adam model and a GRU-AdaGrad model, the GRU-SGD model is a neural network model using an SGD gradient descent algorithm, the GRU-Adam model is a neural network model using an Adam gradient descent algorithm, the GRU-AdaGrad model is a neural network model using an AdaGrad gradient descent algorithm, the GRU-SGD model, the GRU-Adam model and the GRU-AdaGrad model are used for predicting respectively, and data predicted by each model are added to obtain predicted network flow data through averaging; the method specifically comprises the steps of S1-S5, S1, obtaining historical network flow data; s2, determining training data and verification data in historical network flow data; s3, bringing training data into each GRU neural network model for training; s4, predicting the verification data through three GRU neural network models; s5, adding the predicted data to average to obtain predicted network flow data; the S3 specifically comprises S31-S35 steps, wherein S31 comprises three GRU neural network models which are a GRU-SGD model, a GRU-Adam model and a GRU-AdaGrad model, the GRU-SGD model is a neural network model using an SGD gradient descent algorithm, the GRU-Adam model is a neural network model using an Adam gradient descent algorithm, the GRU-AdaGrad model is a neural network model using an AdaGrad gradient descent algorithm, training data are input into each GRU neural network model respectively, and the training data are transmitted forward in the GRU neural network models; s32, calculating a loss function of time; s33, using reverse chain type derivation, and sequentially iterating until the loss function is converged; s34, updating the GRU-SGD model by using an SGD gradient descent algorithm, updating the GRU-Adam model by using an Adam gradient descent algorithm, and updating the GRU-AdaGrad model by using an AdaGrad gradient descent algorithm; s34, the steps include S341 to S343 specifically, S341 and GRU-SGD model use SGD gradient descent algorithm to calculate the descent quantity of each parameter, so as to update; s342, calculating the reduction amount of each parameter by using an Adam gradient reduction algorithm through the GRU-Adam model, and updating; s343, calculating the descending amount of each parameter by using an AdaGrad gradient descending algorithm through the GRU-AdaGrad model, and updating; and S35, repeating the steps from S31 to S34, continuously updating until the loss function is less than 0.2, and finishing the model training.
2. The method of claim 1, wherein the method comprises the following steps: the method is run on a server basis.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911343425.1A CN111030889B (en) | 2019-12-24 | 2019-12-24 | Network traffic prediction method based on GRU model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911343425.1A CN111030889B (en) | 2019-12-24 | 2019-12-24 | Network traffic prediction method based on GRU model |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111030889A CN111030889A (en) | 2020-04-17 |
CN111030889B true CN111030889B (en) | 2022-11-01 |
Family
ID=70211860
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201911343425.1A Active CN111030889B (en) | 2019-12-24 | 2019-12-24 | Network traffic prediction method based on GRU model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111030889B (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111932010B (en) * | 2020-08-10 | 2023-09-22 | 重庆大学 | Shared bicycle flow prediction method based on riding context information |
CN111970206A (en) * | 2020-08-21 | 2020-11-20 | 北京浪潮数据技术有限公司 | FC network flow control method, device and related components |
CN113094860B (en) * | 2021-04-29 | 2023-09-01 | 北京邮电大学 | Industrial control network flow modeling method based on attention mechanism |
CN113746696A (en) * | 2021-08-02 | 2021-12-03 | 中移(杭州)信息技术有限公司 | Network flow prediction method, equipment, storage medium and device |
US20230114810A1 (en) * | 2021-10-11 | 2023-04-13 | Samsung Electronics Co., Ltd. | Method of communication traffic prediction via continual learning with knowledge distillation, and an apparatus for the same |
CN117060984B (en) * | 2023-10-08 | 2024-01-09 | 中国人民解放军战略支援部队航天工程大学 | Satellite network flow prediction method based on empirical mode decomposition and BP neural network |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109062901A (en) * | 2018-08-14 | 2018-12-21 | 第四范式(北京)技术有限公司 | Neural network training method and device and name entity recognition method and device |
CN109325624A (en) * | 2018-09-28 | 2019-02-12 | 国网福建省电力有限公司 | A kind of monthly electric power demand forecasting method based on deep learning |
CN109799533A (en) * | 2018-12-28 | 2019-05-24 | 中国石油化工股份有限公司 | A kind of method for predicting reservoir based on bidirectional circulating neural network |
CN109816095A (en) * | 2019-01-14 | 2019-05-28 | 湖南大学 | Based on the network flow prediction method for improving gating cycle neural network |
CN109889391A (en) * | 2019-03-13 | 2019-06-14 | 南京理工大学 | A kind of network short term traffic forecasting method based on built-up pattern |
WO2019208998A1 (en) * | 2018-04-27 | 2019-10-31 | 한국과학기술원 | Gru-based cell structure design robust to missing data and noise in time series data in recurrent neural network |
-
2019
- 2019-12-24 CN CN201911343425.1A patent/CN111030889B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019208998A1 (en) * | 2018-04-27 | 2019-10-31 | 한국과학기술원 | Gru-based cell structure design robust to missing data and noise in time series data in recurrent neural network |
CN109062901A (en) * | 2018-08-14 | 2018-12-21 | 第四范式(北京)技术有限公司 | Neural network training method and device and name entity recognition method and device |
CN109325624A (en) * | 2018-09-28 | 2019-02-12 | 国网福建省电力有限公司 | A kind of monthly electric power demand forecasting method based on deep learning |
CN109799533A (en) * | 2018-12-28 | 2019-05-24 | 中国石油化工股份有限公司 | A kind of method for predicting reservoir based on bidirectional circulating neural network |
CN109816095A (en) * | 2019-01-14 | 2019-05-28 | 湖南大学 | Based on the network flow prediction method for improving gating cycle neural network |
CN109889391A (en) * | 2019-03-13 | 2019-06-14 | 南京理工大学 | A kind of network short term traffic forecasting method based on built-up pattern |
Also Published As
Publication number | Publication date |
---|---|
CN111030889A (en) | 2020-04-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111030889B (en) | Network traffic prediction method based on GRU model | |
CN110738360B (en) | Method and system for predicting residual life of equipment | |
CN108900346B (en) | Wireless network flow prediction method based on LSTM network | |
CN110705692B (en) | Nonlinear dynamic industrial process product prediction method of space-time attention network | |
CN110245801A (en) | A kind of Methods of electric load forecasting and system based on combination mining model | |
CN107480440A (en) | A kind of method for predicting residual useful life for modeling of being degenerated at random based on two benches | |
CN111815053B (en) | Prediction method and system for industrial time sequence data | |
CN110852515A (en) | Water quality index prediction method based on mixed long-time and short-time memory neural network | |
CN116757534A (en) | Intelligent refrigerator reliability analysis method based on neural training network | |
CN116303786B (en) | Block chain financial big data management system based on multidimensional data fusion algorithm | |
CN112765894B (en) | K-LSTM-based aluminum electrolysis cell state prediction method | |
CN115587666A (en) | Load prediction method and system based on seasonal trend decomposition and hybrid neural network | |
CN114166509A (en) | Motor bearing fault prediction method | |
CN111461551A (en) | Electric submersible pump fault early warning method based on deep learning and SPC (statistical computer) criterion | |
CN113095484A (en) | Stock price prediction method based on LSTM neural network | |
CN114694379B (en) | Traffic flow prediction method and system based on self-adaptive dynamic graph convolution | |
CN115600105A (en) | Water body missing data interpolation method and device based on MIC-LSTM | |
CN115308558A (en) | Method and device for predicting service life of CMOS (complementary Metal oxide semiconductor) device, electronic equipment and medium | |
CN114330815A (en) | Ultra-short-term wind power prediction method and system based on improved GOA (generic object oriented architecture) optimized LSTM (least Square TM) | |
CN113159395A (en) | Deep learning-based sewage treatment plant water inflow prediction method and system | |
CN116628444A (en) | Water quality early warning method based on improved meta-learning | |
CN115759343A (en) | E-LSTM-based user electric quantity prediction method and device | |
CN115936236A (en) | Method, system, equipment and medium for predicting energy consumption of cigarette factory | |
CN114970674A (en) | Time sequence data concept drift adaptation method based on relevance alignment | |
CN113570129A (en) | Method for predicting strip steel pickling concentration and computer readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |