CN115018038B - Network interface flow prediction method based on sparrow optimization width learning system - Google Patents

Network interface flow prediction method based on sparrow optimization width learning system Download PDF

Info

Publication number
CN115018038B
CN115018038B CN202210247884.5A CN202210247884A CN115018038B CN 115018038 B CN115018038 B CN 115018038B CN 202210247884 A CN202210247884 A CN 202210247884A CN 115018038 B CN115018038 B CN 115018038B
Authority
CN
China
Prior art keywords
sparrow
parameters
network
learning system
width learning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210247884.5A
Other languages
Chinese (zh)
Other versions
CN115018038A (en
Inventor
李少波
李笑瑜
周鹏
陈光林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guizhou University
Original Assignee
Guizhou University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guizhou University filed Critical Guizhou University
Priority to CN202210247884.5A priority Critical patent/CN115018038B/en
Publication of CN115018038A publication Critical patent/CN115018038A/en
Application granted granted Critical
Publication of CN115018038B publication Critical patent/CN115018038B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/004Artificial life, i.e. computing arrangements simulating life
    • G06N3/006Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/14Network analysis or design
    • H04L41/147Network analysis or design for predicting network behaviour
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a network interface flow prediction method based on a sparrow optimization width learning system, which comprises the following steps: the obtained network interface flow data determine a prediction period T, and the network flow at the moment T is predicted by utilizing the network flow information of [ T-12, T-1 ]; initializing network parameters; randomly generating p groups of shrinkage coefficients and regularization coefficients in the value range of the parameters as initial super parameters; respectively using p groups of initial super-parameters to automatically train a width learning system model to generate initial fitness; (5) optimizing the super-parameters by utilizing a sparrow optimization algorithm; (6) Training a width learning system by using the updated super parameters, and updating the fitness value; (7) Judging the maximum iteration times, outputting network parameters corresponding to the optimal adaptability, training a width learning model by using the network parameters, and establishing a network interface flow prediction model; otherwise, returning to the step (5). The invention can predict the accuracy of the network flow and reduce the influence of the network super-parameters on the prediction result.

Description

Network interface flow prediction method based on sparrow optimization width learning system
Technical Field
The invention belongs to the technical field of network interface flow prediction, and relates to a network interface flow prediction method based on a sparrow optimization width learning system.
Background
Most of the existing network traffic prediction models are deep learning models. The amount of network parameters for deep learning models is typically large, consuming more time and computing resources to train the model. The width learning system can complete training of the model in a shorter time, but the super parameters have a larger influence on the performance of the network model. The current network parameter adjustment mode is mostly manual parameter adjustment. This method is relatively dependent on the experience of the researcher and requires repeated training of the model to make adjustments, which is time consuming.
Disclosure of Invention
The invention aims to solve the technical problems that: a network interface flow prediction method based on a sparrow optimization width learning system is provided to solve the technical problems in the prior art.
The technical scheme adopted by the invention is as follows: a network interface flow prediction method based on a sparrow optimization width learning system comprises the following steps:
(1) For the obtained network interface flow data, determining a prediction period T, and predicting the network flow at the moment T by using the network flow information of [ T-12, T-1 ];
(2) Initializing network parameters: the method comprises the steps of population number, seeker proportion and maximum iteration times of a sparrow optimization algorithm; the method comprises the steps of a contraction coefficient value range, a regularization coefficient value range, the number of windows of a feature mapping layer, the number of nodes in a single window and the number of nodes of an enhancement layer in a width learning system;
(3) Randomly generating p groups of shrinkage coefficients and regularization coefficients as initial super parameters in the value range of the parameters, wherein p is the number of sparrows;
(4) Respectively using p groups of initial super parameters and network interface flow data to automatically train a width learning system model to generate initial fitness;
(5) Optimizing the super parameters by utilizing a sparrow optimization algorithm;
(6) The adaptability value is updated by utilizing the network interface flow data and the updated super-parameter training width learning system;
(7) Judging whether the maximum iteration times are reached, if so, outputting network parameters corresponding to the optimal adaptability, training a width learning model by using the network parameters, and establishing a network interface flow prediction model; otherwise, returning to the step (5).
The method for optimizing the super-parameters by utilizing the sparrow optimization algorithm in the step (5) comprises the following steps:
1) Taking the p groups of initial super parameters generated in the step (3) as initial positions of p sparrows, and calculating the number of the cable seekers, wherein the calculation formula is as follows:
pNum=p×p_scale
wherein pNum is the number of explorers in the population, p is the population size, and p_scale is the ratio of the explorers;
2) Updating the seeker position, and calculating the following formula:
wherein X is a super parameter to be optimized, namely the sparrow position,a value representing the jth dimension of the ith sparrow in the t-th iteration; r is R 2 Indicating a warning value in the range of 0,1]Is a random number which is uniformly distributed; ST represents a safety threshold, Q is a random number matrix in normal distribution, L is a 1 x d dimension matrix, and d is the population size; when R is 2 When ST is less than the threshold, the current position is safe, and sparrows search food in the accessories; otherwise, the current position is dangerous, and the seeker needs to guide the sparrow group to search a new place to search food;
3) Updating the follower position, and calculating the following formula:
wherein A is + =A T (AA T ) -1 A is a 1 xd dimensional matrix, each latitude value is from [ -1,1]Randomly generating;indicating the position of sparrow in the worst state, when i is less than n/2, indicating that the follower does not get food, the state of the sparrow is bad, and the sparrow needs to go to other places where more food is obtained; otherwise, continuing to search for food nearby the seeker;
4) For sparrows found to be dangerous as a scouter and their locations updated, the calculation formula is as follows:
wherein f i And f g The fitness and the optimal fitness of the ith sparrow are respectively f w Is the worst fitness; beta is a random number in standard normal distribution, K is a value of [ -1,1]Uniform random numbers within the range, epsilon is a small value (0 < epsilon < 10) -10 ) To prevent the denominator from being 0.
The invention has the beneficial effects that: compared with the prior art, the invention optimizes two super parameters of the regularization parameter and the contraction coefficient in the width learning system by using the sparrow optimization algorithm, thereby avoiding the complicated manual parameter adjustment process; and training a width learning system by utilizing the optimal super parameters obtained by optimization, and improving the accuracy of network flow prediction.
Drawings
FIG. 1 is a flow chart of a width learning system algorithm;
FIG. 2 is a flow chart of a sparrow search algorithm;
FIG. 3 is a pseudo code diagram of the SSA-BLS algorithm;
FIG. 4 is a flowchart of the SSA-BLS algorithm;
fig. 5 is a graph of predicted results of a traffic data set of a core network in a certain city in europe;
FIG. 6 is a graph of predicted outcome for a British academic backbone network traffic dataset;
FIG. 7 is a graph of the prediction results of the enterprise cloud platform switch interface traffic data set;
FIG. 8 is a diagram of the breadth-learning system at LSTM runtime.
Detailed Description
The invention will be further described with reference to specific examples.
Preliminary knowledge: the width learning system is a novel random weight neural network. Compared with the traditional neural network, the neural network is based on a random vector function linked neural network, and is mainly used for solving the problems of large calculation amount and high calculation cost of deep learning. The width learning system contains 3 layers, namely an input layer, a hidden layer and an output layer. The hidden layer is composed of a feature mapping layer and an enhancement node layer. The hidden layer of the width learning system is of a single-layer structure and consists of a feature mapping layer and an enhancement node layer.
The algorithm flow of the width learning system is as follows. Let training data be X ε R N×M The sample comprises N samples, each sample has M dimensions, and the corresponding label is Y epsilon R N×C . By n feature mapping functions(i=1,., n) mapping training data X into n sets of feature maps, each set of maps yielding K i A node in which the ith group of features maps Z i The calculation method is shown in formula (1).
\*MERGEFORMAT Z i =φ i (XW eiei ) \*MERGEFORMAT (1)
Wherein W is ei And beta ei Is a randomly generated feature map weight matrix and bias matrix. In the practical application process, phi i Often a non-linear mapping function.
Representing n groups of feature mapping nodes obtained by the feature mapping layer as Z in =(Z 1 ,Z 2 ,…,Z n ) Will Z in Is connected with the enhanced node layer and passes through the activation function ζ of the enhanced node layer j Will Z in Mapping into m groups of enhanced nodes, each group consisting of q nodes, then the j-th group of enhanced nodes H j May be expressed as.
H j =ζ j (Z in W hjhj ) \*MERGEFORMAT (2)
Wherein W is hj Is a randomly generated enhancement layer node weight matrix, beta hj Is the corresponding bias matrix; activation function ζ j Different nonlinear activation functions may be selected to adequately extract characteristic information of the input data.
Similarly, m groups of enhanced nodes H j Denoted as H jm =(H 1 ,H 2 ,…,H m ) Will Z in And H is jm The combination is expressed as:
A=(Z in |H jm ) \*MERGEFORMAT (3)
the output of the width learning system is then:
Y=(Z in |H jm )W=AW \*MERGEFORMAT (4)
wherein, W is the output layer connection weight matrix, and there are:
W=A + Y \*MERGEFORMAT (5)
A + is a pseudo-inverse matrix of the matrix A, and the model training is completed by calculating a connection weight matrix W by solving pseudo-inverse:
wherein lambda > 0 is a regularization coefficient; i is the identity matrix.
Compared with the traditional nerve, the width learning system has two main characteristics: firstly, the sparse self-encoder is utilized to refine random features of input data into sparse and compact feature sets, and important features are explored through a sparse feature learning model, so that the features of the input data are better described, and the operation efficiency is improved; secondly, aiming at the problem that in the deep learning system, if the network model cannot reach the required precision, the number of network layers is increased or the structure is changed, and then the training model is readjusted, so that the calculated amount and the calculated time are increased, the width learning system adopts incremental learning, and the model is dynamically adjusted by adding the enhancement nodes, so that the training time is greatly shortened, and a certain precision can be ensured.
An algorithm flow chart of the width learning system is shown in fig. 1.
The sparrow search algorithm is a novel group intelligent optimization algorithm and is based on the search and anti-predation behaviors of sparrows on foods. Sparrow search algorithms divide sparrows into seekers and followers and formulate the following rules for sparrow exercise.
1) Sparrows in a population are classified into explorers and followers according to the degree of fitness of the sparrows. The fitness reflects the merits of sparrow predation sites. The better the sparrow position with higher fitness, the better the food can be found.
2) Sparrows with high fitness are seekers, and other sparrows act as followers. The seeker is responsible for investigating the food-rich site and directing the follower to the foraging site and direction. The follower can search for the best seeker at the predatory location and then find food around it.
3) The fitness value of the sparrow individuals is dynamically changed, so that the identities of the seeker and the follower can be mutually changed, but the proportion of the seeker is kept unchanged.
4) The lower the fitness value of the follower, the worse its predation position. These followers may randomly fly to other places to find food.
5) A certain proportion of individuals are randomly selected from the sparrow population as scouts and are responsible for monitoring the safety of the surrounding environment. When predators are found, the scout will give an alarm, and when the alarm value is greater than the safety value, the explorer will take the follower to a safer area to find food.
(6) Sparrows at the edges of the population will quickly shift to a safe area to obtain a better location when the hazard is perceived, while sparrows at the center will randomly move.
The algorithm flow of SSA is as follows, and the algorithm flow chart is shown in fig. 2.
The first step: parameter initialization mainly comprises setting the number of sparrows, the proportion of explorers, the positions of sparrows, the maximum iteration number and the like.
Step 2: judging whether the current position of the population is safe or not, and updating the position of the explorer by using a formula (7).
Wherein,the value of the j dimension of the i-th sparrow in the t-th iteration is represented. R is R 2 Indicating a warning value in the range of 0,1]Is a random number which is uniformly distributed. ST represents a safety threshold value, and the value range is [0.5,1.0]Q is a random number in normal distribution, L is a 1 x d dimension matrix, and d is the population size. When R is 2 When ST is less than the current position, the current position is safe, and sparrow groups search for food; otherwise, the current position is threatened, and the cable is detectedThe user needs to guide the sparrow group to find new places to find food.
And a third step of: and judging the state of the follower, and updating the position of the follower according to the formula (8).
Wherein,indicating the position of sparrow in worst condition, A + =A T (AA T ) -1 A is a 1 xd-dimensional matrix, each latitude value is from [ -1,1]Is randomly generated. When i is less than n/2, the follower does not get food, the state is poor, and the follower needs to go to other places where more food can be obtained; otherwise, continue to find food in the vicinity of the seeker.
Fourth step: part of sparrows in the population are found dangerous and become alertors, and the positions of the alerters are determined according to the formula (9).
Wherein f i And f g The fitness and the optimal fitness of the ith sparrow are respectively f w Is the worst fitness. Beta is a random number in standard normal distribution, K is a value of [ -1,1]The uniform random number in the range, epsilon, is a small value to prevent the denominator from being 0.
Step 5: updating the sparrow fitness.
Step 6: judging whether the iteration stop condition is met, and if the iteration stop condition is not met, repeating the steps 2 to 5.
Example 1: as shown in fig. 1-8, a network interface flow prediction method based on a sparrow optimization width learning system, the method comprises the following steps:
(1) For the obtained network interface flow data, determining a prediction period T, and predicting the network flow at the moment T by using the network flow information of [ T-12, T-1 ];
(2) Initializing network parameters: the method comprises the steps of population number, seeker proportion and maximum iteration times of a sparrow optimization algorithm; the method comprises the steps of a contraction coefficient value range, a regularization coefficient value range, the number of windows of a feature mapping layer, the number of nodes in a single window and the number of nodes of an enhancement layer in a width learning system;
(3) Randomly generating p groups of shrinkage coefficients and regularization coefficients as initial super parameters in the value range of the parameters, wherein p is the number of sparrows;
(4) Respectively using p groups of initial super parameters and network interface flow data to automatically train a width learning system model to generate initial fitness;
(5) The sparrow optimization algorithm is utilized to optimize the super parameters, and the detailed steps are as follows:
1) Taking the p groups of initial super parameters generated in the step (3) as initial positions of p sparrows, and calculating the number of the cable seekers, wherein the calculation formula is as follows:
pNum=p×p_scale
wherein pNum is the number of explorers in the population, p is the population size, and p_scale is the ratio of the explorers;
2) Updating the seeker position, and calculating the following formula:
wherein X is a super parameter to be optimized, namely the sparrow position,a value representing the jth dimension of the ith sparrow in the t-th iteration; r is R 2 Indicating a warning value in the range of 0,1]Is a random number which is uniformly distributed; ST represents a safety threshold, Q is a random number matrix in normal distribution, L is a 1 x d dimension matrix, and d is the population size; when R is 2 When ST is less than the threshold, the current position is safe, and sparrows search food in the accessories; otherwise, the current position is dangerous, and the seeker needs to guide the sparrow group to search a new place to search food;
3) Updating the follower position, and calculating the following formula:
wherein A is + =A T (AA T ) -1 A is a 1 xd dimensional matrix, each latitude value is from [ -1,1]Randomly generating;indicating the position of sparrow in the worst state, when i is less than n/2, indicating that the follower does not get food, the state of the sparrow is bad, and the sparrow needs to go to other places where more food is obtained; otherwise, continuing to search for food nearby the seeker;
4) For sparrows found to be dangerous as a scouter and their locations updated, the calculation formula is as follows:
wherein f i And f g The fitness and the optimal fitness of the ith sparrow are respectively f w Is the worst fitness. Beta is a random number in standard normal distribution, K is a value of [ -1,1]A uniform random number within the range, ε being a small value to prevent the denominator from being 0;
(6) The adaptability value is updated by utilizing the network interface flow data and the updated super-parameter training width learning system;
(7) Judging whether the maximum iteration times are reached, if so, outputting network parameters corresponding to the optimal adaptability, training a width learning model by using the network parameters, and establishing a network interface flow prediction model; otherwise, returning to the step (5).
In order to predict the accuracy of network flow and reduce the influence of network superparameters on a prediction result, two superparameters of a width learning system, namely a contraction coefficient (r) and a regularization coefficient (lambda), are optimized by utilizing a sparrow search algorithm, and an optimal superparameter output by optimization is utilized to establish a network flow prediction model. I call this method SSA-BLS with pseudo code and algorithm flow diagrams shown in FIGS. 3 and 4.
The experimental data set uses the European core network traffic data set of a certain city and the uk academic backbone network traffic data set.
European core network traffic data set: the data is from the uk academic backbone network traffic dataset: the data set collected aggregate traffic in bits for the uk academic network backbone from 30 minutes at 9 of 11 months 2004 to 11 minutes at 27 of 1 month 2005, with a sampling interval of five minutes.
Taking data of 7 months 1 to 25 days in 2005 in a European core network traffic data set as a training set and taking data of 7 months 26 to 28 days in 7 months as a test set; the data of 1 month, 1 day and 24 days of 2005 in the uk academic backbone network flow data set are used as training sets, and the data of 1 month, 25 days and 1 month, 27 days are used as test sets.
1) Parameters and evaluation index
The SSA-BLS parameters are configured as follows. The population size is 50, the producer accounts for 20%, the maximum iteration number is 5, and the dimension is 2; the number of windows of the mapping layer is 10, the number of nodes in each window of the mapping layer is 10, the number of nodes of the enhancement layer is 50, and the values of the shrinkage coefficient (r) and the regularization coefficient (lambda) are respectively in the range of [0.09,0.999999 ]]And [2 ] -30 ,2 -35 ]。
MSE, RMSE, MAE, MAPE, MA is used as an evaluation index, and the calculation modes are shown in formulas (10), (11), (12), (13) and (14).
Where n is the total number of samples,as predicted value, y i Is a true value. The smaller MSE, RMSE, MAE, MAPE the better the MAPE the closer to 100% the better the model predictive performance.
2) Results and discussion
The experiments used the SSA-BLS model and [ T-12, T-1]]To predict the flow value at time T. In order to verify the performance of the SSA-BLS model, the SSA-BLS model is compared with BLS, ELM, SCN, RVFL with similar structures, with the RVFL deformation dRVFL and an LSTM model commonly used in network traffic prediction, each model is independently operated for 100 times, and the average value of the evaluation indexes of the operation results is taken as a final result. Wherein, the values of the contraction coefficient r and the regularization coefficient lambda of the BLS model are respectively from {0.1,0.5,0.9,0.99,0.9999,0.99999} and {2 } -30 ,2 -20 ,2 -10 0.5,1,5,10, the rest parameters are the same as SSA-BLS model; the number of nodes of the maximum hidden layer of the SCN model is 250, and the number of the maximum candidate nodes is 100; the regularization coefficient of RVFL is 1e-3, and the number of hidden layer nodes is 100; the dRVFL parameter is the same as RVFL; the LSTM model contains 12 hidden layers per layer, with a training learning rate of 1e-2, a batch_size of 64, and an epoch of 15. The predicted performance of each model on the test set of the two data sets is shown in tables 1-2.
Table 1 experimental results of flow data set of core network of certain city in europe
MSE RMSE MAE MAPE MA
SSA-BLS 0.0159047 0.1261069 0.0937315 0.0294284 97.057155%
BLS 0.0781227 0.2551322 0.1878021 0.0571019 94.289801%
SCN 0.0154907 0.1244372 0.0934485 0.0295662 97.043378%
RVFL 0.0254023 0.1593589 0.1208186 0.0388347 96.116525%
dRVFL 0.0227553 0.1507691 0.1135728 0.0367191 96.328085%
ELM 0.1394488 0.3686439 0.2710739 0.0780252 92.197470%
LSTM 0.0781441 0.2502372 0.1884968 0.0517535 94.824642%
Table 2 experimental results of the uk academic backbone network traffic dataset
Table 3 Experimental results of enterprise cloud platform switch interface flow data set
MSE RMSE MAE MAPE MA
SSA-BLS 0.0000734 0.0082991 0.0063628 0.0021407 99.785924%
BLS 0.0103714 0.0811761 0.0563080 0.0176119 98.238804%
SCN 0.0001742 0.0130396 0.0067544 0.0021857 99.781427%
RVFL 0.0361230 0.1899801 0.1288009 0.0400804 95.991952%
dRVFL 0.0327578 0.1807739 0.1277397 0.0403579 95.964208%
ELM 0.0579519 0.2382928 0.1327614 0.0400340 95.996599%
LSTM 0.0283041 0.1057008 0.0759722 0.0238097 97.619024%
Fig. 5 and 6 are graphs of SSA-BLS model versus other models for predicted versus actual values on a training set of common data sets.
In addition, to better verify the predictive performance of the SSA-BLS model, the model is applied to a private traffic dataset. The private traffic data set is the actual ingress traffic data from the switch interface of a business 2021, 10 th, 5 th to 18 th. The model was validated using data from day 5 of year 10 to day 16 of year 10 of the private dataset as the training set and data from day 17 to day 18 of year 10 as the test set.
Because the sampling intervals of the enterprise switch interface traffic data are not equal, resampling is performed first: the mean value of the interface flow over 5 minutes is calculated and if no flow data is present over 5 minutes, the previous value is used for filling. Meanwhile, the original data has a great abnormal flow value, and the data is smoothed by utilizing spectrum smoothing (spectral smoother) in order to reduce the influence of the abnormal value on prediction. The SSA-BLS model and the comparative model parameters are as above, and the predicted performance of each model is shown in Table 3.
FIG. 7 is a graph of predicted versus actual values of the SSA-BLS model versus other models on a training set in a private dataset. From tables 1, 2 and 3, it can be seen that the SSA-BLS model has better ultra-short term predictive performance than other models in both the uk academic backbone network traffic data set and the private data set, and also has higher model accuracy in the european core network traffic data set in a certain city. In terms of the accuracy of single-step prediction of network traffic, the prediction accuracy of the SSA-BLS model is higher than that of other models in a British academic backbone network traffic data set and an enterprise cloud platform exchanger interface traffic data set, and the prediction performance of the SSA-BLS model in a European core network traffic data set is only slightly inferior to that of an SCN model, but the SSA-BLS model has higher accuracy overall, so that the model can select better network super-parameters and better time sequence characteristics of capturing traffic. In terms of time consumption, FIG. 8 is the time spent by the breadth-learning system training once to run 1 epoch with the LSTM model. Wherein, dataset1, dataset2 and dataset3 respectively represent a UK academic main trunk network traffic data set, a European core network traffic data set and an enterprise cloud platform switch
Interface traffic data sets. As can be clearly seen in fig. 8, the width learning system can complete training in a shorter time, and the greater the data volume, the greater the temporal advantage of the BLS.
The invention provides a width learning model based on a sparrow optimization algorithm, which optimizes two super parameters of a contraction coefficient (r) and a regularization coefficient (lambda) in the width learning model by utilizing the sparrow optimization algorithm, and trains the model by utilizing the output optimal super parameters, so that the influence of the super parameters on the model is reduced, and the accuracy of the model is improved. The model avoids the complicated manual parameter adjustment process, and the better network super-parameter combination is selected by utilizing an algorithm, so that the performance of the BLS is optimal. The SSA-BLS model is applied to the field of short-term prediction of network traffic, and two public data sets of the network traffic and a real data set of the interface traffic of an enterprise cloud platform network switch are selected for experiments. In order to verify the model effect, the SSA-BLS model is compared with models such as BLS, and experiments show that the SSA-BLS model can select better super parameters to enable the prediction accuracy of network flow to reach more than 97%.
The foregoing is merely illustrative of the present invention, and the scope of the present invention is not limited thereto, and any person skilled in the art can easily think about variations or substitutions within the scope of the present invention, and therefore, the scope of the present invention shall be defined by the scope of the appended claims.

Claims (2)

1. A network interface flow prediction method based on a sparrow optimization width learning system is characterized by comprising the following steps of: the method comprises the following steps:
(1) For the obtained network interface flow data, determining a prediction period T, and predicting the network flow at the moment T by using the network flow information of [ T-12, T-1 ];
(2) Initializing network parameters: the method comprises the steps of population number, seeker proportion and maximum iteration times of a sparrow optimization algorithm; the method comprises the steps of a contraction coefficient value range, a regularization coefficient value range, the number of windows of a feature mapping layer, the number of nodes in a single window and the number of nodes of an enhancement layer in a width learning system;
(3) Randomly generating p groups of shrinkage coefficients and regularization coefficients as initial super parameters in the value range of the parameters, wherein p is the number of sparrows;
(4) Respectively using p groups of initial super parameters and network interface flow data to automatically train a width learning system model to generate initial fitness;
(5) Optimizing the super parameters by utilizing a sparrow optimization algorithm;
(6) Training a width learning system by utilizing the network interface flow data and the updated super parameters, and updating the fitness value;
(7) Judging whether the maximum iteration times are reached, if so, outputting network parameters corresponding to the optimal adaptability, training a width learning model by using the network parameters, and establishing a network interface flow prediction model; otherwise, returning to the step (5).
2. The network interface traffic prediction method based on the sparrow optimization width learning system according to claim 1, wherein the method comprises the following steps: the method for optimizing the super-parameters by utilizing the sparrow optimization algorithm in the step (5) comprises the following steps:
1) Taking the p groups of initial super parameters generated in the step (3) as initial positions of p sparrows, and calculating the number of explorers, wherein the calculation formula is as follows:
pNum=p×p_scale
wherein pNum is the number of explorers in the population, p is the population size, and p_scale is the ratio of the explorers;
2) Updating the seeker position, and calculating the following formula:
wherein X is a super parameter to be optimized, namely the sparrow position,a value representing the jth dimension of the ith sparrow in the t-th iteration; r is R 2 Representation ofWarning value, the value range is [0, 1]]Is a random number which is uniformly distributed; ST represents a safety threshold, Q is a random number matrix in normal distribution, L is a 1 x d dimension matrix, and d is the population size; when R is 2 When ST is less than the threshold, the current position is safe, and sparrows search food in the accessories; otherwise, the current position is dangerous, and the seeker needs to guide the sparrow group to search a new place to search food;
3) Updating the follower position, and calculating the following formula:
wherein A is + =A T (AA T ) -1 A is a 1 xd dimensional matrix, each latitude value is from [ -1,1]Randomly generating;indicating the position of sparrow in the worst state, when i is less than n/2, indicating that the follower does not get food, the state of the sparrow is bad, and the sparrow needs to go to other places where more food is obtained; otherwise, continue to find food in the vicinity of the seeker;
4) For sparrows found to be dangerous as a scouter and their locations updated, the calculation formula is as follows:
wherein f i And f g The fitness and the optimal fitness of the ith sparrow are respectively f w Is the worst fitness; beta is a random number in standard normal distribution, K is a value of [ -1,1]Uniform random number in the range, epsilon value range is 0 < epsilon < 10 -10
CN202210247884.5A 2022-03-14 2022-03-14 Network interface flow prediction method based on sparrow optimization width learning system Active CN115018038B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210247884.5A CN115018038B (en) 2022-03-14 2022-03-14 Network interface flow prediction method based on sparrow optimization width learning system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210247884.5A CN115018038B (en) 2022-03-14 2022-03-14 Network interface flow prediction method based on sparrow optimization width learning system

Publications (2)

Publication Number Publication Date
CN115018038A CN115018038A (en) 2022-09-06
CN115018038B true CN115018038B (en) 2024-03-05

Family

ID=83066479

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210247884.5A Active CN115018038B (en) 2022-03-14 2022-03-14 Network interface flow prediction method based on sparrow optimization width learning system

Country Status (1)

Country Link
CN (1) CN115018038B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012168582A (en) * 2011-02-09 2012-09-06 Ntt Docomo Inc Potential class analysis device, potential class analysis method, and program
WO2021007812A1 (en) * 2019-07-17 2021-01-21 深圳大学 Deep neural network hyperparameter optimization method, electronic device and storage medium
CN112653142A (en) * 2020-12-18 2021-04-13 武汉大学 Wind power prediction method and system for optimizing depth transform network
CN114021689A (en) * 2021-09-16 2022-02-08 湖州师范学院 Chaotic search optimization method for self-adaptive neural network traffic flow prediction

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012168582A (en) * 2011-02-09 2012-09-06 Ntt Docomo Inc Potential class analysis device, potential class analysis method, and program
WO2021007812A1 (en) * 2019-07-17 2021-01-21 深圳大学 Deep neural network hyperparameter optimization method, electronic device and storage medium
CN112653142A (en) * 2020-12-18 2021-04-13 武汉大学 Wind power prediction method and system for optimizing depth transform network
CN114021689A (en) * 2021-09-16 2022-02-08 湖州师范学院 Chaotic search optimization method for self-adaptive neural network traffic flow prediction

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
一种优化的RBF神经网络模型用于网络流量预测;余健;郭平;;计算机应用与软件;20081215(12);39-41+51 *

Also Published As

Publication number Publication date
CN115018038A (en) 2022-09-06

Similar Documents

Publication Publication Date Title
Luk et al. A study of optimal model lag and spatial inputs to artificial neural network for rainfall forecasting
CN111639748A (en) Watershed pollutant flux prediction method based on LSTM-BP space-time combination model
CN110650153B (en) Industrial control network intrusion detection method based on focus loss deep neural network
CN111726243B (en) Method and device for predicting node state
CN103886218A (en) Lake and reservoir algal bloom predicating method based on multielement nonstationary time series analysis and neural network and support vector machine compensation
CN114626512A (en) High-temperature disaster forecasting method based on directed graph neural network
CN109583588B (en) Short-term wind speed prediction method and system
CN108280998A (en) Short-time Traffic Flow Forecasting Methods based on historical data dynamic select
CN111355633A (en) Mobile phone internet traffic prediction method in competition venue based on PSO-DELM algorithm
CN114694379B (en) Traffic flow prediction method and system based on self-adaptive dynamic graph convolution
CN113554466A (en) Short-term power consumption prediction model construction method, prediction method and device
CN113411216A (en) Network flow prediction method based on discrete wavelet transform and FA-ELM
Sun et al. Precipitation analysis and forecasting using singular spectrum analysis with artificial neural networks
Kalteh Enhanced monthly precipitation forecasting using artificial neural network and singular spectrum analysis conjunction models
CN112784493A (en) Geographic space prediction method and system based on self-adaptive deep Q network
CN113177673B (en) Air conditioner cold load prediction optimization method, system and equipment
Hoang et al. A novel time series prediction approach based on a hybridization of least squares support vector regression and swarm intelligence
CN115018038B (en) Network interface flow prediction method based on sparrow optimization width learning system
CN117313990A (en) Air conditioner load prediction method and system based on meta-integrated learning and electronic equipment
CN116911178A (en) Method and system for predicting capacity of small and medium-sized reservoirs based on weather forecast
Peng et al. Meteorological satellite operation prediction using a BiLSTM deep learning model
Wang et al. ARIMA-BP integrated intelligent algorithm for China’s consumer price index forecasting and its applications
CN114139783A (en) Wind power short-term power prediction method and device based on nonlinear weighted combination
CN112884186A (en) Transformer substation network power supply load prediction method considering DG and electric heating load
Xiong et al. Time series prediction of wind speed based on SARIMA and LSTM

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant