CN111967679A - Ionized layer total electron content forecasting method based on TCN model - Google Patents

Ionized layer total electron content forecasting method based on TCN model Download PDF

Info

Publication number
CN111967679A
CN111967679A CN202010842098.0A CN202010842098A CN111967679A CN 111967679 A CN111967679 A CN 111967679A CN 202010842098 A CN202010842098 A CN 202010842098A CN 111967679 A CN111967679 A CN 111967679A
Authority
CN
China
Prior art keywords
tec
tcn
model
data
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010842098.0A
Other languages
Chinese (zh)
Other versions
CN111967679B (en
Inventor
唐丝语
黄智�
陈伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Normal University
Original Assignee
Jiangsu Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu Normal University filed Critical Jiangsu Normal University
Priority to CN202010842098.0A priority Critical patent/CN111967679B/en
Publication of CN111967679A publication Critical patent/CN111967679A/en
Application granted granted Critical
Publication of CN111967679B publication Critical patent/CN111967679B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S19/00Satellite radio beacon positioning systems; Determining position, velocity or attitude using signals transmitted by such systems
    • G01S19/01Satellite radio beacon positioning systems transmitting time-stamped messages, e.g. GPS [Global Positioning System], GLONASS [Global Orbiting Navigation Satellite System] or GALILEO
    • G01S19/03Cooperating elements; Interaction or communication between different cooperating elements or between cooperating elements and receivers
    • G01S19/07Cooperating elements; Interaction or communication between different cooperating elements or between cooperating elements and receivers providing data for correcting measured positioning data, e.g. DGPS [differential GPS] or ionosphere corrections
    • G01S19/072Ionosphere corrections
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • G06F30/27Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0639Performance analysis of employees; Performance analysis of enterprise or organisation operations
    • G06Q10/06393Score-carding, benchmarking or key performance indicator [KPI] analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Human Resources & Organizations (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Economics (AREA)
  • General Engineering & Computer Science (AREA)
  • Strategic Management (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Development Economics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Computational Linguistics (AREA)
  • Marketing (AREA)
  • Tourism & Hospitality (AREA)
  • Quality & Reliability (AREA)
  • Game Theory and Decision Science (AREA)
  • Educational Administration (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Business, Economics & Management (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Operations Research (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Geometry (AREA)
  • Computer Hardware Design (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)

Abstract

The invention discloses an ionospheric total electron content forecasting method based on a TCN (tool tracking network) model, and relates to the technical field of ionospheric total electron content forecasting. The forecasting method comprises the following steps: firstly, acquiring data, dividing the data into a training set and a test set, and carrying out normalization processing; then, determining a topological structure of the TCN neural network, initializing the weight of the TCN neural network, and adjusting parameters in the model; secondly, inputting the training set into an ionized layer TEC forecasting model in batches, calculating an output error of the effective historical length, and meanwhile, reversely propagating the error to update TCN prediction model parameters until the optimal parameters are found, and finishing training; and finally, inputting the data of the test set, obtaining a TEC prediction result, comparing the prediction result with the observation data, and evaluating the effectiveness of the prediction model. Due to the TCN convolution characteristic, the method can better capture the space-time characteristic of the electron content of the ionized layer, improve the accuracy of the total electron content forecast of the ionized layer to a certain extent and enhance the real-time property.

Description

Ionized layer total electron content forecasting method based on TCN model
Technical Field
The invention relates to the technical field of ionosphere total electron content prediction, in particular to an ionosphere total electron content prediction method based on a convolutional neural network TCN model.
Background
The ionosphere, an important component of the earth's atmosphere, is located above 60km above ground to the magnetic top layer and contains a large number of free electrons and ions that affect the propagation of radio waves.
Total Electron Content (TEC) is one of the important parameters of the ionosphere, and determines the degree of influence of the ionosphere on satellite signals. Therefore, the method has important significance for correcting radio wave propagation and researching ionosphere theory.
When the satellite signal frequency is known, only the TEC on the propagation path needs to be determined, and the ionospheric delay can be calculated. Dual-band or multi-band users can form a linear combination of radio layer delays based on satellite observations, eliminating or attenuating ionospheric delays to the maximum. However, because the price of the dual-frequency or multi-frequency GNSS receiver is much higher than that of the single-frequency receiver, in practical application, many single-frequency users are still used, and the single-frequency users generally cannot obtain the ionospheric delay through self-measured data, and usually need to evaluate and correct the ionospheric delay by means of an ionospheric TEC model.
Currently, research on a prediction model of an Ionosphere TEC has gradually evolved from traditional statistical and computational economics models, such as modeling methods of an International Reference Ionosphere (IRI), an exponential smoothing model, an autoregressive moving average model ARMA, a similar prediction method, an autocorrelation analysis method, and the like, to a neural network model which is more advantageous and more capable of learning in terms of processing nonlinear input and output relationships and time-varying problems. Particularly, in recent years, with rapid development of computer hardware technology and artificial intelligence technology, the deep learning network replaces the traditional neural network represented by BP, and has been applied to the field of spatial weather forecast to a certain extent. The long-time memory network LSTM and the variant GRU thereof show good performance in ionosphere TEC prediction due to good sequence learning and sequence conversion capability. However, LSTM and GRU mainly perform sequential processing over time, and cannot completely solve the problem of gradient disappearance or gradient explosion, and meanwhile, a large amount of memory is required to store partial results of a plurality of units, the calculation time is long, and the real-time requirement is difficult to meet in actual engineering application. In addition, the ionosphere TEC is not only related to temporal variation, but also closely related to spatial position. Therefore, the temporal-spatial variation characteristics need to be further extracted to achieve more accurate and rapid prediction results.
Disclosure of Invention
In view of the above, the invention discloses a method for forecasting total electron content of an ionosphere based on a TCN model, which not only can provide ionosphere refraction correction parameters for a large number of single-frequency GNSS users, but also can provide a large amount of data support for ionosphere space environment analysis and research.
The ionospheric total electron content forecasting method based on the TCN model comprises the following steps:
the method comprises the following steps: acquiring ionized layer TEC data, a solar radiation flux value F10.7 and a geomagnetic activity index Ap, and converting the TEC into a high-time resolution sequence.
Step two: and dividing the converted TEC time sequence, the solar radiation flux value F10.7 and the geomagnetic activity index Ap into a training set and a testing set, and then carrying out zero-mean standardization processing on data in the training set.
Step three: determining the topological structure of the TCN neural network, initializing the weight of the TCN neural network, and adjusting the parameters in the model.
Step four: inputting the training set into an ionized layer TEC forecasting model in batches, calculating an output error of an effective historical length, and updating a TCN forecasting model weight parameter by back propagation of the error until the maximum iteration times or the accuracy and the error rate reach the optimal values;
step five: and comparing the training results, if the error is within the set range, ending the training, executing the step six, if the error exceeds the set range, executing the step three, updating the TCN network parameters, performing the training again until the optimal parameters are found, and determining the optimal model.
Step six: after the model training is finished, the trained model is recorded as TCNTECInputting test set data into TCNTECIn the model, the output result is recorded as Y0Then Y is put0And carrying out Z-score denormalization to obtain a final TEC prediction result.
Step seven: and comparing the prediction result with the observation data, and evaluating the effectiveness of the model by adopting three performance indexes, namely MAPE, MAD and RMSE.
Preferably, in the step one, the TEC high resolution sequence conversion method is: converting the downloaded ionized layer data into VTEC data through a spherical harmonic function, eliminating abnormal data which do not conform to the normal range of the electron content, and filling ionized layer TEC data through a spline interpolation method to obtain a TEC sequence with the time resolution of 1 h; f10.7 and Ap are processed into a data format consistent with TEC time resolution.
Preferably, in the second step, the TEC, the solar radiation flux F10.7, and the geomagnetic activity index Ap are normalized in the following manner: recording TEC time sequence as X0={x1,x2,x3,...,xn}, training set Xtr={x1,x2,...xm}: test set Xte={x1,x2,...xl4: 1; after the data set is divided, the element x in the training set is subjected tomCarrying out zero mean value standardization to obtain a standardized training set of X'tr={x′1,x′2,...x′tThe normalization formula is:
Figure BDA0002641805950000031
in the formula: mu is the mean value of the original TEC time sequence, sigma is the standard error of the original TEC time sequence, and t is more than or equal to 1 and less than or equal to m; the division ratio of the solar radiation flux value F10.7 and the geomagnetic activity index Ap to the training set and the test set is also 4:1, and the zero-mean standardization process and formula and the TEC data zero-mean standardThe same applies.
Preferably, in step three, the TCN network is designed as a three-layer network structure, including an input layer, a hidden layer, and an output layer; the input layer comprises input parameters of a geomagnetic index Ap, a solar activity index F10.7 and an electron content TEC; the hidden layer uses causal convolution and expansion convolution as standard convolution layers, every two convolution layers and identity mapping are packaged into a residual module, then the residual module is used for stacking a deep layer network, and the last layers use full convolution layers to replace full connection layers; the output layer output target is the ionospheric electron content value 24 hours in the future.
Preferably, in step three, the weight for initializing the TCN network is: the number of iterations is 50, the dilation factor d is 1,2,4, the convolution kernel size is 3, the learning rate size is 0.001, and the number of hidden layer neurons is 20.
Preferably, in step four, the training set training step is: s1, dividing single batch of input data into two parts, wherein the output of a residual error module is formed by adding a series of changes F of the two parts of input data; s2, performing one-dimensional convolution operation on the input X to finally obtain the output of the residual error module, wherein the output of the residual error module is calculated by the next residual error module; and S3, repeating the steps S1-S2, and obtaining an output result of the TCN when the last residual error module is calculated.
Preferably, in S1, the input data is subjected to a hole causal convolution, then to a nonlinear process by activating the function Relu, and a Dropout layer is added at the same time, and the same operation is performed twice, so as to obtain the output of the residual module.
Preferably, in step seven, the three performance indexes of MAPE, MAD, and RMSE are:
Figure BDA0002641805950000041
Figure BDA0002641805950000042
wherein, YiIonospheric TEC predicted value, Q, for hour iiThe observed value of the ionized layer TEC in the ith hour is shown, and n is the length of the time period; when the three performance index values are smaller, theThe closer the predicted value and the true value are, the better the fitting degree is, and the smaller the error between the predicted value and the true value is.
Compared with the prior art, the ionosphere total electron content forecasting method based on the TCN model has the advantages that:
(1) due to causal characteristics between layers of the TCN convolutional network, the space-time characteristics of the electron content of the ionized layer can be better captured, the accuracy of the total electron content prediction of the ionized layer is improved to a certain extent, and the instantaneity is enhanced.
(2) The added layer-skipping convolution of the residual convolution uses a larger expansion coefficient to change the size of a receptive field to provide more flexibility, can better control the memory length of a prediction model, and further improves the prediction accuracy of the total electron content of the ionosphere.
(3) The time directions of the reverse propagation path and the sequence of the TCN are different, the problem of gradient disappearance or gradient explosion in the network is solved, and the method is superior to the traditional feedforward network.
(4) The sharing characteristic of the convolution kernel enables the operation to be fast and less in internal memory under the condition of processing the data with large total electron content for many years, and the electron content forecasting speed is accelerated.
Drawings
For a clearer explanation of the embodiments or technical solutions of the present invention, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for a person skilled in the art to obtain other drawings based on these drawings without creative efforts.
FIG. 1 is an overall flow chart of the present invention.
FIG. 2 is a diagram illustrating a residual error model in the TCN model of the present invention.
FIG. 3 shows the calculation process of F (x) in the TCN network of the present invention.
Detailed Description
The following provides a brief description of embodiments of the present invention with reference to the accompanying drawings. It is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all embodiments, and all other embodiments obtained by those skilled in the art based on the embodiments of the present invention without any inventive work belong to the protection scope of the present invention.
Fig. 1-3 illustrate a preferred embodiment of the present invention, which is parsed in detail.
Fig. 1 shows a method for forecasting total electron content in an ionosphere based on a TCN model, which includes the following steps:
the method comprises the following steps: acquiring ionized layer TEC data, a solar radiation flux value F10.7 and a geomagnetic activity index Ap, and converting the TEC into a time sequence with high time resolution by adopting an interpolation technology. Specifically, TEC data is derived from european orbit determination center CODE, F10.7 is derived from gordad space flight center by NASA, Ap is derived from geomagnetic and space magnetic data analysis centers. And (3) performing expansion calculation on the downloaded ionized layer data through a spherical harmonic function to obtain the vertical ionized layer content, eliminating abnormal data which do not conform to the normal range of the electron content, and filling ionized layer TEC data through a spline interpolation method to obtain a TEC sequence with the time resolution of 1 h. F10.7 and Ap are processed into a data format consistent with TEC time resolution.
Step two: converting TEC time sequence X after conversion0={x1,x2,x3,...,xnDivide into training set Xtr={x1,x2,...xmAnd test set Xte={x1,x2,...xlThe ratio is 4: 1. After the data set is divided, the element x in the training set is subjected tomCarrying out zero mean value standardization to obtain a standardized training set of X'tr={x′1,x′2,...x′tThe normalization formula is:
Figure BDA0002641805950000061
in the formula: mu is the mean value of the original TEC time sequence, sigma is the standard error of the original TEC time sequence, and t is more than or equal to 1 and less than or equal to m. The training set and test set division ratio of the solar radiation flux value F10.7 and the geomagnetic activity index Ap is also 4: 1.The zero mean normalization process is the same as the equation for zero mean normalization of TEC data.
Step three: determining the topological structure of the TCN neural network, initializing the weight of the TCN network, adjusting parameters in the model, setting parameters such as the number of iterations as 50, the expansion factor d as 1,2 and 4, the convolution kernel size as 3, the learning rate as 0.001 and the number of hidden layer neurons as 20. Specifically, the TCN network is designed as a three-layer network structure including an input layer, a hidden layer, and an output layer. Because the electron content of the ionized layer is closely related to geomagnetic activity and solar activity, the geomagnetic index Ap, the solar activity index F10.7 and the electron content TEC are selected as input parameters by the input layer. The hidden layer uses causal convolution and expansion convolution as standard convolution layers, every two convolution layers and identity mapping are packaged into a residual module, then the residual module is used for stacking a deep layer network, and the last layers use full convolution layers to replace full connection layers. The output layer output target is the ionospheric electron content value 24 hours in the future.
Using the causal convolutional CNN model, the sequence problem can be translated into: according to x1,x2,...xnTo predict y1,y2,…ynDefinition of causal convolution, filter F ═ F1,f2,...fk}, sequence X0={x1,x2,x3,...,xnAt xtThe causal convolution of (a) is:
Figure BDA0002641805950000071
in the causal convolution model of the prediction model, the last two nodes of the input layer are assumed to be x respectivelyt-1,xtThe last node of the first hidden layer is ytFilter F ═ F1,f2) According to the formula, there is yt=f1xt-1+f2xt
Hole convolution is to add holes to the standard convolution to increase the field of view, so that each convolution output contains a larger range of information, at xtThe convolution of a hole at a diaglationatrate equal to d is:
Figure BDA0002641805950000072
where d is the dilation coefficient, k is the convolution kernel size, and (s-d · i) indicates which upper neuron is used when calculating the lower neurons. The TCN model adds the structural characteristic of hole convolution, expands the receptive field by skipping some existing pixels, and the number of the skipped pixels is (contrast-1). In a cavity convolution model of the TEC forecasting model, the last five nodes of the first hidden layer are assumed to be x respectivelyt-4,xt-3,xt-2,xt-1,xtThe last node of the hidden layer of the second layer is ytFilter F ═ F1,f2,f3) According to the formula, there is yt=f1xt-2d+f2xt-d+f3xt(d=2)。
The TCN network adds a layer jump connection of residual convolution and a convolution operation of 1 x 1 to improve accuracy. The residual convolution solves the training problem of a deep network through a special connection mode, and the number of network layers is greatly increased. The output of the residual block combines the input information and the output information of the input convolution operation, so that the information can be transmitted in a cross-layer mode, and a residual module in the TCN framework is shown in FIG. 2. In the invention, a residual error layer is constructed to replace one layer of convolution, and the input is subjected to hole convolution, weight normalization, an activation function and Dropout. Where WeighNorm and Dropout regularize the network.
Step four: and inputting the training set into the ionosphere TEC forecasting model in batches, calculating the output error of the effective historical length, and automatically adjusting the weight and the bias parameter by the model according to the output error in the process until the maximum iteration times or the accuracy and the error rate reach the optimal values. Specifically, the training set comprises the following training steps: s1, dividing single batch of input data into two parts, wherein the output of a residual error module is formed by adding a series of changes F of the two parts of input data. And in the F calculation process, as shown in fig. 3, the input data is subjected to the hole causal convolution, then the nonlinear processing is performed through the activation function Relu, and meanwhile, a Dropout layer is added, and the same operation is performed twice, so that the output of the residual error module is finally obtained. And S2, performing one-dimensional convolution operation on the input X. 1 × 1 convolution is used for reducing dimensions, if a feature jump layer of a lower layer is connected to an upper layer, the number of corresponding feature maps of each Cell is inconsistent, so that the addition operation of jump layer feature maps similar to Resnet cannot be directly performed, in order to make the number of two layers and the number of feature maps consistent, 1 × 1 convolution is used for performing a dimension reduction operation, and finally the output of the residual error module is obtained, and the output of the residual error module is calculated through a next residual error module; and S3, repeating the steps S1-S2, and obtaining an output result of the TCN when the last residual error module is calculated.
Step five: calculating errors by utilizing the square of the difference, accumulating errors obtained by multiple batches of data, comparing the errors with an error threshold value, taking the fact that the Loss function is infinitely close to 0 and the accuracy function is infinitely close to 1 as a model optimization target, finishing training if the errors are within a set range, executing a sixth step, executing a third step if the errors are beyond the set range, continuously updating parameters such as network weight and the like through reverse error propagation, and training again until the optimal parameters are found out to determine the optimal model.
Step six: after the model training is finished, the trained model is recorded as TCNTECInputting test set data into TCNTECIn the model, the output result is recorded as Y0Then Y is put0And carrying out Z-score denormalization to obtain a final TEC prediction result.
Step seven: and comparing the prediction result with the observation data, and evaluating the effectiveness of the model by adopting three performance indexes, namely MAPE, MAD and RMSE, wherein the smaller the three performance index values are, the closer the predicted value and the true value are, the better the fitting degree is, and the smaller the error between the predicted value and the true value is. The performance index is correspondingly defined as:
Figure BDA0002641805950000091
Figure BDA0002641805950000092
Figure BDA0002641805950000093
wherein, YiIonospheric TEC predicted value, Q, for hour iiThe observed value of the ionized layer TEC in the ith hour is shown, and n is the length of the time period.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (8)

1. An ionospheric total electron content forecasting method based on a TCN model is characterized by comprising the following steps:
the method comprises the following steps: acquiring ionized layer TEC data, a solar radiation flux value F10.7 and a geomagnetic activity index Ap, and converting the TEC into a high-time resolution sequence;
step two: dividing the converted TEC time sequence, the solar radiation flux value F10.7 and the geomagnetic activity index Ap into a training set and a testing set, and then carrying out zero-mean standardization processing on data in the training set;
step three: determining a topological structure of the TCN neural network, initializing the weight of the TCN neural network, and adjusting parameters in the model;
step four: inputting the training set into an ionized layer TEC forecasting model in batches, calculating an output error of an effective historical length, and updating a TCN forecasting model weight parameter by back propagation of the error until the maximum iteration times or the accuracy and the error rate reach the optimal values;
step five: comparing training results, if the error is within a set range, ending the training, executing the step six, if the error exceeds the set range, executing the step three, updating TCN network parameters, performing the training again until the optimal parameters are found, and determining an optimal model;
step six: after the model training is finished, the trained model is recorded as TCNTECInputting test set data into TCNTECIn the model, the output result is recorded as Y0Then Y is put0Performing Z-score denormalization to obtain a final TEC prediction result;
step seven: and comparing the prediction result with the observation data, and evaluating the effectiveness of the model by adopting three performance indexes, namely MAPE, MAD and RMSE.
2. The method for forecasting the total electron content of the ionized layer based on the TCN model according to claim 1, wherein in the first step, the TEC high resolution sequence conversion mode is as follows: converting the downloaded ionized layer data into VTEC data through a spherical harmonic function, eliminating abnormal data which do not conform to the normal range of the electron content, and filling ionized layer TEC data through a spline interpolation method to obtain a TEC sequence with the time resolution of 1 h; f10.7 and Ap are processed into a data format consistent with TEC time resolution.
3. The method for forecasting the total electron content in the ionosphere based on the TCN model according to claim 1, wherein in the second step, the TEC, the solar radiation flux F10.7, and the geomagnetic activity index Ap are normalized by: recording TEC time sequence as X0={x1,x2,x3,...,xn}, training set Xtr={x1,x2,...xm}: test set Xte={x1,x2,...xl4: 1; after the data set is divided, the element x in the training set is subjected tomCarrying out zero mean value standardization to obtain a standardized training set of X'tr={x′1,x′2,...x′tThe normalization formula is:
Figure FDA0002641805940000021
in the formula: mu is the mean value of the original TEC time sequence, sigma is the standard error of the original TEC time sequence, and t is more than or equal to 1 and less than or equal to m; the division ratio of the solar radiation flux value F10.7 and the geomagnetic activity index Ap to the training set and the test set is also 4:1, and the zero-mean standardization process and formula are the same as those of TEC data zero-mean standardization.
4. The method for forecasting the total electron content of the ionized layer based on the TCN model according to claim 1, wherein in the third step, the TCN network is designed into a three-layer network structure comprising an input layer, a hidden layer and an output layer; the input layer comprises input parameters of a geomagnetic index Ap, a solar activity index F10.7 and an electron content TEC; the hidden layer uses causal convolution and expansion convolution as standard convolution layers, every two convolution layers and identity mapping are packaged into a residual module, then the residual module is used for stacking a deep layer network, and the last layers use full convolution layers to replace full connection layers; the output layer output target is the ionospheric electron content value 24 hours in the future.
5. The method for forecasting the total electron content of the ionized layer based on the TCN model as claimed in claim 4, wherein in the third step, the weight for initializing the TCN network is as follows: the number of iterations is 50, the dilation factor d is 1,2,4, the convolution kernel size is 3, the learning rate size is 0.001, and the number of hidden layer neurons is 20.
6. The method for forecasting the total electron content of the ionosphere based on the TCN model according to claim 5, wherein in the fourth step, the training set training step is as follows: s1, dividing single batch of input data into two parts, wherein the output of a residual error module is formed by adding a series of changes F of the two parts of input data; s2, performing one-dimensional convolution operation on the input X to finally obtain the output of the residual error module, wherein the output of the residual error module is calculated by the next residual error module; and S3, repeating the steps S1-S2, and obtaining an output result of the TCN when the last residual error module is calculated.
7. The method for predicting the total electron content of the ionosphere based on the TCN model according to claim 6, wherein the input data in S1 is subjected to a hole causal convolution, then to a nonlinear treatment by activating a function Relu, and a Dropout layer is added at the same time, and the output of the residual error module is obtained after two identical operations.
8. The method for forecasting the total electron content of the ionized layer based on the TCN model according to claim 1, wherein in the seventh step, the three performance indexes of MAPE, MAD and RMSE are respectively:
Figure FDA0002641805940000031
Figure FDA0002641805940000032
wherein, YiIonospheric TEC predicted value, Q, for hour iiThe observed value of the ionized layer TEC in the ith hour is shown, and n is the length of the time period; when the three performance index values are smaller, the predicted value is closer to the true value, the fitting degree is better, and the error between the predicted value and the true value is smaller.
CN202010842098.0A 2020-08-20 2020-08-20 Ionosphere total electron content forecasting method based on TCN model Active CN111967679B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010842098.0A CN111967679B (en) 2020-08-20 2020-08-20 Ionosphere total electron content forecasting method based on TCN model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010842098.0A CN111967679B (en) 2020-08-20 2020-08-20 Ionosphere total electron content forecasting method based on TCN model

Publications (2)

Publication Number Publication Date
CN111967679A true CN111967679A (en) 2020-11-20
CN111967679B CN111967679B (en) 2023-12-08

Family

ID=73388691

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010842098.0A Active CN111967679B (en) 2020-08-20 2020-08-20 Ionosphere total electron content forecasting method based on TCN model

Country Status (1)

Country Link
CN (1) CN111967679B (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112734137A (en) * 2021-01-27 2021-04-30 国网电力科学研究院武汉能效测评有限公司 Short-term building power load prediction method and system based on intelligent electric meter data
CN112862159A (en) * 2021-01-13 2021-05-28 中铁第四勘察设计院集团有限公司 Method, device, equipment and storage medium for predicting total electron content of ionized layer
CN113379107A (en) * 2021-05-26 2021-09-10 江苏师范大学 Regional ionized layer TEC forecasting method based on LSTM and GCN
CN113625144A (en) * 2021-08-11 2021-11-09 北京信息科技大学 IGBT fault prediction method and system
CN115951430A (en) * 2023-03-15 2023-04-11 数字太空(北京)智能技术研究院有限公司 Geomagnetic index forecasting method based on LSTM
CN116050600A (en) * 2022-12-28 2023-05-02 中国电子科技集团公司第三十八研究所 CNN-GA-BP-based combined model spare part demand prediction method and system
CN116341356A (en) * 2022-12-19 2023-06-27 湖北珞珈实验室 Ionosphere total electron content prediction system and method with additional constraint station
CN116882594A (en) * 2023-09-06 2023-10-13 数字太空(北京)智能技术研究院有限公司 Geomagnetic ap index medium-term forecasting method based on limited recursion and deep learning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110728411A (en) * 2019-10-18 2020-01-24 河海大学 High-low altitude area combined rainfall prediction method based on convolutional neural network
CN110942194A (en) * 2019-11-27 2020-03-31 徐州上若科技有限公司 Wind power prediction error interval evaluation method based on TCN
CN111540199A (en) * 2020-04-21 2020-08-14 浙江省交通规划设计研究院有限公司 High-speed traffic flow prediction method based on multi-mode fusion and graph attention machine mechanism

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110728411A (en) * 2019-10-18 2020-01-24 河海大学 High-low altitude area combined rainfall prediction method based on convolutional neural network
CN110942194A (en) * 2019-11-27 2020-03-31 徐州上若科技有限公司 Wind power prediction error interval evaluation method based on TCN
CN111540199A (en) * 2020-04-21 2020-08-14 浙江省交通规划设计研究院有限公司 High-speed traffic flow prediction method based on multi-mode fusion and graph attention machine mechanism

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112862159A (en) * 2021-01-13 2021-05-28 中铁第四勘察设计院集团有限公司 Method, device, equipment and storage medium for predicting total electron content of ionized layer
CN112734137A (en) * 2021-01-27 2021-04-30 国网电力科学研究院武汉能效测评有限公司 Short-term building power load prediction method and system based on intelligent electric meter data
CN112734137B (en) * 2021-01-27 2022-12-16 国网电力科学研究院武汉能效测评有限公司 Short-term building power load prediction method and system based on intelligent electric meter data
CN113379107A (en) * 2021-05-26 2021-09-10 江苏师范大学 Regional ionized layer TEC forecasting method based on LSTM and GCN
CN113379107B (en) * 2021-05-26 2023-12-19 江苏师范大学 Regional ionosphere TEC forecasting method based on LSTM and GCN
CN113625144A (en) * 2021-08-11 2021-11-09 北京信息科技大学 IGBT fault prediction method and system
CN116341356A (en) * 2022-12-19 2023-06-27 湖北珞珈实验室 Ionosphere total electron content prediction system and method with additional constraint station
CN116341356B (en) * 2022-12-19 2024-01-12 湖北珞珈实验室 Ionosphere total electron content prediction system and method with additional constraint station
CN116050600A (en) * 2022-12-28 2023-05-02 中国电子科技集团公司第三十八研究所 CNN-GA-BP-based combined model spare part demand prediction method and system
CN115951430A (en) * 2023-03-15 2023-04-11 数字太空(北京)智能技术研究院有限公司 Geomagnetic index forecasting method based on LSTM
CN116882594A (en) * 2023-09-06 2023-10-13 数字太空(北京)智能技术研究院有限公司 Geomagnetic ap index medium-term forecasting method based on limited recursion and deep learning
CN116882594B (en) * 2023-09-06 2023-12-01 数字太空(北京)智能技术研究院有限公司 Geomagnetic ap index medium-term forecasting method based on limited recursion and deep learning

Also Published As

Publication number Publication date
CN111967679B (en) 2023-12-08

Similar Documents

Publication Publication Date Title
CN111967679A (en) Ionized layer total electron content forecasting method based on TCN model
CN113379107B (en) Regional ionosphere TEC forecasting method based on LSTM and GCN
CN112766600B (en) Urban area crowd flow prediction method and system
CN112488396A (en) Wavelet transform-based electric power load prediction method of Holt-Winters and LSTM combined model
CN111144644B (en) Short-term wind speed prediction method based on variation variance Gaussian process regression
CN114091766B (en) CEEMDAN-LSTM-based space load prediction method
CN116316591A (en) Short-term photovoltaic power prediction method and system based on hybrid bidirectional gating cycle
CN113553755B (en) Power system state estimation method, device and equipment
CN113281790A (en) Beidou satellite clock error forecasting method and device
CN115587666A (en) Load prediction method and system based on seasonal trend decomposition and hybrid neural network
CN114358389A (en) Short-term power load prediction method combining VMD decomposition and time convolution network
CN111817805A (en) Method, device and medium for adjusting channel propagation model parameters
CN113836823A (en) Load combination prediction method based on load decomposition and optimized bidirectional long-short term memory network
Manry et al. Signal processing using the multilayer perceptron
CN116484189A (en) ERA5 precipitation product downscaling method based on deep learning
CN113984198B (en) Shortwave radiation prediction method and system based on convolutional neural network
CN117114168A (en) Long-time-scale sea surface temperature intelligent forecasting method
CN116502774A (en) Time sequence prediction method based on time sequence decomposition and Legend projection
CN116152206A (en) Photovoltaic output power prediction method, terminal equipment and storage medium
CN113723006B (en) LS-SVM (least squares-support vector machine) -based single-station earth change magnetic field modeling prediction method and system
CN113221450B (en) Space-time prediction method and system for sparse non-uniform time sequence data
CN114819264A (en) Photovoltaic power station irradiance ultra-short term prediction method based on space-time dependence and storage medium
CN114004421B (en) Traffic data missing value interpolation method based on space-time integrated learning
Zhao et al. A quantized kernel least-mean-square algorithm based on echo state network for online time-series prediction
CN114912577A (en) Wind power plant short-term wind speed prediction method combining VMD and attention mechanism

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant