CN113033776A - Time sequence prediction method combining global cavity convolution and local identification characterization - Google Patents

Time sequence prediction method combining global cavity convolution and local identification characterization Download PDF

Info

Publication number
CN113033776A
CN113033776A CN202110262391.4A CN202110262391A CN113033776A CN 113033776 A CN113033776 A CN 113033776A CN 202110262391 A CN202110262391 A CN 202110262391A CN 113033776 A CN113033776 A CN 113033776A
Authority
CN
China
Prior art keywords
time sequence
time
local identification
time series
context
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110262391.4A
Other languages
Chinese (zh)
Inventor
陈天翼
金苍宏
董腾然
吴明晖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University City College ZUCC
Original Assignee
Zhejiang University City College ZUCC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University City College ZUCC filed Critical Zhejiang University City College ZUCC
Priority to CN202110262391.4A priority Critical patent/CN113033776A/en
Publication of CN113033776A publication Critical patent/CN113033776A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a time sequence prediction method combining global cavity convolution and local identification representation, which comprises the following steps: extracting the characteristics of the time sequence segments to generate corresponding representative characteristic vectors; based on the obtained vectors, a clustering label is given to each time sequence segment through a clustering method so as to identify different types of time sequences; then, extracting local identification representations (Shapelets) corresponding to different time sequence categories through time sequence representation learning, performing multiple identifiable context fusion processing on each time sequence segment and the obtained representations, and stacking the time sequence segments and the obtained representations in a channel direction to enable a native time sequence to be advanced into an identifiable context; and performing time sequence prediction training on the obtained context through time hole convolution with residual connection. The invention integrates time convolution and provides a novel time sequence representation extraction and fusion mode to enhance the influence of different bit importance in time sequence on the prediction result, so that the precision of the prediction result is effectively improved.

Description

Time sequence prediction method combining global cavity convolution and local identification characterization
Technical Field
The invention belongs to the field of data analysis, and relates to a time series prediction method combining global cavity convolution and local identification characterization.
Background
The time sequence data is used for describing the characteristics of the change of the object along with the time, and the research on the time sequence can help people to know the historical development mode of the object and predict the future change trend of the object, such as stock price fluctuation, road and vehicle flow prediction, user behavior analysis and the like. The essential characteristics of the sequence can be drawn according to the mean, variance, covariance and the like of the sequence, and the sequence is divided into a stationary time sequence, a non-stationary time sequence and the like. Deep learning methods such as recurrent neural networks and time convolution networks cannot take into account the problems of perception domain limitation and local feature loss. Therefore, there is a need to provide a new fusion method that can overcome the above-mentioned difficulties by fusing local features into the hole convolution.
Disclosure of Invention
In order to overcome the defects of the existing method, one of the purposes of the invention is to provide a time series prediction method combining global hole convolution and local identification characterization, which can improve the prediction accuracy and reduce the prediction error.
One of the purposes of the invention is realized by adopting the following technical scheme:
a time series prediction method combining global cavity convolution and local identification characterization is characterized by comprising the following steps:
and extracting the characteristics of the time sequence segments to obtain representative characteristic vector expressions corresponding to different time sequence segments, wherein the extracted characteristics comprise but are not limited to autocorrelation coefficients, mean value changes and average second derivative centers. The obtained features are extracted to describe the features of repetitive patterns, sequence fluctuation, numerical mutation, information from the future and the past, and the like contained in the time sequence, so that feature vector representations of different time sequences are obtained.
Further, dividing the time sequence into a plurality of different clusters by an unsupervised clustering method according to the obtained feature vector so as to realize the division of different time sequence categories;
further, the time sequence characteristics are learned according to the obtained time sequence category labels, generalization corresponding to different time sequence categories is extracted through optimization of the objective function, and the generalization is used as a recognizable similarity measurement standard;
further, a sliding window with a fixed size is set to sample the time sequence to obtain the micro context, and similarity calculation is performed according to different criteria with each previously obtained similarity metric, including a value square deviation, a value average deviation and a dot product ratio. A higher similarity indicates that more distinguishable information is contained in the context. Splicing the three similarity vectors obtained by calculation with the original time sequence in the channel direction, so that the identifiable characteristic information of the original time sequence is distributed into a plurality of different channels;
further, the perceptron of the model is increased by the stacked several layers of time convolution, so as to ensure that the model can capture the context distinguishing information for a longer time in the training process. By the micro context importance feature, the model can have the capability of predicting based on the weights of different points, so that the importance of the context is enhanced. Hole convolution can be interpreted as the convolution of a down-sampled version of the lower layer features, reducing the resolution of combining information from distant history and the future. By increasing the expansion rate of the hole convolution, it can be made to use more information more efficiently.
Further, according to the time series prediction model, the processed time series data is predicted and a result is output, and the method comprises the following steps:
according to a univariate time sequence, clustering is carried out through time sequence feature extraction to obtain a corresponding clustering label, local identification representations of the time sequence are obtained through the label through unsupervised training, then the representations are used as measurement standards, similarity is calculated according to different standards with micro-context obtained through sliding window sampling, importance features of different point positions are obtained, superposition is carried out in a channel direction, finally fusion and learning are carried out on the features through time hole convolution with residual connection, and therefore results are predicted and output.
Compared with the prior art, the invention has the beneficial effects that:
the invention integrates a point location weight obtaining and fusing method based on micro-context, can systematically realize the local identification representation of point locations through local identification representation, thereby solving the problem that the specific point location environment is difficult to identify in time sequence prediction; the model uses the time convolution with the holes to enhance the global capturing capability of the time sequence, and meanwhile, the model is combined with the point importance, so that not only can the overall trend be considered, but also the local characteristics can be adapted, the prediction result is more accurate, and meanwhile, the model has better robustness and interpretability.
Drawings
FIG. 1 is a schematic diagram of a model structure of a time series prediction method combining global hole convolution and local identification characterization according to the present invention.
Detailed Description
The present invention will now be described in more detail with reference to the accompanying drawings, in which the description of the invention is given by way of illustration and not of limitation. The various embodiments may be combined with each other to form other embodiments not shown in the following description.
As shown in fig. 1, the method for predicting a time series by combining global hole convolution and local identification representation according to the present invention includes the following steps:
performing feature extraction on the time sequence segments so as to generate a corresponding representative feature vector for each time sequence segment;
based on the obtained vectors, a clustering label is given to each time sequence segment through a clustering method so as to identify different types of time sequences;
extracting local identification representations (Shapelets) corresponding to different time sequence categories through time sequence representation learning, performing multi-identification context fusion and fusion on each time sequence segment and the obtained representations, and stacking the time sequence segments and the representations in a channel direction to enable a native time sequence to be advanced into an identification context;
performing the processing according to the time sequence segments in the training set and the verification set respectively, and performing time sequence prediction training on the obtained context through time hole convolution with residual connection;
and importing the test set into a pre-training model, predicting the processed time sequence segment and outputting a result.
The following describes the method for predicting a time series by combining global hole convolution and local identification representation according to the present invention in further detail with reference to the following embodiments.
Extracting the features of the time sequence segments to obtain representative feature vector expressions corresponding to different time sequence segments, wherein the extracted features are as follows:
autocorrelation coefficient:
Figure RE-GDA0003070779520000031
wherein L represents the total length of the time series T; t ist,t∈[1,L]The value of the time sequence at the time t is obtained; sigma2Is the variance of the time series; μ is the mean of the time series; l is a hysteresis coefficient. The autocorrelation coefficients can help to find repetitive patterns contained in the timing, such as periodic signals that are masked by noise, and the like.
Mean change:
Figure RE-GDA0003070779520000032
wherein L represents the total length of the time series T; t ist,t∈[1,L]The value of the time sequence at the time t is taken. Mean change was calculated from the mean of the differences between subsequences, and sequence fluctuations could be captured and mutations identified in sequence values.
Center of average second derivative:
Figure RE-GDA0003070779520000041
wherein L represents the total length of the time series T; t ist,t∈[1,L]The value of the time sequence at the time t is taken. Providing a more accurate approximation of the derivative by averaging the second derivative center approximations facilitates summarizing information from both the future and the past contained in the time series.
The extracted features are used as a representative feature vector of the original time series, so that the features of a certain time series can be expressed as:
Figure RE-GDA0003070779520000042
wherein, ViRepresenting a characteristic vector corresponding to the ith time sequence; e.g. of the typeijRepresenting the jth characteristic corresponding to the ith time sequence; m represents the number of extracted features.
In this embodiment, N time series are divided into K different clusters C by a K-Means clustering method according to the obtained feature vector, so as to obtain cluster labels with different time sequences. The optimization target is as follows:
Figure RE-GDA0003070779520000043
wherein, mujIs the average vector of all samples in cluster C; n is the total number of time series.
In this embodiment, the local identification representations corresponding to different time sequence categories are extracted through the obtained category labels, and are used as the identifiable similarity measurement standard. The regularized time series characterization learning objective function is defined as f, the goal of time series characterization learning is to learn several optimal locally recognized characterizations P and a linear hyperplane H to minimize the objective f:
Figure RE-GDA0003070779520000044
wherein λ isHIs a regularization coefficient; n isfThe number of the extracted time series characteristics; etajCorresponding to the jth feature in the extracted features of the current time sequence; xijJ-th feature in the features extracted corresponding to the residual time sequence;
in this embodiment, the micro contexts are obtained by sampling the time sequence according to a fixed sliding window, and then the distinguishable similarity metric is respectively calculated for each micro context.
First of all, for a time sequence T ═ T according to a given sliding window size W1,T2,...,TL-1,TLSampling is carried out, and a micro context can be defined as
Figure RE-GDA0003070779520000051
Where | W | is the size of the sliding window. Similarity calculation is performed with the previously obtained metrics during the sliding sampling, and when the similarity between the micro context and the local identification representation is higher, the micro context contains more identifiable information. Different criteria are thus required to obtain a characterization of micro-context importance:
value squared offset (VSD):
Figure RE-GDA0003070779520000052
wherein, CiA micro context obtained by sampling a certain section in the time sequence; piE.P is the ith local identification representation serving as the measurement standard; i Ci-Pi||2Is a vector Ci-PiThe second norm. The value squared offset is used to measure the mean squared deviation between two vectors.
Mean shift of Value (VMD):
Figure RE-GDA0003070779520000053
wherein, | | Ci-Pi||1Is a vector Ci-PiThe first norm. The mean deviation of the values is used to measure the mean absolute deviation between the two vectors.
Dot Product Ratio (DPR):
Figure RE-GDA0003070779520000054
wherein, CijAs a micro context CiThe jth value of (d); pijA jth value characterizing an ith local identification; and | W | is the size of the sliding window. The dot product ratio is used to measure the ratio of dot products between two vectors. Then, toThe over-stacking way will splice the calculated vectors:
Figure RE-GDA0003070779520000055
wherein the content of the first and second substances,
Figure RE-GDA0003070779520000056
representing a vector splicing operation in the channel direction;
Figure RE-GDA0003070779520000057
representing a time series after the superposition;
Figure RE-GDA0003070779520000058
representing a squared offset vector of values;
Figure RE-GDA0003070779520000059
representing a value average offset vector;
Figure RE-GDA00030707795200000510
representing a dot product ratio vector. Thus, the distinguishable feature information of the original time series is distributed into a plurality of different channels.
This embodiment employs a time-hole convolution with residual concatenation. Hole convolution can be interpreted as the convolution of a down-sampled version of the lower layer features, reducing the resolution of combining information from distant history and the future. By increasing the expansion rate of the hole convolution, it can be made to use more information more efficiently. The hole convolution is defined as:
Figure RE-GDA0003070779520000061
wherein the content of the first and second substances,
Figure RE-GDA0003070779520000062
represents the intermediate state of the convolutional layer at time t; is a convolution operation; w is represented atFixed weights for filters of layer l;
Figure RE-GDA0003070779520000063
presentation pair
Figure RE-GDA0003070779520000064
Carrying out down-sampling operation; dlRepresents the expansion rate of the hole convolution;
Figure RE-GDA00030707795200000611
is the coefficient of expansion; k is the maximum length of the vector.
Figure RE-GDA0003070779520000065
Wherein l is a layer in the calculation in the time convolution; ShapeletJoint is the local identification characterization fusion step described previously; dlConvolving the dilation rate in the current layer with time;
Figure RE-GDA0003070779520000066
weights in the ith layer for time convolution; ReLU is a nonlinear activation function that can be expressed as:
Figure RE-GDA0003070779520000067
in the algorithm, the receptive field of the model is increased by stacking the layers of the time convolution, so that the model can capture longer-time context distinguishing information in the training process. By the micro context importance feature, the model can have the capability of predicting based on the weights of different points, so that the importance of the context is enhanced.
The present embodiment trains the model using a quantile loss function. Selecting a quantile penalty instead of a squared error enables a gradient descent based learning algorithm to learn a specified quantile instead of a mean, which would be beneficial for obtaining better robustness against outliers in the prediction process. Meanwhile, quantile loss may help to better explain variables other than the mean of the data, thereby helping to understand the results of the non-normal distribution and the non-linear relationship with the predictive variables.
For a given quantile ρ ∈ (0; 1), the real value y of the time seriestAnd rho-quantile prediction
Figure RE-GDA0003070779520000068
The ρ -quantile loss at this time can be defined as:
Figure RE-GDA0003070779520000069
wherein the content of the first and second substances,
Figure RE-GDA00030707795200000610
is defined as:
Figure RE-GDA0003070779520000071
and finally, performing the processing according to the time sequence segments in the training set and the verification set respectively, and performing time sequence prediction training on the obtained context through time hole convolution with residual connection. After training is finished, the test set can be led into a pre-training model, processed time sequence segments are predicted, and results are output.
Various other modifications and changes may be made by those skilled in the art based on the above-described technical solutions and concepts, and all such modifications and changes should fall within the scope of the claims of the present invention.
The technical effects of the present invention will be described below by comparing the global cavity convolution and local identification representation time series depth model with other various methods.
Common data set: electric (hourly Electricity time series for 370 customers), Solar Energy (Solar data sampled every 10 minutes in alabama in 2016), Traffic (hourly road occupancy for the san francisco highway, between 0 and 1), Taxi (30 minutes records the number of taxis in different blocks of new york city), Wikipedia (records daily online visits per ten thousand Wikipedia pages per day), M4 (from M4 contests and contains different time aggregation lengths, day, month, quarter, and year, corresponding to M4-D, M4-M, M4-Q, and M4-Y, respectively).
Other comparison methods:
ARIMA: the differential integration moving average autoregressive model is a typical time series model and consists of three parts: AR model (autoregressive model) and MA model (moving average model), and order I of the difference;
NTPS: the prediction of the next timestamp is the previous value of the time series, and sampling is carried out according to exponential distribution;
MLP: the forward propagation neural network comprises 3 hidden layers;
WaveNet: the core is an expanded causal convolutional layer which allows neural networks to correctly process time sequences and process long-term dependencies without causing model complexity explosion, alleviating long-term challenges of learning in a large number of time steps;
LSTM: a time cycle neural network can solve the long-term dependence problem of the general cycle neural network;
MQCNN: a sequence-to-sequence neural network architecture using LSTM as an encoder and decoder;
transformer: a sequence-to-sequence neural network architecture with a self-attention mechanism;
TCN: a network structure capable of processing time sequence data can deduce new information at a plurality of time points in the future according to the sequence of each point position in a known sequence;
DeepAR: the time sequence prediction method based on deep learning can conveniently take additional characteristics into consideration, and the prediction target is the probability distribution of the value of the sequence at each time step
NBeats: deep neural networks with interpretability are implemented based on forward and backward residual linkage and a very deep full-link layer stack.
The present embodiment employs a normalized quantile loss sum as an evaluation index that can reduce bias in making error measurements, thereby facilitating comparisons between different data sets or methods of different scales. It is defined as:
Figure RE-GDA0003070779520000081
wherein y is the true value of the time sequence;
Figure RE-GDA0003070779520000082
is the predicted value output by the model. We calculate the results when ρ is 0.5 and 0.9, which are denoted as P50QL and P90QL, respectively.
TABLE 1 time series prediction method comparison
Figure RE-GDA0003070779520000083
The results of the experiment are shown in table 1, with results that are best in performance being shown in bold and results that are suboptimal in performance being underlined. The method of the present invention is superior to other comparison methods in most cases.

Claims (7)

1. A time series prediction method combining global cavity convolution and local identification characterization is characterized by comprising the following steps:
extracting the characteristics of the time sequence segments to obtain representative characteristic vector expressions corresponding to different time sequence segments, and marking cluster labels by a clustering method;
extracting local identification representations (Shapelets) corresponding to different time sequence categories through time sequence representation learning, and taking the local identification representations (Shapelets) as identifiable similarity measurement standards; then, performing multiple distinguishable context fusion processing on the obtained characterization and sequence, and stacking in the channel direction to evolve into distinguishable contexts;
performing the processing according to the training set and the time sequence segments in the verification set respectively, and performing time sequence prediction training on the obtained context through global hole convolution, namely time hole convolution with residual connection;
and importing the test set into a pre-training model, predicting the processed time sequence segment and outputting a result.
2. The method for predicting time series by combining convolution of global hole and local identification characterization according to claim 1, wherein the extracting of the time series segment feature includes the following indexes:
autocorrelation coefficient:
Figure FDA0002969685390000011
wherein L represents the total length of the time series T; t ist,t∈[1,L]The value of the time sequence at the time t is obtained; sigma2Is the variance of the time series; μ is the mean of the time series; l is a hysteresis coefficient; the autocorrelation coefficients can help to find the repetitive patterns contained in the timing;
mean change:
Figure FDA0002969685390000012
wherein L represents the total length of the time series T; t ist,t∈[1,L]The value of the time sequence at the time t is obtained; the mean change is obtained by calculating the mean of the differences between the subsequences, and can capture sequence fluctuation and identify the mutation on the sequence value;
center of average second derivative:
Figure FDA0002969685390000013
wherein L represents the total length of the time series T; t ist,t∈[1,L]The value of the time sequence at the time t is obtained; providing a more accurate approximation of the derivative by averaging the second derivative center approximations facilitates summarizing information from both the future and the past contained in the time series.
3. The method of claim 2, wherein the global hole convolution and local identification characterization are combined to predict the time series, and the method comprises: in order to realize the identification of the clustering labels of the segments of different time sequences, the extracted features are used as a representative feature vector of the original time sequence, so that the features of a certain time sequence can be expressed as follows through the vector:
Figure FDA0002969685390000021
wherein, ViRepresenting the feature vector corresponding to the ith time series, eijRepresenting the jth feature corresponding to the ith time sequence, wherein m represents the number of extracted features;
dividing N time sequences into K different clusters C by a K-Means clustering method according to the obtained feature vectors to obtain cluster labels with different time sequences, wherein the optimization goal is as follows:
Figure FDA0002969685390000022
wherein, mujIs the average vector of all samples in cluster C; n is the total number of time series.
4. The method of claim 1, wherein the global hole convolution and local identification characterization are combined to predict the time series, and the method comprises: extracting local identification representations corresponding to different time sequence categories through time sequence representation learning, taking the local identification representations as distinguishable similarity measurement standards, defining a regularized time sequence representation learning target function as gamma, and learning a plurality of optimal local identification representations P and a linear hyperplane H to minimize the target gamma:
Figure FDA0002969685390000023
wherein λ isHFor regularizing coefficients, nfThe number of the extracted time series characteristics; etajCorresponding to the jth feature in the extracted features of the current time sequence; xijAnd j-th feature in the extracted features corresponding to the residual time sequence.
5. The method of claim 1, wherein the global hole convolution and local identification characterization are combined to predict the time series, and the method comprises: sampling according to a fixed sliding window in a time sequence to obtain micro contexts, and then performing multiple distinguishable context fusion processing on the obtained local identification representations and each micro context respectively, wherein the multiple distinguishable context fusion processing specifically comprises the following steps:
first of all, for a time sequence T ═ T according to a given sliding window size W1,T2,...,TL-1,TLSampling is carried out, and a micro context can be defined as
Figure FDA0002969685390000024
Wherein | W | is the size of the sliding window; during the sliding sampling, similarity calculation is carried out with the previously obtained measurement standard, when the similarity of the micro context and the local identification representation is higher, the micro context contains more identifiable information, and therefore different standards are needed to obtain the characteristic representation of the micro context importance:
value squared offset (VSD):
Figure FDA0002969685390000031
wherein, CiA micro context obtained by sampling a certain section in the time sequence; piE.P is the ith local identification representation serving as the measurement standard; i Ci-Pi||2Is a vector Ci-PiA second norm of; the value square offset is used to measure the mean square deviation between two vectors;
mean shift of Value (VMD):
Figure FDA0002969685390000032
wherein, | | Ci-Pi||1Is a vector Ci-PiThe mean deviation of the values used to measure the mean absolute deviation between the two vectors;
dot Product Ratio (DPR):
Figure FDA0002969685390000033
wherein, CijAs a micro context CiThe jth value of (d); pijA jth value characterizing an ith local identification; l W is the size of the sliding window; the dot product ratio is used for measuring the ratio of dot products between two vectors; then, the calculated vectors are spliced by means of superposition:
Figure FDA0002969685390000034
wherein the content of the first and second substances,
Figure FDA0002969685390000035
representing a vector splicing operation in the channel direction;
Figure FDA0002969685390000036
representing a time series after the superposition;
Figure FDA0002969685390000038
representing a squared offset vector of values;
Figure FDA0002969685390000039
representing a value average offset vector;
Figure FDA00029696853900000310
representing a dot product ratio vector, the distinguishable feature information of the original time series is distributed into a plurality of different channels.
6. The method of claim 1, wherein the global hole convolution and local identification characterization are combined to predict the time series, and the method comprises: global hole convolution can be interpreted as convolution of a downsampled version of the lower layer features, reducing the resolution of combining information from distant histories and the future; by increasing the expansion rate of the global hole convolution, more information can be more effectively utilized; the global hole convolution is defined as:
Figure FDA0002969685390000037
wherein the content of the first and second substances,
Figure FDA0002969685390000041
indicating the convolutional layer at time t
Figure FDA0002969685390000044
An intermediate state of (1); is a convolution operation; w is represented at
Figure FDA0002969685390000045
Fixed weights for filters of a layer;
Figure FDA0002969685390000042
presentation pair
Figure FDA0002969685390000043
Carrying out down-sampling operation; dlRepresents the expansion rate of the global hole convolution; τ is the coefficient of expansion; k is the maximum length of the vector;
increasing the receptive field of the model by stacking the number of layers of the time convolution so as to ensure that the model can capture longer-time context distinguishing information in the training process; by the micro context importance feature, the model can have the capability of predicting based on the weights of different points, so that the importance of the context is enhanced.
7. The method of claim 6, wherein the global hole convolution and local identification characterization are combined to predict the time series, and the method comprises: according to the time sequence of the univariate, clustering is carried out through time sequence characteristic extraction to obtain a corresponding clustering label, and therefore the label obtains local identification representation of the time sequence through unsupervised training; then, taking the representations as measurement standards, calculating similarity with a micro context obtained by sampling through a sliding window according to different standards, acquiring importance characteristics of different point positions, and superposing the importance characteristics in a channel direction; finally, the features are fused and learned through time hole convolution with residual connection, and therefore prediction and result output are conducted.
CN202110262391.4A 2021-03-10 2021-03-10 Time sequence prediction method combining global cavity convolution and local identification characterization Pending CN113033776A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110262391.4A CN113033776A (en) 2021-03-10 2021-03-10 Time sequence prediction method combining global cavity convolution and local identification characterization

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110262391.4A CN113033776A (en) 2021-03-10 2021-03-10 Time sequence prediction method combining global cavity convolution and local identification characterization

Publications (1)

Publication Number Publication Date
CN113033776A true CN113033776A (en) 2021-06-25

Family

ID=76469426

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110262391.4A Pending CN113033776A (en) 2021-03-10 2021-03-10 Time sequence prediction method combining global cavity convolution and local identification characterization

Country Status (1)

Country Link
CN (1) CN113033776A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114239718A (en) * 2021-12-15 2022-03-25 杭州电子科技大学 High-precision long-term time sequence prediction method based on multivariate time sequence data analysis
CN115114345A (en) * 2022-04-02 2022-09-27 腾讯科技(深圳)有限公司 Feature representation extraction method, device, equipment, storage medium and program product
CN117407681A (en) * 2023-12-15 2024-01-16 江苏为恒智能科技有限公司 Time sequence data prediction model establishment method based on vector clustering

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114239718A (en) * 2021-12-15 2022-03-25 杭州电子科技大学 High-precision long-term time sequence prediction method based on multivariate time sequence data analysis
CN114239718B (en) * 2021-12-15 2024-03-01 杭州电子科技大学 High-precision long-term time sequence prediction method based on multi-element time sequence data analysis
CN115114345A (en) * 2022-04-02 2022-09-27 腾讯科技(深圳)有限公司 Feature representation extraction method, device, equipment, storage medium and program product
CN115114345B (en) * 2022-04-02 2024-04-09 腾讯科技(深圳)有限公司 Feature representation extraction method, device, equipment, storage medium and program product
CN117407681A (en) * 2023-12-15 2024-01-16 江苏为恒智能科技有限公司 Time sequence data prediction model establishment method based on vector clustering
CN117407681B (en) * 2023-12-15 2024-03-22 江苏为恒智能科技有限公司 Time sequence data prediction model establishment method based on vector clustering

Similar Documents

Publication Publication Date Title
CN113033776A (en) Time sequence prediction method combining global cavity convolution and local identification characterization
Li et al. A new flood forecasting model based on SVM and boosting learning algorithms
CN112561165A (en) Multidimensional time series data prediction method based on combined model
CN110674858B (en) Traffic public opinion detection method based on space-time correlation and big data mining
CN106709588B (en) Prediction model construction method and device and real-time prediction method and device
CN111598325A (en) Traffic speed prediction method based on hierarchical clustering and hierarchical attention mechanism
CN112487822A (en) Cross-modal retrieval method based on deep learning
CN112508265A (en) Time and activity multi-task prediction method and system for business process management
CN112507479B (en) Oil drilling machine health state assessment method based on manifold learning and softmax
CN114503124A (en) Multi-level prediction for processing time series data
CN116227716A (en) Multi-factor energy demand prediction method and system based on Stacking
CN115017970A (en) Migration learning-based gas consumption behavior anomaly detection method and system
CN110659767A (en) Stock trend prediction method based on LSTM-CNN deep learning model
Fu et al. MCA-DTCN: A novel dual-task temporal convolutional network with multi-channel attention for first prediction time detection and remaining useful life prediction
CN117094835A (en) Multi-target group classification method for social media content
CN116579342A (en) Electric power marketing named entity identification method based on dual-feature combined extraction
CN116304941A (en) Ocean data quality control method and device based on multi-model combination
CN112650949B (en) Regional POI (point of interest) demand identification method based on multi-source feature fusion collaborative filtering
Jasim et al. Characteristics of data mining by classification educational dataset to improve student’s evaluation
CN113657533A (en) Multi-element time sequence segmentation clustering method for space-time scene construction
Rameh et al. Designing a hybrid model for stock marketing prediction based on LSTM and transfer learning
CN111767402B (en) Limited domain event detection method based on counterstudy
Watts et al. Local score dependent model explanation for time dependent covariates
CN112989105A (en) Music structure analysis method and system
Kim et al. Similarity-based historical input selection to predict irregular holiday traffics in real-time

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210625

RJ01 Rejection of invention patent application after publication