CN117786374B - Multivariate time sequence anomaly detection method and system based on graph neural network - Google Patents
Multivariate time sequence anomaly detection method and system based on graph neural network Download PDFInfo
- Publication number
- CN117786374B CN117786374B CN202410221484.6A CN202410221484A CN117786374B CN 117786374 B CN117786374 B CN 117786374B CN 202410221484 A CN202410221484 A CN 202410221484A CN 117786374 B CN117786374 B CN 117786374B
- Authority
- CN
- China
- Prior art keywords
- time series
- input
- sequence
- output
- anomaly detection
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 95
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 33
- 238000012549 training Methods 0.000 claims abstract description 60
- 238000000034 method Methods 0.000 claims abstract description 48
- 230000008569 process Effects 0.000 claims abstract description 43
- 238000010276 construction Methods 0.000 claims abstract description 4
- 239000011159 matrix material Substances 0.000 claims description 63
- 230000007246 mechanism Effects 0.000 claims description 52
- 239000010410 layer Substances 0.000 claims description 39
- 238000013527 convolutional neural network Methods 0.000 claims description 32
- 230000004927 fusion Effects 0.000 claims description 32
- 238000010606 normalization Methods 0.000 claims description 26
- 230000006870 function Effects 0.000 claims description 20
- 230000010339 dilation Effects 0.000 claims description 11
- 230000002159 abnormal effect Effects 0.000 claims description 7
- 230000004913 activation Effects 0.000 claims description 7
- 238000011176 pooling Methods 0.000 claims description 4
- 239000002356 single layer Substances 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 8
- 238000004590 computer program Methods 0.000 description 7
- 238000010200 validation analysis Methods 0.000 description 6
- 238000012545 processing Methods 0.000 description 4
- 238000012360 testing method Methods 0.000 description 4
- 238000004422 calculation algorithm Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 101001121408 Homo sapiens L-amino-acid oxidase Proteins 0.000 description 1
- 102100026388 L-amino-acid oxidase Human genes 0.000 description 1
- 101100012902 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) FIG2 gene Proteins 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000002950 deficient Effects 0.000 description 1
- 238000007429 general method Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000010845 search algorithm Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
Landscapes
- Testing And Monitoring For Control Systems (AREA)
Abstract
本发明公开了一种基于图神经网络的多变量时序异常检测方法及系统,方法包括:获取待检测的时序输入数据,将时序输入数据输入至预设的时序异常检测模型进行时序异常判断;所述时序异常检测模型的构建过程包括:获取样本集合后利用样本集合对时序异常检测模型进行训练;将单位样本x转化为同时融合了时域与空域关系的融合时间序列;将融合时间序列输入至编码器输出中间特征;将融合时间序列和中间特征输入至解码器获取训练输出序列,基于训练输出序列计算训练损失值并对时序异常检测模型的参数进行优化,输出训练后的时序异常检测模型;本发明能够更全面地分析多变量时间序列数据,从而提高异常检测的准确性和可靠性。
The present invention discloses a multivariate time series anomaly detection method and system based on graph neural network. The method comprises: obtaining time series input data to be detected, inputting the time series input data into a preset time series anomaly detection model to judge the time series anomaly; the construction process of the time series anomaly detection model comprises: obtaining a sample set and then training the time series anomaly detection model with the sample set; converting the unit sample x into a fused time series that simultaneously fuses the relationship between the time domain and the spatial domain ; The fused time series Input to encoder output intermediate features ; The fused time series and intermediate features The training output sequence is input to the decoder, the training loss value is calculated based on the training output sequence, the parameters of the time series anomaly detection model are optimized, and the trained time series anomaly detection model is output; the present invention can more comprehensively analyze multivariate time series data, thereby improving the accuracy and reliability of anomaly detection.
Description
技术领域Technical Field
本发明属于数据时序异常检测领域,具体涉及基于图神经网络的多变量时序异常检测方法及系统。The present invention belongs to the field of data time series anomaly detection, and specifically relates to a multivariate time series anomaly detection method and system based on graph neural network.
背景技术Background Art
多变量时序异常检测在许多领域如金融、工业制造和网络安全等方面具有广泛的应用,其主要目标是识别时间序列数据中的异常模式,以便及早发现潜在问题并采取相应的措施。传统的异常检测方法通常基于统计学或规则,但在处理复杂的多维时序数据时,这些方法往往面临限制。Multivariate time series anomaly detection has wide applications in many fields such as finance, industrial manufacturing, and network security. Its main goal is to identify abnormal patterns in time series data so that potential problems can be discovered early and appropriate measures can be taken. Traditional anomaly detection methods are usually based on statistics or rules, but these methods often face limitations when dealing with complex multidimensional time series data.
近年来,基于深度学习的方法在多变量时序异常检测领域取得了显著的进展。特别是,结合了图神经网络和编码器-解码器结构的算法,提供了更强大的时空建模能力。图神经网络能够捕捉数据中的复杂空间关系,同时编码器-解码器结构能够有效地学习数据的时空表示,并用于重构和异常检测。In recent years, deep learning-based methods have made significant progress in the field of multivariate time series anomaly detection. In particular, algorithms that combine graph neural networks and encoder-decoder structures provide more powerful spatiotemporal modeling capabilities. Graph neural networks can capture complex spatial relationships in data, while encoder-decoder structures can effectively learn spatiotemporal representations of data for reconstruction and anomaly detection.
这一新兴技术背后的关键思想是将多维时序数据视为图结构,其中节点表示不同的特征或维度,边表示它们之间的空间关系。通过图神经网络,可以有效地对空间关系进行学习,从而改善异常检测的准确性和鲁棒性。此外,编码器-解码器结构有助于捕获数据的内在表示,使其成为一种适用于各种应用场景的通用方法。The key idea behind this emerging technology is to view multidimensional time series data as a graph structure, where nodes represent different features or dimensions and edges represent the spatial relationships between them. Through graph neural networks, spatial relationships can be effectively learned, thereby improving the accuracy and robustness of anomaly detection. In addition, the encoder-decoder structure helps capture the intrinsic representation of the data, making it a general method suitable for a variety of application scenarios.
基于图神经网络和编码器解码器结构的多变量时序异常检测算法代表了先进的技术趋势,但是现有技术依然存在的两个问题:一是时间序列重构过程难以同时考虑时间序列的空间关系和时域关系;二是难以准确定位瑕疵数据的异常点。The multivariate time series anomaly detection algorithm based on graph neural network and encoder-decoder structure represents an advanced technological trend, but there are still two problems with the existing technology: first, it is difficult to simultaneously consider the spatial relationship and time domain relationship of the time series during the time series reconstruction process; second, it is difficult to accurately locate the abnormal points of defective data.
发明内容Summary of the invention
本发明提供了一种基于图神经网络的多变量时序异常检测方法及系统,能够更全面地分析多变量时间序列数据,从而提高异常检测的准确性和可靠性。The present invention provides a multivariate time series anomaly detection method and system based on graph neural network, which can more comprehensively analyze multivariate time series data, thereby improving the accuracy and reliability of anomaly detection.
为达到上述目的,本发明所采用的技术方案是:In order to achieve the above object, the technical solution adopted by the present invention is:
本发明第一方面提供了一种基于图神经网络的多变量时序异常检测方法,包括:The first aspect of the present invention provides a multivariate time series anomaly detection method based on a graph neural network, comprising:
获取待检测的时序输入数据,将时序输入数据输入至预设的时序异常检测模型获得检测输出序列,基于检测输出序列和时序输入数据计算时序异常分数,通过比较时序异常分数与预设的时序异常判断阈值输出判断结果;Obtain the time series input data to be detected, input the time series input data into a preset time series anomaly detection model to obtain a detection output sequence, calculate the time series anomaly score based on the detection output sequence and the time series input data, and output the judgment result by comparing the time series anomaly score with a preset time series anomaly judgment threshold;
所述时序异常检测模型的构建过程包括:The construction process of the time series anomaly detection model includes:
获取样本集合后利用样本集合对时序异常检测模型进行训练,所述时序异常检测模型包括时序重构模块、编码器、解码器;After obtaining the sample set, the time series anomaly detection model is trained using the sample set, wherein the time series anomaly detection model includes a time series reconstruction module, an encoder, and a decoder;
基于样本集合中单位样本构建图结构G(V,E);将单位样本输入至时域扩张卷积获得时间序列;将时间序列和图结构G(V,E)同时输入至图注意力影响网络,获得融合时间序列;Construct a graph structure G(V,E) based on the unit samples in the sample set; input the unit samples into the time domain dilation convolution to obtain the time series ; The time series And the graph structure G(V,E) are simultaneously input into the graph attention influence network to obtain the fused time series ;
将融合时间序列输入至编码器输出中间特征;将融合时间序列和中间特征输入至解码器获取训练输出序列,基于训练输出序列计算训练损失值,根据训练损失值对时序异常检测模型的参数进行优化;重复迭代时序异常检测模型的训练过程直至损失值收敛,设定时序异常判断阈值并输出训练后的时序异常检测模型。Fusion of time series Input to encoder output intermediate features ; The fused time series and intermediate features Input into the decoder to obtain the training output sequence, calculate the training loss value based on the training output sequence, and optimize the parameters of the timing anomaly detection model according to the training loss value; repeat the training process of the timing anomaly detection model until the loss value converges, set the timing anomaly judgment threshold and output the trained timing anomaly detection model.
优选的,基于样本集合中单位样本构建图结构G(V,E)的过程包括:Preferably, the process of constructing the graph structure G(V, E) based on the unit samples in the sample set includes:
以单位样本为输入数据,构建节点信息,其中节点,表示为可学习的权重矩阵,表示为时间序列的单位属性特征表示,T为序列长度,N为特征数;Unit Sample For input data, construct node information , where the node , Represented as a learnable weight matrix, It is expressed as a unit attribute feature representation of a time series, where T is the length of the sequence and N is the number of features;
构建节点和节点之间的边信息;其中,连接边,,\为节点编号去除操作;Build Node and nodes The side information between ; Among them, the connecting edge , , \ is the node number removal operation;
根据边信息选择每个节点相关性最高的K个连接边,基于节点信息V和边信息E构建图结构G(V, E)。According to the edge information, the K connecting edges with the highest relevance for each node are selected, and the graph structure G(V, E) is constructed based on the node information V and the edge information E.
优选的,将单位样本输入至时域扩张卷积获得时间序列的过程包括:Preferably, the unit sample is input into the time domain dilation convolution to obtain the time series The process includes:
将单位样本x拆分为N组长度为T的一维序列后输入至时域扩张卷积,其中,对应的卷积核为;膨胀因子为df;感受野为;k表示为卷积核层数,用以抽象出更深层次的时域特征表示;Split the unit sample x into N groups of one-dimensional sequences of length T Then input to the time domain dilation convolution, where The corresponding convolution kernel is ; the expansion factor is df ; the receptive field is ; k represents the number of convolution kernel layers, which is used to abstract deeper time domain feature representation;
最后,在序列位置t、卷积核长度为K、特征通道n上施加空洞卷积操作得到时间序列,定义为。Finally, a dilated convolution operation is applied to the sequence position t, the convolution kernel length K, and the feature channel n to obtain the time series , defined as .
优选的,将时间序列和图结构G(V,E)输入至图注意力影响网络获得融合时间序列的过程包括:Preferably, the time series And the graph structure G(V,E) is input into the graph attention influence network to obtain the fusion time series The process includes:
对时间序列按时域维度抽取子序列后再按属性维度分割成;For time series Extract subsequences according to the time domain dimension Then split it into ;
计算各连接边的权重,其中:表示为可学习的权重矩阵,为单层全连接层,基于各连接边的权重计算注意力系数,表达公式为:Calculate the weight of each connecting edge ,in: Represented as a learnable weight matrix, It is a single-layer fully connected layer, based on the weight of each connection edge Calculate the attention coefficient, the expression formula is:
; ;
其中:为激活函数,表示矩阵连接,为权重向量;表示为目标节点的邻接节点集合;in: is the activation function, represents matrix connection, is the weight vector; Represented as the target node The set of adjacent nodes of ;
根据注意力系数对目标节点进行更新获得新节点,表达公式为:According to the attention coefficient, the target node Update to get new nodes , the expression formula is:
; ;
其中,表示多层感知机;表示为注意力系数;通过时空关系学习获得新节点信息;使经过了时域学习后的时间序列处理为经过了空域关系学习后的时间序列。in, represents a multi-layer perceptron; Expressed as attention coefficient; new node information is obtained through spatiotemporal relationship learning ; Make the time series after time domain learning Processed into a time series after spatial domain relationship learning .
优选的,将融合时间序列输入至编码器输出中间特征的过程包括:Preferably, the fused time series Input to encoder output intermediate features The process includes:
将融合时间序列输入至编码器,所述编码器依次包括多分支注意力机制、层归一化B1、前馈神经网络和层归一化B2;Fusion of time series Input to the encoder, which includes a multi-branch attention mechanism, layer normalization B1, a feedforward neural network and layer normalization B2 in sequence;
将融合时间序列输入至多分支注意力机制获得特征序列output;将特征序列output输入至层归一化B1获得时间序列,表达公式为:Fusion of time series Input the multi-branch attention mechanism to obtain the feature sequence output; input the feature sequence output to the layer normalization B1 to obtain the time series , the expression formula is:
; ;
将时间序列输入至前馈神经网络获得时间序列,表达公式为:The time series Input to the feedforward neural network to obtain the time series , the expression formula is:
; ;
将时间序列输入至层归一化B2获得中间特征,表达公式为:The time series Input to layer normalization B2 to obtain intermediate features , the expression formula is:
; ;
公式中,表示为归一化函数;表示为前馈神经网络。In the formula, Expressed as a normalized function; Represented as a feed-forward neural network.
优选的,将融合时间序列输入至多分支注意力机制获得特征序列output的过程包括:Preferably, the fused time series The process of inputting into the multi-branch attention mechanism to obtain the feature sequence output includes:
所述多分支注意力机制包括Vaswani自注意力机制、密集综合注意力机制和动态卷积神经网络;The multi-branch attention mechanism includes Vaswani self-attention mechanism, dense integrated attention mechanism and dynamic convolutional neural network;
将融合时间序列输入至Vaswani自注意力机制获得特征序列,表达公式为:Fusion of time series Input to Vaswani self-attention mechanism to obtain feature sequence , the expression formula is:
; ;
公式中,为Vaswani Self Attention机制;WQ表示为权重矩阵,用以将输入序列线性映射至查询矩阵;WK表示为权重矩阵,用以将输入序列线性映射至键矩阵;WV表示为权重矩阵,用以将输入序列线性映射至值矩阵;d att 为注意力机制输出序列的特征维度;In the formula, is the Vaswani Self Attention mechanism; W Q is the weight matrix used to linearly map the input sequence to the query matrix; W K is the weight matrix used to linearly map the input sequence to the key matrix; W V is the weight matrix used to linearly map the input sequence to the value matrix; d att is the feature dimension of the output sequence of the attention mechanism;
将融合时间序列输入至密集综合注意力机制获得特征序列output2,表达公式为:Fusion of time series Input to the dense comprehensive attention mechanism to obtain the feature sequence output 2 , expressed as:
; ;
; ;
公式中,W1表示为可学习权重矩阵;W2表示为可学习权重矩阵;b1表示为可学习偏差参数;b2表示为可学习偏差参数;In the formula, W1 represents the learnable weight matrix; W2 represents the learnable weight matrix; b1 represents the learnable bias parameter; b2 represents the learnable bias parameter;
将融合时间序列输入至动态卷积神经网络获得特征序列output3,表达公式为:Fusion of time series Input into the dynamic convolutional neural network to obtain the feature sequence output 3 , expressed as:
预设卷积神经网络,利用融合时间序列对卷积神经网络训练学习注意力权重,表达公式为:Preset Convolutional Neural Network , using the fused time series Convolutional Neural Network Training to learn attention weights , the expression formula is:
; ;
其中,AvgPool为平均池化层,FC为全连接层,ReLU为非线性激活函数,Softmax为归一化函数;WAvg表示为可学习权重矩阵;Among them, AvgPool is the average pooling layer, FC is the fully connected layer, ReLU is the nonlinear activation function, and Softmax is the normalization function; W Avg represents the learnable weight matrix;
预设卷积神经网络ConvNet参数,将注意力权重输入至卷积神经网络ConvNet获得特征序列output3,表达公式为:Preset Convolutional Neural Network ConvNet parameters , the attention weight Input to the convolutional neural network ConvNet to obtain the feature sequence output 3 , expressed as:
; ;
; ;
将特征序列output1、特征序列output2和特征序列output3进行加权求和获得特征序列output;表达公式为:The feature sequence output 1 , feature sequence output 2 and feature sequence output 3 are weighted and summed to obtain the feature sequence output; the expression formula is:
; ;
其中,为可学习权重。in, are learnable weights.
优选的,将融合时间序列和中间特征输入至解码器获取训练输出序列的过程包括:Preferably, the fused time series and intermediate features The process of inputting to the decoder to obtain the training output sequence includes:
将融合时间序列输入至解码器的自注意力机制获得中间特征,表达公式为:;Fusion of time series The self-attention mechanism input to the decoder obtains intermediate features , the expression formula is: ;
公式中,为Vaswani Self Attention机制;WQ表示为权重矩阵,用以将输入序列线性映射至查询矩阵;WK表示为权重矩阵,用以将输入序列线性映射至键矩阵;WV表示为权重矩阵,用以将输入序列线性映射至值矩阵;d att 为注意力机制输出序列的特征维度;T为序列长度;In the formula, is the Vaswani Self Attention mechanism; W Q is the weight matrix used to linearly map the input sequence to the query matrix; W K is the weight matrix used to linearly map the input sequence to the key matrix; W V is the weight matrix used to linearly map the input sequence to the value matrix; d att is the feature dimension of the output sequence of the attention mechanism; T is the sequence length;
将中间特征输入至解码器的层归一化D1获得中间特征,表达公式为:The intermediate features The layer normalization D1 input to the decoder obtains the intermediate features , the expression formula is:
; ;
将中间特征和中间特征输入至解码器的交叉注意力机制获得中间特征,表达公式为:The intermediate features and intermediate features The cross attention mechanism input to the decoder obtains intermediate features , the expression formula is:
; ;
将中间特征输入至解码器的层归一化D2获得中间特征,表达公式为:The intermediate features The layer normalization D2 input to the decoder obtains the intermediate features , the expression formula is:
; ;
将中间特征输入至解码器的前馈神经网络FFN获得训练输出序列,表达公式为:The intermediate features The feedforward neural network FFN input to the decoder obtains the training output sequence, which is expressed as:
; ;
公式中,表示为训练输出序列,表示为归一化函数;表示为前馈神经网络。In the formula, Represented as the training output sequence, Expressed as a normalized function; Represented as a feed-forward neural network.
优选的,基于训练输出序列计算训练损失值的过程包括:Preferably, the process of calculating the training loss value based on the training output sequence includes:
; ;
公式中,表示为训练输出序列中重构信号,表示为单位样本中输入信号。In the formula, Represented as the reconstructed signal in the training output sequence, Represents the input signal in unit samples.
优选的,基于检测输出序列和时序输入数据计算时序异常分数,通过比较时序异常分数与预设的时序异常判断阈值输出判断结果的过程包括:Preferably, the process of calculating the time series anomaly score based on the detection output sequence and the time series input data, and outputting the judgment result by comparing the time series anomaly score with a preset time series anomaly judgment threshold comprises:
; ;
; ;
公式中,表示为时序异常分数;表示为时序输入数据;表示为检测输出序列数据,表示为判断结果;anomal表示为数据异常;nomal表示为数据正常。In the formula, It is expressed as a time series anomaly score; Represented as time series input data; Represents the detection output sequence data, Indicates the judgment result; anomal indicates abnormal data; and nomal indicates normal data.
本发明第二方面提供了一种基于图神经网络的多变量时序异常检测系统,包括:A second aspect of the present invention provides a multivariate time series anomaly detection system based on a graph neural network, comprising:
检测单元,用于获取待检测的时序输入数据,将时序输入数据输入至预设的时序异常检测模型获得检测输出序列,基于检测输出序列和时序输入数据计算时序异常分数,通过比较时序异常分数与预设的时序异常判断阈值输出判断结果;A detection unit is used to obtain time series input data to be detected, input the time series input data into a preset time series anomaly detection model to obtain a detection output sequence, calculate a time series anomaly score based on the detection output sequence and the time series input data, and output a judgment result by comparing the time series anomaly score with a preset time series anomaly judgment threshold;
获取单元,获取样本集合后利用样本集合对时序异常检测模型进行训练,所述时序异常检测模型包括时序重构模块、编码器、解码器;基于样本集合中单位样本构建图结构G(V,E);将单位样本输入至时域扩张卷积获得时间序列;将时间序列和图结构G(V,E)同时输入至图注意力影响网络,获得融合时间序列;The acquisition unit acquires the sample set and uses the sample set to train the time series anomaly detection model, wherein the time series anomaly detection model includes a time series reconstruction module, an encoder, and a decoder; constructs a graph structure G(V, E) based on the unit samples in the sample set; inputs the unit samples into the time domain dilation convolution to obtain the time series ; The time series And the graph structure G(V,E) are simultaneously input into the graph attention influence network to obtain the fused time series ;
训练单元,用于将融合时间序列输入至编码器输出中间特征;将融合时间序列和中间特征输入至解码器获取训练输出序列,基于训练输出序列计算训练损失值,根据训练损失值对时序异常检测模型的参数进行优化;重复迭代时序异常检测模型的训练过程直至损失值收敛,设定时序异常判断阈值并输出训练后的时序异常检测模型。Training unit, used to fuse time series Input to encoder output intermediate features ; The fused time series and intermediate features Input into the decoder to obtain the training output sequence, calculate the training loss value based on the training output sequence, and optimize the parameters of the timing anomaly detection model according to the training loss value; repeat the training process of the timing anomaly detection model until the loss value converges, set the timing anomaly judgment threshold and output the trained timing anomaly detection model.
与现有技术相比,本发明的有益效果:Compared with the prior art, the present invention has the following beneficial effects:
本发明基于图注意力影响网络和时域扩张卷积网络处理具有时间和空间关系的多维数据,并采用编码器解码器结构重新构建时间序列,通过比较输入序列与重建序列之间的差异,可以有效判定异常状态;本发明能够更全面地分析多变量时间序列数据,从而提高异常检测的准确性和可靠性。The present invention processes multidimensional data with time and space relationships based on a graph attention influence network and a time-domain dilated convolutional network, and reconstructs the time series using an encoder-decoder structure. By comparing the difference between the input sequence and the reconstructed sequence, the abnormal state can be effectively determined; the present invention can more comprehensively analyze multivariate time series data, thereby improving the accuracy and reliability of anomaly detection.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1是本实施例1提供了一种基于图神经网络的多变量时序异常检测方法的流程图;FIG1 is a flow chart of a multivariate time series anomaly detection method based on a graph neural network provided in Embodiment 1;
图2是本实施例1提供了一种基于图神经网络的多变量时序异常检测方法的模型图;FIG2 is a model diagram of a multivariate time series anomaly detection method based on a graph neural network provided in Embodiment 1;
图3是本实施例1提供了时空关系学习模块图。FIG. 3 is a diagram of a spatiotemporal relationship learning module provided in the first embodiment.
具体实施方式DETAILED DESCRIPTION
下面结合附图对本发明作进一步描述。以下实施例仅用于更加清楚地说明本发明的技术方案,而不能以此来限制本发明的保护范围。The present invention will be further described below in conjunction with the accompanying drawings. The following embodiments are only used to more clearly illustrate the technical solution of the present invention, and cannot be used to limit the protection scope of the present invention.
实施例1Example 1
如图1至图3所示,本实施提供了一种基于图神经网络的多变量时序异常检测方法,包括:As shown in Figures 1 to 3, this implementation provides a multivariate time series anomaly detection method based on graph neural network, including:
获取待检测的时序输入数据,将时序输入数据输入至预设的时序异常检测模型获得检测输出序列,基于检测输出序列和时序输入数据计算时序异常分数,通过比较时序异常分数与预设的时序异常判断阈值输出判断结果的过程包括:The process of obtaining time series input data to be detected, inputting the time series input data into a preset time series anomaly detection model to obtain a detection output sequence, calculating a time series anomaly score based on the detection output sequence and the time series input data, and outputting a judgment result by comparing the time series anomaly score with a preset time series anomaly judgment threshold includes:
; ;
; ;
公式中,表示为时序异常分数;表示为时序输入数据;表示为检测输出序列数据,表示为判断结果;anomal表示为数据异常;nomal表示为数据正常。In the formula, It is expressed as a time series anomaly score; Represented as time series input data; Represents the detection output sequence data, Indicates the judgment result; anomal indicates abnormal data; and nomal indicates normal data.
所述时序异常检测模型的构建过程包括:The construction process of the time series anomaly detection model includes:
获取样本集合样本集划分为“训练集”、“验证集”和“测试集”。其中,“训练集”包含无瑕疵数据和无标签样本,用于训练预测模型;“验证集”为有标签样本,其中包含瑕疵数据和无瑕疵数据,用于寻找合理的阈值以及调试异常评估器;“测试集”为有标签样本,其中包含瑕疵数据和无瑕疵数据,用于测试预测模型和异常评估器的泛化性能。Get the sample set. The sample set is divided into "training set", "validation set" and "test set". Among them, the "training set" contains flawless data and unlabeled samples, which are used to train the prediction model; the "validation set" is a labeled sample, which contains flawed data and flawless data, which is used to find a reasonable threshold and debug the anomaly evaluator; the "test set" is a labeled sample, which contains flawed data and flawless data, which is used to test the generalization performance of the prediction model and anomaly evaluator.
此外,定义训练集中的单位样本为,其中,表示为单位样本的时序数据,,N为特征数,T为序列长度,t为当前样本起始点;验证集和测试集中的有标签单位样本定义为,表示为样本标签,。In addition, the unit samples in the training set are defined as ,in, Represented as time series data of unit samples, , N is the number of features, T is the sequence length, and t is the starting point of the current sample; the labeled unit samples in the validation set and the test set are defined as , Represented as sample label, .
利用样本集合对时序异常检测模型进行训练,所述时序异常检测模型包括编码器和解码器;Using the sample set to train a time series anomaly detection model, the time series anomaly detection model includes an encoder and a decoder;
基于样本集合中单位样本构建图结构G(V,E)的过程包括:The process of constructing the graph structure G(V,E) based on the unit samples in the sample set includes:
以单位样本为输入数据,构建节点信息,其中节点,表示为可学习的权重矩阵,表示为时间序列的单位属性特征表示,Unit Sample For input data, construct node information , where the node , Represented as a learnable weight matrix, Represented as a unit attribute feature representation of a time series,
构建节点和节点之间的边信息;其中,连接边,,\为节点编号去除操作;Build Node and nodes The side information between ; Among them, the connecting edge , , \ is the node number removal operation;
根据边信息选择每个节点相关性最高的K个连接边,基于节点信息V和边信息E构建图结构G(V, E)。According to the edge information, the K connecting edges with the highest relevance for each node are selected, and the graph structure G(V, E) is constructed based on the node information V and the edge information E.
将单位样本输入至时域扩张卷积获得时间序列的过程包括:Input the unit sample into the time domain dilation convolution to obtain the time series The process includes:
将单位样本x拆分为N组长度为T的一维序列后输入至时域扩张卷积,其中,对应的卷积核为;膨胀因子为df;感受野为;k表示为卷积核层数,用以抽象出更深层次的时域特征表示;Split the unit sample x into N groups of one-dimensional sequences of length T Then input to the time domain dilation convolution, where The corresponding convolution kernel is ; the expansion factor is df ; the receptive field is ; k represents the number of convolution kernel layers, which is used to abstract deeper time domain feature representation;
最后,在序列位置t、卷积核长度为K、特征通道n上施加空洞卷积操作得到时间序列,定义为。Finally, a dilated convolution operation is applied to the sequence position t, the convolution kernel length K, and the feature channel n to obtain the time series , defined as .
将时间序列和图结构G(V,E)输入至图注意力影响网络获得融合时间序列的过程包括:The time series And the graph structure G(V,E) is input into the graph attention influence network to obtain the fusion time series The process includes:
对时间序列按时域维度抽取子序列后再按属性维度分割成;For time series Extract subsequences according to the time domain dimension Then split it into ;
计算各连接边的权重,其中:表示为可学习的权重矩阵,为单层全连接层,基于各连接边的权重计算注意力系数,表达公式为:Calculate the weight of each connecting edge ,in: Represented as a learnable weight matrix, It is a single-layer fully connected layer, based on the weight of each connection edge Calculate the attention coefficient, the expression formula is:
; ;
其中:为激活函数,表示矩阵连接,为权重向量;表示为目标节点的邻接节点集合;in: is the activation function, represents matrix connection, is the weight vector; Represented as the target node The set of adjacent nodes of ;
根据注意力系数对目标节点进行更新获得新节点,表达公式为:According to the attention coefficient, the target node Update to get new nodes , the expression formula is:
; ;
其中,表示多层感知机;表示为注意力系数;通过时空关系学习获得新节点信息;使经过了时域学习后的时间序列处理为经过了空域关系学习后的时间序列;所述采用图注意力影响网络,通过计算边权重,灵活分配目标节点与源节点之间的注意力系数,并基于筛选后的邻接节点和注意力系数,更新目标节点。in, represents a multi-layer perceptron; Expressed as attention coefficient; new node information is obtained through spatiotemporal relationship learning ; Make the time series after time domain learning Processed into a time series after spatial domain relationship learning The graph attention influence network is adopted to flexibly allocate the attention coefficient between the target node and the source node by calculating the edge weight, and update the target node based on the screened adjacent nodes and attention coefficients.
将融合时间序列输入至编码器输出中间特征的过程包括:Fusion of time series Input to encoder output intermediate features The process includes:
将融合时间序列输入至编码器,所述编码器依次包括多分支注意力机制、层归一化B1、前馈神经网络和层归一化B2;Fusion of time series Input to the encoder, which includes a multi-branch attention mechanism, layer normalization B1, a feedforward neural network and layer normalization B2 in sequence;
将融合时间序列输入至多分支注意力机制获得特征序列output的过程包括:Fusion of time series The process of inputting into the multi-branch attention mechanism to obtain the feature sequence output includes:
所述多分支注意力机制包括Vaswani自注意力机制、密集综合注意力机制和动态卷积神经网络;The multi-branch attention mechanism includes Vaswani self-attention mechanism, dense integrated attention mechanism and dynamic convolutional neural network;
将融合时间序列输入至Vaswani自注意力机制获得特征序列,表达公式为:Fusion of time series Input to Vaswani self-attention mechanism to obtain feature sequence , the expression formula is:
; ;
公式中,为Vaswani Self Attention机制;WQ表示为权重矩阵,用以将输入序列线性映射至查询矩阵;WK表示为权重矩阵,用以将输入序列线性映射至键矩阵;WV表示为权重矩阵,用以将输入序列线性映射至值矩阵;d att 为注意力机制输出序列的特征维度;In the formula, is the Vaswani Self Attention mechanism; W Q is the weight matrix used to linearly map the input sequence to the query matrix; W K is the weight matrix used to linearly map the input sequence to the key matrix; W V is the weight matrix used to linearly map the input sequence to the value matrix; d att is the feature dimension of the output sequence of the attention mechanism;
将融合时间序列输入至密集综合注意力机制获得特征序列output2,表达公式为:Fusion of time series Input to the dense comprehensive attention mechanism to obtain the feature sequence output 2 , expressed as:
; ;
; ;
公式中,W1表示为可学习权重矩阵;W2表示为可学习权重矩阵;b1表示为可学习偏差参数;b2表示为可学习偏差参数;In the formula, W1 represents the learnable weight matrix; W2 represents the learnable weight matrix; b1 represents the learnable bias parameter; b2 represents the learnable bias parameter;
将融合时间序列输入至动态卷积神经网络获得特征序列output3,表达公式为:Fusion of time series Input into the dynamic convolutional neural network to obtain the feature sequence output 3 , expressed as:
预设卷积神经网络,利用融合时间序列对卷积神经网络训练学习注意力权重,表达公式为:Preset Convolutional Neural Network , using the fused time series Convolutional Neural Network Training to learn attention weights , the expression formula is:
; ;
其中,AvgPool为平均池化层,FC为全连接层,ReLU为非线性激活函数,Softmax为归一化函数;WAvg表示为可学习权重矩阵;Among them, AvgPool is the average pooling layer, FC is the fully connected layer, ReLU is the nonlinear activation function, and Softmax is the normalization function; W Avg represents the learnable weight matrix;
预设卷积神经网络ConvNet参数,将注意力权重输入至卷积神经网络ConvNet获得特征序列output3,表达公式为:Preset Convolutional Neural Network ConvNet parameters , the attention weight Input to the convolutional neural network ConvNet to obtain the feature sequence output 3 , expressed as:
; ;
; ;
将特征序列output1、特征序列output2和特征序列output3进行加权求和获得特征序列output;表达公式为:The feature sequence output 1 , feature sequence output 2 and feature sequence output 3 are weighted and summed to obtain the feature sequence output; the expression formula is:
;其中,为可学习权重。 ;in, are learnable weights.
所述采用多分支注意力机制将“Vaswani注意力机制”、“动态卷积网络”和“密集综合注意力”整合,通过可学习权重将这三种注意力合并为整体,以使各变种注意力机制发挥各自优势。The multi-branch attention mechanism is used to integrate the "Vaswani attention mechanism", "dynamic convolutional network" and "dense integrated attention", and these three types of attention are merged into a whole through learnable weights, so that each variant of the attention mechanism can play its own advantages.
将特征序列output输入至层归一化B1获得时间序列,表达公式为:Input the feature sequence output to the layer normalization B1 to obtain the time series , the expression formula is:
; ;
将时间序列输入至前馈神经网络获得时间序列,表达公式为:The time series Input to the feedforward neural network to obtain the time series , the expression formula is:
; ;
将时间序列输入至层归一化B2获得中间特征,表达公式为:The time series Input to layer normalization B2 to obtain intermediate features , the expression formula is:
; ;
公式中,表示为归一化函数;表示为前馈神经网络。In the formula, Expressed as a normalized function; Represented as a feed-forward neural network.
将融合时间序列和中间特征输入至解码器获取训练输出序列的过程包括:Fusion of time series and intermediate features The process of inputting to the decoder to obtain the training output sequence includes:
将融合时间序列输入至解码器的自注意力机制获得中间特征,表达公式为:Fusion of time series The self-attention mechanism input to the decoder obtains intermediate features , the expression formula is:
; ;
公式中,为Vaswani Self Attention机制;WQ表示为权重矩阵,用以将输入序列线性映射至查询矩阵;WK表示为权重矩阵,用以将输入序列线性映射至键矩阵;WV表示为权重矩阵,用以将输入序列线性映射至值矩阵;d att 为注意力机制输出序列的特征维度;In the formula, is the Vaswani Self Attention mechanism; W Q is the weight matrix used to linearly map the input sequence to the query matrix; W K is the weight matrix used to linearly map the input sequence to the key matrix; W V is the weight matrix used to linearly map the input sequence to the value matrix; d att is the feature dimension of the output sequence of the attention mechanism;
将中间特征输入至解码器的层归一化D1获得中间特征,表达公式为:The intermediate features The layer normalization D1 input to the decoder obtains the intermediate features , the expression formula is:
; ;
将中间特征和中间特征输入至解码器的交叉注意力机制获得中间特征,表达公式为:The intermediate features and intermediate features The cross attention mechanism input to the decoder obtains intermediate features , the expression formula is:
; ;
将中间特征输入至解码器的层归一化D2获得中间特征,表达公式为:The intermediate features The layer normalization D2 input to the decoder obtains the intermediate features , the expression formula is:
; ;
将中间特征输入至解码器的前馈神经网络FFN获得训练输出序列,表达公式为:The intermediate features The feedforward neural network FFN input to the decoder obtains the training output sequence, which is expressed as:
; ;
公式中,表示为训练输出序列。In the formula, is represented as the training output sequence.
基于训练输出序列计算训练损失值,表达公式为:The training loss value is calculated based on the training output sequence, and the expression formula is:
; ;
公式中,表示为训练输出序列中重构信号,表示为单位样本中输入信号。In the formula, Represented as the reconstructed signal in the training output sequence, Represents the input signal in unit samples.
根据训练损失值对时序异常检测模型的参数进行优化;重复迭代时序异常检测模型的训练过程直至损失值收敛。The parameters of the time series anomaly detection model are optimized according to the training loss value; the training process of the time series anomaly detection model is iterated repeatedly until the loss value converges.
基于验证集和网格搜索算法来调整时序异常判断阈值,为了对每个特征维度(共N个)进行时序异常判断阈值调整,首先,对N个不同维度的特征计算标准差,第n维特征的标准差计算公式:Based on the validation set and grid search algorithm, the time series anomaly judgment threshold is adjusted. In order to adjust the time series anomaly judgment threshold for each feature dimension (N in total), first, the standard deviation of the features of N different dimensions is calculated. The standard deviation calculation formula of the n-th dimension feature is:
; ;
其中,,表示为验证集中单位样本数量;表示为验证集中的单位样本in, , It is expressed as the number of unit samples in the validation set; Represented as a unit sample in the validation set
N个标准差共享可变参数,两者相乘作为N个特征维度的时序异常判断阈值,表达公式为:N standard deviations share a variable parameter , the two are multiplied as the time series anomaly judgment threshold of N feature dimensions, and the expression formula is:
; ;
本实施例基于图注意力影响网络和时域扩张卷积网络处理具有时间和空间关系的多维数据,并采用编码器解码器结构重新构建时间序列,通过比较输入序列与重建序列之间的差异,可以有效判定异常状态,能够更全面地分析多变量时间序列数据,从而提高异常检测的准确性和可靠性。This embodiment processes multidimensional data with temporal and spatial relationships based on a graph attention influence network and a time-domain dilated convolutional network, and reconstructs the time series using an encoder-decoder structure. By comparing the difference between the input sequence and the reconstructed sequence, the abnormal state can be effectively determined, and multivariate time series data can be analyzed more comprehensively, thereby improving the accuracy and reliability of anomaly detection.
实施例2Example 2
本实施例提供了一种基于图神经网络的多变量时序异常检测系统,所述多变量时序异常检测系统可以应用于实施例1所述的多变量时序异常检测方法,多变量时序异常检测系统包括:This embodiment provides a multivariate time series anomaly detection system based on a graph neural network. The multivariate time series anomaly detection system can be applied to the multivariate time series anomaly detection method described in Example 1. The multivariate time series anomaly detection system includes:
检测单元,用于获取待检测的时序输入数据,将时序输入数据输入至预设的时序异常检测模型获得检测输出序列,基于检测输出序列和时序输入数据计算时序异常分数,通过比较时序异常分数与预设的时序异常判断阈值输出判断结果;A detection unit is used to obtain time series input data to be detected, input the time series input data into a preset time series anomaly detection model to obtain a detection output sequence, calculate a time series anomaly score based on the detection output sequence and the time series input data, and output a judgment result by comparing the time series anomaly score with a preset time series anomaly judgment threshold;
获取单元,获取样本集合后利用样本集合对时序异常检测模型进行训练,所述时序异常检测模型包括时序重构模块、编码器、解码器;基于样本集合中单位样本构建图结构G(V,E);将单位样本输入至时域扩张卷积获得时间序列;将时间序列和图结构G(V,E)同时输入至图注意力影响网络,获得融合时间序列;The acquisition unit acquires the sample set and uses the sample set to train the time series anomaly detection model, wherein the time series anomaly detection model includes a time series reconstruction module, an encoder, and a decoder; constructs a graph structure G(V, E) based on the unit samples in the sample set; inputs the unit samples into the time domain dilation convolution to obtain the time series ; The time series And the graph structure G(V,E) are simultaneously input into the graph attention influence network to obtain the fused time series ;
训练单元,用于将融合时间序列输入至编码器输出中间特征;将融合时间序列和中间特征输入至解码器获取训练输出序列,基于训练输出序列计算训练损失值,根据训练损失值对时序异常检测模型的参数进行优化;重复迭代时序异常检测模型的训练过程直至损失值收敛,设定时序异常判断阈值并输出训练后的时序异常检测模型。Training unit, used to fuse time series Input to encoder output intermediate features ; The fused time series and intermediate features Input into the decoder to obtain the training output sequence, calculate the training loss value based on the training output sequence, and optimize the parameters of the timing anomaly detection model according to the training loss value; repeat the training process of the timing anomaly detection model until the loss value converges, set the timing anomaly judgment threshold and output the trained timing anomaly detection model.
本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。Those skilled in the art will appreciate that the embodiments of the present application may be provided as methods, systems, or computer program products. Therefore, the present application may adopt the form of a complete hardware embodiment, a complete software embodiment, or an embodiment combining software and hardware. Moreover, the present application may adopt the form of a computer program product implemented on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) containing computer-usable program codes.
本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present application is described with reference to the flowcharts and/or block diagrams of the methods, devices (systems), and computer program products according to the embodiments of the present application. It should be understood that each process and/or box in the flowchart and/or block diagram, as well as the combination of the processes and/or boxes in the flowchart and/or block diagram, can be implemented by computer program instructions. These computer program instructions can be provided to a processor of a general-purpose computer, a special-purpose computer, an embedded processor, or other programmable data processing device to generate a machine, so that the instructions executed by the processor of the computer or other programmable data processing device generate a device for implementing the functions specified in one process or multiple processes in the flowchart and/or one box or multiple boxes in the block diagram.
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing device to operate in a specific manner, so that the instructions stored in the computer-readable memory produce a manufactured product including an instruction device that implements the functions specified in one or more processes in the flowchart and/or one or more boxes in the block diagram.
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions may also be loaded onto a computer or other programmable data processing device so that a series of operational steps are executed on the computer or other programmable device to produce a computer-implemented process, whereby the instructions executed on the computer or other programmable device provide steps for implementing the functions specified in one or more processes in the flowchart and/or one or more boxes in the block diagram.
以上所述仅是本发明的优选实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本发明技术原理的前提下,还可以做出若干改进和变形,这些改进和变形也应视为本发明的保护范围。The above is only a preferred embodiment of the present invention. It should be pointed out that for ordinary technicians in this technical field, several improvements and modifications can be made without departing from the technical principles of the present invention. These improvements and modifications should also be regarded as the scope of protection of the present invention.
Claims (8)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410221484.6A CN117786374B (en) | 2024-02-28 | 2024-02-28 | Multivariate time sequence anomaly detection method and system based on graph neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410221484.6A CN117786374B (en) | 2024-02-28 | 2024-02-28 | Multivariate time sequence anomaly detection method and system based on graph neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN117786374A CN117786374A (en) | 2024-03-29 |
CN117786374B true CN117786374B (en) | 2024-05-14 |
Family
ID=90385777
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202410221484.6A Active CN117786374B (en) | 2024-02-28 | 2024-02-28 | Multivariate time sequence anomaly detection method and system based on graph neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117786374B (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118428420B (en) * | 2024-05-15 | 2025-02-07 | 太原理工大学 | Long-term time series prediction method based on front-back symmetric decomposition of ESAttention |
CN119292867B (en) * | 2024-09-27 | 2025-04-08 | 山东省人工智能研究院 | Cloud system anomaly detection method for graph comparison learning |
CN118963774B (en) * | 2024-10-18 | 2025-03-14 | 浙江大学 | Scheduling method based on submodel for AI compiler |
CN119357707A (en) * | 2024-12-27 | 2025-01-24 | 湖南大学 | Time series similarity measurement method and terminal based on twin neural network |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114065862A (en) * | 2021-11-18 | 2022-02-18 | 南京航空航天大学 | Anomaly detection method and system for multidimensional time series data |
CN115587335A (en) * | 2022-09-15 | 2023-01-10 | 中国电力科学研究院有限公司 | Outlier detection model training method, outlier detection method and system |
CN116776269A (en) * | 2023-04-23 | 2023-09-19 | 成都信息工程大学 | Traffic anomaly detection method based on graph convolution neural network self-encoder |
CN116796272A (en) * | 2023-06-21 | 2023-09-22 | 复旦大学 | A Transformer-based multivariate time series anomaly detection method |
CN117076936A (en) * | 2023-10-16 | 2023-11-17 | 北京理工大学 | Time sequence data anomaly detection method based on multi-head attention model |
CN117290800A (en) * | 2023-11-24 | 2023-12-26 | 华东交通大学 | Timing sequence anomaly detection method and system based on hypergraph attention network |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3958185B1 (en) * | 2020-06-26 | 2025-06-18 | Tata Consultancy Services Limited | Neural networks for handling variable-dimensional time series data |
CN114626512B (en) * | 2022-05-17 | 2022-09-06 | 南京信息工程大学 | A high temperature disaster prediction method based on directed graph neural network |
-
2024
- 2024-02-28 CN CN202410221484.6A patent/CN117786374B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114065862A (en) * | 2021-11-18 | 2022-02-18 | 南京航空航天大学 | Anomaly detection method and system for multidimensional time series data |
CN115587335A (en) * | 2022-09-15 | 2023-01-10 | 中国电力科学研究院有限公司 | Outlier detection model training method, outlier detection method and system |
CN116776269A (en) * | 2023-04-23 | 2023-09-19 | 成都信息工程大学 | Traffic anomaly detection method based on graph convolution neural network self-encoder |
CN116796272A (en) * | 2023-06-21 | 2023-09-22 | 复旦大学 | A Transformer-based multivariate time series anomaly detection method |
CN117076936A (en) * | 2023-10-16 | 2023-11-17 | 北京理工大学 | Time sequence data anomaly detection method based on multi-head attention model |
CN117290800A (en) * | 2023-11-24 | 2023-12-26 | 华东交通大学 | Timing sequence anomaly detection method and system based on hypergraph attention network |
Non-Patent Citations (3)
Title |
---|
Multivariate time-series anomaly detection via temporal convolutional and graph attention networks;He Qiang等;《Journal of Intelligent & Fuzzy Systems》;20230403;第44卷(第4期);5953-5962 * |
基于启发式时空图神经网络的多变量时序异常检测;姜羽等;《中国科学:信息科学》;20230919;第53卷(第9期);1784-1801 * |
基于深度学习的时间序列异常检测方法研究;郑育靖;《中国优秀硕士学位论文全文数据库 (基础科学辑)》;20220115(第1期);A002-453 * |
Also Published As
Publication number | Publication date |
---|---|
CN117786374A (en) | 2024-03-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN117786374B (en) | Multivariate time sequence anomaly detection method and system based on graph neural network | |
Hou et al. | Remaining useful life estimation using deep convolutional generative adversarial networks based on an autoencoder scheme | |
Ji et al. | A novel deep learning approach for anomaly detection of time series data | |
CN114386521A (en) | Anomaly detection method, system, device and storage medium for time series data | |
CN112257263B (en) | Equipment residual life prediction system based on self-attention mechanism | |
CN117034099A (en) | A system log anomaly detection method | |
CN118332291B (en) | Multi-sensor data fault prediction method for aircraft | |
CN117216525A (en) | Sparse graph attention soft measurement modeling method based on CNN-LKA | |
Lin et al. | Hierarchical context representation and self-adaptive thresholding for multivariate anomaly detection | |
Liu et al. | An anomaly detection method based on double encoder–decoder generative adversarial networks | |
Li et al. | Multiscale feature extension enhanced deep global–local attention network for remaining useful life prediction | |
CN119024216B (en) | Energy storage battery life prediction method and device based on improved TFT | |
Liao et al. | DPDGAD: A Dual-Process Dynamic Graph-based Anomaly Detection for multivariate time series analysis in cyber-physical systems | |
CN117439800B (en) | A network security situation prediction method, system and device | |
Zhu et al. | Auto-starting semisupervised-learning-based identification of synchrophasor data anomalies | |
JP7392366B2 (en) | Optimal solution acquisition program, optimal solution acquisition method, and information processing device | |
CN118350284A (en) | Space-time feature fusion-based prediction method for residual service life of complex equipment | |
CN118018258A (en) | Industrial Internet of things abnormality detection method based on time-space gating graph attention network | |
CN117421666A (en) | Robust multivariate time sequence anomaly detection model | |
CN117743964A (en) | Method, system and medium for predicting remaining service life of equipment based on graph data | |
CN117785522A (en) | Method and system for performing root cause analysis using a trained machine learning model | |
Wang et al. | Integrating Feature Engineering with Deep Learning to Conduct Diagnostic and Predictive Analytics for Turbofan Engines | |
Liu et al. | Structural-Guided Interaction and Attention-Enhancing Compensation Network for Machine Remaining Useful Life Prediction | |
CN112612260A (en) | Deep self-coding network process monitoring method based on dynamic estimation | |
Wang et al. | Attention-based Representation Learning for Time Series with Principal and Residual Space Monitoring |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |