CN118035889A - Method, device, equipment, storage medium and product for fault classification of rotary equipment - Google Patents

Method, device, equipment, storage medium and product for fault classification of rotary equipment Download PDF

Info

Publication number
CN118035889A
CN118035889A CN202410173763.XA CN202410173763A CN118035889A CN 118035889 A CN118035889 A CN 118035889A CN 202410173763 A CN202410173763 A CN 202410173763A CN 118035889 A CN118035889 A CN 118035889A
Authority
CN
China
Prior art keywords
fault
attention
graph
output information
head self
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202410173763.XA
Other languages
Chinese (zh)
Inventor
闫循石
时振刚
周燕
赵晶晶
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN202410173763.XA priority Critical patent/CN118035889A/en
Publication of CN118035889A publication Critical patent/CN118035889A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/2431Multiple classes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0499Feedforward networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Testing And Monitoring For Control Systems (AREA)

Abstract

The application relates to a fault classification method, a fault classification device, a fault classification storage medium and a fault classification product for rotary equipment. The method comprises the following steps: extracting characteristics of fault information of the rotary equipment to be detected to obtain fault characteristics of the fault information; inputting fault characteristics into a target multi-scale multi-head self-attention force diagram convolutional neural network GCN model to obtain fault types of the rotating equipment to be tested; the target multi-scale multi-head self-attention force-seeking convolutional neural network GCN model is obtained by training an initial multi-scale multi-head self-attention GCN model based on a fusion graph feature sample, wherein the fusion graph feature sample is obtained by fusing target output information and a plurality of graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on a fault information sample, and the graph feature sample is determined by the fault information sample, an adjacent matrix and the initial multi-scale multi-head self-attention GCN model. By adopting the method, the fault detection precision can be improved.

Description

旋转设备的故障分类方法、装置、设备、存储介质和产品Fault classification method, device, equipment, storage medium and product for rotating equipment

技术领域Technical Field

本申请涉及机器学习技术领域,特别是涉及一种旋转设备的故障分类方法、装置、设备、存储介质和产品。The present application relates to the field of machine learning technology, and in particular to a method, apparatus, device, storage medium and product for fault classification of rotating equipment.

背景技术Background technique

常见的旋转设备如电机、发动机等,在工作过程中容易产生故障,因此,有必要对旋转设备进行故障检测。Common rotating equipment such as motors and engines are prone to failure during operation. Therefore, it is necessary to perform fault detection on rotating equipment.

随着机器学习技术的发展,支持向量机、图卷积神经网络等学习模型越来越多地应用于旋转设备的故障检测中。With the development of machine learning technology, learning models such as support vector machines and graph convolutional neural networks are increasingly used in fault detection of rotating equipment.

传统的图卷积神经网络在进行训练时,由于使用单个图网络进行检测,难以从有限的训练样本中提取出足够的信息,因而训练得到的图卷积神经网络模型的故障检测精度较低。When training traditional graph convolutional neural networks, since a single graph network is used for detection, it is difficult to extract sufficient information from limited training samples. Therefore, the fault detection accuracy of the trained graph convolutional neural network model is low.

发明内容Summary of the invention

基于此,有必要针对上述技术问题,提供一种能够提高故障检测精度的旋转设备的故障分类方法、装置、设备、介质和产品。Based on this, it is necessary to provide a fault classification method, device, equipment, medium and product for rotating equipment that can improve the fault detection accuracy in response to the above technical problems.

第一方面,本申请提供了一种旋转设备的故障分类方法。方法包括:In a first aspect, the present application provides a method for fault classification of rotating equipment. The method comprises:

对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征;Extract features of fault information of the rotating equipment to be tested to obtain fault features of the fault information;

将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别;目标多尺度多头自注意力图卷积神经网络GCN模型是基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到的,融合图特征样本为对目标输出信息和多个图特征样本进行融合得到的,目标输出信息为初始多尺度多头自注意力GCN模型基于故障信息样本确定,图特征样本为基于故障信息样本、邻接矩阵和初始多尺度多头自注意力GCN模型确定的。The fault features are input into the target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested; the target multi-scale multi-head self-attention graph convolutional neural network GCN model is trained by the initial multi-scale multi-head self-attention GCN model based on the fused graph feature samples, the fused graph feature samples are obtained by fusing the target output information and multiple graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on the fault information samples, and the graph feature samples are determined based on the fault information samples, the adjacency matrix and the initial multi-scale multi-head self-attention GCN model.

在其中一个实施例中,初始多尺度多头自注意力GCN模型包括多个第一子模型和多层感知器;方法还包括:In one embodiment, the initial multi-scale multi-head self-attention GCN model includes multiple first sub-models and a multi-layer perceptron; the method further includes:

针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型,得到第一子模型输出的图特征样本;For each first sub-model, the fault information sample and the adjacency matrix corresponding to each first sub-model are input into the first sub-model to obtain a graph feature sample output by the first sub-model;

将故障信息样本输入至多层感知器,得到第一输出信息;目标输出信息包括第一输出信息;Inputting the fault information sample into the multilayer perceptron to obtain first output information; the target output information includes the first output information;

根据第一输出信息和各图特征样本得到融合图特征样本;Obtaining a fused graph feature sample according to the first output information and each graph feature sample;

根据融合图特征样本对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型。The initial multi-scale multi-head self-attention GCN model is trained according to the fusion graph feature samples to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

在其中一个实施例中,第一子模型包括第一图卷积层和第二图卷积层,初始多尺度多头自注意力GCN模型包括连接层和线性层;根据第一输出信息和各图特征样本得到融合图特征样本,包括:In one embodiment, the first sub-model includes a first graph convolution layer and a second graph convolution layer, and the initial multi-scale multi-head self-attention GCN model includes a connection layer and a linear layer; and obtaining a fused graph feature sample according to the first output information and each graph feature sample includes:

针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型的第一图卷积层,得到第一中间图特征;For each first sub-model, input the fault information sample and the adjacency matrix corresponding to each first sub-model into the first graph convolution layer of the first sub-model to obtain a first intermediate graph feature;

将各第一子模型的第一图卷积层得到的第一中间图特征输入至连接层,得到第二中间图特征;Inputting the first intermediate graph features obtained by the first graph convolution layer of each first sub-model into the connection layer to obtain the second intermediate graph features;

将第二中间图特征输入至线性层得到第二输出信息;Inputting the second intermediate graph feature into the linear layer to obtain second output information;

根据第一输出信息、第二输出信息和各图特征样本得到融合图特征样本;目标输出信息包括第一输出信息和第二输出信息。A fused graph feature sample is obtained according to the first output information, the second output information and each graph feature sample; the target output information includes the first output information and the second output information.

在其中一个实施例中,根据第一输出信息、第二输出信息和各图特征样本得到融合图特征样本;目标输出信息包括第一输出信息和第二输出信息,包括:In one embodiment, a fused graph feature sample is obtained according to the first output information, the second output information and each graph feature sample; the target output information includes the first output information and the second output information, including:

根据第一输出信息、第二输出信息、各图特征样本、第一输出信息对应的第一权重、第二输出信息对应的第二权重、各图特征样本对应的第三权重,得到第一输出信息、第二输出信息和各图特征样本的加权值;Obtaining weighted values of the first output information, the second output information, and each graph feature sample according to the first output information, the second output information, each graph feature sample, a first weight corresponding to the first output information, a second weight corresponding to the second output information, and a third weight corresponding to each graph feature sample;

基于加权值得到融合图特征样本。The fusion graph feature samples are obtained based on the weighted values.

在其中一个实施例中,融合图特征样本的数量为多个;初始多尺度多头自注意力GCN模型包括多头自注意力融合模型和全连接层;根据融合图特征样本对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型,包括:In one embodiment, the number of fused graph feature samples is multiple; the initial multi-scale multi-head self-attention GCN model includes a multi-head self-attention fusion model and a fully connected layer; the initial multi-scale multi-head self-attention GCN model is trained according to the fused graph feature samples to obtain a target multi-scale multi-head self-attention graph convolutional neural network GCN model, including:

利用多头自注意力融合模型对各融合图特征样本进行连接得到图特征向量;Use the multi-head self-attention fusion model to connect the fusion graph feature samples to obtain the graph feature vector;

利用全连接层基于图特征向量得到预测故障类别;Use the fully connected layer to get the predicted fault category based on the graph feature vector;

根据预测故障类别与故障信息样本对应的真实故障类别之间的误差,对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型。According to the error between the predicted fault category and the actual fault category corresponding to the fault information sample, the initial multi-scale multi-head self-attention GCN model is trained to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

在其中一个实施例中,根据预测故障类别与故障信息样本对应的真实故障类别之间的误差,对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型,包括:In one embodiment, according to the error between the predicted fault category and the actual fault category corresponding to the fault information sample, the initial multi-scale multi-head self-attention GCN model is trained to obtain a target multi-scale multi-head self-attention graph convolutional neural network GCN model, including:

根据预测故障类别与故障信息样本对应的真实故障类别之间的误差确定损失值;Determine the loss value according to the error between the predicted fault category and the actual fault category corresponding to the fault information sample;

在损失值大于预设阈值的情况下,根据损失值调整初始多尺度多头自注意力GCN模型的参数得到中间GCN模型,以对初始多尺度多头自注意力GCN模型进行训练,直至损失值小于预设阈值,并将小于预设阈值的损失值对应的中间GCN模型作为目标多尺度多头自注意力图卷积神经网络GCN模型。When the loss value is greater than the preset threshold, the parameters of the initial multi-scale multi-head self-attention GCN model are adjusted according to the loss value to obtain the intermediate GCN model, so as to train the initial multi-scale multi-head self-attention GCN model until the loss value is less than the preset threshold, and the intermediate GCN model corresponding to the loss value less than the preset threshold is used as the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

第二方面,本申请还提供了一种旋转设备的故障分类装置。装置包括:In a second aspect, the present application also provides a fault classification device for rotating equipment. The device comprises:

特征提取模块,用于对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征;A feature extraction module is used to extract features from the fault information of the rotating equipment to be tested, and obtain the fault features of the fault information;

故障分类模块,用于将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别;目标多尺度多头自注意力图卷积神经网络GCN模型是基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到的,融合图特征样本为对目标输出信息和多个图特征样本进行融合得到的,目标输出信息为初始多尺度多头自注意力GCN模型基于故障信息样本确定,图特征样本为基于故障信息样本、邻接矩阵和初始多尺度多头自注意力GCN模型确定的。A fault classification module is used to input fault features into a target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested; the target multi-scale multi-head self-attention graph convolutional neural network GCN model is obtained by training the initial multi-scale multi-head self-attention GCN model based on fused graph feature samples, the fused graph feature samples are obtained by fusing the target output information and multiple graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on the fault information samples, and the graph feature samples are determined based on the fault information samples, the adjacency matrix and the initial multi-scale multi-head self-attention GCN model.

第三方面,本申请还提供了一种计算机设备。计算机设备包括存储器和处理器,存储器存储有计算机程序,处理器执行计算机程序时实现上述第一方面任一项的方法的步骤。In a third aspect, the present application further provides a computer device, which includes a memory and a processor, wherein the memory stores a computer program, and the processor implements the steps of any one of the methods in the first aspect when executing the computer program.

第四方面,本申请还提供了一种计算机可读存储介质。计算机可读存储介质,其上存储有计算机程序,计算机程序被处理器执行时实现上述第一方面任一项的方法的步骤。In a fourth aspect, the present application further provides a computer-readable storage medium, wherein a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the steps of any one of the methods in the first aspect are implemented.

第五方面,本申请还提供了一种计算机程序产品。计算机程序产品,包括计算机程序,该计算机程序被处理器执行时实现上述第一方面任一项的方法的步骤。In a fifth aspect, the present application further provides a computer program product, including a computer program, which implements the steps of any one of the methods in the first aspect when executed by a processor.

上述旋转设备的故障分类方法、装置、设备、存储介质和产品,对目标输出信息和多个图特征样本进行融合得到融合图特征样本,再基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到目标多尺度多头自注意力图卷积神经网络GCN模型,由于融合图特征样本包括目标输出信息和多个图特征样本,包含的信息比传统技术中单一的图特征样本所包含的信息更丰富,因此,基于信息更丰富的融合图特征样本训练得到的目标多尺度多头自注意力图卷积神经网络GCN模型的预测精度也就更高,进而,能提高故障检测的精度。The above-mentioned fault classification method, device, equipment, storage medium and product of rotating equipment fuse the target output information and multiple graph feature samples to obtain a fused graph feature sample, and then based on the fused graph feature sample, train the initial multi-scale multi-head self-attention GCN model to obtain a target multi-scale multi-head self-attention graph convolutional neural network GCN model. Since the fused graph feature sample includes the target output information and multiple graph feature samples, the information contained is richer than the information contained in a single graph feature sample in traditional technology. Therefore, the prediction accuracy of the target multi-scale multi-head self-attention graph convolutional neural network GCN model obtained by training based on the fused graph feature sample with richer information is higher, thereby improving the accuracy of fault detection.

附图说明BRIEF DESCRIPTION OF THE DRAWINGS

图1为一个实施例中旋转设备的故障分类方法的应用环境图;FIG1 is an application environment diagram of a fault classification method for rotating equipment in one embodiment;

图2为一个实施例中旋转设备的故障分类方法的流程示意图;FIG2 is a schematic flow chart of a method for classifying faults of rotating equipment in one embodiment;

图3为一个实施例中构图过程的示意图;FIG3 is a schematic diagram of a patterning process in one embodiment;

图4为一个实施例中K=2时对应的模型和K=4时对应的模型的示意图;FIG4 is a schematic diagram of a model corresponding to when K=2 and a model corresponding to when K=4 in one embodiment;

图5为一个实施例中训练得到目标多尺度多头自注意力图卷积神经网络GCN模型的流程示意图;FIG5 is a schematic diagram of a process of training a target multi-scale multi-head self-attention graph convolutional neural network GCN model in one embodiment;

图6为一个实施例中根据融合图特征样本对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型的流程示意图;FIG6 is a schematic diagram of a process of training an initial multi-scale multi-head self-attention GCN model according to a fusion graph feature sample to obtain a target multi-scale multi-head self-attention graph convolutional neural network GCN model in one embodiment;

图7为一个示例性的实施例中一种旋转设备的故障分类方法的流程示意图;FIG7 is a schematic flow chart of a method for classifying faults of rotating equipment in an exemplary embodiment;

图8为一个示例性的实施例中一种目标多尺度多头自注意力图卷积神经网络GCN模型的示意图;FIG8 is a schematic diagram of a target multi-scale multi-head self-attention graph convolutional neural network GCN model in an exemplary embodiment;

图9为一个示例性的实施例中一种多头自注意力融合模型的示意图;FIG9 is a schematic diagram of a multi-head self-attention fusion model in an exemplary embodiment;

图10为一个实施例中一种旋转设备的故障分类装置的示意图;FIG10 is a schematic diagram of a fault classification device for a rotating device according to an embodiment;

图11为一个实施例中服务器的内部结构图;FIG11 is a diagram showing the internal structure of a server in one embodiment;

图12为一个实施例中终端的内部结构图。FIG. 12 is a diagram showing the internal structure of a terminal in one embodiment.

具体实施方式Detailed ways

为了使本申请的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本申请进行进一步详细说明。应当理解,此处描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。In order to make the purpose, technical solution and advantages of the present application more clearly understood, the present application is further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are only used to explain the present application and are not used to limit the present application.

本申请实施例提供的旋转设备的故障分类方法,可以应用于如图1所示的应用环境中。其中,计算机设备102对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征;将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别;目标多尺度多头自注意力图卷积神经网络GCN模型是基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到的,融合图特征样本为对目标输出信息和多个图特征样本进行融合得到的,目标输出信息为初始多尺度多头自注意力GCN模型基于故障信息样本确定,图特征样本为基于故障信息样本、邻接矩阵和初始多尺度多头自注意力GCN模型确定的。其中,计算机设备102可以但不限于是各种个人计算机、笔记本电脑、智能手机、平板电脑、物联网设备和便携式可穿戴设备,物联网设备可为智能音箱、智能电视、智能空调、智能车载设备等。便携式可穿戴设备可为智能手表、智能手环、头戴设备等。The fault classification method for rotating equipment provided in the embodiment of the present application can be applied in the application environment shown in FIG1. The computer device 102 extracts features from the fault information of the rotating equipment to be tested to obtain the fault features of the fault information; the fault features are input into the target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested; the target multi-scale multi-head self-attention graph convolutional neural network GCN model is obtained by training the initial multi-scale multi-head self-attention GCN model based on the fusion graph feature samples, the fusion graph feature samples are obtained by fusing the target output information and multiple graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on the fault information samples, and the graph feature samples are determined based on the fault information samples, the adjacency matrix and the initial multi-scale multi-head self-attention GCN model. The computer device 102 can be, but is not limited to, various personal computers, laptops, smart phones, tablet computers, Internet of Things devices and portable wearable devices, and the Internet of Things devices can be smart speakers, smart TVs, smart air conditioners, smart car-mounted devices, etc. The portable wearable device can be a smart watch, a smart bracelet, a head-mounted device, etc.

在一个实施例中,如图2所示,提供了一种旋转设备的故障分类方法,以该方法应用于图1中的计算机设备102为例进行说明,包括以下步骤:In one embodiment, as shown in FIG. 2 , a method for fault classification of rotating equipment is provided, and the method is applied to the computer device 102 in FIG. 1 as an example for description, including the following steps:

步骤202,对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征。Step 202: extracting features from the fault information of the rotating equipment to be tested, and obtaining fault features of the fault information.

其中,待测旋转设备可以是电机、发动机等设备中的任意一种,本实施例对此不作限定。The rotating device to be tested may be any one of a motor, an engine, and the like, and this embodiment does not limit this.

可选地,首先,利用安装在待测旋转设备中的振动传感器、电流传感器等传感器获取待测旋转设备的故障信号,对待测旋转设备的故障信号进行切分,得到待测旋转设备的故障信息;然后,再对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征。其中,特征提取的方法可以是快速傅里叶变换的方法和归一化的方法,也可以是其它方法,本实施例对此不作限定。Optionally, first, a vibration sensor, a current sensor or other sensor installed in the rotating device to be tested is used to obtain a fault signal of the rotating device to be tested, and the fault signal of the rotating device to be tested is segmented to obtain fault information of the rotating device to be tested; then, feature extraction is performed on the fault information of the rotating device to be tested to obtain fault features of the fault information. The feature extraction method may be a fast Fourier transform method and a normalization method, or other methods, which are not limited in this embodiment.

步骤204,将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别;目标多尺度多头自注意力图卷积神经网络GCN模型是基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到的,融合图特征样本为对目标输出信息和多个图特征样本进行融合得到的,目标输出信息为初始多尺度多头自注意力GCN模型基于故障信息样本确定,图特征样本为基于故障信息样本、邻接矩阵和初始多尺度多头自注意力GCN模型确定的。Step 204, input the fault feature into the target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested; the target multi-scale multi-head self-attention graph convolutional neural network GCN model is based on the fusion graph feature samples, and the initial multi-scale multi-head self-attention GCN model is trained, the fusion graph feature samples are obtained by fusing the target output information and multiple graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on the fault information samples, and the graph feature samples are determined based on the fault information samples, the adjacency matrix and the initial multi-scale multi-head self-attention GCN model.

可选地,由于故障信息样本之间不存在显示的关联关系,使用初始多尺度多头自注意力GCN模型时需要构图。可以视某一故障信息样本作为中心节点,利用KNN(K-nearestNeighbor,K近邻)算法计算该中心节点与其它故障信息样本的两两距离,将距离该中心节点最近的K个故障信息样本作为与该中心节点有关联的近邻节点,使用邻接矩阵A来描述图上各故障信息样本之间的关联关系,邻接矩阵A的公式如公式(1)所示。Optionally, since there is no explicit association between fault information samples, graph composition is required when using the initial multi-scale multi-head self-attention GCN model. A certain fault information sample can be regarded as a central node, and the KNN (K-nearestNeighbor) algorithm is used to calculate the pairwise distance between the central node and other fault information samples. The K fault information samples closest to the central node are regarded as the neighboring nodes associated with the central node, and the adjacency matrix A is used to describe the association between the fault information samples on the graph. The formula of the adjacency matrix A is shown in formula (1).

(1) (1)

公式(1)中,xi为中心节点,xj为与中心节点有关联的近邻节点。In formula (1), xi is the central node, and xj is the neighboring node associated with the central node.

可选地,可以取不同的K值以构造不同的图,构造的图的数量可以是大于或者等于2的整数,本实施例对此不作限定。假设构造的图的数量为2,K的取值分别为2和4,其构图过程如图3所示。K=2时,将故障信息样本X1作为中心节点,与X1相关联的近邻节点为X2和X3,邻接矩阵为A1。K=4时,将故障信息样本X5作为中心节点,与X5相关联的近邻节点为X1、X2、X3、X4,邻接矩阵为A2。Optionally, different K values may be taken to construct different graphs, and the number of constructed graphs may be an integer greater than or equal to 2, which is not limited in this embodiment. Assuming that the number of constructed graphs is 2, the values of K are 2 and 4 respectively, and the graph construction process is shown in FIG3. When K=2, the fault information sample X1 is taken as the central node, the neighboring nodes associated with X1 are X2 and X3 , and the adjacency matrix is A1. When K=4, the fault information sample X5 is taken as the central node, the neighboring nodes associated with X5 are X1 , X2 , X3 , and X4 , and the adjacency matrix is A2.

图4展示了K=2时对应的模型和K=4时对应的模型的示意图,K=2时对应的模型表示为第一子模型1,第一子模型1包括两个图卷积层,分别为图卷积层1-1和图卷积层1-2,K=4时对应的模型表示为第一子模型2,第一子模型2包括两个图卷积层,分别为图卷积层2-1和图卷积层2-2。Figure 4 shows schematic diagrams of the model corresponding to K=2 and the model corresponding to K=4. The model corresponding to K=2 is represented as the first sub-model 1, and the first sub-model 1 includes two graph convolution layers, namely graph convolution layer 1-1 and graph convolution layer 1-2. The model corresponding to K=4 is represented as the first sub-model 2, and the first sub-model 2 includes two graph convolution layers, namely graph convolution layer 2-1 and graph convolution layer 2-2.

将故障信息样本和邻接矩阵A1输入至第一子模型1,得到图特征样本1,将故障信息样本和邻接矩阵A2输入至第一子模型2,得到图特征样本2。将故障信息样本输入至初始多尺度多头自注意力GCN模型中的目标网络中,得到目标输出信息。其中,初始多尺度多头自注意力GCN模型中的目标网络可以是多层感知器,或者是由连接层和线性层组成的子网络,或者是多层感知器和该子网络,也即,目标输出信息可以是多层感知器输出的第一输出信息,或者,目标输出信息也可以是该子网络输出的第二输出信息,或者,目标输出信息可以包括第一输出信息和第二输出信息。本实施例对此不作限定。Input the fault information sample and the adjacency matrix A1 into the first sub-model 1 to obtain the graph feature sample 1, and input the fault information sample and the adjacency matrix A2 into the first sub-model 2 to obtain the graph feature sample 2. Input the fault information sample into the target network in the initial multi-scale multi-head self-attention GCN model to obtain the target output information. Among them, the target network in the initial multi-scale multi-head self-attention GCN model can be a multi-layer perceptron, or a sub-network composed of a connection layer and a linear layer, or a multi-layer perceptron and the sub-network, that is, the target output information can be the first output information output by the multi-layer perceptron, or the target output information can also be the second output information output by the sub-network, or the target output information can include the first output information and the second output information. This embodiment is not limited to this.

对目标输出信息和多个图特征样本进行融合得到融合图特征样本,基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练,得到目标多尺度多头自注意力图卷积神经网络GCN模型。其中,这里的融合可以是加权求和,也可以是其它计算方法,本实施例对此不作限定。The target output information and multiple graph feature samples are fused to obtain fused graph feature samples. Based on the fused graph feature samples, the initial multi-scale multi-head self-attention GCN model is trained to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model. The fusion here can be weighted summation or other calculation methods, which is not limited in this embodiment.

将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型,得到待测旋转设备的故障类别。The fault features are input into the target multi-scale multi-head self-attention graph convolutional neural network (GCN) model to obtain the fault category of the rotating equipment to be tested.

上述旋转设备的故障分类方法中,对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征;将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别;目标多尺度多头自注意力图卷积神经网络GCN模型是基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到的,融合图特征样本为对目标输出信息和多个图特征样本进行融合得到的,目标输出信息为初始多尺度多头自注意力GCN模型基于故障信息样本确定,图特征样本为基于故障信息样本、邻接矩阵和初始多尺度多头自注意力GCN模型确定的。其中,对目标输出信息和多个图特征样本进行融合得到融合图特征样本,再基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到目标多尺度多头自注意力图卷积神经网络GCN模型,由于融合图特征样本包括目标输出信息和多个图特征样本,包含的信息比传统技术中单一的图特征样本所包含的信息更丰富,因此,基于信息更丰富的融合图特征样本训练得到的目标多尺度多头自注意力图卷积神经网络GCN模型的预测精度也就更高,进而,能提高故障检测的精度。In the above-mentioned fault classification method for rotating equipment, feature extraction is performed on the fault information of the rotating equipment to be tested to obtain the fault features of the fault information; the fault features are input into the target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested; the target multi-scale multi-head self-attention graph convolutional neural network GCN model is obtained by training the initial multi-scale multi-head self-attention GCN model based on the fused graph feature samples, the fused graph feature samples are obtained by fusing the target output information and multiple graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on the fault information samples, and the graph feature samples are determined based on the fault information samples, the adjacency matrix and the initial multi-scale multi-head self-attention GCN model. Among them, the target output information and multiple graph feature samples are fused to obtain fused graph feature samples, and then based on the fused graph feature samples, the initial multi-scale multi-head self-attention GCN model is trained to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model. Since the fused graph feature samples include the target output information and multiple graph feature samples, the information contained is richer than the information contained in a single graph feature sample in traditional technology. Therefore, the target multi-scale multi-head self-attention graph convolutional neural network GCN model trained based on the fused graph feature samples with richer information has higher prediction accuracy, thereby improving the accuracy of fault detection.

在一个实施例中,初始多尺度多头自注意力GCN模型包括多个第一子模型和多层感知器,训练得到目标多尺度多头自注意力图卷积神经网络GCN模型的流程如图5所示,包括:In one embodiment, the initial multi-scale multi-head self-attention GCN model includes multiple first sub-models and multi-layer perceptrons. The process of training to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model is shown in FIG5, including:

步骤502,针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型,得到第一子模型输出的图特征样本。Step 502: For each first sub-model, the fault information sample and the adjacency matrix corresponding to each first sub-model are input into the first sub-model to obtain the graph feature sample output by the first sub-model.

可选地,假设初始多尺度多头自注意力GCN模型包括2个第一子模型,分别为第一子模型1和第一子模型2,将故障信息样本和第一子模型1对应的邻接矩阵输入至第一子模型1,得到第一子模型输出的图特征样本1。将故障信息样本和第一子模型2对应的邻接矩阵输入至第一子模型2,得到第一子模型输出的图特征样本2。Optionally, assuming that the initial multi-scale multi-head self-attention GCN model includes two first sub-models, namely, first sub-model 1 and first sub-model 2, the fault information sample and the adjacency matrix corresponding to the first sub-model 1 are input into the first sub-model 1 to obtain the graph feature sample 1 output by the first sub-model. The fault information sample and the adjacency matrix corresponding to the first sub-model 2 are input into the first sub-model 2 to obtain the graph feature sample 2 output by the first sub-model.

步骤504,将故障信息样本输入至多层感知器,得到第一输出信息;目标输出信息包括第一输出信息。Step 504: input the fault information sample into the multilayer perceptron to obtain first output information; the target output information includes the first output information.

可选地,将故障信息样本输入至多层感知器,得到第一输出信息,第一输出信息如公式(2)所示。Optionally, the fault information sample is input into a multilayer perceptron to obtain first output information, and the first output information is shown in formula (2).

(2) (2)

公式(2)中,为故障信息样本对应的特征,/>为第一输出信息。In formula (2), is the feature corresponding to the fault information sample, /> It is the first output information.

步骤506,根据第一输出信息和各图特征样本得到融合图特征样本。Step 506: Obtain a fused graph feature sample according to the first output information and each graph feature sample.

可选地,确定第一输出信息对应的第一权重,各图特征样本对应的第三权重,对第一权重与第一输出信息的乘积、各图特征样本与第三权重的乘积进行加和,将加和得到的结果作为融合图特征样本。Optionally, determine a first weight corresponding to the first output information and a third weight corresponding to each graph feature sample, add the product of the first weight and the first output information and the product of each graph feature sample and the third weight, and use the summed result as the fused graph feature sample.

步骤508,根据融合图特征样本对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型。Step 508, training the initial multi-scale multi-head self-attention GCN model according to the fusion graph feature samples to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

可选地,融合图特征样本的数量可以为一个,也可以为多个,本实施例对此不作限定。若融合图特征样本的数量为一个,利用初始多尺度多头自注意力GCN模型中的全连接层基于融合图特征样本得到预测故障类别,根据预测故障类别与故障信息样本对应的真实故障类别之间的误差,对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型。Optionally, the number of fusion graph feature samples may be one or more, which is not limited in this embodiment. If the number of fusion graph feature samples is one, the predicted fault category is obtained based on the fusion graph feature samples using the fully connected layer in the initial multi-scale multi-head self-attention GCN model, and the error between the predicted fault category and the actual fault category corresponding to the fault information sample is used to train the initial multi-scale multi-head self-attention GCN model to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

本实施例中,针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型,得到第一子模型输出的图特征样本;将故障信息样本输入至多层感知器,得到第一输出信息;目标输出信息包括第一输出信息;根据第一输出信息和各图特征样本得到融合图特征样本;根据融合图特征样本对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型。其中,对第一输出信息和多个图特征样本进行融合得到融合图特征样本,再基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到目标多尺度多头自注意力图卷积神经网络GCN模型,由于融合图特征样本包括第一输出信息和多个图特征样本,包含的信息比传统技术中单一的图特征样本所包含的信息更丰富,因此,基于信息更丰富的融合图特征样本训练得到的目标多尺度多头自注意力图卷积神经网络GCN模型的精度也就更高。In this embodiment, for each first sub-model, the fault information sample and the adjacency matrix corresponding to each first sub-model are input into the first sub-model to obtain the graph feature sample output by the first sub-model; the fault information sample is input into the multi-layer perceptron to obtain the first output information; the target output information includes the first output information; the fused graph feature sample is obtained according to the first output information and each graph feature sample; the initial multi-scale multi-head self-attention GCN model is trained according to the fused graph feature sample to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model. Among them, the first output information and multiple graph feature samples are fused to obtain the fused graph feature sample, and then based on the fused graph feature sample, the initial multi-scale multi-head self-attention GCN model is trained to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model. Since the fused graph feature sample includes the first output information and multiple graph feature samples, the information contained is richer than the information contained in the single graph feature sample in the traditional technology. Therefore, the accuracy of the target multi-scale multi-head self-attention graph convolutional neural network GCN model obtained by training based on the more information-rich fused graph feature sample is also higher.

在一个实施例中,第一子模型包括第一图卷积层和第二图卷积层,初始多尺度多头自注意力GCN模型包括连接层和线性层;根据第一输出信息和各图特征样本得到融合图特征样本,包括:In one embodiment, the first sub-model includes a first graph convolution layer and a second graph convolution layer, and the initial multi-scale multi-head self-attention GCN model includes a connection layer and a linear layer; obtaining a fused graph feature sample according to the first output information and each graph feature sample includes:

针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型的第一图卷积层,得到第一中间图特征;For each first sub-model, input the fault information sample and the adjacency matrix corresponding to each first sub-model into the first graph convolution layer of the first sub-model to obtain a first intermediate graph feature;

将各第一子模型的第一图卷积层得到的第一中间图特征输入至连接层,得到第二中间图特征;Inputting the first intermediate graph features obtained by the first graph convolution layer of each first sub-model into the connection layer to obtain the second intermediate graph features;

将第二中间图特征输入至线性层得到第二输出信息。The second intermediate graph feature is input into the linear layer to obtain the second output information.

可选地,假设初始多尺度多头自注意力GCN模型包括2个第一子模型,分别为第一子模型1和第一子模型2,将故障信息样本和第一子模型1对应的邻接矩阵输入至第一子模型1的第一图卷积层,得到第一中间图特征1,将故障信息样本和第一子模型2对应的邻接矩阵输入至第一子模型2的第一图卷积层,得到第一中间图特征2,将第一中间图特征1和第一中间图特征2输入至连接层,得到第二中间图特征,将第二中间图特征输入至线性层得到第二输出信息,第二输出信息的公式如公式(3)所示。Optionally, assuming that the initial multi-scale multi-head self-attention GCN model includes two first sub-models, namely the first sub-model 1 and the first sub-model 2, the fault information sample and the adjacency matrix corresponding to the first sub-model 1 are input into the first graph convolution layer of the first sub-model 1 to obtain the first intermediate graph feature 1, the fault information sample and the adjacency matrix corresponding to the first sub-model 2 are input into the first graph convolution layer of the first sub-model 2 to obtain the first intermediate graph feature 2, the first intermediate graph feature 1 and the first intermediate graph feature 2 are input into the connection layer to obtain the second intermediate graph feature, the second intermediate graph feature is input into the linear layer to obtain the second output information, and the formula of the second output information is shown in formula (3).

(3) (3)

公式(3)中,为第一中间图特征1,/>为第一中间图特征2,W和b为可训练参数。In formula (3), The first intermediate figure feature 1, /> is the first intermediate image feature 2, W and b are trainable parameters.

根据第一输出信息、第二输出信息和各图特征样本得到融合图特征样本;目标输出信息包括第一输出信息和第二输出信息。A fused graph feature sample is obtained according to the first output information, the second output information and each graph feature sample; the target output information includes the first output information and the second output information.

可选地,可以对第一输出信息、第二输出信息和各图特征样本进行加权求和,将加权求和的结果作为融合图特征样本。Optionally, a weighted sum may be performed on the first output information, the second output information, and each graph feature sample, and the result of the weighted sum may be used as a fused graph feature sample.

本实施例中,针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型的第一图卷积层,得到第一中间图特征;将各第一子模型的第一图卷积层得到的第一中间图特征输入至连接层,得到第二中间图特征;将第二中间图特征输入至线性层得到第二输出信息;根据第一输出信息、第二输出信息和各图特征样本得到融合图特征样本;目标输出信息包括第一输出信息和第二输出信息。其中,融合图特征样本包括第一输出信息、第二输出信息和各图特征样本,包含的信息比传统技术中单一的图特征样本所包含的信息更丰富。In this embodiment, for each first sub-model, the fault information sample and the adjacency matrix corresponding to each first sub-model are input into the first graph convolution layer of the first sub-model to obtain the first intermediate graph feature; the first intermediate graph feature obtained by the first graph convolution layer of each first sub-model is input into the connection layer to obtain the second intermediate graph feature; the second intermediate graph feature is input into the linear layer to obtain the second output information; the fused graph feature sample is obtained according to the first output information, the second output information and each graph feature sample; the target output information includes the first output information and the second output information. Among them, the fused graph feature sample includes the first output information, the second output information and each graph feature sample, and the information contained is richer than the information contained in a single graph feature sample in the traditional technology.

在一个实施例中,根据第一输出信息、第二输出信息和各图特征样本得到融合图特征样本;目标输出信息包括第一输出信息和第二输出信息,包括:In one embodiment, a fused graph feature sample is obtained according to the first output information, the second output information and each graph feature sample; the target output information includes the first output information and the second output information, including:

根据第一输出信息、第二输出信息、各图特征样本、第一输出信息对应的第一权重、第二输出信息对应的第二权重、各图特征样本对应的第三权重,得到第一输出信息、第二输出信息和各图特征样本的加权值;Obtaining weighted values of the first output information, the second output information, and each graph feature sample according to the first output information, the second output information, each graph feature sample, a first weight corresponding to the first output information, a second weight corresponding to the second output information, and a third weight corresponding to each graph feature sample;

基于加权值得到融合图特征样本。The fusion graph feature samples are obtained based on the weighted values.

可选地,计算第一输出信息对应的第一权重、第二输出信息对应的第二权重、各图特征样本对应的第三权重的公式如公式(4)、(5)所示。Optionally, formulas for calculating a first weight corresponding to the first output information, a second weight corresponding to the second output information, and a third weight corresponding to each graph feature sample are shown in formulas (4) and (5).

(4) (4)

(5) (5)

公式(4)、(5)中,假设图特征样本的数量为2,则i的取值为1、2、3、4,Z1表示图特征样本1,Z2表示图特征样本2,Z3表示第一输出信息,Z4表示第二输出信息,和/>为各图特征样本对应的第三权重,/>为第一输出信息对应的第一权重,/>为第二输出信息对应的第二权重。其中,/>的加和为1。In formulas (4) and (5), assuming that the number of graph feature samples is 2, the value of i is 1, 2, 3, and 4, Z1 represents graph feature sample 1, Z2 represents graph feature sample 2, Z3 represents the first output information, and Z4 represents the second output information. and/> is the third weight corresponding to each graph feature sample,/> is the first weight corresponding to the first output information, /> is the second weight corresponding to the second output information. The sum of is 1.

根据第一输出信息、第二输出信息、各图特征样本、第一输出信息对应的第一权重、第二输出信息对应的第二权重、各图特征样本对应的第三权重,得到第一输出信息、第二输出信息和各图特征样本的加权值,计算公式如公式(6)所示,将加权值作为融合特征样本。According to the first output information, the second output information, each graph feature sample, the first weight corresponding to the first output information, the second weight corresponding to the second output information, and the third weight corresponding to each graph feature sample, the weighted value of the first output information, the second output information, and each graph feature sample is obtained. The calculation formula is shown in formula (6), and the weighted value is used as the fusion feature sample.

(6) (6)

公式(6)中,表示融合图特征样本。In formula (6), Represents the fusion graph feature sample.

本实施例中,根据第一输出信息、第二输出信息、各图特征样本、第一输出信息对应的第一权重、第二输出信息对应的第二权重、各图特征样本对应的第三权重,得到第一输出信息、第二输出信息和各图特征样本的加权值;基于加权值得到融合图特征样本。通过给第一输出信息、第二输出信息、各图特征样本赋予不同的权重,可以自适应地融合第一输出信息、第二输出信息和各图特征样本,在丰富信息的同时,避免深层次网络导致的过平滑现象。In this embodiment, the weighted values of the first output information, the second output information, and each graph feature sample are obtained according to the first output information, the second output information, each graph feature sample, the first weight corresponding to the first output information, the second weight corresponding to the second output information, and the third weight corresponding to each graph feature sample; and the fused graph feature sample is obtained based on the weighted value. By assigning different weights to the first output information, the second output information, and each graph feature sample, the first output information, the second output information, and each graph feature sample can be adaptively fused, thereby enriching the information and avoiding the over-smoothing phenomenon caused by the deep network.

在一个实施例中,融合图特征样本的数量为多个;初始多尺度多头自注意力GCN模型包括多头自注意力融合模型和全连接层;根据融合图特征样本对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型,流程如图6所示,包括:In one embodiment, the number of fused graph feature samples is multiple; the initial multi-scale multi-head self-attention GCN model includes a multi-head self-attention fusion model and a fully connected layer; the initial multi-scale multi-head self-attention GCN model is trained according to the fused graph feature samples to obtain a target multi-scale multi-head self-attention graph convolutional neural network GCN model, and the process is shown in FIG6, including:

步骤602,利用多头自注意力融合模型对各融合图特征样本进行连接得到图特征向量。Step 602: Use a multi-head self-attention fusion model to connect each fused graph feature sample to obtain a graph feature vector.

可选地,为了使融合图特征样本的表征能力更强,可设置多组可训练参数,再基于公式(4)、(5)和(6),计算得到多个融合图特征样本/>,其中,。利用多头自注意力融合模型对多个融合图特征样本进行连接,如公式(7)所示,得到图特征向量。Optionally, in order to make the representation ability of the fusion graph feature samples stronger, multiple sets of trainable parameters can be set , and then based on formulas (4), (5) and (6), multiple fusion graph feature samples are calculated./> ,in, The multi-head self-attention fusion model is used to connect multiple fusion graph feature samples, as shown in formula (7), to obtain the graph feature vector.

(7) (7)

公式(7)中,为图特征向量。In formula (7), is the graph feature vector.

步骤604,利用全连接层基于图特征向量得到预测故障类别。Step 604: Use the fully connected layer to obtain the predicted fault category based on the graph feature vector.

可选地,利用全连接层对图特征向量进行线性变换和Softmax函数的处理,得到预测故障类别,如公式(8)所示。Optionally, the fully connected layer is used to perform linear transformation and Softmax function processing on the graph feature vector to obtain the predicted fault category, as shown in formula (8).

(8) (8)

公式(8)中,表示图特征向量,W和b表示可训练参数,y表示预测故障类别。In formula (8), represents the graph feature vector, W and b represent trainable parameters, and y represents the predicted fault category.

步骤606,根据预测故障类别与故障信息样本对应的真实故障类别之间的误差,对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型。Step 606, according to the error between the predicted fault category and the actual fault category corresponding to the fault information sample, the initial multi-scale multi-head self-attention GCN model is trained to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

可选地,利用损失函数描述预测故障类别与故障信息样本对应的真实故障类别之间的误差,根据该误差对应的损失值,对初始多尺度多头自注意力GCN模型进行训练,直到损失值满足预设条件,得到目标多尺度多头自注意力图卷积神经网络GCN模型。其中,这里的损失函数可以是交叉熵,也可以是其它函数,本实施例对此不作限定。这里的预设条件可以是损失值小于预设阈值,也可以是损失值等于预设阈值,本实施例对此不作限定。Optionally, a loss function is used to describe the error between the predicted fault category and the actual fault category corresponding to the fault information sample. According to the loss value corresponding to the error, the initial multi-scale multi-head self-attention GCN model is trained until the loss value meets the preset condition, and the target multi-scale multi-head self-attention graph convolutional neural network GCN model is obtained. Among them, the loss function here can be cross entropy or other functions, which is not limited in this embodiment. The preset condition here can be that the loss value is less than the preset threshold, or that the loss value is equal to the preset threshold, which is not limited in this embodiment.

本实施例中,利用多头自注意力融合模型对各融合图特征样本进行连接得到图特征向量;利用全连接层基于图特征向量得到预测故障类别;根据预测故障类别与故障信息样本对应的真实故障类别之间的误差,对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型。通过引入多头自注意力融合模型,使得融合图特征样本的表征能力更强,进一步提高目标多尺度多头自注意力图卷积神经网络GCN模型的精度。In this embodiment, the multi-head self-attention fusion model is used to connect each fusion graph feature sample to obtain a graph feature vector; the fully connected layer is used to obtain the predicted fault category based on the graph feature vector; according to the error between the predicted fault category and the real fault category corresponding to the fault information sample, the initial multi-scale multi-head self-attention GCN model is trained to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model. By introducing the multi-head self-attention fusion model, the representation ability of the fusion graph feature sample is made stronger, and the accuracy of the target multi-scale multi-head self-attention graph convolutional neural network GCN model is further improved.

在一个实施例中,根据预测故障类别与故障信息样本对应的真实故障类别之间的误差,对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型,包括:In one embodiment, according to the error between the predicted fault category and the actual fault category corresponding to the fault information sample, the initial multi-scale multi-head self-attention GCN model is trained to obtain a target multi-scale multi-head self-attention graph convolutional neural network GCN model, including:

根据预测故障类别与故障信息样本对应的真实故障类别之间的误差确定损失值。The loss value is determined based on the error between the predicted fault category and the actual fault category corresponding to the fault information sample.

可选地,利用交叉熵描述预测故障类别与故障信息样本对应的真实故障类别之间的误差,根据交叉熵确定损失值。Optionally, cross entropy is used to describe the error between the predicted fault category and the actual fault category corresponding to the fault information sample, and the loss value is determined according to the cross entropy.

在损失值大于预设阈值的情况下,根据损失值调整初始多尺度多头自注意力GCN模型的参数得到中间GCN模型,以对初始多尺度多头自注意力GCN模型进行训练,直至损失值小于预设阈值,并将小于预设阈值的损失值对应的中间GCN模型作为目标多尺度多头自注意力图卷积神经网络GCN模型。When the loss value is greater than the preset threshold, the parameters of the initial multi-scale multi-head self-attention GCN model are adjusted according to the loss value to obtain the intermediate GCN model, so as to train the initial multi-scale multi-head self-attention GCN model until the loss value is less than the preset threshold, and the intermediate GCN model corresponding to the loss value less than the preset threshold is used as the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

可选地,若损失值大于预设阈值,则调整初始多尺度多头自注意力GCN模型的参数得到中间GCN模型,基于故障信息样本、中间GCN模型中各第一子模型对应的邻接矩阵和中间GCN模型中各第一子模型,确定中间GCN模型中各第一子模型输出的图特征样本,基于故障信息样本、中间GCN模型,确定中间GCN模型对应的目标输出信息,对中间GCN模型中各第一子模型输出的图特征样本和中间GCN模型对应的目标输出信息进行融合,得到中间GCN模型对应的融合特征样本,利用多头自注意力融合模型对多个中间GCN模型对应的融合特征样本进行连接得到中间GCN模型对应的图特征向量,利用全连接层基于中间GCN模型对应的图特征向量得到中间GCN模型对应的预测故障类别,根据中间GCN模型对应的预测故障类别与故障信息样本对应的真实故障类别之间的误差确定新的损失值,若该新的损失值小于预设阈值,则将中间GCN模型作为目标多尺度多头自注意力图卷积神经网络GCN模型。Optionally, if the loss value is greater than a preset threshold, the parameters of the initial multi-scale multi-head self-attention GCN model are adjusted to obtain an intermediate GCN model, and based on the fault information sample, the adjacency matrix corresponding to each first sub-model in the intermediate GCN model, and each first sub-model in the intermediate GCN model, the graph feature samples output by each first sub-model in the intermediate GCN model are determined, and based on the fault information sample and the intermediate GCN model, the target output information corresponding to the intermediate GCN model is determined, and the graph feature samples output by each first sub-model in the intermediate GCN model and the target output information corresponding to the intermediate GCN model are fused to obtain a fused feature sample corresponding to the intermediate GCN model, and the fused feature samples corresponding to the multiple intermediate GCN models are connected using the multi-head self-attention fusion model to obtain a graph feature vector corresponding to the intermediate GCN model, and the predicted fault category corresponding to the intermediate GCN model is obtained based on the graph feature vector corresponding to the intermediate GCN model using a fully connected layer, and a new loss value is determined according to the error between the predicted fault category corresponding to the intermediate GCN model and the true fault category corresponding to the fault information sample, and if the new loss value is less than the preset threshold, the intermediate GCN model is used as the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

本实施例中,根据预测故障类别与故障信息样本对应的真实故障类别之间的误差确定损失值;在损失值大于预设阈值的情况下,根据损失值调整初始多尺度多头自注意力GCN模型的参数得到中间GCN模型,以对初始多尺度多头自注意力GCN模型进行训练,直至损失值小于预设阈值,并将小于预设阈值的损失值对应的中间GCN模型作为目标多尺度多头自注意力图卷积神经网络GCN模型。通过比较损失值与预设阈值的大小关系,不断调整初始多尺度多头自注意力GCN模型的参数,以使模型的预测值更接近于真实值,提高目标多尺度多头自注意力图卷积神经网络GCN模型的精度。In this embodiment, the loss value is determined according to the error between the predicted fault category and the real fault category corresponding to the fault information sample; when the loss value is greater than the preset threshold, the parameters of the initial multi-scale multi-head self-attention GCN model are adjusted according to the loss value to obtain an intermediate GCN model, so as to train the initial multi-scale multi-head self-attention GCN model until the loss value is less than the preset threshold, and the intermediate GCN model corresponding to the loss value less than the preset threshold is used as the target multi-scale multi-head self-attention graph convolutional neural network GCN model. By comparing the size relationship between the loss value and the preset threshold, the parameters of the initial multi-scale multi-head self-attention GCN model are continuously adjusted to make the predicted value of the model closer to the true value, thereby improving the accuracy of the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

在一个示例性的实施例中,提供了一种旋转设备的故障分类方法,流程如图7所示,包括:In an exemplary embodiment, a method for fault classification of rotating equipment is provided, and the process is shown in FIG7 , including:

步骤701,针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型,得到第一子模型输出的图特征样本。Step 701: for each first sub-model, input the fault information sample and the adjacency matrix corresponding to each first sub-model into the first sub-model to obtain the graph feature sample output by the first sub-model.

步骤702,将故障信息样本输入至多层感知器,得到第一输出信息;目标输出信息包括第一输出信息。Step 702: Input the fault information sample into a multilayer perceptron to obtain first output information; the target output information includes the first output information.

步骤703,针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型的第一图卷积层,得到第一中间图特征。Step 703: For each first sub-model, the fault information sample and the adjacency matrix corresponding to each first sub-model are input into the first graph convolution layer of the first sub-model to obtain the first intermediate graph feature.

步骤704,将各第一子模型的第一图卷积层得到的第一中间图特征输入至连接层,得到第二中间图特征。Step 704: input the first intermediate graph features obtained by the first graph convolution layer of each first sub-model into the connection layer to obtain the second intermediate graph features.

步骤705,将第二中间图特征输入至线性层得到第二输出信息。Step 705: Input the second intermediate graph feature into the linear layer to obtain second output information.

步骤706,根据第一输出信息、第二输出信息、各图特征样本、第一输出信息对应的第一权重、第二输出信息对应的第二权重、各图特征样本对应的第三权重,得到第一输出信息、第二输出信息和各图特征样本的加权值。Step 706, obtaining weighted values of the first output information, the second output information and each graph feature sample according to the first output information, the second output information, each graph feature sample, the first weight corresponding to the first output information, the second weight corresponding to the second output information and the third weight corresponding to each graph feature sample.

步骤707,基于加权值得到融合图特征样本。Step 707: Obtain fusion graph feature samples based on the weighted values.

步骤708,利用多头自注意力融合模型对各融合图特征样本进行连接得到图特征向量。Step 708: Use the multi-head self-attention fusion model to connect the fused graph feature samples to obtain a graph feature vector.

步骤709,利用全连接层基于图特征向量得到预测故障类别。Step 709: Use the fully connected layer to obtain the predicted fault category based on the graph feature vector.

步骤710,根据预测故障类别与故障信息样本对应的真实故障类别之间的误差确定损失值。Step 710: determine a loss value according to an error between the predicted fault category and the actual fault category corresponding to the fault information sample.

步骤711,在损失值大于预设阈值的情况下,根据损失值调整初始多尺度多头自注意力GCN模型的参数得到中间GCN模型,以对初始多尺度多头自注意力GCN模型进行训练,直至损失值小于预设阈值,并将小于预设阈值的损失值对应的中间GCN模型作为目标多尺度多头自注意力图卷积神经网络GCN模型。Step 711, when the loss value is greater than the preset threshold, the parameters of the initial multi-scale multi-head self-attention GCN model are adjusted according to the loss value to obtain an intermediate GCN model, so as to train the initial multi-scale multi-head self-attention GCN model until the loss value is less than the preset threshold, and the intermediate GCN model corresponding to the loss value less than the preset threshold is used as the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

步骤712,对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征。Step 712: extract features from the fault information of the rotating equipment to be tested, and obtain fault features of the fault information.

步骤713,将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别。Step 713, input the fault features into the target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested.

上述旋转设备的故障分类方法,对目标输出信息和多个图特征样本进行融合得到融合图特征样本,再基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到目标多尺度多头自注意力图卷积神经网络GCN模型,由于融合图特征样本包括目标输出信息和多个图特征样本,包含的信息比传统技术中单一的图特征样本所包含的信息更丰富,因此,基于信息更丰富的融合图特征样本训练得到的目标多尺度多头自注意力图卷积神经网络GCN模型的预测精度也就更高,进而,能提高故障检测的精度。In the above-mentioned fault classification method for rotating equipment, the target output information and multiple graph feature samples are fused to obtain fused graph feature samples, and then based on the fused graph feature samples, the initial multi-scale multi-head self-attention GCN model is trained to obtain a target multi-scale multi-head self-attention graph convolutional neural network GCN model. Since the fused graph feature samples include the target output information and multiple graph feature samples, the information contained is richer than the information contained in a single graph feature sample in traditional technology. Therefore, the prediction accuracy of the target multi-scale multi-head self-attention graph convolutional neural network GCN model obtained by training based on the fused graph feature samples with richer information is higher, thereby improving the accuracy of fault detection.

在一个示例性的实施例中,提供了一种目标多尺度多头自注意力图卷积神经网络GCN模型的示意图,如图8所示。将故障信息样本、邻接矩阵1和邻接矩阵2作为输入,经过目标多尺度多头自注意力图卷积神经网络GCN模型中各模块的处理,得到预测故障类别。In an exemplary embodiment, a schematic diagram of a target multi-scale multi-head self-attention graph convolutional neural network GCN model is provided, as shown in Figure 8. The fault information sample, adjacency matrix 1 and adjacency matrix 2 are taken as inputs, and the predicted fault category is obtained after being processed by each module in the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

在一个示例性的实施例中,提供了一种多头自注意力融合模型的示意图,如图9所示。将图特征样本1、将图特征样本2、第一输出信息和第二输出信息作为输入,经过线性层、Tanh函数和Softmax函数的处理,得到图特征向量。In an exemplary embodiment, a schematic diagram of a multi-head self-attention fusion model is provided, as shown in Figure 9. Graph feature sample 1, graph feature sample 2, first output information, and second output information are taken as input, and processed by a linear layer, a Tanh function, and a Softmax function to obtain a graph feature vector.

应该理解的是,虽然如上的各实施例所涉及的流程图中的各个步骤按照箭头的指示依次显示,但是这些步骤并不是必然按照箭头指示的顺序依次执行。除非本文中有明确的说明,这些步骤的执行并没有严格的顺序限制,这些步骤可以以其它的顺序执行。而且,如上的各实施例所涉及的流程图中的至少一部分步骤可以包括多个步骤或者多个阶段,这些步骤或者阶段并不必然是在同一时刻执行完成,而是可以在不同的时刻执行,这些步骤或者阶段的执行顺序也不必然是依次进行,而是可以与其它步骤或者其它步骤中的步骤或者阶段的至少一部分轮流或者交替地执行。It should be understood that, although the steps in the flowcharts involved in the above embodiments are displayed in sequence according to the indication of the arrows, these steps are not necessarily executed in sequence according to the order indicated by the arrows. Unless there is a clear explanation in this article, the execution of these steps is not strictly limited in order, and these steps can be executed in other orders. Moreover, at least a part of the steps in the flowcharts involved in the above embodiments may include multiple steps or multiple stages, and these steps or stages are not necessarily executed at the same time, but can be executed at different times, and the execution order of these steps or stages is not necessarily to be carried out in sequence, but can be executed in turn or alternately with other steps or at least a part of the steps or stages in other steps.

基于同样的发明构思,本申请实施例还提供了一种用于实现上述所涉及的旋转设备的故障分类方法的旋转设备的故障分类装置。该装置所提供的解决问题的实现方案与上述方法中所记载的实现方案相似,故下面所提供的一个或多个旋转设备的故障分类装置实施例中的具体限定可以参见上文中对于旋转设备的故障分类方法的限定,在此不再赘述。Based on the same inventive concept, the embodiment of the present application also provides a rotating equipment fault classification device for implementing the above-mentioned rotating equipment fault classification method. The implementation scheme for solving the problem provided by the device is similar to the implementation scheme recorded in the above-mentioned method, so the specific limitations in the embodiments of the fault classification device for one or more rotating equipment provided below can refer to the limitations of the rotating equipment fault classification method above, and will not be repeated here.

在一个实施例中,如图10所示,提供了一种旋转设备的故障分类装置1000,包括:特征提取模块1020、故障分类模块1040,其中:In one embodiment, as shown in FIG. 10 , a fault classification device 1000 for rotating equipment is provided, comprising: a feature extraction module 1020 and a fault classification module 1040, wherein:

特征提取模块1020,用于对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征;The feature extraction module 1020 is used to extract features from the fault information of the rotating equipment to be tested, and obtain the fault features of the fault information;

故障分类模块1040,用于将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别;目标多尺度多头自注意力图卷积神经网络GCN模型是基于融合图特征样本,对初始图卷积神经网络GCN模型训练得到的,融合图特征样本为对目标输出信息和多个图特征样本进行融合得到的,目标输出信息为初始多尺度多头自注意力GCN模型基于故障信息样本确定,图特征样本为基于故障信息样本、邻接矩阵和初始多尺度多头自注意力GCN模型确定的。The fault classification module 1040 is used to input the fault features into the target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested; the target multi-scale multi-head self-attention graph convolutional neural network GCN model is obtained by training the initial graph convolutional neural network GCN model based on the fused graph feature samples, the fused graph feature samples are obtained by fusing the target output information and multiple graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on the fault information samples, and the graph feature samples are determined based on the fault information samples, the adjacency matrix and the initial multi-scale multi-head self-attention GCN model.

在一个实施例中,旋转设备的故障分类装置1000,还包括:In one embodiment, the fault classification device 1000 for rotating equipment further includes:

第一训练模块,用于针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型,得到第一子模型输出的图特征样本;将故障信息样本输入至多层感知器,得到第一输出信息;目标输出信息包括第一输出信息;根据第一输出信息和各图特征样本得到融合图特征样本;根据融合图特征样本对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型。The first training module is used to input the fault information samples and the adjacency matrix corresponding to each first sub-model into the first sub-model for each first sub-model to obtain the graph feature samples output by the first sub-model; input the fault information samples into the multi-layer perceptron to obtain the first output information; the target output information includes the first output information; obtain the fused graph feature samples according to the first output information and each graph feature sample; train the initial multi-scale multi-head self-attention GCN model according to the fused graph feature samples to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

在一个实施例中,旋转设备的故障分类装置1000,还包括:In one embodiment, the fault classification device 1000 for rotating equipment further includes:

第二训练模块,用于针对各第一子模型,将故障信息样本和各第一子模型对应的邻接矩阵输入至第一子模型的第一图卷积层,得到第一中间图特征;将各第一子模型的第一图卷积层得到的第一中间图特征输入至连接层,得到第二中间图特征;将第二中间图特征输入至线性层得到第二输出信息;根据第一输出信息、第二输出信息和各图特征样本得到融合图特征样本;目标输出信息包括第一输出信息和第二输出信息。The second training module is used to input the fault information sample and the adjacency matrix corresponding to each first sub-model into the first graph convolution layer of the first sub-model for each first sub-model to obtain the first intermediate graph feature; input the first intermediate graph feature obtained by the first graph convolution layer of each first sub-model into the connection layer to obtain the second intermediate graph feature; input the second intermediate graph feature into the linear layer to obtain the second output information; obtain the fused graph feature sample according to the first output information, the second output information and each graph feature sample; the target output information includes the first output information and the second output information.

在一个实施例中,旋转设备的故障分类装置1000,还包括:In one embodiment, the fault classification device 1000 for rotating equipment further includes:

第三训练模块,用于根据第一输出信息、第二输出信息、各图特征样本、第一输出信息对应的第一权重、第二输出信息对应的第二权重、各图特征样本对应的第三权重,得到第一输出信息、第二输出信息和各图特征样本的加权值;基于加权值得到融合图特征样本。The third training module is used to obtain the weighted values of the first output information, the second output information and each graph feature sample according to the first output information, the second output information, each graph feature sample, the first weight corresponding to the first output information, the second weight corresponding to the second output information, and the third weight corresponding to each graph feature sample; and obtain the fused graph feature sample based on the weighted value.

在一个实施例中,旋转设备的故障分类装置1000,还包括:In one embodiment, the fault classification device 1000 for rotating equipment further includes:

第四训练模块,用于利用多头自注意力融合模型对各融合图特征样本进行连接得到图特征向量;利用全连接层基于图特征向量得到预测故障类别;根据预测故障类别与故障信息样本对应的真实故障类别之间的误差,对初始多尺度多头自注意力GCN模型进行训练得到目标多尺度多头自注意力图卷积神经网络GCN模型。The fourth training module is used to use the multi-head self-attention fusion model to connect each fused graph feature sample to obtain a graph feature vector; use the fully connected layer to obtain the predicted fault category based on the graph feature vector; according to the error between the predicted fault category and the actual fault category corresponding to the fault information sample, the initial multi-scale multi-head self-attention GCN model is trained to obtain the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

在一个实施例中,旋转设备的故障分类装置1000,还包括:In one embodiment, the fault classification device 1000 for rotating equipment further includes:

第五训练模块,用于根据预测故障类别与故障信息样本对应的真实故障类别之间的误差确定损失值;在损失值大于预设阈值的情况下,根据损失值调整初始多尺度多头自注意力GCN模型的参数得到中间GCN模型,以对初始多尺度多头自注意力GCN模型进行训练,直至损失值小于预设阈值,并将小于预设阈值的损失值对应的中间GCN模型作为目标多尺度多头自注意力图卷积神经网络GCN模型。The fifth training module is used to determine the loss value according to the error between the predicted fault category and the actual fault category corresponding to the fault information sample; when the loss value is greater than a preset threshold, the parameters of the initial multi-scale multi-head self-attention GCN model are adjusted according to the loss value to obtain an intermediate GCN model, so as to train the initial multi-scale multi-head self-attention GCN model until the loss value is less than the preset threshold, and the intermediate GCN model corresponding to the loss value less than the preset threshold is used as the target multi-scale multi-head self-attention graph convolutional neural network GCN model.

上述旋转设备的故障分类装置中的各个模块可全部或部分通过软件、硬件及其组合来实现。上述各模块可以硬件形式内嵌于或独立于计算机设备中的处理器中,也可以以软件形式存储于计算机设备中的存储器中,以便于处理器调用执行以上各个模块对应的操作。Each module in the above-mentioned fault classification device for rotating equipment can be implemented in whole or in part by software, hardware or a combination thereof. Each module can be embedded in or independent of a processor in a computer device in the form of hardware, or can be stored in a memory in a computer device in the form of software, so that the processor can call and execute operations corresponding to each module.

在一个实施例中,提供了一种计算机设备,该计算机设备可以是服务器,其内部结构图可以如图11所示。该计算机设备包括通过系统总线连接的处理器、存储器和网络接口。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质和内存储器。该非易失性存储介质存储有操作系统、计算机程序和数据库。该内存储器为非易失性存储介质中的操作系统和计算机程序的运行提供环境。该计算机设备的数据库用于存储数据。该计算机设备的网络接口用于与外部的终端通过网络连接通信。该计算机程序被处理器执行时以实现一种旋转设备的故障分类方法。In one embodiment, a computer device is provided, which may be a server, and its internal structure diagram may be as shown in FIG11. The computer device includes a processor, a memory, and a network interface connected via a system bus. The processor of the computer device is used to provide computing and control capabilities. The memory of the computer device includes a non-volatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, a computer program, and a database. The internal memory provides an environment for the operation of the operating system and the computer program in the non-volatile storage medium. The database of the computer device is used to store data. The network interface of the computer device is used to communicate with an external terminal via a network connection. When the computer program is executed by the processor, a fault classification method for rotating equipment is implemented.

在一个实施例中,提供了一种计算机设备,该计算机设备可以是终端,其内部结构图可以如图12所示。该计算机设备包括通过系统总线连接的处理器、存储器、通信接口、显示屏和输入装置。其中,该计算机设备的处理器用于提供计算和控制能力。该计算机设备的存储器包括非易失性存储介质、内存储器。该非易失性存储介质存储有操作系统和计算机程序。该内存储器为非易失性存储介质中的操作系统和计算机程序的运行提供环境。该计算机设备的通信接口用于与外部的终端进行有线或无线方式的通信,无线方式可通过WIFI、移动蜂窝网络、NFC(近场通信)或其他技术实现。该计算机程序被处理器执行时以实现一种旋转设备的故障分类方法。该计算机设备的显示屏可以是液晶显示屏或者电子墨水显示屏,该计算机设备的输入装置可以是显示屏上覆盖的触摸层,也可以是计算机设备外壳上设置的按键、轨迹球或触控板,还可以是外接的键盘、触控板或鼠标等。In one embodiment, a computer device is provided, which may be a terminal, and its internal structure diagram may be shown in FIG12. The computer device includes a processor, a memory, a communication interface, a display screen, and an input device connected through a system bus. Among them, the processor of the computer device is used to provide computing and control capabilities. The memory of the computer device includes a non-volatile storage medium and an internal memory. The non-volatile storage medium stores an operating system and a computer program. The internal memory provides an environment for the operation of the operating system and the computer program in the non-volatile storage medium. The communication interface of the computer device is used to communicate with an external terminal in a wired or wireless manner, and the wireless manner may be implemented through WIFI, a mobile cellular network, NFC (near field communication) or other technologies. When the computer program is executed by the processor, a fault classification method for a rotating device is implemented. The display screen of the computer device may be a liquid crystal display screen or an electronic ink display screen, and the input device of the computer device may be a touch layer covered on the display screen, or a key, a trackball or a touch pad set on the housing of the computer device, or an external keyboard, touch pad or mouse, etc.

本领域技术人员可以理解,图11和图12中示出的结构,仅仅是与本申请方案相关的部分结构的框图,并不构成对本申请方案所应用于其上的计算机设备的限定,具体的计算机设备可以包括比图中所示更多或更少的部件,或者组合某些部件,或者具有不同的部件布置。Those skilled in the art will understand that the structures shown in FIG. 11 and FIG. 12 are merely block diagrams of partial structures related to the scheme of the present application, and do not constitute a limitation on the computer device to which the scheme of the present application is applied. The specific computer device may include more or fewer components than those shown in the figures, or combine certain components, or have a different arrangement of components.

在一个实施例中,提供了一种计算机设备,包括存储器和处理器,存储器中存储有计算机程序,该处理器执行计算机程序时实现以下步骤:In one embodiment, a computer device is provided, including a memory and a processor, wherein a computer program is stored in the memory, and when the processor executes the computer program, the following steps are implemented:

对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征;Extract features of fault information of the rotating equipment to be tested to obtain fault features of the fault information;

将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别;目标多尺度多头自注意力图卷积神经网络GCN模型是基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到的,融合图特征样本为对目标输出信息和多个图特征样本进行融合得到的,目标输出信息为初始多尺度多头自注意力GCN模型基于故障信息样本确定,图特征样本为基于故障信息样本、邻接矩阵和初始多尺度多头自注意力GCN模型确定的。The fault features are input into the target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested; the target multi-scale multi-head self-attention graph convolutional neural network GCN model is trained by the initial multi-scale multi-head self-attention GCN model based on the fused graph feature samples, the fused graph feature samples are obtained by fusing the target output information and multiple graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on the fault information samples, and the graph feature samples are determined based on the fault information samples, the adjacency matrix and the initial multi-scale multi-head self-attention GCN model.

在一个实施例中,还提供了一种计算机设备,包括存储器和处理器,存储器中存储有计算机程序,该处理器执行计算机程序时实现上述各方法实施例中的步骤。In one embodiment, a computer device is further provided, including a memory and a processor, wherein a computer program is stored in the memory, and the processor implements the steps in the above method embodiments when executing the computer program.

在一个实施例中,提供了一种计算机可读存储介质,其上存储有计算机程序,计算机程序被处理器执行时实现以下步骤:In one embodiment, a computer readable storage medium is provided, on which a computer program is stored, and when the computer program is executed by a processor, the following steps are implemented:

对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征;Extract features of fault information of the rotating equipment to be tested to obtain fault features of the fault information;

将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别;目标多尺度多头自注意力图卷积神经网络GCN模型是基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到的,融合图特征样本为对目标输出信息和多个图特征样本进行融合得到的,目标输出信息为初始多尺度多头自注意力GCN模型基于故障信息样本确定,图特征样本为基于故障信息样本、邻接矩阵和初始多尺度多头自注意力GCN模型确定的。The fault features are input into the target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested; the target multi-scale multi-head self-attention graph convolutional neural network GCN model is trained by the initial multi-scale multi-head self-attention GCN model based on the fused graph feature samples, the fused graph feature samples are obtained by fusing the target output information and multiple graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on the fault information samples, and the graph feature samples are determined based on the fault information samples, the adjacency matrix and the initial multi-scale multi-head self-attention GCN model.

在一个实施例中,提供了一种计算机可读存储介质,其上存储有计算机程序,该计算机程序被处理器执行时实现上述各方法实施例中的步骤。In one embodiment, a computer-readable storage medium is provided, on which a computer program is stored. When the computer program is executed by a processor, the steps in the above-mentioned method embodiments are implemented.

在一个实施例中,提供了一种计算机程序产品,包括计算机程序,该计算机程序被处理器执行时实现以下步骤:In one embodiment, a computer program product is provided, comprising a computer program, which, when executed by a processor, implements the following steps:

对待测旋转设备的故障信息进行特征提取,得到故障信息的故障特征;Extract features of fault information of the rotating equipment to be tested to obtain fault features of the fault information;

将故障特征输入至目标多尺度多头自注意力图卷积神经网络GCN模型得到待测旋转设备的故障类别;目标多尺度多头自注意力图卷积神经网络GCN模型是基于融合图特征样本,对初始多尺度多头自注意力GCN模型训练得到的,融合图特征样本为对目标输出信息和多个图特征样本进行融合得到的,目标输出信息为初始多尺度多头自注意力GCN模型基于故障信息样本确定,图特征样本为基于故障信息样本、邻接矩阵和初始多尺度多头自注意力GCN模型确定的。The fault features are input into the target multi-scale multi-head self-attention graph convolutional neural network GCN model to obtain the fault category of the rotating equipment to be tested; the target multi-scale multi-head self-attention graph convolutional neural network GCN model is trained by the initial multi-scale multi-head self-attention GCN model based on the fused graph feature samples, the fused graph feature samples are obtained by fusing the target output information and multiple graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on the fault information samples, and the graph feature samples are determined based on the fault information samples, the adjacency matrix and the initial multi-scale multi-head self-attention GCN model.

在一个实施例中,提供了一种计算机程序产品,包括计算机程序,该计算机程序被处理器执行时实现上述各方法实施例中的步骤。In one embodiment, a computer program product is provided, including a computer program, which implements the steps in the above method embodiments when executed by a processor.

本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,的计算机程序可存储于一非易失性计算机可读取存储介质中,该计算机程序在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、数据库或其它介质的任何引用,均可包括非易失性和易失性存储器中的至少一种。非易失性存储器可包括只读存储器(Read-OnlyMemory,ROM)、磁带、软盘、闪存、光存储器、高密度嵌入式非易失性存储器、阻变存储器(ReRAM)、磁变存储器(Magnetoresistive Random Access Memory,MRAM)、铁电存储器(Ferroelectric Random Access Memory,FRAM)、相变存储器(Phase Change Memory,PCM)、石墨烯存储器等。易失性存储器可包括随机存取存储器(Random Access Memory,RAM)或外部高速缓冲存储器等。作为说明而非局限,RAM可以是多种形式,比如静态随机存取存储器(Static Random Access Memory,SRAM)或动态随机存取存储器(Dynamic RandomAccess Memory,DRAM)等。本申请所提供的各实施例中所涉及的数据库可包括关系型数据库和非关系型数据库中至少一种。非关系型数据库可包括基于区块链的分布式数据库等,不限于此。本申请所提供的各实施例中所涉及的处理器可为通用处理器、中央处理器、图形处理器、数字信号处理器、可编程逻辑器、基于量子计算的数据处理逻辑器等,不限于此。Those skilled in the art can understand that all or part of the processes in the above-mentioned embodiments can be completed by instructing the relevant hardware through a computer program, and the computer program can be stored in a non-volatile computer-readable storage medium. When the computer program is executed, it can include the processes of the embodiments of the above-mentioned methods. Among them, any reference to the memory, database or other medium used in the embodiments provided in the present application can include at least one of non-volatile and volatile memory. Non-volatile memory can include read-only memory (ROM), magnetic tape, floppy disk, flash memory, optical memory, high-density embedded non-volatile memory, resistive random access memory (ReRAM), magnetoresistive random access memory (MRAM), ferroelectric random access memory (FRAM), phase change memory (PCM), graphene memory, etc. Volatile memory can include random access memory (RAM) or external cache memory, etc. As an illustration and not limitation, RAM can be in various forms, such as static random access memory (SRAM) or dynamic random access memory (DRAM). The database involved in each embodiment provided in this application may include at least one of a relational database and a non-relational database. Non-relational databases may include distributed databases based on blockchains, etc., but are not limited to this. The processor involved in each embodiment provided in this application may be a general-purpose processor, a central processing unit, a graphics processor, a digital signal processor, a programmable logic device, a data processing logic device based on quantum computing, etc., but are not limited to this.

以上实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。The technical features of the above embodiments may be arbitrarily combined. To make the description concise, not all possible combinations of the technical features in the above embodiments are described. However, as long as there is no contradiction in the combination of these technical features, they should be considered to be within the scope of this specification.

以上实施例仅表达了本申请的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对本申请专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本申请构思的前提下,还可以做出若干变形和改进,这些都属于本申请的保护范围。因此,本申请的保护范围应以所附权利要求为准。The above embodiments only express several implementation methods of the present application, and the descriptions thereof are relatively specific and detailed, but they cannot be understood as limiting the scope of the present application. It should be pointed out that, for a person of ordinary skill in the art, several variations and improvements can be made without departing from the concept of the present application, and these all belong to the protection scope of the present application. Therefore, the protection scope of the present application shall be subject to the attached claims.

Claims (10)

1. A method of fault classification for a rotating device, the method comprising:
extracting characteristics of fault information of the rotary equipment to be detected to obtain fault characteristics of the fault information;
Inputting fault characteristics into a target multi-scale multi-head self-attention force diagram convolutional neural network GCN model to obtain fault types of the rotating equipment to be tested; the target multi-scale multi-head self-attention force-seeking convolutional neural network GCN model is obtained by training an initial multi-scale multi-head self-attention GCN model based on a fusion graph feature sample, wherein the fusion graph feature sample is obtained by fusing target output information and a plurality of graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on a fault information sample, and the graph feature sample is determined by the fault information sample, an adjacent matrix and the initial multi-scale multi-head self-attention GCN model.
2. The method of claim 1, wherein the initial multi-scale multi-headed self-attention GCN model comprises a plurality of first sub-models and a multi-layer perceptron; the method further comprises the steps of:
Inputting fault information samples and adjacent matrixes corresponding to the first sub-models into the first sub-models aiming at the first sub-models to obtain graph feature samples output by the first sub-models;
inputting the fault information sample into a multi-layer sensor to obtain first output information; the target output information includes first output information;
Obtaining a fusion graph characteristic sample according to the first output information and each graph characteristic sample;
training the initial multi-scale multi-head self-attention GCN model according to the fusion map feature sample to obtain a target multi-scale multi-head self-attention force diagram convolutional neural network GCN model.
3. The method of claim 2, wherein the first sub-model comprises a first graph convolution layer and a second graph convolution layer, and the initial multi-scale multi-headed self-attention GCN model comprises a connection layer and a linear layer; obtaining a fusion graph feature sample according to the first output information and each graph feature sample, including:
Aiming at each first sub-model, inputting a fault information sample and an adjacent matrix corresponding to each first sub-model into a first graph convolution layer of the first sub-model to obtain a first intermediate graph characteristic;
Inputting first intermediate graph features obtained by the first graph convolution layers of the first sub-models into the connecting layer to obtain second intermediate graph features;
inputting the second intermediate graph characteristics into a linear layer to obtain second output information;
obtaining a fusion graph characteristic sample according to the first output information, the second output information and each graph characteristic sample; the target output information includes first output information and second output information.
4. A method according to claim 3, wherein a fused graph feature sample is obtained from the first output information, the second output information and each graph feature sample; the target output information includes first output information and second output information, including:
Obtaining the first output information, the second output information and the weighted value of each graph feature sample according to the first output information, the second output information, the first weight corresponding to the first output information, the second weight corresponding to the second output information and the third weight corresponding to each graph feature sample;
and obtaining a fusion map feature sample based on the weighted value.
5. The method according to any one of claims 2-4, wherein the number of fusion map feature samples is a plurality; the initial multi-scale multi-head self-attention GCN model comprises a multi-head self-attention fusion model and a full connection layer; training an initial multi-scale multi-head self-attention GCN model according to the fusion map feature sample to obtain a target multi-scale multi-head self-attention force diagram convolutional neural network GCN model, wherein the method comprises the following steps:
Connecting the fusion map feature samples by utilizing a multi-head self-attention fusion model to obtain map feature vectors;
obtaining a predicted fault category based on the graph feature vector by using the full connection layer;
And training the initial multi-scale multi-head self-attention GCN model according to the error between the predicted fault category and the real fault category corresponding to the fault information sample to obtain a target multi-scale multi-head self-attention force diagram convolutional neural network GCN model.
6. The method of claim 5, wherein training the initial multi-scale multi-headed self-attention GCN model to obtain the target multi-scale multi-headed self-attention seeking convolutional neural network GCN model based on errors between the predicted fault category and the true fault category corresponding to the fault information samples, comprises:
Determining a loss value according to an error between the predicted fault category and a real fault category corresponding to the fault information sample;
Under the condition that the loss value is larger than a preset threshold value, adjusting parameters of the initial multi-scale multi-head self-attention GCN model according to the loss value to obtain an intermediate GCN model, training the initial multi-scale multi-head self-attention GCN model until the loss value is smaller than the preset threshold value, and taking the intermediate GCN model corresponding to the loss value smaller than the preset threshold value as a target multi-scale multi-head self-attention power seeking convolutional neural network GCN model.
7. A fault classification device for a rotating apparatus, the device comprising:
the feature extraction module is used for extracting features of fault information of the rotary equipment to be detected to obtain fault features of the fault information;
The fault classification module is used for inputting fault characteristics into a target multi-scale multi-head self-attention force diagram convolutional neural network GCN model to obtain fault types of the rotating equipment to be detected; the target multi-scale multi-head self-attention force-seeking convolutional neural network GCN model is obtained by training an initial multi-scale multi-head self-attention GCN model based on a fusion graph feature sample, wherein the fusion graph feature sample is obtained by fusing target output information and a plurality of graph feature samples, the target output information is determined by the initial multi-scale multi-head self-attention GCN model based on a fault information sample, and the graph feature sample is determined by the fault information sample, an adjacent matrix and the initial multi-scale multi-head self-attention GCN model.
8. A computer device comprising a memory and a processor, the memory storing a computer program, characterized in that the processor implements the steps of the method of any one of claims 1 to 6 when the computer program is executed.
9. A computer readable storage medium, on which a computer program is stored, characterized in that the computer program, when being executed by a processor, carries out the steps of the method of any one of claims 1 to 6.
10. A computer program product comprising a computer program, characterized in that the computer program, when being executed by a processor, realizes the steps of the method of any one of claims 1 to 6.
CN202410173763.XA 2024-02-07 2024-02-07 Method, device, equipment, storage medium and product for fault classification of rotary equipment Pending CN118035889A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410173763.XA CN118035889A (en) 2024-02-07 2024-02-07 Method, device, equipment, storage medium and product for fault classification of rotary equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410173763.XA CN118035889A (en) 2024-02-07 2024-02-07 Method, device, equipment, storage medium and product for fault classification of rotary equipment

Publications (1)

Publication Number Publication Date
CN118035889A true CN118035889A (en) 2024-05-14

Family

ID=90983614

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202410173763.XA Pending CN118035889A (en) 2024-02-07 2024-02-07 Method, device, equipment, storage medium and product for fault classification of rotary equipment

Country Status (1)

Country Link
CN (1) CN118035889A (en)

Similar Documents

Publication Publication Date Title
CN113887845B (en) Extreme event prediction method, device, equipment and storage medium
CN116304885B (en) A method, device and equipment for event recognition based on graph node embedding
CN117033039A (en) Fault detection method, device, computer equipment and storage medium
CN116452333A (en) Construction method of abnormal transaction detection model, abnormal transaction detection method and device
CN115827977A (en) Training method and device of user preference prediction model and computer equipment
CN118035889A (en) Method, device, equipment, storage medium and product for fault classification of rotary equipment
CN116894802A (en) Image enhancement method, device, computer equipment and storage medium
CN116127195A (en) Package recommendation method, device, computer equipment, storage medium and program product
CN114648678B (en) Challenge sample detection method, apparatus, computer device, and storage medium
CN116579669B (en) Reliability assessment method, device, computer equipment and storage medium thereof
CN116011236B (en) Control element rapid modeling method for solving electric power time domain simulation differential algebraic equation
CN117829964A (en) A sequence recommendation method, device and medium based on graph neural network
CN119518770A (en) Power generation prediction method, device, computer equipment and storage medium
CN119228121A (en) Model weight information updating method, apparatus, computer device, readable storage medium, and program product
CN115630973A (en) User data processing method, device, computer equipment and storage medium
CN118916627A (en) State prediction method, state prediction apparatus, computer device, storage medium, and computer program product
CN117559433A (en) Transient stability prediction method and device for power system and computer equipment
CN114491291A (en) User preference prediction method, terminal and storage medium
CN115496185A (en) Research on Graph Enhanced Capsule Network for Serialized Recommendation
CN116245637A (en) Variable generation method, device, computer equipment and storage medium
CN118839070A (en) Determination method, application method and related device of information propagation prediction model
CN117436972A (en) Resource object recommendation method, device, computer equipment and storage medium
CN117437994A (en) Prediction methods, devices, equipment and storage media for dynamic properties of amorphous alloys
CN119416022A (en) Health state assessment method, device, computer equipment, medium and product of electrochemical energy storage system
CN117976238A (en) Infectious disease trend prediction method, infectious disease trend prediction device, infectious disease trend prediction computer device, and infectious disease trend prediction storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination