WO2023159336A1 - Deep autoregressive network based prediction method for stalling and surging of axial-flow compressor - Google Patents

Deep autoregressive network based prediction method for stalling and surging of axial-flow compressor Download PDF

Info

Publication number
WO2023159336A1
WO2023159336A1 PCT/CN2022/077168 CN2022077168W WO2023159336A1 WO 2023159336 A1 WO2023159336 A1 WO 2023159336A1 CN 2022077168 W CN2022077168 W CN 2022077168W WO 2023159336 A1 WO2023159336 A1 WO 2023159336A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
data
prediction
layer
deepar
Prior art date
Application number
PCT/CN2022/077168
Other languages
French (fr)
Chinese (zh)
Inventor
李英顺
弓子勤
孙希明
全福祥
Original Assignee
大连理工大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 大连理工大学 filed Critical 大连理工大学
Priority to PCT/CN2022/077168 priority Critical patent/WO2023159336A1/en
Publication of WO2023159336A1 publication Critical patent/WO2023159336A1/en

Links

Images

Classifications

    • FMECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
    • F04POSITIVE - DISPLACEMENT MACHINES FOR LIQUIDS; PUMPS FOR LIQUIDS OR ELASTIC FLUIDS
    • F04DNON-POSITIVE-DISPLACEMENT PUMPS
    • F04D27/00Control, e.g. regulation, of pumps, pumping installations or pumping systems specially adapted for elastic fluids
    • F04D27/001Testing thereof; Determination or simulation of flow characteristics; Stall or surge detection, e.g. condition monitoring
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T90/00Enabling technologies or technologies with a potential or indirect contribution to GHG emissions mitigation

Definitions

  • the invention relates to a method for predicting the stall and surge probability of an axial flow compressor based on a deep autoregressive network based on an attention mechanism, and belongs to the technical field of aeroengine modeling and simulation.
  • Aeroengines are the "crown jewels" in the history of human industry, reflecting the highest level of technology in a country.
  • the compressor is a key component of a high-performance aero-engine. It increases the air pressure through the high-speed rotation of the blades and limits the stable working range of the engine while providing a high-pressure ratio. It plays a vital role in the stability and safety of the aero-engine. Surge and rotating stall are two important manifestations of compressor gas flow instability faults.
  • the main feature of compressor surge is the occurrence of airflow interruption.
  • the airflow oscillates along the axis of the compressor with low frequency (several Hz or more than a dozen Hz) and high amplitude. In severe cases, flow blockage or even reverse flow occurs. Once the surge occurs, it will cause very serious damage to the aero-engine.
  • Rotating stall is an unstable flow phenomenon that can significantly degrade aeroengine performance. A large number of studies have shown that rotating stall is the precursor of surge, and surge is the consequence of the extreme development of rotating stall. Therefore, fast and accurate prediction of rotating stall has become an urgent problem to be solved in the field of aeroengines.
  • the traditional algorithms based on the time-domain characteristics of the pressure signal mainly include: short-term energy method, autocorrelation function method, variance analysis method, rate of change method, pressure difference method, statistical characteristic method, etc., traditional surge algorithms based on the frequency-domain characteristics of the pressure signal
  • the detection algorithms mainly include: spectrum analysis method, wavelet analysis method, frequency domain amplitude method and so on.
  • the present invention provides an axial compressor stall surge probability prediction based on an attention-based deep autoregressive network (TPA-DeepAR, Temporal Pattern Attention DeepAutoregressive Recurrent Networks) method.
  • TPA-DeepAR Temporal Pattern Attention DeepAutoregressive Recurrent Networks
  • a method for predicting stall and surge of an axial flow compressor with a deep autoregressive network specifically a method for predicting stall and surge of an axial flow compressor with a deep autoregressive network based on an attention mechanism, comprising the following steps:
  • the aero-engine surge data is preprocessed, including the following steps:
  • test data set is divided into a training data set
  • the Gaussian layer is composed of two fully connected layers, and the hidden vector output by the attention layer
  • the output of the two fully connected layers of the Gaussian layer is the parameter ⁇ and the parameter ⁇ respectively, so the output of the Gaussian layer will determine a Gaussian distribution, so that the model achieves the purpose of fitting the Gaussian distribution;
  • the model outputs the parameters ⁇ and ⁇ of the predicted Gaussian distribution during forward propagation.
  • the traditional loss function for regression cannot handle ⁇ , ⁇ , and y_true (the real label of the sample). Therefore, the loss function adopted is as follows:
  • n represents the number of samples
  • y_true is known, and represents the real label of the sample
  • ⁇ and ⁇ are the parameters of the Gaussian distribution predicted by the model
  • the likelihood function describes the distribution formed by the parameters ⁇ and ⁇
  • y_true appears The size of the probability of the sample point.
  • the network parameters are learned by maximizing the log likelihood function, that is, the distribution formed by the parameters ⁇ and ⁇ can have the maximum probability of the sample point y_true, and the loss function of the corresponding model training can be determined as -lnL( ⁇ , ⁇ 2 ) .
  • step S4.3 Use the preliminary prediction model to test on the verification set obtained in step S1, obtain the F2 evaluation index, adjust the parameters of the TPA-DeepAR model according to the F2 index, confusion matrix and ROC curve to achieve better performance, and save the performance of each evaluation index Optimal TPA-DeepAR prediction model;
  • P is the precision rate (precision), which indicates the proportion of the samples that are classified as positive classes that are actually positive classes:
  • TP is the number of true cases
  • FP is the number of false positive cases
  • R is the recall rate (recall), which means that among all the samples that are actually positive classes, the proportion that is correctly judged as positive class:
  • FN is the number of false negative cases.
  • the first to fourth quadrants of the table are TP, FP, FN, and TN respectively.
  • TN is the number of true negative cases.
  • the prediction method provided by the present invention learns the time-correlation characteristics of the compressor dynamic pressure experimental data, captures the small stall precursor signal, calculates and outputs the prediction probability of surge, and gives a warning signal in time whether the surge occurs.
  • the prediction method uses the attention mechanism to select relevant dimensions for attention weighting, which can effectively capture the characteristics of the experimental data to achieve accurate prediction of the probability of surge, and improve the prediction stability and accuracy.
  • the method outputs multiple quantiles of predicted probability, which is convenient for the system to carry out early warning according to different quantiles.
  • the method can judge whether the surge occurs according to the surge probability output in real time, and feed back to the engine control system in time, so as to adjust the engine running state and gain time for the active control method of the compressor.
  • Figure 1 is a flow chart of an axial compressor stall surge prediction method based on an attention mechanism based on a deep autoregressive network
  • Fig. 2 is a flow chart of data preprocessing
  • FIG. 3 is a structural diagram of the TPA-DeepAR model
  • Figure 4 is a structural diagram of the attention mechanism
  • Figure 5 is the prediction results of the TPA-DeepAR model on the test data, in which (a) is the dynamic pressure p2 at the tip of the second-stage stator changing with time, and (b) is the surge prediction probability given by the TPA-DeepAR model The change graph over time, (c) is the early warning signal given by the TPA-DeepAR model;
  • the background of the present invention is the surge experimental data of a certain type of aeroengine, and the process flow of the axial compressor stall surge prediction method based on the deep autoregressive network of the attention mechanism is shown in FIG. 1 .
  • FIG. 2 is a flow chart of data preprocessing, and the steps of data preprocessing are as follows:
  • test data set is divided into a training data set
  • FIG. 3 is a structural diagram of the TPA-DeepAR model.
  • the Gaussian layer is composed of two fully connected layers, and the hidden vector As the input of the Gaussian layer, the outputs of the two fully connected layers are parameter ⁇ and parameter ⁇ respectively, and the output of the Gaussian layer will determine a Gaussian distribution, so that the model achieves the purpose of fitting the Gaussian distribution;
  • Figure 4 is a structural diagram of the attention layer
  • the hidden vector ⁇ h t-w+1, h t-w+2 ,...,h t ⁇ of each time step of the sample is obtained, and the hidden vector of each hidden vector
  • the row vector of the hidden state matrix represents the state of a single dimension at all time steps, that is, the vector composed of all time steps of the same dimension.
  • the column vector of the hidden state matrix represents the state of a single time step, that is, the vector composed of all dimensions at the same time step.
  • W a is the weight.
  • h t and v t are spliced and input into a fully connected layer to obtain a new hidden vector as output;
  • W h and W v are weights.
  • n represents the number of samples
  • y_true is known, and represents the real label of the sample
  • ⁇ and ⁇ are the parameters of the Gaussian distribution predicted by the model
  • the likelihood function describes the distribution formed by the parameters ⁇ and ⁇
  • the sample y_true appears The size of the probability of a point.
  • the network parameters are learned by maximizing the log likelihood function, that is, the distribution formed by the parameters ⁇ and ⁇ can have the maximum probability of the sample point y_true, and the loss function of the corresponding model training can be determined as -lnL( ⁇ , ⁇ 2 ) .
  • step S4.3 Use the preliminary prediction model to test on the verification set obtained in step S1, obtain the F2 evaluation index, adjust the parameters of the TPA-DeepAR model according to the F2 index, confusion matrix and ROC curve to achieve better performance, and save the performance of each evaluation index Optimal TPA-DeepAR prediction model;
  • P is the precision rate (precision), which indicates the proportion of the samples that are classified as positive classes that are actually positive classes:
  • TP is the number of true cases
  • FP is the number of false positive cases
  • R is the recall rate (recall), which means that among all the samples that are actually positive classes, the proportion that is correctly judged as positive class:
  • FN is the number of false negative cases.
  • the first to fourth quadrants of the table are TP, FP, FN, and TN respectively.
  • TN is the number of true negative cases.
  • Figure 5 is the prediction result of the TPA-DeepAR prediction model on the test data, where (a) is the change of the dynamic pressure p2 at the tip of the second-stage stator with time Figure (b) is the change of surge prediction probability over time given by the TPA-DeepAR prediction model, and (c) is the early warning signal given by the TPA-DeepAR prediction model based on the prediction probability.
  • the steps to perform real-time prediction on test data are as follows:
  • test set data is the dynamic pressure data of the tip position of the secondary stator, from It can be seen from Figure (a) that a downward-developing protrusion began to appear at 7.48s, which was in the initial disturbance stage of the stall. With the development of the stall disturbance, it began to fluctuate violently at 7.826s, and completely developed into a stall surge .

Abstract

The present invention relates to the technical field of aeroengine modeling and simulation, and provides a deep autoregressive network based prediction method for stalling and surging of an axial-flow compressor. The method comprises: first using surge experiment data of a certain type of an aeroengine, selecting and preprocessing the data, and dividing the data into a training set and a test set; then building and training a deep autoregressive network model based on an attention mechanism, performing real-time prediction on the test set by using the finally trained model, and giving model loss and evaluation indexes; and finally, performing real-time prediction on test data by using a prediction model, and giving a change trend of the surge probability along with time according to a time sequence. According to the present invention, the attention mechanism is used to effectively capture the characteristics of the experimental data to accurately predict the surge probability, so that the prediction stability and accuracy can be improved; and the active control performance of the engine can be improved, and certain universality is achieved.

Description

一种深度自回归网络的轴流压气机失速喘振预测方法A Deep Autoregressive Network Based Axial Compressor Stall Surge Prediction Method 技术领域technical field
本发明涉及一种基于注意力机制的深度自回归网络的轴流压气机失速喘振概率预测方法,属于航空发动机建模与仿真技术领域。The invention relates to a method for predicting the stall and surge probability of an axial flow compressor based on a deep autoregressive network based on an attention mechanism, and belongs to the technical field of aeroengine modeling and simulation.
背景技术Background technique
航空发动机是人类工业史上“皇冠上的明珠”,体现了一个国家的科技最高水平。压气机是高性能航空发动机的关键部件,它通过叶片高速旋转提高空气压力并且在提供高压比的同时也限制了发动机的稳定工作范围,它对于航空发动机的稳定性和安全性起着至关重要的作用,喘振和旋转失速是压气机气体流动不稳定故障的两种重要表现形式。Aeroengines are the "crown jewels" in the history of human industry, reflecting the highest level of technology in a country. The compressor is a key component of a high-performance aero-engine. It increases the air pressure through the high-speed rotation of the blades and limits the stable working range of the engine while providing a high-pressure ratio. It plays a vital role in the stability and safety of the aero-engine. Surge and rotating stall are two important manifestations of compressor gas flow instability faults.
压气机喘振的主要特征是产生气流中断现象,气流沿压气机轴线方向发生低频率(几赫兹或十几赫兹)、高振幅的振荡,严重时发生流动阻塞甚至倒流。喘振一旦发生,会对航空发动机产生非常严重的损害。旋转失速是一种不稳定流动现象,它会显著降低航空发动机性能。大量研究表明,旋转失速是喘振的先兆,喘振是旋转失速极度发展的后果,因此对旋转失速进行快速准确的预测成为航空发动机领域要迫切解决的难题。The main feature of compressor surge is the occurrence of airflow interruption. The airflow oscillates along the axis of the compressor with low frequency (several Hz or more than a dozen Hz) and high amplitude. In severe cases, flow blockage or even reverse flow occurs. Once the surge occurs, it will cause very serious damage to the aero-engine. Rotating stall is an unstable flow phenomenon that can significantly degrade aeroengine performance. A large number of studies have shown that rotating stall is the precursor of surge, and surge is the consequence of the extreme development of rotating stall. Therefore, fast and accurate prediction of rotating stall has become an urgent problem to be solved in the field of aeroengines.
当前,国内外的压气机旋转失速故障检测和判别方法有两种:一种是通过建立模型的方法,对压气机进行主动控制,当压气机出现喘振先兆时抑制压气机的扰动继续发生,防止进入喘振状态。第二种是根据压气机压力信号的时域特征或频域特征进行喘振预测算法研究。其中基于压力信号时域特征的传统算法主要有:短时能量法、自相关函数法、方差分析法、变化率法、压差法、统计特征法等,基于压力信号频域特征的传统喘振检测算法主要有:频谱分析法、小波分析法、频域幅值法等。At present, there are two methods for the detection and identification of compressor rotating stall faults at home and abroad: one is to actively control the compressor through the method of building a model, and suppress the disturbance of the compressor from continuing to occur when the compressor has a precursor to surge; Prevent from entering a surge state. The second is to study the surge prediction algorithm based on the time-domain or frequency-domain characteristics of the compressor pressure signal. Among them, the traditional algorithms based on the time-domain characteristics of the pressure signal mainly include: short-term energy method, autocorrelation function method, variance analysis method, rate of change method, pressure difference method, statistical characteristic method, etc., traditional surge algorithms based on the frequency-domain characteristics of the pressure signal The detection algorithms mainly include: spectrum analysis method, wavelet analysis method, frequency domain amplitude method and so on.
发明内容Contents of the invention
针对现有技术中准确性低,可靠性差的问题,本发明提供一种基于注意力机制的深度自回归网络(TPA-DeepAR,Temporal Pattern Attention DeepAutoregressive Recurrent Networks)的轴流压气机失速喘振概率预测方法。Aiming at the problems of low accuracy and poor reliability in the prior art, the present invention provides an axial compressor stall surge probability prediction based on an attention-based deep autoregressive network (TPA-DeepAR, Temporal Pattern Attention DeepAutoregressive Recurrent Networks) method.
为了达到上述目的,本发明采用的技术方案如下:In order to achieve the above object, the technical scheme adopted in the present invention is as follows:
一种深度自回归网络的轴流压气机失速喘振预测方法,具体为一种基于注意力机制的深度自回归网络的轴流压气机失速喘振预测方法,包括以下步骤:A method for predicting stall and surge of an axial flow compressor with a deep autoregressive network, specifically a method for predicting stall and surge of an axial flow compressor with a deep autoregressive network based on an attention mechanism, comprising the following steps:
S1.对航空发动机喘振数据进行预处理,包括以下步骤:S1. The aero-engine surge data is preprocessed, including the following steps:
S1.1获取某型号航空发动机喘振实验数据,剔除实验数据中由于传感器故障产生的无效数据;S1.1 Acquire the experimental data of a certain type of aero-engine surge, and eliminate the invalid data due to sensor failure in the experimental data;
S1.2对剩余有效数据依次进行降采样处理和滤波处理;S1.2 Perform down-sampling processing and filtering processing on the remaining valid data in sequence;
S1.3对滤波处理后的数据进行归一化、平滑化处理;S1.3 normalize and smooth the filtered data;
S1.4为保证测试结果的客观性,将实验数据划分为测试数据集和训练数据集;S1.4 In order to ensure the objectivity of the test results, the experimental data is divided into a test data set and a training data set;
S1.5通过时间窗切分训练数据集,每个时间窗覆盖的数据点组成一个样本,并将训练数据集按4:1的比例划分为训练集和验证集;S1.5 Segment the training data set by time windows, the data points covered by each time window form a sample, and divide the training data set into a training set and a verification set in a ratio of 4:1;
S2.构建基于注意力机制的深度自回归网络模型,即TPA-DeepAR模型,包括以下步骤:S2. Construct a deep autoregressive network model based on the attention mechanism, that is, the TPA-DeepAR model, including the following steps:
S2.1将每个样本维度调整为(w,1),作为TPA-DeepAR模型的输入,其中w代表时间窗长度;S2.1 Adjust each sample dimension to (w,1) as the input of the TPA-DeepAR model, where w represents the length of the time window;
S2.2搭建嵌入层将输入样本的维度从(w,1)转换为(w,m),m为指定的维度,将样本的特征从一维分散到m个维度;S2.2 Build an embedding layer to convert the dimension of the input sample from (w, 1) to (w, m), m is the specified dimension, and disperse the characteristics of the sample from one dimension to m dimensions;
S2.3搭建LSTM层,将嵌入层的输出作为LSTM层的输入,LSTM层输出w个隐藏向量{h t-w+1,h t-w+2,……,h t},每一个隐藏向量的维度为m。 S2.3 Build the LSTM layer, use the output of the embedding layer as the input of the LSTM layer, and the LSTM layer outputs w hidden vectors {h t-w+1, h t-w+2 ,...,h t }, each hidden The dimension of the vector is m.
S2.4搭建注意力层,LSTM层输出的w个隐藏向量{h t-w+1,h t-w+2,……,h t}作为注意力层的输入,经过注意力层对相关维度加权,最终输出一个隐藏向量
Figure PCTCN2022077168-appb-000001
S2.4 Build the attention layer. The w hidden vectors {h t-w+1, h t-w+2 ,...,h t } output by the LSTM layer are used as the input of the attention layer, and the relevant Dimensionally weighted, and finally output a hidden vector
Figure PCTCN2022077168-appb-000001
S2.5搭建高斯层,高斯层由两个全连接层组成,将注意力层输出的隐藏向量
Figure PCTCN2022077168-appb-000002
作为高斯层的输入,高斯层的两个全连接层的输出分别为参数μ和参数σ,因此高斯层的输出会确定一个高斯分布,这样模型就实现了拟合高斯分布的目的;
S2.5 Build a Gaussian layer, the Gaussian layer is composed of two fully connected layers, and the hidden vector output by the attention layer
Figure PCTCN2022077168-appb-000002
As the input of the Gaussian layer, the output of the two fully connected layers of the Gaussian layer is the parameter μ and the parameter σ respectively, so the output of the Gaussian layer will determine a Gaussian distribution, so that the model achieves the purpose of fitting the Gaussian distribution;
S2.6采用拟合的高斯分布进行多次随机采样,得到预测点的数据,并依据这些采样点得到预测点的不同分位数以实现概率预测;S2.6 Use the fitted Gaussian distribution to perform random sampling multiple times to obtain the data of the prediction points, and obtain different quantiles of the prediction points based on these sampling points to realize probability prediction;
S3.构建S2中提到的注意力层:S3. Build the attention layer mentioned in S2:
S3.1注意力层的输入为LSTM层的输出{h t-w+1,h t-w+2,……,h t},输入数据的维度为(w,m),除了最后一个隐藏向量ht以外,将其他w-1个隐藏向量组成隐状态矩阵H={h t-w+1,h t-w+2,……,h t-1}; S3.1 The input of the attention layer is the output of the LSTM layer {h t-w+1, h t-w+2 ,...,h t }, the dimension of the input data is (w, m), except for the last hidden Except for the vector ht, other w-1 hidden vectors form a hidden state matrix H={h t-w+1, h t-w+2 ,...,h t-1 };
S3.2采用k个卷积核捕捉H的信号模式得到H C矩阵,增强模型对特征的学习能力。 S3.2 Use k convolution kernels to capture the signal pattern of H to obtain the H C matrix to enhance the model's ability to learn features.
S3.3隐藏向量h t与H C矩阵通过得分函数进行相似度计算得到注意力权重α i,利用注意力权重α i对H C每一行加权求和,得到向量v tS3.3 Calculate the similarity between the hidden vector h t and the H C matrix through the scoring function to obtain the attention weight α i , use the attention weight α i to weight and sum each row of H C to obtain the vector v t ;
S3.4最后将h t和v t拼接,输入一个全接连层得到一个新的隐藏向量输出
Figure PCTCN2022077168-appb-000003
S3.4 Finally, h t and v t are spliced and input into a fully connected layer to obtain a new hidden vector output
Figure PCTCN2022077168-appb-000003
S4.TPA-DeepAR模型损失函数及评价指标:S4.TPA-DeepAR model loss function and evaluation indicators:
S4.1针对TPA-DeepAR模型,模型在前向传播时输出的是预测高斯分布的参数μ和σ,传统的用于回归的损失函数无法处理μ,σ,y_true(样本的真实标签)三者的关系,因此采用的损失函数具体如下:S4.1 For the TPA-DeepAR model, the model outputs the parameters μ and σ of the predicted Gaussian distribution during forward propagation. The traditional loss function for regression cannot handle μ, σ, and y_true (the real label of the sample). Therefore, the loss function adopted is as follows:
假设样本服从高斯分布y_true~(μ,σ 2),则其似然函数为: Assuming that the sample obeys the Gaussian distribution y_true~(μ, σ 2 ), then its likelihood function is:
Figure PCTCN2022077168-appb-000004
Figure PCTCN2022077168-appb-000004
其对数似然函数为:Its log-likelihood function is:
Figure PCTCN2022077168-appb-000005
Figure PCTCN2022077168-appb-000005
其中,n表示样本个数,y_true是已知的,表示样本的真实标签,μ和σ是模型预测的高斯分布的参数,似然函数描述的是对于参数μ和σ形成的分布,出现y_true这个样本点的概率的大小。Among them, n represents the number of samples, y_true is known, and represents the real label of the sample, μ and σ are the parameters of the Gaussian distribution predicted by the model, and the likelihood function describes the distribution formed by the parameters μ and σ, and y_true appears The size of the probability of the sample point.
因此通过最大化对数似然函数来学习网络参数,即参数μ和σ形成的分布可以最大概率的出现y_true这个样本点,相应的模型训练的损失函数可以确定为-lnL(μ,σ 2)。 Therefore, the network parameters are learned by maximizing the log likelihood function, that is, the distribution formed by the parameters μ and σ can have the maximum probability of the sample point y_true, and the loss function of the corresponding model training can be determined as -lnL(μ,σ 2 ) .
S4.2基于损失函数,在步骤S1得到的训练集上对TPA-DeepAR模型进行权重更新,最终生成模型的初步预测模型。S4.2 Based on the loss function, update the weight of the TPA-DeepAR model on the training set obtained in step S1, and finally generate a preliminary prediction model of the model.
S4.3采用初步预测模型在步骤S1得到的验证集上进行测试,获取F2评价指标,根据F2指标,混淆矩阵以及ROC曲线调整TPA-DeepAR模型参数,以达到更优,保存各项评价指标表现最优的TPA-DeepAR预测模型;S4.3 Use the preliminary prediction model to test on the verification set obtained in step S1, obtain the F2 evaluation index, adjust the parameters of the TPA-DeepAR model according to the F2 index, confusion matrix and ROC curve to achieve better performance, and save the performance of each evaluation index Optimal TPA-DeepAR prediction model;
其中,所述的F2指标为:Among them, the F2 index mentioned is:
Figure PCTCN2022077168-appb-000006
Figure PCTCN2022077168-appb-000006
其中,P为精确率(precision),表示被分为正类的样本中实际为正类的比例:
Figure PCTCN2022077168-appb-000007
其中,TP为真正例数,FP为假正例数,R为召回率(recall),表示在所有实际为正类的样本中,被正确地判断为正类的比例:
Figure PCTCN2022077168-appb-000008
其中,FN为假负例数。
Among them, P is the precision rate (precision), which indicates the proportion of the samples that are classified as positive classes that are actually positive classes:
Figure PCTCN2022077168-appb-000007
Among them, TP is the number of true cases, FP is the number of false positive cases, and R is the recall rate (recall), which means that among all the samples that are actually positive classes, the proportion that is correctly judged as positive class:
Figure PCTCN2022077168-appb-000008
Among them, FN is the number of false negative cases.
将TP,FP,TN,FN四个指标一起呈现在2*2表格中,就会得到混淆矩阵,表格的第一象限到第四象限分别为TP,FP,FN,TN。其中,TN为真负例数。Present the four indicators of TP, FP, TN, and FN together in a 2*2 table, and a confusion matrix will be obtained. The first to fourth quadrants of the table are TP, FP, FN, and TN respectively. Among them, TN is the number of true negative cases.
得到混淆矩阵后,矩阵第二象限和第四象限的数值越大越好,反之,第一象限和第三象限的数值越小越好。After obtaining the confusion matrix, the larger the values in the second and fourth quadrants of the matrix, the better. Conversely, the smaller the values in the first and third quadrants, the better.
在所有实际为负例的样本中,被错误地判断为正例的比例为FPR:FPR=FP/(FP+TN)。以FPR为横轴,R为纵轴,得到ROC曲线。所述的ROC曲线越靠近左上角,TPA-DeepAR模型的召回率越高,其假正例和假反例总数最少,预测效果越好。Among all the samples that are actually negative examples, the proportion that is wrongly judged as positive examples is FPR: FPR=FP/(FP+TN). With FPR as the horizontal axis and R as the vertical axis, the ROC curve is obtained. The closer the ROC curve is to the upper left corner, the higher the recall rate of the TPA-DeepAR model, the least number of false positives and false negatives, and the better the prediction effect.
S5.采用最终TPA-DeepAR预测模型对测试集进行实时预测:S5. Use the final TPA-DeepAR prediction model to make real-time predictions on the test set:
S5.1.按照预处理的步骤对测试集数据进行预处理,调整测试集数据维度后输入到已经训练好的TPA-DeepAR模型中进行测试;S5.1. Preprocess the test set data according to the preprocessing steps, adjust the dimension of the test set data and input it into the trained TPA-DeepAR model for testing;
S5.2.依照时间顺序,用TPA-DeepAR预测模型给出每个测试集样本的喘振预测概率,得到测试集样本的实时喘振概率。S5.2. According to the time sequence, use the TPA-DeepAR prediction model to give the surge prediction probability of each test set sample, and obtain the real-time surge probability of the test set sample.
本发明的有益效果为:The beneficial effects of the present invention are:
本发明所提供的预测方法对压气机动压实验数据进行时间相关性特征的学习,捕捉其中微小失速先兆信号,计算输出喘振预测概率,并及时给出喘振是否发生的警示信号。与传统方法对比,该预测方法采用注意力机制选择相关维度进行注意力加权,能有效的捕捉实验数据的特征实现对喘振概率的准确预测,提高了预测稳定性和精确度。同时该方法输出预测概率的多个分位数,方便系统根据不同的分位数进行预警。该方法可以根据实时输出的喘振概率判断喘振是否发生,及时反馈给发动机控制系统,从而调整发动机运行状态,为压气机主动控制方法争取时间。The prediction method provided by the present invention learns the time-correlation characteristics of the compressor dynamic pressure experimental data, captures the small stall precursor signal, calculates and outputs the prediction probability of surge, and gives a warning signal in time whether the surge occurs. Compared with the traditional method, the prediction method uses the attention mechanism to select relevant dimensions for attention weighting, which can effectively capture the characteristics of the experimental data to achieve accurate prediction of the probability of surge, and improve the prediction stability and accuracy. At the same time, the method outputs multiple quantiles of predicted probability, which is convenient for the system to carry out early warning according to different quantiles. The method can judge whether the surge occurs according to the surge probability output in real time, and feed back to the engine control system in time, so as to adjust the engine running state and gain time for the active control method of the compressor.
附图说明Description of drawings
图1为基于注意力机制的深度自回归网络的轴流压气机失速喘振预测方法流程图;Figure 1 is a flow chart of an axial compressor stall surge prediction method based on an attention mechanism based on a deep autoregressive network;
图2为数据预处理流程图;Fig. 2 is a flow chart of data preprocessing;
图3为TPA-DeepAR模型结构图;Figure 3 is a structural diagram of the TPA-DeepAR model;
图4为注意力机制结构图;Figure 4 is a structural diagram of the attention mechanism;
图5为TPA-DeepAR模型在测试数据上的预测结果图,其中(a)为二级静子尖部动压p 2随时间变化图,(b)为TPA-DeepAR模型给出的喘振预测概率随时间的变化图,(c)为TPA-DeepAR模型给出的预警信号; Figure 5 is the prediction results of the TPA-DeepAR model on the test data, in which (a) is the dynamic pressure p2 at the tip of the second-stage stator changing with time, and (b) is the surge prediction probability given by the TPA-DeepAR model The change graph over time, (c) is the early warning signal given by the TPA-DeepAR model;
具体实施方式Detailed ways
下面结合附图对本发明作进一步说明,本发明依托背景为某型号航空发动机喘振实验数据,基于注意力机制的深度自回归网络的轴流压气机失速喘振预测方法流程如图1所示。The present invention will be further described below in conjunction with the accompanying drawings. The background of the present invention is the surge experimental data of a certain type of aeroengine, and the process flow of the axial compressor stall surge prediction method based on the deep autoregressive network of the attention mechanism is shown in FIG. 1 .
图2为数据预处理流程图,数据预处理步骤如下:Figure 2 is a flow chart of data preprocessing, and the steps of data preprocessing are as follows:
S1.对航空发动机喘振数据进行预处理。S1. Preprocessing the aero-engine surge data.
S1.1获取某型号航空发动机喘振实验数据,剔除实验数据中由于传感器故障产生的无效数据;实验数据共16组,每组实验包含10个测量点所测量的从正常到喘振共10s的动态压力数值,传感器测量频率为6kHz,10个测量点分别位于:进口导向叶片静子尖部、零级静子尖部、一级静子尖部(周向三个)、二级静子尖部、三级静子尖部、四级静子尖部、五级静子尖部、出口壁面;S1.1 Acquire the experimental data of a certain type of aeroengine surge, and eliminate the invalid data due to sensor failure in the experimental data; there are 16 groups of experimental data, each group of experiments includes 10 measurement points from normal to surge for a total of 10s Dynamic pressure value, sensor measurement frequency is 6kHz, 10 measurement points are located at: inlet guide vane stator tip, zero-stage stator tip, first-stage stator tip (three in the circumferential direction), second-stage stator tip, third-stage stator tip Ministry, the tip of the fourth-level stator, the tip of the fifth-level stator, and the exit wall;
S1.2对剩余有效数据依次进行降采样处理和滤波处理;S1.2 Perform down-sampling processing and filtering processing on the remaining valid data in sequence;
S1.3对滤波处理后的数据进行归一化、平滑化处理;S1.3 normalize and smooth the filtered data;
S1.4为保证测试结果的客观性,将实验数据划分为测试数据集和训练数据集;S1.4 In order to ensure the objectivity of the test results, the experimental data is divided into a test data set and a training data set;
S1.5通过时间窗切分训练数据集,每个时间窗覆盖的数据点组成一个样本,并将训练数据集按4:1的比例划分为训练集和验证集;S1.5 Segment the training data set by time windows, the data points covered by each time window form a sample, and divide the training data set into a training set and a verification set in a ratio of 4:1;
图3为TPA-DeepAR模型结构图。Figure 3 is a structural diagram of the TPA-DeepAR model.
S2.构建TPA-DeepAR模型的步骤如下:S2. The steps to construct the TPA-DeepAR model are as follows:
S2.1将每个样本维度调整为(w,1),作为TPA-DeepAR模型的输入,其中w代表时间窗长度;S2.1 Adjust each sample dimension to (w,1) as the input of the TPA-DeepAR model, where w represents the length of the time window;
S2.2搭建嵌入层将输入样本的维度从(w,1)转换为(w,m),m为指定的维度,将样本的特征从一维分散到m个维度;S2.2 Build an embedding layer to convert the dimension of the input sample from (w, 1) to (w, m), m is the specified dimension, and disperse the characteristics of the sample from one dimension to m dimensions;
S2.3搭建LSTM层,将嵌入层的输出作为LSTM层的输入,LSTM层输出w个隐藏向量{h t-w+1,h t-w+2,……,h t},每一个隐藏向量的维度为m; S2.3 Build the LSTM layer, use the output of the embedding layer as the input of the LSTM layer, and the LSTM layer outputs w hidden vectors {h t-w+1, h t-w+2 ,...,h t }, each hidden The dimension of the vector is m;
S2.4在最后一个时间步的隐藏向量h t输出后,添加注意力层,LSTM层输出的w个隐藏向量{h t-w+1,h t-w+2,……,h t}作为注意力层的输入,注意力层对这些隐藏向量的m个维度添加注意力,选择相关维度加权,更好的捕捉隐藏向量的特征,最终输出一个新的隐藏向量
Figure PCTCN2022077168-appb-000009
S2.4 After the hidden vector h t of the last time step is output, add an attention layer, w hidden vectors output by the LSTM layer {h t-w+1, h t-w+2 ,...,h t } As the input of the attention layer, the attention layer adds attention to the m dimensions of these hidden vectors, selects the relevant dimension weights, better captures the characteristics of the hidden vector, and finally outputs a new hidden vector
Figure PCTCN2022077168-appb-000009
S2.5搭建高斯层,高斯层由两个全连接层组成,将隐藏向量
Figure PCTCN2022077168-appb-000010
作为高斯层的输入,两个全连接层的输出分别为参数μ和参数σ,高斯层的输出会确定一个高斯分布,这样模型就实现了拟合高斯分布的目的;
S2.5 Build a Gaussian layer, the Gaussian layer is composed of two fully connected layers, and the hidden vector
Figure PCTCN2022077168-appb-000010
As the input of the Gaussian layer, the outputs of the two fully connected layers are parameter μ and parameter σ respectively, and the output of the Gaussian layer will determine a Gaussian distribution, so that the model achieves the purpose of fitting the Gaussian distribution;
S2.6采用拟合的高斯分布进行多次随机采样,得到预测点的数据,并依据这些采样点可以得到预测点的不同分位数以实现概率预测,本发明采用预测点的0.5分位数作为输出的喘振概率;S2.6 Use the fitted Gaussian distribution to carry out random sampling multiple times to obtain the data of the prediction points, and according to these sampling points, different quantiles of the prediction points can be obtained to realize the probability prediction. The present invention adopts the 0.5 quantile of the prediction points Surge probability as output;
图4为注意力层的结构图Figure 4 is a structural diagram of the attention layer
S3.构建注意力层的步骤如下:S3. The steps to construct the attention layer are as follows:
S3.1原始序列经过嵌入层和LSTM层处理后,得到样本每个时间步的隐藏向量{h t-w+1,h t-w+2,……,h t},每个隐藏向量的维度为m,除了最后一个隐藏向量h t以外,将其他w-1个隐藏向量组成隐状态矩阵H={h t-w+1,h t-w+2,……,h t-1}; S3.1 After the original sequence is processed by the embedding layer and the LSTM layer, the hidden vector {h t-w+1, h t-w+2 ,...,h t } of each time step of the sample is obtained, and the hidden vector of each hidden vector The dimension is m, except for the last hidden vector h t , the other w-1 hidden vectors form a hidden state matrix H={h t-w+1, h t-w+2 ,...,h t-1 } ;
隐状态矩阵的行向量代表单个维度在所有时间步下的状态,即同一维度的所有时间步构成的向量。The row vector of the hidden state matrix represents the state of a single dimension at all time steps, that is, the vector composed of all time steps of the same dimension.
隐状态矩阵的列向量代表单个时间步的状态,即同一时间步下所有维度构成的向量。The column vector of the hidden state matrix represents the state of a single time step, that is, the vector composed of all dimensions at the same time step.
S3.2利用卷积捕获可变的信号模式形成矩阵H CS3.2 Use convolution to capture variable signal patterns to form matrix H C ;
Figure PCTCN2022077168-appb-000011
Figure PCTCN2022077168-appb-000011
卷积配置为k个卷积核,w为时间窗长度,卷积核尺寸为1×T(T代表注意力所覆盖的范围,令T=w-1),将上述卷积核沿隐状态矩阵H的行向量计算卷积,提取该变量在该卷积核范围内的时间模式矩阵
Figure PCTCN2022077168-appb-000012
表示H矩阵的第i个行向量和第j个卷积核作用的结果值。
The convolution is configured as k convolution kernels, w is the length of the time window, and the size of the convolution kernel is 1×T (T represents the range covered by the attention, let T=w-1), the above convolution kernel along the hidden state The row vector calculation convolution of the matrix H extracts the time pattern matrix of the variable within the scope of the convolution kernel
Figure PCTCN2022077168-appb-000012
Indicates the result value of the i-th row vector of the H matrix and the j-th convolution kernel.
S3.3隐藏向量h t与H C矩阵通过得分函数(scoring function)进行相似度计算得到注意力权重α i,选择得分函数为: S3.3 The similarity between the hidden vector h t and the H C matrix is calculated through the scoring function (scoring function) to obtain the attention weight α i , and the scoring function is selected as:
Figure PCTCN2022077168-appb-000013
Figure PCTCN2022077168-appb-000013
其中,W a为权重。 Among them, W a is the weight.
利用sigmoid进行归一化,得到注意力权重α i,便于选择多维度: Use sigmoid for normalization to get the attention weight α i , which is convenient for selecting multiple dimensions:
Figure PCTCN2022077168-appb-000014
Figure PCTCN2022077168-appb-000014
最后利用注意力权重α i
Figure PCTCN2022077168-appb-000015
每一行加权求和,得到向量v t:
Finally, use the attention weight α i to
Figure PCTCN2022077168-appb-000015
Each row is weighted and summed to get the vector v t :
Figure PCTCN2022077168-appb-000016
Figure PCTCN2022077168-appb-000016
最后将h t和v t拼接,输入一个全接连层得到一个新的隐藏向量
Figure PCTCN2022077168-appb-000017
作为输出;
Finally, h t and v t are spliced and input into a fully connected layer to obtain a new hidden vector
Figure PCTCN2022077168-appb-000017
as output;
Figure PCTCN2022077168-appb-000018
Figure PCTCN2022077168-appb-000018
其中,W h和W v为权重。 Among them, W h and W v are weights.
S4.TPA-DeepAR模型损失函数及评价指标:S4.TPA-DeepAR model loss function and evaluation indicators:
S4.1针对TPA-DeepAR模型,模型在前向传播时输出的是预测高斯分布的μ和σ,传统的用于回归的损失函数无法处理μ,σ,y_true(样本的真实标签)三者的关系,因此采用的损失函数具体如下:S4.1 For the TPA-DeepAR model, the model outputs the predicted Gaussian distribution μ and σ during forward propagation, and the traditional loss function for regression cannot handle μ, σ, y_true (the real label of the sample) Relationship, so the loss function used is as follows:
假设样本服从高斯分布y_true~(μ,σ 2),则其似然函数为: Assuming that the sample obeys the Gaussian distribution y_true~(μ, σ 2 ), then its likelihood function is:
Figure PCTCN2022077168-appb-000019
Figure PCTCN2022077168-appb-000019
其对数似然函数为:Its log-likelihood function is:
Figure PCTCN2022077168-appb-000020
Figure PCTCN2022077168-appb-000020
其中n表示样本个数,y_true是已知的,表示样本的真实标签,μ和σ是模型预测的高斯分布的参数,似然函数描述的是对于参数μ和σ形成的分布,出现y_true这个样本点的概率的大小。Among them, n represents the number of samples, y_true is known, and represents the real label of the sample, μ and σ are the parameters of the Gaussian distribution predicted by the model, and the likelihood function describes the distribution formed by the parameters μ and σ, and the sample y_true appears The size of the probability of a point.
因此通过最大化对数似然函数来学习网络参数,即参数μ和σ形成的分布可以最大概率的出现y_true这个样本点,相应的模型训练的损失函数可以确定为-lnL(μ,σ 2)。 Therefore, the network parameters are learned by maximizing the log likelihood function, that is, the distribution formed by the parameters μ and σ can have the maximum probability of the sample point y_true, and the loss function of the corresponding model training can be determined as -lnL(μ,σ 2 ) .
S4.2基于损失函数,在步骤S1得到的训练集上对TPA-DeepAR模型进行权重更新,最终生成模型的初步预测模型。S4.2 Based on the loss function, update the weight of the TPA-DeepAR model on the training set obtained in step S1, and finally generate a preliminary prediction model of the model.
S4.3采用初步预测模型在步骤S1得到的验证集上进行测试,获取F2评价指标,根据F2指标,混淆矩阵以及ROC曲线调整TPA-DeepAR模型参数,以达到更优,保存各项评价指标表现最优的TPA-DeepAR预测模型;S4.3 Use the preliminary prediction model to test on the verification set obtained in step S1, obtain the F2 evaluation index, adjust the parameters of the TPA-DeepAR model according to the F2 index, confusion matrix and ROC curve to achieve better performance, and save the performance of each evaluation index Optimal TPA-DeepAR prediction model;
其中,所述的F2指标为:Among them, the F2 index mentioned is:
Figure PCTCN2022077168-appb-000021
Figure PCTCN2022077168-appb-000021
其中,P为精确率(precision),表示被分为正类的样本中实际为正类的比例:
Figure PCTCN2022077168-appb-000022
其中,TP为真正例数,FP为假正例数,R为召回率(recall),表示在所有实际为正类的样本中,被正确地判断为正类的比例:
Figure PCTCN2022077168-appb-000023
其中,FN为假负例数。
Among them, P is the precision rate (precision), which indicates the proportion of the samples that are classified as positive classes that are actually positive classes:
Figure PCTCN2022077168-appb-000022
Among them, TP is the number of true cases, FP is the number of false positive cases, and R is the recall rate (recall), which means that among all the samples that are actually positive classes, the proportion that is correctly judged as positive class:
Figure PCTCN2022077168-appb-000023
Among them, FN is the number of false negative cases.
将TP,FP,TN,FN四个指标一起呈现在2*2表格中,就会得到混淆矩阵,表格的第一象限到第四象限分别为TP,FP,FN,TN。Present the four indicators of TP, FP, TN, and FN together in a 2*2 table, and a confusion matrix will be obtained. The first to fourth quadrants of the table are TP, FP, FN, and TN respectively.
其中,TN为真负例数。得到混淆矩阵后,矩阵第二象限和第四象限的数值越大越好,反之,第一象限和第三象限的数值越小越好。Among them, TN is the number of true negative cases. After obtaining the confusion matrix, the larger the values in the second and fourth quadrants of the matrix, the better. Conversely, the smaller the values in the first and third quadrants, the better.
在所有实际为负例的样本中,被错误地判断为正例的比例为FPR:FPR=FP/(FP+TN)。以FPR为横轴,R为纵轴,得到ROC曲线。所述的ROC曲线越靠近左上角,TPA-DeepAR模型的召回率越高,其假正例和假反例总数最少,预测效果越好。Among all the samples that are actually negative examples, the proportion that is wrongly judged as positive examples is FPR: FPR=FP/(FP+TN). With FPR as the horizontal axis and R as the vertical axis, the ROC curve is obtained. The closer the ROC curve is to the upper left corner, the higher the recall rate of the TPA-DeepAR model, the least number of false positives and false negatives, and the better the prediction effect.
S5.采用最终TPA-DeepAR预测模型对测试集进行实时预测;图5为TPA-DeepAR预测模型在测试数据上的预测结果图,其中(a)为二级静子尖部动压p 2随时间变化图,(b)为TPA-DeepAR预测模型给出的喘振预测概率随时间的变化图,(c)为TPA-DeepAR预测模型根据预测概率给出的预警信号。在测试数据上进行实时预测的步骤如下: S5. Use the final TPA-DeepAR prediction model to make real-time predictions on the test set; Figure 5 is the prediction result of the TPA-DeepAR prediction model on the test data, where (a) is the change of the dynamic pressure p2 at the tip of the second-stage stator with time Figure (b) is the change of surge prediction probability over time given by the TPA-DeepAR prediction model, and (c) is the early warning signal given by the TPA-DeepAR prediction model based on the prediction probability. The steps to perform real-time prediction on test data are as follows:
S5.1按照预处理的步骤对测试集数据进行预处理,调整测试集数据维度后输入到已经训练好的TPA-DeepAR模型中;测试集数据为二级静子尖部位置的动态压力数据,从图(a)中可以看出,7.48s开始出现了一个向下发展的突尖,处于失速初始扰动阶段,随着失速扰动的发展,在7.826s开始有剧烈的波动,彻底发展为失速喘振。S5.1 Preprocess the test set data according to the preprocessing steps, adjust the dimension of the test set data and input it into the trained TPA-DeepAR model; the test set data is the dynamic pressure data of the tip position of the secondary stator, from It can be seen from Figure (a) that a downward-developing protrusion began to appear at 7.48s, which was in the initial disturbance stage of the stall. With the development of the stall disturbance, it began to fluctuate violently at 7.826s, and completely developed into a stall surge .
S5.2依照时间顺序,用TPA-DeepAR预测模型给出每个测试集数据的喘振预测概率;观察 图(b),可以看到预测概率曲线在7.488s左右识别出初始扰动,喘振概率迅速上升,随后保持着较高的喘振概率,直到7.68s左右原始动压数据恢复到平稳状态,喘振概率曲线也迅速回落,之后伴随原始动压数据的波动再次上升。当初始扰动发生后,旋转失速和喘振大概率会发生,一旦发生就会产生非常严重的影响,因此为喘振概率预测曲线设定阈值,当超过阈值后给出预警信号,实现在初始扰动阶段就做出预警。因此TPA-DeepAR预测模型可以及时对初始扰动阶段的微小变化做出反应,并根据扰动的发展输出喘振概率值。S5.2 In chronological order, use the TPA-DeepAR prediction model to give the surge prediction probability of each test set data; observe the figure (b), you can see that the prediction probability curve identifies the initial disturbance at around 7.488s, and the surge probability It rose rapidly, and then maintained a high surge probability until the original dynamic pressure data returned to a stable state at about 7.68s, and the surge probability curve also fell rapidly, and then rose again with the fluctuation of the original dynamic pressure data. When the initial disturbance occurs, the high probability of rotating stall and surge will occur. Once it occurs, it will have a very serious impact. Therefore, a threshold is set for the surge probability prediction curve. When the threshold is exceeded, an early warning signal is given to realize the initial disturbance. early warning stage. Therefore, the TPA-DeepAR prediction model can respond to small changes in the initial disturbance stage in time, and output the surge probability value according to the development of the disturbance.
以上所述实施例仅表达本发明的实施方式,但并不能因此而理解为对本发明专利的范围的限制,应当指出,对于本领域的技术人员来说,在不脱离本发明构思的前提下,还可以做出若干变形和改进,这些均属于本发明的保护范围。The above-mentioned embodiment only expresses the implementation mode of the present invention, but can not therefore be interpreted as the limitation of the scope of the patent of the present invention, it should be pointed out that, for those skilled in the art, under the premise of not departing from the concept of the present invention, Several modifications and improvements can also be made, all of which belong to the protection scope of the present invention.

Claims (3)

  1. 一种深度自回归网络的轴流压气机失速喘振预测方法,其特征在于,包括以下步骤:A deep autoregressive network axial flow compressor stall surge prediction method is characterized in that it comprises the following steps:
    S1.对航空发动机喘振数据进行预处理,将实验数据划分为测试数据集和训练数据集后,再将训练数据集按比例划分为训练集和验证集;S1. Preprocess the aero-engine surge data, divide the experimental data into a test data set and a training data set, and then divide the training data set into a training set and a verification set in proportion;
    S2.构建基于注意力机制的深度自回归网络模型,即TPA-DeepAR模型,包括以下步骤:S2. Construct a deep autoregressive network model based on the attention mechanism, that is, the TPA-DeepAR model, including the following steps:
    S2.1将每个样本维度调整为(w,1),作为TPA-DeepAR模型的输入,其中w代表时间窗长度;S2.1 Adjust each sample dimension to (w,1) as the input of the TPA-DeepAR model, where w represents the length of the time window;
    S2.2搭建嵌入层将输入样本的维度从(w,1)转换为(w,m),m为指定的维度,将样本的特征从一维分散到m个维度;S2.2 Build an embedding layer to convert the dimension of the input sample from (w, 1) to (w, m), m is the specified dimension, and disperse the characteristics of the sample from one dimension to m dimensions;
    S2.3搭建LSTM层,将嵌入层的输出作为LSTM层的输入,LSTM层输出w个隐藏向量{h t-w+1,h t-w+2,……,h t},每一个隐藏向量的维度为m; S2.3 Build the LSTM layer, use the output of the embedding layer as the input of the LSTM layer, and the LSTM layer outputs w hidden vectors {h t-w+1, h t-w+2 ,...,h t }, each hidden The dimension of the vector is m;
    S2.4搭建注意力层,LSTM层输出的w个隐藏向量{h t-w+1,h t-w+2,……,h t}作为注意力层的输入,经过注意力层对相关维度加权,最终输出一个隐藏向量
    Figure PCTCN2022077168-appb-100001
    S2.4 Build the attention layer. The w hidden vectors {h t-w+1 , h t-w+2 ,...,h t } output by the LSTM layer are used as the input of the attention layer, and the relevant Dimensionally weighted, and finally output a hidden vector
    Figure PCTCN2022077168-appb-100001
    S2.5搭建高斯层,所述高斯层由两个全连接层组成,将注意力层输出的隐藏向量
    Figure PCTCN2022077168-appb-100002
    作为高斯层的输入,高斯层的两个全连接层的输出分别为参数μ和参数σ,因此高斯层的输出会确定一个高斯分布,这样模型能够实现拟合高斯分布的目的;
    S2.5 Build a Gaussian layer, the Gaussian layer is composed of two fully connected layers, and the hidden vector output by the attention layer
    Figure PCTCN2022077168-appb-100002
    As the input of the Gaussian layer, the output of the two fully connected layers of the Gaussian layer is the parameter μ and the parameter σ respectively, so the output of the Gaussian layer will determine a Gaussian distribution, so that the model can achieve the purpose of fitting the Gaussian distribution;
    S2.6采用拟合的高斯分布进行多次随机采样,得到预测点的数据,并依据这些采样点得到预测点的不同分位数以实现概率预测;S2.6 Use the fitted Gaussian distribution to perform random sampling multiple times to obtain the data of the prediction points, and obtain different quantiles of the prediction points based on these sampling points to realize probability prediction;
    S3.构建S2中所述的注意力层:S3. Build the attention layer described in S2:
    S3.1注意力层的输入为LSTM层的输出{h t-w+1,h t-w+2,……,h t},输入数据的维度为(w,m),除了最后一个隐藏向量ht以外,将其他w-1个隐藏向量组成隐状态矩阵H={h t-w+1,h t-w+2,……,h t-1}; S3.1 The input of the attention layer is the output of the LSTM layer {h t-w+1 ,h t-w+2 ,...,h t }, the dimension of the input data is (w,m), except for the last hidden Except for the vector ht, other w-1 hidden vectors form a hidden state matrix H={h t-w+1 ,h t-w+2 ,...,h t-1 };
    S3.2采用k个卷积核捕捉H的信号模式得到H C矩阵,增强模型对特征的学习能力; S3.2 Use k convolution kernels to capture the signal pattern of H to obtain the H C matrix to enhance the model's ability to learn features;
    S3.3隐藏向量h t与H C矩阵通过得分函数进行相似度计算得到注意力权重α i,利用注意力权重α i对H C每一行加权求和,得到向量v tS3.3 Calculate the similarity between the hidden vector h t and the H C matrix through the scoring function to obtain the attention weight α i , use the attention weight α i to weight and sum each row of H C to obtain the vector v t ;
    S3.4最后将h t和v t拼接,输入一个全接连层得到一个新的隐藏向量输出
    Figure PCTCN2022077168-appb-100003
    S3.4 Finally, h t and v t are spliced and input into a fully connected layer to obtain a new hidden vector output
    Figure PCTCN2022077168-appb-100003
    S4.TPA-DeepAR模型损失函数及评价指标:S4.TPA-DeepAR model loss function and evaluation indicators:
    S4.1针对TPA-DeepAR模型,模型在前向传播时输出的是预测高斯分布的参数μ和σ,采用的损失函数具体如下:S4.1 For the TPA-DeepAR model, the model outputs the parameters μ and σ of the predicted Gaussian distribution during forward propagation, and the loss function used is as follows:
    假设样本服从高斯分布y_true~(μ,σ 2),则其似然函数为: Assuming that the sample obeys the Gaussian distribution y_true~(μ, σ 2 ), then its likelihood function is:
    Figure PCTCN2022077168-appb-100004
    Figure PCTCN2022077168-appb-100004
    其对数似然函数为:Its log-likelihood function is:
    Figure PCTCN2022077168-appb-100005
    Figure PCTCN2022077168-appb-100005
    其中,n表示样本个数,y_true是已知的,表示样本的真实标签,μ和σ是模型预测的高斯分布的参数,似然函数描述的是对于参数μ和σ形成的分布,出现y_true这个样本点的概率的大小;Among them, n represents the number of samples, y_true is known, and represents the real label of the sample, μ and σ are the parameters of the Gaussian distribution predicted by the model, and the likelihood function describes the distribution formed by the parameters μ and σ, and y_true appears The size of the probability of the sample point;
    因此通过最大化对数似然函数来学习网络参数,即参数μ和σ形成的分布可以最大概率的出现y_true这个样本点,相应的模型训练的损失函数可以确定为-lnL(μ,σ 2); Therefore, the network parameters are learned by maximizing the log likelihood function, that is, the distribution formed by the parameters μ and σ can have the maximum probability of the sample point y_true, and the loss function of the corresponding model training can be determined as -lnL(μ,σ 2 ) ;
    S4.2基于损失函数,在步骤S1得到的训练集上对TPA-DeepAR模型进行权重更新,最终生成模型的初步预测模型;S4.2 Based on the loss function, update the weight of the TPA-DeepAR model on the training set obtained in step S1, and finally generate a preliminary prediction model of the model;
    S4.3采用初步预测模型在步骤S1得到的验证集上进行测试,获取F2评价指标,根据F2指标,混淆矩阵以及ROC曲线调整TPA-DeepAR模型参数,以达到更优,保存各项评价指标表现最优的TPA-DeepAR预测模型;S4.3 Use the preliminary prediction model to test on the verification set obtained in step S1, obtain the F2 evaluation index, adjust the parameters of the TPA-DeepAR model according to the F2 index, confusion matrix and ROC curve to achieve better performance, and save the performance of each evaluation index Optimal TPA-DeepAR prediction model;
    S5.采用最终TPA-DeepAR预测模型对测试集进行实时预测:S5. Use the final TPA-DeepAR prediction model to make real-time predictions on the test set:
    S5.1.按照预处理的步骤对测试集数据进行预处理,调整测试集数据维度后输入到已经训练好的TPA-DeepAR模型中进行测试;S5.1. Preprocess the test set data according to the preprocessing steps, adjust the dimension of the test set data and input it into the trained TPA-DeepAR model for testing;
    S5.2.依照时间顺序,采用TPA-DeepAR预测模型给出每个测试集样本的喘振预测概率,得到测试集样本的实时喘振概率。S5.2. According to the time sequence, use the TPA-DeepAR prediction model to give the surge prediction probability of each test set sample, and obtain the real-time surge probability of the test set sample.
  2. 根据权利要求1所述的一种深度自回归网络的轴流压气机失速喘振预测方法,其特征在于,所述步骤S1对航空发动机喘振数据进行预处理具体如下:A method for predicting stall and surge of an axial flow compressor with a deep autoregressive network according to claim 1, wherein said step S1 preprocesses the surge data of an aeroengine specifically as follows:
    S1.1获取某型号航空发动机喘振实验数据,剔除实验数据中由于传感器故障产生的无效数据;S1.1 Acquire the experimental data of a certain type of aero-engine surge, and eliminate the invalid data due to sensor failure in the experimental data;
    S1.2对剩余有效数据依次进行降采样处理和滤波处理;S1.2 Perform down-sampling processing and filtering processing on the remaining valid data in sequence;
    S1.3对滤波处理后的数据进行归一化、平滑化处理;S1.3 normalize and smooth the filtered data;
    S1.4为保证测试结果的客观性,将实验数据划分为测试数据集和训练数据集;S1.4 In order to ensure the objectivity of the test results, the experimental data is divided into a test data set and a training data set;
    S1.5通过时间窗切分训练数据集,每个时间窗覆盖的数据点组成一个样本,并将训练数据集按4:1的比例划分为训练集和验证集。S1.5 Segment the training data set by time windows, the data points covered by each time window form a sample, and divide the training data set into a training set and a verification set at a ratio of 4:1.
  3. 根据权利要求2所述的一种深度自回归网络的轴流压气机失速喘振预测方法,其特征在于,所述步骤S4.3中:A deep autoregressive network axial compressor stall surge prediction method according to claim 2, characterized in that, in the step S4.3:
    所述的F2指标为:The F2 indicators described are:
    Figure PCTCN2022077168-appb-100006
    Figure PCTCN2022077168-appb-100006
    其中,P为精确率,表示被分为正类的样本中实际为正类的比例:
    Figure PCTCN2022077168-appb-100007
    其中,TP为真正例数,FP为假正例数,R为召回率,表示在所有实际为正类的样本中,被正确地判断为正类的比例:
    Figure PCTCN2022077168-appb-100008
    其中,FN为假负例数;
    Among them, P is the accuracy rate, indicating the proportion of the samples that are classified as positive classes that are actually positive classes:
    Figure PCTCN2022077168-appb-100007
    Among them, TP is the number of true cases, FP is the number of false positive cases, and R is the recall rate, which means that among all the samples that are actually positive classes, the proportion that is correctly judged as positive class:
    Figure PCTCN2022077168-appb-100008
    Among them, FN is the number of false negative cases;
    将TP,FP,TN,FN四个指标一起呈现在2*2表格中,就会得到混淆矩阵,表格的第一象限到第四象限分别为TP,FP,FN,TN;其中,TN为真负例数;Present the four indicators of TP, FP, TN, and FN together in a 2*2 table, and a confusion matrix will be obtained. The first to fourth quadrants of the table are TP, FP, FN, and TN respectively; among them, TN is true number of negative cases;
    得到混淆矩阵后,矩阵第二象限和第四象限的数值越大越好,反之,第一象限和第三象限的数值越小越好;After obtaining the confusion matrix, the larger the values in the second and fourth quadrants of the matrix, the better; conversely, the smaller the values in the first and third quadrants, the better;
    在所有实际为负例的样本中,被错误地判断为正例的比例为FPR:FPR=FP/(FP+TN);以FPR为横轴,R为纵轴,得到ROC曲线;所述的ROC曲线越靠近左上角,TPA-DeepAR模型的召回率越高,其假正例和假反例总数最少,预测效果越好。Among all the samples that are actually negative examples, the proportion that is wrongly judged as positive examples is FPR: FPR=FP/(FP+TN); with FPR as the horizontal axis and R as the vertical axis, the ROC curve is obtained; the The closer the ROC curve is to the upper left corner, the higher the recall rate of the TPA-DeepAR model, the least number of false positives and false negatives, and the better the prediction effect.
PCT/CN2022/077168 2022-02-22 2022-02-22 Deep autoregressive network based prediction method for stalling and surging of axial-flow compressor WO2023159336A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/077168 WO2023159336A1 (en) 2022-02-22 2022-02-22 Deep autoregressive network based prediction method for stalling and surging of axial-flow compressor

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/077168 WO2023159336A1 (en) 2022-02-22 2022-02-22 Deep autoregressive network based prediction method for stalling and surging of axial-flow compressor

Publications (1)

Publication Number Publication Date
WO2023159336A1 true WO2023159336A1 (en) 2023-08-31

Family

ID=87764293

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/077168 WO2023159336A1 (en) 2022-02-22 2022-02-22 Deep autoregressive network based prediction method for stalling and surging of axial-flow compressor

Country Status (1)

Country Link
WO (1) WO2023159336A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117077544A (en) * 2023-10-13 2023-11-17 北京宝隆泓瑞科技有限公司 Oil-gas separator outlet pressure prediction method and device and electronic equipment
CN117575046A (en) * 2024-01-15 2024-02-20 中煤科工开采研究院有限公司 Multi-hydraulic support load model training and multi-hydraulic support load prediction method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111737910A (en) * 2020-06-10 2020-10-02 大连理工大学 Axial flow compressor stall surge prediction method based on deep learning
CN112580267A (en) * 2021-01-13 2021-03-30 南京航空航天大学 Aero-engine surge prediction method based on multi-branch feature fusion network
CN113125161A (en) * 2021-04-13 2021-07-16 浙江大学 Gas turbine radial air inlet complex distortion digital twin test device and control method
US20210294818A1 (en) * 2020-03-19 2021-09-23 Cisco Technology, Inc. Extraction of prototypical trajectories for automatic classification of network kpi predictions
CN113836817A (en) * 2021-10-09 2021-12-24 大连理工大学 Axial flow compressor rotating stall prediction method based on stacked long-short term memory network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210294818A1 (en) * 2020-03-19 2021-09-23 Cisco Technology, Inc. Extraction of prototypical trajectories for automatic classification of network kpi predictions
CN111737910A (en) * 2020-06-10 2020-10-02 大连理工大学 Axial flow compressor stall surge prediction method based on deep learning
CN112580267A (en) * 2021-01-13 2021-03-30 南京航空航天大学 Aero-engine surge prediction method based on multi-branch feature fusion network
CN113125161A (en) * 2021-04-13 2021-07-16 浙江大学 Gas turbine radial air inlet complex distortion digital twin test device and control method
CN113836817A (en) * 2021-10-09 2021-12-24 大连理工大学 Axial flow compressor rotating stall prediction method based on stacked long-short term memory network

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117077544A (en) * 2023-10-13 2023-11-17 北京宝隆泓瑞科技有限公司 Oil-gas separator outlet pressure prediction method and device and electronic equipment
CN117077544B (en) * 2023-10-13 2024-01-05 北京宝隆泓瑞科技有限公司 Oil-gas separator outlet pressure prediction method and device and electronic equipment
CN117575046A (en) * 2024-01-15 2024-02-20 中煤科工开采研究院有限公司 Multi-hydraulic support load model training and multi-hydraulic support load prediction method

Similar Documents

Publication Publication Date Title
WO2021248746A1 (en) Axial flow compressor stall and surge prediction method based on deep learning
WO2021135630A1 (en) Rolling bearing fault diagnosis method based on grcmse and manifold learning
WO2023159336A1 (en) Deep autoregressive network based prediction method for stalling and surging of axial-flow compressor
WO2023056614A1 (en) Method for predicting rotating stall of axial flow compressor on the basis of stacked long short-term memory network
WO2023010658A1 (en) Time dilated convolutional network-based method for alerting of rotating stall in compressor
WO2020181998A1 (en) Method for detecting mixed sound event on basis of factor decomposition of supervised variational encoder
CN113836817B (en) Axial flow compressor rotating stall prediction method based on stacked long-term and short-term memory network
CN112629854B (en) Bearing fault classification method based on neural network attention mechanism
CN113255848A (en) Water turbine cavitation sound signal identification method based on big data learning
CN102937320B (en) Health protection method used for intelligent air conditioner
CN112052871B (en) Rocket engine health diagnosis method and system based on support vector machine
CN111680875A (en) Unmanned aerial vehicle state risk fuzzy comprehensive evaluation method based on probability baseline model
CN114062850B (en) Double-threshold power grid early fault detection method
CN110701087A (en) Axial flow compressor pneumatic instability detection method based on single-classification overrun learning machine
CN114548555B (en) Axial flow compressor stall surge prediction method based on deep autoregressive network
Zhang et al. Optimization of hmm based on adaptive GAPSO and its application in fault diagnosis of rolling bearing
Tong et al. Fault prediction of marine diesel engine based on time series and support vector machine
CN114708885A (en) Fan fault early warning method based on sound signals
CN110826587A (en) Improved weighted support vector machine-based turboshaft engine fault detection method
US20240133391A1 (en) Prediction method for stall and surging of axial-flow compressor based on deep autoregressive network
CN110232221A (en) Dam Crack influence factor dynamic Contribution Rate method
Luo et al. Rolling bearing diagnosis based on adaptive probabilistic PCA and the enhanced morphological filter
CN116010884A (en) Fault diagnosis method of SSA-LightGBM oil-immersed transformer based on principal component analysis
CN115659323A (en) Intrusion detection method based on information entropy theory and convolution neural network
Mi et al. A nonparametric cumulative sum-based fault detection method for rolling bearings using high-level extended isolated forest

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 18014573

Country of ref document: US

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22927633

Country of ref document: EP

Kind code of ref document: A1