CN113780138B - Self-adaptive robustness VOCs gas leakage detection method, system and storage medium - Google Patents

Self-adaptive robustness VOCs gas leakage detection method, system and storage medium Download PDF

Info

Publication number
CN113780138B
CN113780138B CN202111013939.8A CN202111013939A CN113780138B CN 113780138 B CN113780138 B CN 113780138B CN 202111013939 A CN202111013939 A CN 202111013939A CN 113780138 B CN113780138 B CN 113780138B
Authority
CN
China
Prior art keywords
data
training
dimensional
neural network
vocs
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111013939.8A
Other languages
Chinese (zh)
Other versions
CN113780138A (en
Inventor
曹洋
谭几方
康宇
夏秀山
许镇义
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN202111013939.8A priority Critical patent/CN113780138B/en
Publication of CN113780138A publication Critical patent/CN113780138A/en
Application granted granted Critical
Publication of CN113780138B publication Critical patent/CN113780138B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Examining Or Testing Airtightness (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a self-adaptive robustness VOCs gas leakage detection method, a system and a storage medium, which comprises the following steps of acquiring infrared video data and carrying out preprocessing operation; extracting one-dimensional time sequence characteristic data of pixel points with a certain length from infrared video data, and training a one-dimensional convolutional neural network classifier; training parameter alpha of prior gamma distribution by using one-dimensional convolutional neural network classifier and EVT algorithm with output value led into Bayesian framework 0 And beta 0 (ii) a And inputting related parameters, adjusting a threshold value through a self-adaptive algorithm, and outputting a prediction result. The method fully utilizes the time-space characteristics of pixel points in VOCs gas regions in infrared video data to carry out pre-screening on the infrared video image by using the convolutional neural network, optimizes and adjusts the screening threshold value through an extreme value theory in a Bayesian framework, approaches the right tail part of a probability density function of a fraction by using index distribution, and uses gamma conjugate prior learned from training data, so that the variability of error rate can be reduced and the overall performance can be improved.

Description

自适应鲁棒性VOCs气体泄漏检测方法、系统及存储介质Adaptive robust VOCs gas leak detection method, system and storage medium

技术领域technical field

本发明涉及环境检测领域中VOCs气体泄漏检测技术领域,具体涉及一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法、系统及存储介质。The invention relates to the technical field of VOCs gas leak detection in the field of environmental detection, in particular to an adaptive robust VOCs gas leak detection method, system and storage medium based on extreme value theory.

背景技术Background technique

近年来,随着石油化工行业的迅速发展,生产安全问题也越来越重要。如挥发性有机化合物(VOCs)的泄露会导致癌症、出生缺陷和生殖影响等人类健康问题。VOCs还有导致了臭氧的形成,臭氧是烟雾的主要来源,也是城市地区和炼油厂和化工厂附近地区呼吸系统疾病的主要原因之一,因此,对VOCs的检测与治理已成为当前空气处理问题的一个焦点。In recent years, with the rapid development of the petrochemical industry, the issue of production safety has become more and more important. The release of volatile organic compounds (VOCs) can lead to human health problems such as cancer, birth defects and reproductive effects. VOCs also lead to the formation of ozone, which is the main source of smog and one of the main causes of respiratory diseases in urban areas and areas near oil refineries and chemical plants. Therefore, the detection and treatment of VOCs has become a current air treatment problem. a focus.

考虑到VOCs气体对红外光的吸收,导致红外视频数据上VOCs泄漏区域的颜色较周围区域深(白热模式)。而红外视频数据会受到光照、气温、湿度和气候等因素的影响。成像条件复杂使得检测VOCs气体泄漏的可靠不高。综上,我们提出了一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法,抽取红外视频数据中的像素时空信息进行泄漏预筛查,使用贝叶斯框架内的极值理论(EVT)来优化调整筛查阈值,通过用指数分布(广义帕累托分布的特殊情况)逼近分数的概率密度函数的右尾部,并使用从训练数据中学习的伽马共轭先验,可以降低错误率的可变性并提高整体性能。旨在实现VOCs气体泄漏鲁棒性检测,从而适应各类成像条件的VOCs气体泄漏检测。Considering the absorption of infrared light by VOCs gas, the color of the VOCs leakage area on the infrared video data is darker than the surrounding area (white-hot mode). Infrared video data will be affected by factors such as light, temperature, humidity and climate. The complex imaging conditions make the detection of VOCs gas leaks less reliable. In summary, we propose an adaptive and robust VOCs gas leak detection method based on extreme value theory, extracting pixel spatiotemporal information in infrared video data for leak pre-screening, using extreme value theory in the Bayesian framework ( EVT) to optimally adjust the screening threshold, by approximating the right tail of the probability density function of the fraction with an exponential distribution (a special case of the generalized Pareto distribution), and using a gamma conjugate prior learned from the training data, which can be reduced Variability in error rates and improve overall performance. It aims to achieve robust detection of VOCs gas leakage, so as to adapt to VOCs gas leakage detection of various imaging conditions.

发明内容SUMMARY OF THE INVENTION

本发明提出的一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法,可以实现在成像条件复杂的情况下,可靠地进行VOCs泄漏检测。An adaptive robust VOCs gas leak detection method based on extreme value theory proposed by the invention can realize reliable VOCs leak detection under complex imaging conditions.

为实现上述目的,本发明采用了以下技术方案:To achieve the above object, the present invention has adopted the following technical solutions:

包括以下步骤,Include the following steps,

步骤1:获取红外视频数据中VOCs泄漏区和无泄漏区的数据进行预处理操作;Step 1: Obtain the data of the VOCs leakage area and the non-leakage area in the infrared video data for preprocessing;

步骤2:从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器;Step 2: Extract one-dimensional time series feature data of a certain length of pixels from the infrared video data, and train a one-dimensional convolutional neural network classifier;

步骤3:多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0Step 3: Sample the spatiotemporal features of several pixels from the infrared video data multiple times, use a one-dimensional convolutional neural network classifier, import the output values into the EVT algorithm in the Bayesian framework, and train the parameters of the prior gamma distribution α 0 and β 0 ;

步骤4:输入相关参数,通过自适应算法调整阈值,输出预测结果;Step 4: Input the relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result;

其中,所述步骤2:从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器;具体包括如下细分步骤S21至S23:Wherein, the step 2: extracting a certain length of pixel point one-dimensional time series feature data from the infrared video data, and training a one-dimensional convolutional neural network classifier; specifically includes the following subdivision steps S21 to S23:

步骤S21:从具有VOCs泄漏的切分场景视频帧的VOCs气体区域每8*8或16*16的块中提取一个像素,形成若干具有长度L的像素点一维时序泄露数据(XL,1),L为场景帧数量,其中1代表此数据来自存在VOCs泄漏区域,且XL=[x′1 x′2 ... x′L]T;同时也从不具有VOCs泄漏的切分场景中VOCs气体区域以同样方式提取出若干具有相同长度的像素点一维时序正常数据(XL,0),其中0代表此数据来自正常区域;Step S21: Extract one pixel from every 8*8 or 16*16 block of the VOCs gas region of the segmented scene video frame with VOCs leakage to form one-dimensional time series leakage data of several pixels with length L (X L , 1 ), L is the number of scene frames, where 1 means that the data comes from an area with VOCs leakage, and XL = [x′ 1 x′ 2 ... x′ L ] T ; at the same time, there is never a segmented scene with no VOCs leakage In the VOCs gas area in the same way, several one-dimensional time series normal data (X L , 0) of pixels with the same length are extracted, where 0 means that this data comes from the normal area;

步骤S22:首先对提取得到的像素点一维时序数据XL进行数值归一化,使像素点一维时序数据XL每项元素满足0≤x′i≤255,i=1,2,...,L,随后对每一一维时序数据元素x′i进行零均值化;再分别对处理完成后的两类数据进行切分,80%作为训练数据,20%作为验证数据;Step S22: First, perform numerical normalization on the extracted one-dimensional time series data XL of pixels, so that each element of the one-dimensional time series data XL of pixels satisfies 0≤x′ i≤255 , i =1, 2, . .., L, and then zero-average each one-dimensional time series data element x′ i ; then divide the two types of data after processing, 80% as training data, and 20% as verification data;

步骤S23:使用处理完成的训练数据训练一维卷积神经网络分类器,一维卷积神经网络分类器的输入为像素点一维时序数据XL,输出为D(XL),其中D(XL)∈(0,1),当分类器在验证数据集上分类准确率达到98%以上后停止训练,从而得到一维卷积神经网络分类模型;Step S23: Use the processed training data to train a one-dimensional convolutional neural network classifier, the input of the one-dimensional convolutional neural network classifier is the one-dimensional time series data XL of pixels, and the output is D( XL ), where D( X L )∈(0, 1), stop training when the classification accuracy of the classifier on the validation data set reaches more than 98%, thus obtaining a one-dimensional convolutional neural network classification model;

所述步骤3:多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0,具体包括如下细分步骤S31至S34:Step 3: Sample the spatiotemporal features of several pixels from the infrared video data multiple times, use a one-dimensional convolutional neural network classifier, import the output value into the EVT algorithm in the Bayesian framework, and train the parameter α of the prior gamma distribution 0 and β 0 , specifically including the following subdivision steps S31 to S34:

步骤S31:从待测切分场景视频帧的暗部每8*8或16*16的块中随机提取一个像素,得到K个有长度L的像素点一维时序数据XL,送入一阶段一维卷积神经网络分类器,得到输出D(XL),其中D(XL)∈(0,1);Step S31: randomly extract a pixel from every 8*8 or 16*16 block of the dark part of the video frame of the segmented scene to be measured, obtain K one-dimensional time series data XL of pixels with length L , and send it into a stage-1 dimensional convolutional neural network classifier, and get the output D( XL ), where D( XL )∈(0,1);

步骤S32:将K个XL=[x′1 x′2 ... x′L]T作为数据x,K个D(XL)为对应的标签序列y,构造EVT训练算法数据集T;Step S32: Construct the EVT training algorithm data set T by using K pieces of XL =[x' 1 x' 2 ... x' L ] T as data x, and K pieces of D( XL ) as corresponding label sequences y;

步骤S33:从数据集T中挑选出负样本g={xi|yi=0},根据右尾部概率pu查找负样本g的上限阈值u,通过上限阈值取出右尾部t,更新未被标记为异常的数据的充分统计量n和s;Step S33: Select the negative sample g={x i |y i =0} from the data set T, find the upper threshold u of the negative sample g according to the right tail probability p u , take out the right tail t through the upper threshold, and update the Sufficient statistics n and s for data marked as anomalies;

步骤S34:调整先验伽马分布的参数α0和β0,表示为Step S34: Adjust the parameters α 0 and β 0 of the prior gamma distribution, expressed as

α0=1+w0 α 0 =1+w 0

Figure GDA0003749844670000031
Figure GDA0003749844670000031

其中,w0是分配给训练集的样本计数的权重;where w 0 is the weight assigned to the sample count of the training set;

所述步骤4:输入相关参数,通过自适应算法调整阈值,输出预测结果,具体包括如下细分步骤S41至S43:Described step 4: input relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result, which specifically includes the following subdivision steps S41 to S43:

步骤S41:将待测切分场景视频帧,使用一维卷积神经网络分类器,构造数据集;根据步骤S33的方式,找出数据集的上限阈值,并取出右尾部t1的所有样本执行KolmogorovSmirnov测试,以发现和消除异常,KolmogorovSmirnov测试方法为Step S41: Divide the video frame of the scene to be tested, and use a one-dimensional convolutional neural network classifier to construct a data set; according to the method of step S33, find out the upper limit threshold of the data set, and take out all the samples of the right tail t1 to execute KolmogorovSmirnov test to find and eliminate anomalies, the KolmogorovSmirnov test method is

Figure GDA0003749844670000041
Figure GDA0003749844670000041

其中in

Figure GDA0003749844670000042
Figure GDA0003749844670000042

Figure GDA0003749844670000043
Figure GDA0003749844670000043

Figure GDA0003749844670000044
Figure GDA0003749844670000044

计算得到Dn,然后,移除最大的样本,并使用剩余的样本计算Dn;不断循环,最后选择其中使得Dn最小的样本的标签值记为

Figure GDA0003749844670000045
Calculate D n , then remove the largest sample, and use the remaining samples to calculate Dn; keep looping, and finally select the label value of the sample that makes D n the smallest and record it as
Figure GDA0003749844670000045

步骤S42:在去除异常后,选择以

Figure GDA0003749844670000046
为阈值,提取出右尾部,使用训练期间的先验估计来计算整个序列的后验,更新为α1和β1;Step S42: After removing the abnormality, select the
Figure GDA0003749844670000046
is the threshold value, extract the right tail, use the prior estimation during training to calculate the posterior of the entire sequence, and update it to α 1 and β 1 ;

步骤S43:将S42中计算得到的后验作为先验,设置以样本xj为中心的窗口W,表示为

Figure GDA0003749844670000047
根据右尾部概率pu查找窗口的上限阈值u2,提取右尾部,调整α、β和
Figure GDA0003749844670000051
通过计算得到yj,表示为Step S43: Take the posterior calculated in S42 as the prior, and set the window W with the sample x j as the center, expressed as
Figure GDA0003749844670000047
Find the upper threshold u2 of the window according to the right tail probability p u , extract the right tail, adjust α, β and
Figure GDA0003749844670000051
y j is obtained by calculation, which is expressed as

Figure GDA0003749844670000052
Figure GDA0003749844670000052

其中pf为目标出错率,不断循环获得所有样本的标签值yjwhere p f is the target error rate, and the label value y j of all samples is continuously obtained in a loop.

进一步的,上述步骤1:获取存在VOCs泄漏和无泄漏的红外视频数据并对进行数据预处理,具体包括如下细分步骤S11至S12:Further, the above step 1: obtaining infrared video data with and without leakage of VOCs and performing data preprocessing, which specifically includes the following subdivided steps S11 to S12:

步骤S11:获取存在VOCs泄漏和无泄漏的红外视频数据;Step S11: Acquire infrared video data with and without leakage of VOCs;

步骤S12:对红外视频数据进行随机旋转、帧尺寸归一化、场景切分这些预处理操作。Step S12: Perform preprocessing operations such as random rotation, frame size normalization, and scene segmentation on the infrared video data.

另一方面,本发明还公开一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测系统,包括以下单元,On the other hand, the present invention also discloses an adaptive and robust VOCs gas leak detection system based on extreme value theory, comprising the following units:

数据获取和处理单元,用于获取红外视频数据中VOCs泄漏区和无泄漏区的数据进行预处理操作;The data acquisition and processing unit is used to acquire the data of the VOCs leakage area and the non-leakage area in the infrared video data for preprocessing;

一维网络结构训练单元,用于从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器;The one-dimensional network structure training unit is used to extract one-dimensional time series feature data of a certain length of pixels from infrared video data, and train a one-dimensional convolutional neural network classifier;

参数确定单元,用于多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0The parameter determination unit is used to sample the spatiotemporal features of several pixels from the infrared video data multiple times. It uses a one-dimensional convolutional neural network classifier, and the output value is imported into the EVT algorithm in the Bayesian framework to train the parameters of the prior gamma distribution. α 0 and β 0 ;

预测单元,用于输入相关参数,通过自适应算法调整阈值,输出预测结果;The prediction unit is used to input relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result;

其中,所述一维网络结构训练单元,具体处理步骤如下:The specific processing steps of the one-dimensional network structure training unit are as follows:

步骤S21:从具有VOCs泄漏的切分场景视频帧的VOCs气体区域每8*8或16*16的块中提取一个像素,形成若干具有长度L的像素点一维时序泄露数据(XL,1),L为场景帧数量,其中1代表此数据来自存在VOCs泄漏区域,且XL=[x′1 x′2 ... x′L]T;同时也从不具有VOCs泄漏的切分场景中VOCs气体区域以同样方式提取出若干具有相同长度的像素点一维时序正常数据(XL,0),其中0代表此数据来自正常区域;Step S21: Extract one pixel from every 8*8 or 16*16 block of the VOCs gas region of the segmented scene video frame with VOCs leakage to form one-dimensional time series leakage data of several pixels with length L (X L , 1 ), L is the number of scene frames, where 1 means that the data comes from an area with VOCs leakage, and XL = [x′ 1 x′ 2 ... x′ L ] T ; at the same time, there is never a segmented scene with no VOCs leakage In the VOCs gas area in the same way, several one-dimensional time series normal data (X L , 0) of pixels with the same length are extracted, where 0 means that this data comes from the normal area;

步骤S22:首先对提取得到的像素点一维时序数据XL进行数值归一化,使像素点一维时序数据XL每项元素满足0≤x′i≤255,i=1,2,...,L,随后对每一一维时序数据元素x′i进行零均值化;再分别对处理完成后的两类数据进行切分,80%作为训练数据,20%作为验证数据;Step S22: First, perform numerical normalization on the extracted one-dimensional time series data XL of pixels, so that each element of the one-dimensional time series data XL of pixels satisfies 0≤x′ i≤255 , i =1, 2, . .., L, and then zero-average each one-dimensional time series data element x′ i ; then divide the two types of data after processing, 80% as training data, and 20% as verification data;

步骤S23:使用处理完成的训练数据训练一维卷积神经网络分类器,一维卷积神经网络分类器的输入为像素点一维时序数据XL,输出为D(XL),其中D(XL)∈(0,1),当分类器在验证数据集上分类准确率达到98%以上后停止训练,从而得到一维卷积神经网络分类模型;Step S23: Use the processed training data to train a one-dimensional convolutional neural network classifier, the input of the one-dimensional convolutional neural network classifier is the one-dimensional time series data XL of pixels, and the output is D( XL ), where D( X L )∈(0, 1), stop training when the classification accuracy of the classifier on the validation data set reaches more than 98%, thus obtaining a one-dimensional convolutional neural network classification model;

所述参数确定单元具体处理步骤如下:The specific processing steps of the parameter determination unit are as follows:

步骤S31:从待测切分场景视频帧的暗部每8*8或16*16的块中随机提取一个像素,得到K个有长度L的像素点一维时序数据XL,送入一维卷积神经网络分类器,得到输出D(XL),其中D(XL)∈(0,1);Step S31: randomly extract a pixel from every 8*8 or 16*16 block of the dark part of the video frame of the segmented scene to be measured, obtain K one-dimensional time series data XL of pixels with length L , and send it into a one-dimensional volume Integrate the neural network classifier to get the output D( XL ), where D( XL )∈(0,1);

步骤S32:将K个XL=[x′1 x′2 ... x′L]T作为数据x,K个D(XL)为对应的标签序列y,构造EVT训练算法数据集T;Step S32: Construct the EVT training algorithm data set T by using K pieces of XL =[x' 1 x' 2 ... x' L ] T as data x, and K pieces of D( XL ) as corresponding label sequences y;

步骤S33:从数据集T中挑选出负样本g={xi|yi=0},根据右尾部概率pu查找负样本g的上限阈值u,通过上限阈值取出右尾部t,更新未被标记为异常的数据的充分统计量n和s,Step S33: Select the negative sample g={x i |y i =0} from the data set T, find the upper threshold u of the negative sample g according to the right tail probability p u , take out the right tail t through the upper threshold, and update the Sufficient statistics n and s for data marked as anomalies,

步骤S34:调整先验伽马分布的参数α0和β0,表示为Step S34: Adjust the parameters α 0 and β 0 of the prior gamma distribution, expressed as

α0=1+w0 α 0 =1+w 0

Figure GDA0003749844670000071
Figure GDA0003749844670000071

其中,w0是分配给训练集的样本计数的权重;where w 0 is the weight assigned to the sample count of the training set;

所述预测单元,具体处理步骤如下:The specific processing steps of the prediction unit are as follows:

步骤S41:将待测切分场景视频帧,使用一维卷积神经网络分类器,构造数据集;根据步骤S33的方式,找出数据集的上限阈值,并取出右尾部t1的所有样本执行KolmogorovSmirnov测试,以发现和消除异常,KolmogorovSmirnov测试方法为Step S41: Divide the video frame of the scene to be tested, and use a one-dimensional convolutional neural network classifier to construct a data set; according to the method of step S33, find out the upper limit threshold of the data set, and take out all the samples of the right tail t1 to execute KolmogorovSmirnov test to find and eliminate anomalies, the KolmogorovSmirnov test method is

Figure GDA0003749844670000072
Figure GDA0003749844670000072

其中in

Figure GDA0003749844670000073
Figure GDA0003749844670000073

Figure GDA0003749844670000074
Figure GDA0003749844670000074

Figure GDA0003749844670000081
Figure GDA0003749844670000081

计算得到Dn,然后,移除最大的样本,并使用剩余的样本计算Dn;不断循环,最后选择其中使得Dn最小的样本的标签值记为

Figure GDA0003749844670000082
Calculate D n , then remove the largest sample, and use the remaining samples to calculate Dn; keep looping, and finally select the label value of the sample that makes D n the smallest and record it as
Figure GDA0003749844670000082

步骤S42:在去除异常后,选择以

Figure GDA0003749844670000083
为阈值,提取出右尾部,使用训练期间的先验估计来计算整个序列的后验,更新为α1和β1;Step S42: After removing the abnormality, select the
Figure GDA0003749844670000083
is the threshold value, extract the right tail, use the prior estimation during training to calculate the posterior of the entire sequence, and update it to α 1 and β 1 ;

步骤S43:将S42中计算得到的后验作为先验,设置以样本xj为中心的窗口W,表示为

Figure GDA0003749844670000084
根据右尾部概率pu查找窗口的上限阈值u2,提取右尾部,调整α、β和
Figure GDA0003749844670000085
通过计算得到yj,表示为Step S43: Take the posterior calculated in S42 as the prior, and set the window W with the sample x j as the center, expressed as
Figure GDA0003749844670000084
Find the upper threshold u2 of the window according to the right tail probability p u , extract the right tail, adjust α, β and
Figure GDA0003749844670000085
y j is obtained by calculation, which is expressed as

Figure GDA0003749844670000086
Figure GDA0003749844670000086

其中pf为目标出错率,不断循环获得所有样本的标签值yjwhere p f is the target error rate, and the label value y j of all samples is continuously obtained in a loop.

再一方面,本发明还公开一种计算机可读存储介质,存储有计算机程序,所述计算机程序被处理器执行时,使得所述处理器执行上述基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法的步骤。In another aspect, the present invention also discloses a computer-readable storage medium storing a computer program, which, when executed by a processor, causes the processor to execute the above-mentioned adaptive robustness VOCs gas based on extreme value theory Steps of a leak detection method.

由上述技术方案可知,本发明的基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法和系统,克服现有方法的不足,充分利用红外视频数据中VOCs气体区域像素点的时空特征使用卷积神经网络对红外视频图像进行预筛查,通过贝叶斯框架内的极值理论(EVT)来优化调整筛查阈值,通过用指数分布(广义帕累托分布的特殊情况)逼近分数的概率密度函数的右尾部,并使用从训练数据中学习的伽马共轭先验,可以降低错误率的可变性并提高整体性能。从而在实现VOCs泄漏鲁棒性检测。It can be seen from the above technical solutions that the adaptive and robust VOCs gas leak detection method and system based on the extreme value theory of the present invention overcomes the shortcomings of the existing methods, and makes full use of the spatiotemporal characteristics of the pixels in the VOCs gas region in the infrared video data. Infrared video images are pre-screened by an integrated neural network, and the screening threshold is optimally adjusted by extreme value theory (EVT) within a Bayesian framework, by approximating the probability of the score with an exponential distribution (a special case of the generalized Pareto distribution). The right tail of the density function, and using a gamma-conjugate prior learned from the training data, reduces error rate variability and improves overall performance. So as to realize the robust detection of VOCs leakage.

附图说明Description of drawings

图1为本发明方法总体网络模型示意图;1 is a schematic diagram of the overall network model of the method of the present invention;

图2是本发明的实验结果图。Figure 2 is a graph of the experimental results of the present invention.

具体实施方式Detailed ways

为使本发明实施例的目的、技术方案和优点更加清楚,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本发明一部分实施例,而不是全部的实施例。In order to make the purposes, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments These are some embodiments of the present invention, but not all embodiments.

如图1所示,本实施例所述的基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法,包括以下步骤:As shown in FIG. 1 , the adaptive and robust VOCs gas leak detection method based on extreme value theory described in this embodiment includes the following steps:

步骤1:获取红外视频数据中VOCs泄漏区和无泄漏区的数据进行预处理操作;Step 1: Obtain the data of the VOCs leakage area and the non-leakage area in the infrared video data for preprocessing;

步骤2:从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器。Step 2: Extract one-dimensional time series feature data of a certain length of pixels from the infrared video data, and train a one-dimensional convolutional neural network classifier.

步骤3:多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0Step 3: Sample the spatiotemporal features of several pixels from the infrared video data multiple times, use a one-dimensional convolutional neural network classifier, import the output values into the EVT algorithm in the Bayesian framework, and train the parameters of the prior gamma distribution α 0 and β 0 .

步骤4:输入相关参数,通过自适应算法调整阈值,输出预测结果。Step 4: Input the relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result.

以下具体说明:The following specific instructions:

进一步地,上述步骤1:获取存在VOCs泄漏和无泄漏的红外视频数据并对进行数据预处理。具体包括如下细分步骤S11至S12:Further, the above step 1: obtain infrared video data with and without leakage of VOCs and perform data preprocessing. Specifically, it includes the following subdivision steps S11 to S12:

S11:获取存在VOCs泄漏和无泄漏的红外视频数据。S11: Obtain infrared video data with and without leakage of VOCs.

S12:对红外视频数据进行随机旋转、帧尺寸归一化、场景切分等预处理操作。S12: Perform preprocessing operations such as random rotation, frame size normalization, and scene segmentation on the infrared video data.

进一步地,上述步骤2:从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器。具体包括如下细分步骤S21至S23:Further, in the above step 2: extracting one-dimensional time series feature data of pixels of a certain length from the infrared video data, and training a one-dimensional convolutional neural network classifier. Specifically, it includes the following subdivision steps S21 to S23:

S21:从具有VOCs泄漏的切分场景视频帧的暗部(VOCs气体区域)每8*8或16*16的块中提取一个像素,形成若干具有长度L(场景帧数量,本发明中选取L=160)的像素点一维时序泄露数据(XL,1),其中1代表此数据来自存在VOCs泄漏区域,且XL=[x′1 x′2 ... x′L]T;同时也从不具有VOCs泄漏的切分场景中暗部以同样方式提取出若干具有相同长度的像素点一维时序正常数据(XL,0),其中0代表此数据来自正常区域。S21: Extract one pixel from every 8*8 or 16*16 block of the dark part (VOCs gas region) of the segmented scene video frame with VOCs leakage to form a number of lengths L (the number of scene frames, in the present invention, L= 160) of the pixel point one-dimensional time series leakage data (X L , 1), where 1 represents that the data comes from the leakage area of VOCs, and XL =[x′ 1 x′ 2 ... x′ L ] T ; A number of one-dimensional time series normal data (X L , 0) of pixels with the same length are extracted from the dark part of the segmented scene without VOCs leakage in the same way, where 0 means that the data comes from the normal area.

S22:首先对提取得到的像素点一维时序数据XL进行数值归一化,使其每项元素满足0≤x′i≤255,i=1,2,...,L,随后对每一一维时序数据元素x′i进行零均值化。再分别对处理完成后的两类数据进行切分,80%作为训练数据,20%作为验证数据。S22: First, perform numerical normalization on the extracted one-dimensional time series data XL of pixels, so that each element satisfies 0≤x′ i ≤255, i=1, 2, . . . , L, and then for each element One-dimensional time series data elements x'i are zero-averaged. Then, the two types of data after processing are divided, 80% are used as training data, and 20% are used as verification data.

S23:使用处理完成的训练数据训练一维卷积神经网络分类器,一阶段分类器的输入为像素点一维时序数据XL,输出为得到输出D(XL),其中D(XL)∈(0,1),当分类器在验证数据集上分类准确率达到98%以上后停止训练。从而得到一阶段分类模型如Table1;S23: Use the processed training data to train a one-dimensional convolutional neural network classifier, the input of the one-stage classifier is the one-dimensional time series data XL of pixels, and the output is to obtain the output D( XL ), where D( XL ) ∈ (0, 1), stop training when the classifier achieves more than 98% classification accuracy on the validation dataset. Thus, a one-stage classification model such as Table1 is obtained;

Table 1Table 1

一阶段网络结构One-stage network structure

Figure GDA0003749844670000101
Figure GDA0003749844670000101

Figure GDA0003749844670000111
Figure GDA0003749844670000111

进一步地,上述步骤3:多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0。其中EVT训练算法过程如Table2;Further, the above-mentioned step 3: sampling the spatiotemporal features of several pixels from the infrared video data multiple times, using a one-dimensional convolutional neural network classifier, and importing the output value into the EVT algorithm in the Bayesian framework, training the prior gamma distribution. Parameters α 0 and β 0 . The EVT training algorithm process is shown in Table 2;

具体包括如下细分步骤S31至S34:Specifically, it includes the following subdivision steps S31 to S34:

S31:从待测切分场景视频帧的暗部每8*8或16*16的块中随机提取一个像素,得到K个有长度L(场景帧数量)的像素点一维时序数据XL,送入一阶段一维卷积神经网络,得到输出D(XL),其中D(XL)∈(0,1)。S31: Randomly extract a pixel from every 8*8 or 16*16 block of the dark part of the video frame of the segmented scene to be tested, obtain K one-dimensional time series data XL of pixels with length L (the number of scene frames), and send it to Enter the one-stage one-dimensional convolutional neural network to obtain the output D( XL ), where D( XL )∈(0, 1).

S32:将K个XL=[x′1 x′2 ... x′L]T作为数据x,K个D(XL)为对应的标签序列y,构造EVT训练算法数据集T。S32: Construct the EVT training algorithm data set T by taking K pieces of XL =[x' 1 x' 2 ... x' L ] T as data x, and K pieces of D( XL ) as corresponding label sequences y.

S33:从数据集T中挑选出负样本g={xi|yi=0},并将其代入公式{gi>u}=gpu,该公式表示为根据右尾部概率pu查找负样本g的上限阈值u,其中参数pu是右尾部的概率。通过上限阈值取出右尾部t,更新未被标记为异常的数据的充分统计量n和s,可表示为S33: Select the negative sample g={x i |y i =0} from the data set T, and substitute it into the formula { gi >u}=gp u , which is expressed as finding negative samples according to the right tail probability p u Upper threshold u for sample g, where parameter p u is the probability of the right tail. The right tail t is taken out by the upper threshold, and the sufficient statistics n and s of the data not marked as abnormal are updated, which can be expressed as

nj+1=nj+t nj+1 = nj +t

sj+1=sj+∑ts j+1 =s j +∑t

其中,n0和s0初始化为0。Among them, n 0 and s 0 are initialized to 0.

S34:调整先验伽马分布的参数α0和β0,可表示为S34: Adjust the parameters α 0 and β 0 of the prior gamma distribution, which can be expressed as

α0=1+w0 α 0 =1+w 0

Figure GDA0003749844670000121
Figure GDA0003749844670000121

其中,w0是分配给训练集的样本计数的权重。where w 0 is the weight assigned to the sample count of the training set.

Table 2Table 2

EVT训练算法细节EVT training algorithm details

Figure GDA0003749844670000122
Figure GDA0003749844670000122

进一步地,上述步骤4:输入相关参数,通过自适应算法调整阈值,输出预测结果,其中EVT自适应阈值算法细节如Table 3所示;Further, the above-mentioned step 4: input relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result, wherein the details of the EVT adaptive threshold algorithm are shown in Table 3;

具体包括如下细分步骤S41至S43:Specifically, it includes the following subdivision steps S41 to S43:

S41:根据步骤S33,找出上限阈值,取出尾部的所有样本执行一系列KolmogorovSmirnov(KS)测试,以发现和消除异常。即将待测切分场景视频帧,使用一维卷积神经网络分类器,构造数据集;根据步骤S33的方式,找出数据集的上限阈值,并取出右尾部t1的所有样本执行KS测试,KS测试方法为S41: According to step S33, find the upper threshold, take out all the samples in the tail and perform a series of Kolmogorov Smirnov (KS) tests to find and eliminate abnormalities. The video frame of the scene to be tested is divided, and a one-dimensional convolutional neural network classifier is used to construct a data set; according to the method of step S33, the upper threshold of the data set is found, and all samples in the right tail t1 are taken out to perform the KS test, KS The test method is

Figure GDA0003749844670000131
Figure GDA0003749844670000131

其中in

Figure GDA0003749844670000132
Figure GDA0003749844670000132

Figure GDA0003749844670000133
Figure GDA0003749844670000133

Figure GDA0003749844670000134
Figure GDA0003749844670000134

计算得到Dn或称之为Dn,1。然后,移除最大的样本,并使用剩余的样本计算Dn,2。不断迭代,直到得到

Figure GDA0003749844670000135
最后选择使Dn,i最小的i的值记为
Figure GDA0003749844670000136
S42:在去除异常后,选择以
Figure GDA0003749844670000137
为阈值,根据步骤S33,提取出右尾部,使用训练期间的先验估计来计算整个序列的后验,更新为α1和β1Dn is calculated or referred to as Dn,1 . Then, the largest sample is removed, and Dn,2 is calculated using the remaining samples. Iterate continuously until you get
Figure GDA0003749844670000135
Finally, select the value of i that minimizes D n, i and is recorded as
Figure GDA0003749844670000136
S42: After removing the exception, select to
Figure GDA0003749844670000137
As the threshold, according to step S33, the right tail is extracted, and the posterior of the entire sequence is calculated using the prior estimation during training, and updated to α 1 and β 1 .

S43:将S42中计算的到的后验作为先验。设置以样本为中心的窗口,其表示为

Figure GDA0003749844670000138
根据{wi>u}=wpu查找上限阈值u2,根据S33,提取右尾部,调整α、β和
Figure GDA0003749844670000139
通过计算得到yj,表示为S43: Use the posterior calculated in S42 as the prior. Sets the sample-centered window, which is represented as
Figure GDA0003749844670000138
Find the upper threshold u2 according to { wi > u}=wp u , according to S33, extract the right tail, adjust α, β and
Figure GDA0003749844670000139
y j is obtained by calculation, which is expressed as

Figure GDA0003749844670000141
Figure GDA0003749844670000141

其中pf为目标出错率。不断迭代可以获得调整后的分数yj,即实现自适应阈值。where p f is the target error rate. Continuous iteration can obtain an adjusted score y j , ie, an adaptive threshold is achieved.

Table 3Table 3

EVT自适应阈值算法细节EVT Adaptive Thresholding Algorithm Details

Figure GDA0003749844670000142
Figure GDA0003749844670000142

图2展示了红外视频中存在VOCs泄漏的一帧,该帧在一维卷积神经网络中未被识别为VOCs泄漏帧,而通过EVT阈值自适应算法,提取出右尾部计算yj,其中红框为泄漏的区域。从中可以看出本发明的方法可以有效检出VOCs气体泄漏情况,且在不同场景下,有EVT算法的模型总会比单一维CNN的检测性能高。Figure 2 shows a frame with VOCs leakage in the infrared video. This frame is not recognized as a VOCs leakage frame in the one-dimensional convolutional neural network. Through the EVT threshold adaptive algorithm, the right tail is extracted to calculate y j , where red The box is the leaked area. It can be seen from this that the method of the present invention can effectively detect VOCs gas leakage, and in different scenarios, the model with EVT algorithm will always have higher detection performance than single-dimensional CNN.

综上所述,本发明基于极值理论的自适应鲁棒VOCs气体泄漏检测的方法,使用一维卷积神经网络可以实现对VOCs气体泄漏快速预检测的同时降低计算量,使算法能够适应于性能受限的设备上;采用EVT阈值自适应的方式可以提高红外视频数据VOCs泄漏检测的鲁棒性,减少由光照、温度、气候等因素所造成错误识别的影响。To sum up, the method of the present invention based on the extreme value theory for adaptive and robust VOCs gas leak detection, using a one-dimensional convolutional neural network can achieve rapid pre-detection of VOCs gas leaks while reducing the amount of calculation, so that the algorithm can adapt to On devices with limited performance; adopting the EVT threshold adaptive method can improve the robustness of VOCs leakage detection in infrared video data, and reduce the impact of false identification caused by factors such as illumination, temperature, and climate.

另一方面,本发明还公开一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测系统,包括以下单元,On the other hand, the present invention also discloses an adaptive and robust VOCs gas leak detection system based on extreme value theory, comprising the following units:

数据获取和处理单元,用于获取红外视频数据中VOCs泄漏区和无泄漏区的数据进行预处理操作;The data acquisition and processing unit is used to acquire the data of the VOCs leakage area and the non-leakage area in the infrared video data for preprocessing;

一维网络结构训练单元,用于从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器;The one-dimensional network structure training unit is used to extract one-dimensional time series feature data of a certain length of pixels from infrared video data, and train a one-dimensional convolutional neural network classifier;

参数确定单元,用于多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0The parameter determination unit is used to sample the spatiotemporal features of several pixels from the infrared video data multiple times. It uses a one-dimensional convolutional neural network classifier, and the output value is imported into the EVT algorithm in the Bayesian framework to train the parameters of the prior gamma distribution. α 0 and β 0 ;

预测单元,用于输入相关参数,通过自适应算法调整阈值,输出预测结果。The prediction unit is used to input relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result.

进一步的,所述一维网络结构训练单元,具体处理步骤如下:Further, the specific processing steps of the one-dimensional network structure training unit are as follows:

S21:从具有VOCs泄漏的切分场景视频帧的VOCs气体区域每8*8或16*16的块中提取一个像素,形成若干具有长度L的像素点一维时序泄露数据(XL,1),L为场景帧数量,其中1代表此数据来自存在VOCs泄漏区域,且XL=[x′1 x′2 ... x′L]T;同时也从不具有VOCs泄漏的切分场景中VOCs气体区域以同样方式提取出若干具有相同长度的像素点一维时序正常数据(XL,0),其中0代表此数据来自正常区域;S21: Extract one pixel from every 8*8 or 16*16 block of the VOCs gas region of the segmented scene video frame with VOCs leakage to form one-dimensional time series leakage data of several pixels with length L (X L , 1) , L is the number of scene frames, where 1 means that the data comes from the area with VOCs leakage, and XL = [x′ 1 x′ 2 ... x′ L ] T ; at the same time, it also never has VOCs leakage in the segmentation scene The VOCs gas area extracts several one-dimensional time series normal data (X L , 0) of pixels with the same length in the same way, where 0 represents that the data comes from the normal area;

S22:首先对提取得到的像素点一维时序数据XL进行数值归一化,使像素点一维时序数据XL每项元素满足0≤x′i≤255,i=1,2,...,L,随后对每一一维时序数据元素x′i进行零均值化;再分别对处理完成后的两类数据进行切分,80%作为训练数据,20%作为验证数据;S22: First, perform numerical normalization on the extracted one-dimensional time series data XL of pixels, so that each element of the one-dimensional time series data XL of pixels satisfies 0≤x′ i≤255 , i =1, 2, .. , L, and then zero-average each one-dimensional time series data element x′ i ; then divide the two types of data after processing, 80% as training data, and 20% as verification data;

S23:使用处理完成的训练数据训练一维卷积神经网络分类器,一阶段分类器的输入为像素点一维时序数据XL,输出为D(XL),其中D(XL)∈(0,1),当分类器在验证数据集上分类准确率达到98%以上后停止训练,从而得到一阶段分类模型。S23: Use the processed training data to train a one-dimensional convolutional neural network classifier. The input of the one-stage classifier is the one-dimensional time series data XL of pixels, and the output is D( XL ), where D( XL )∈( 0, 1), when the classification accuracy rate of the classifier on the validation data set reaches more than 98%, the training is stopped, so as to obtain a one-stage classification model.

进一步的,所述参数确定单元具体处理步骤如下:Further, the specific processing steps of the parameter determination unit are as follows:

S31:从待测切分场景视频帧的暗部每8*8或16*16的块中随机提取一个像素,得到K个有长度L的像素点一维时序数据XL,送入一阶段一维卷积神经网络,得到输出D(XL),其中D(XL)∈(0,1);S31: Randomly extract one pixel from every 8*8 or 16*16 block in the dark part of the video frame of the segmented scene to be tested, obtain K one-dimensional time series data XL of pixels with length L , and send it into one-stage one-dimensional Convolutional Neural Network, get the output D( XL ), where D( XL )∈(0,1);

S32:将K个XL=[x′1 x′2 ... x′L]T作为数据x,K个D(XL)为对应的标签序列y,构造EVT训练算法数据集T;S32: Construct the EVT training algorithm dataset T by using K XL = [x' 1 x' 2 ... x' L ] T as data x, and K D(X L ) as corresponding label sequence y;

S33:从数据集T中挑选出负样本g={xi|yi=0},并将其代入公式{gi>u}=gpu查找上限阈值u,其中参数pu是右尾部的概率,通过上限阈值取出右尾部t,更新未被标记为异常的数据的充分统计量n和s,表示为S33: Select the negative sample g={x i |y i =0} from the data set T, and substitute it into the formula { gi >u}=gp u to find the upper threshold u, where the parameter p u is the right tail Probability, taking out the right tail t through the upper threshold, updating the sufficient statistics n and s of the data not marked as abnormal, expressed as

nj+1=nj+t nj+1 = nj +t

sj+1=sj+∑ts j+1 =s j +∑t

其中,n0和s0初始化为0;Among them, n 0 and s 0 are initialized to 0;

S34:调整先验伽马分布的参数α0和β0,表示为S34: Adjust the parameters α 0 and β 0 of the prior gamma distribution, expressed as

α0=1+w0 α 0 =1+w 0

Figure GDA0003749844670000171
Figure GDA0003749844670000171

其中,w0是分配给训练集的样本计数的权重。where w 0 is the weight assigned to the sample count of the training set.

进一步的,所述预测单元,具体处理步骤如下:Further, the specific processing steps of the prediction unit are as follows:

S41:根据步骤S33,找出上限阈值,取出尾部的所有样本即将待测切分场景视频帧,使用一维卷积神经网络,构造数据集;根据步骤S33的方式,找出数据集的上限阈值,并取出右尾部t1的所有样本执行一系列KolmogorovSmirnov(KS)测试,以发现和消除异常,KS为S41: According to step S33, find the upper limit threshold, take out all the samples at the tail, and then divide the video frame of the scene to be tested, and use a one-dimensional convolutional neural network to construct a data set; According to the method of step S33, find the upper limit threshold of the data set , and take out all samples of the right tail t1 to perform a series of KolmogorovSmirnov (KS) tests to find and eliminate anomalies, KS is

Figure GDA0003749844670000172
Figure GDA0003749844670000172

其中in

Figure GDA0003749844670000173
Figure GDA0003749844670000173

Figure GDA0003749844670000174
Figure GDA0003749844670000174

Figure GDA0003749844670000175
Figure GDA0003749844670000175

计算得到Dn或称之为Dn,1,然后,移除最大的样本,并使用剩余的样本计算Dn,2,不断迭代,直到得到

Figure GDA0003749844670000181
最后选择使Dn,i最小的i的值记为
Figure GDA0003749844670000182
Calculate D n or call it D n, 1 , then remove the largest sample and use the remaining samples to calculate D n, 2 , and iterate until you get
Figure GDA0003749844670000181
Finally, select the value of i that minimizes D n, i and is recorded as
Figure GDA0003749844670000182

S42:在去除异常后,选择以

Figure GDA0003749844670000183
为阈值,提取出右尾部,使用训练期间的先验估计来计算整个序列的后验,更新为α1和β1;S42: After removing the exception, select to
Figure GDA0003749844670000183
is the threshold value, extract the right tail, use the prior estimation during training to calculate the posterior of the entire sequence, and update it to α 1 and β 1 ;

S43:将S42中计算得到的后验作为先验,设置以样本xj为中心的窗口W,表示为

Figure GDA0003749844670000184
根据右尾部概率pu查找窗口的上限阈值u2,提取右尾部,调整α、β和
Figure GDA0003749844670000185
通过计算得到yj,表示为S43: Take the posterior calculated in S42 as the prior, and set the window W with the sample x j as the center, expressed as
Figure GDA0003749844670000184
Find the upper threshold u2 of the window according to the right tail probability p u , extract the right tail, adjust α, β and
Figure GDA0003749844670000185
y j is obtained by calculation, which is expressed as

Figure GDA0003749844670000186
Figure GDA0003749844670000186

其中pf为目标出错率,不断迭代获得调整后的分数yj,即实现自适应阈值。where p f is the target error rate, and iteratively obtains the adjusted score y j , that is, the adaptive threshold is realized.

本发明还公开一种计算机可读存储介质,存储有计算机程序,所述计算机程序被处理器执行时,使得所述处理器执行如上述方法的步骤。The present invention also discloses a computer-readable storage medium, which stores a computer program, and when the computer program is executed by a processor, causes the processor to execute the steps of the above method.

可理解的是,本发明实施例提供的系统与本发明实施例提供的方法相对应,相关内容的解释、举例和有益效果可以参考上述方法中的相应部分。It is understandable that the system provided by the embodiment of the present invention corresponds to the method provided by the embodiment of the present invention, and reference may be made to the corresponding part of the above-mentioned method for explanation, examples and beneficial effects of related content.

本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。As will be appreciated by those skilled in the art, the embodiments of the present application may be provided as a method, a system, or a computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.

本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the present application. It will be understood that each process and/or block in the flowchart illustrations and/or block diagrams, and combinations of processes and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to the processor of a general purpose computer, special purpose computer, embedded processor or other programmable data processing device to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing device produce Means for implementing the functions specified in a flow or flow of a flowchart and/or a block or blocks of a block diagram.

这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions The apparatus implements the functions specified in the flow or flow of the flowcharts and/or the block or blocks of the block diagrams.

这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded on a computer or other programmable data processing device to cause a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process such that The instructions provide steps for implementing the functions specified in the flow or blocks of the flowcharts and/or the block or blocks of the block diagrams.

以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。The above embodiments are only used to illustrate the technical solutions of the present invention, but not to limit them; although the present invention has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that: The recorded technical solutions are modified, or some technical features thereof are equivalently replaced; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of the embodiments of the present invention.

Claims (4)

1. A self-adaptive robustness VOCs gas leakage detection method based on an extreme value theory is characterized by comprising the following steps: comprises the following steps of (a) carrying out,
step 1: acquiring data of VOCs leakage areas and non-leakage areas in infrared video data to carry out preprocessing operation;
step 2: extracting one-dimensional time sequence characteristic data of pixel points with a certain length from infrared video data, and training a one-dimensional convolutional neural network classifier;
and step 3: sampling the space-time characteristics of a plurality of pixel points from the infrared video data for multiple times, using a one-dimensional convolution neural network classifier, leading the output value into an EVT algorithm in a Bayesian framework, and training the parameter alpha of prior gamma distribution 0 And beta 0
And 4, step 4: inputting relevant parameters, adjusting a threshold value through a self-adaptive algorithm, and outputting a prediction result;
wherein the step 2: extracting one-dimensional time sequence characteristic data of pixel points with a certain length from infrared video data, and training a one-dimensional convolutional neural network classifier; the method specifically comprises the following subdivision steps S21-S23:
step S21: extracting one pixel from each 8X 8 or 16X 16 block of VOCs gas area of segmented scene video frame with VOCs leakage to form a plurality of pixel point one-dimensional time sequence leakage data (X) with length L L 1), L is the number of scene frames, where 1 represents that the data comes from the area where VOCs are leaking, and X is L =[x′ 1 x′ 2 ...x′ L ] T (ii) a Meanwhile, a plurality of pixel point one-dimensional time sequence normal data (X) with the same length are extracted from VOCs gas regions in segmentation scenes without VOCs leakage in the same way L 0), where 0 represents that the data is from a normal region;
step S22: firstly, one-dimensional time sequence data X of pixel points obtained by extraction is subjected to L Carrying out numerical value normalization to enable one-dimensional time sequence data X of pixel points L Each element satisfies 0 ≦ x' i 1, 2, L followed by x 'for each one-dimensional time series data element' i Carrying out zero equalization; then, the two types of data after the processing are respectively segmented, wherein 80% of the data are used as training data, and 20% of the data are used as verification data;
step S23: training a one-dimensional convolutional neural network classifier by using the processed training data, wherein the input of the one-dimensional convolutional neural network classifier is pixel point one-dimensional time sequence data X L The output is D (X) L ) Wherein D (X) L ) E (0, 1), stopping training when the classification accuracy of the classifier on the verification data set reaches more than 98%, thereby obtaining a one-dimensional convolution neural network classification model;
the step 3: sampling the space-time characteristics of a plurality of pixel points from infrared video data for a plurality of times, using a one-dimensional convolution neural network classifier, leading the output value into an EVT algorithm in a Bayesian framework, and training the parameter alpha of prior gamma distribution 0 And beta 0 Specifically, the method comprises the following subdivision steps S31 to S34:
step S31: randomly extracting one pixel from each 8 × 8 or 16 × 16 block of the dark part of the segmented scene video frame to be detected to obtain K pixel point one-dimensional time sequence data X with length L L Sending the data to a one-stage one-dimensional convolution neural network classifier to obtain an output D (X) L ) Wherein D (X) L )∈(0,1);
Step S32: mixing K X L =[x′ 1 x′ 2 ...x′ L ] T As data X, K D (X) L ) Is corresponding toConstructing an EVT training algorithm data set T by using a label sequence y;
step S33: picking out negative samples g ═ x from the dataset T i |y i 0, according to the right tail probability p u Searching an upper limit threshold u of the negative sample g, taking out the right tail t through the upper limit threshold, and updating sufficient statistics n and s of data which are not marked as abnormal;
step S34: adjusting a parameter alpha of a prior gamma distribution 0 And beta 0 Is shown as
α 0 =1+w 0
Figure FDA0003749844660000031
Wherein, w 0 Is the weight assigned to the sample count of the training set;
the step 4: inputting relevant parameters, adjusting a threshold value through an adaptive algorithm, and outputting a prediction result, wherein the method specifically comprises the following subdivision steps S41-S43:
step S41: constructing a data set by using a one-dimensional convolution neural network classifier on a to-be-detected segmented scene video frame; according to the mode of step S33, an upper threshold value of the data set is found, all samples of the right tail t1 are taken out to carry out a KolmogorovSmirnov test to find and eliminate the abnormity, and the KolmogorovSmirnov test method is that
Figure FDA0003749844660000032
Wherein
Figure FDA0003749844660000033
Figure FDA0003749844660000034
Figure FDA0003749844660000035
Is calculated to obtain D n Then, the largest sample is removed and Dn is calculated using the remaining samples; continuously circulating, and finally selecting D n The label value of the smallest sample is noted
Figure FDA0003749844660000036
Step S42: after removing the anomaly, selecting to
Figure FDA0003749844660000041
For the threshold, the right tail is extracted, the a priori estimate during training is used to calculate the a posteriori of the whole sequence, updated to α 1 And beta 1
Step S43: the posterior obtained by calculation in S42 is used as a prior, and a sample x is set j A window W as a center, denoted by
Figure FDA0003749844660000042
According to the right tail probability p u Searching the upper limit threshold u2 of the window, extracting the right tail part, and adjusting alpha, beta and
Figure FDA0003749844660000043
by calculating to obtain y j Is shown as
Figure FDA0003749844660000044
Wherein p is f Continuously and circularly obtaining the label values y of all samples for the target error rate j
2. The extremum theory-based adaptive robust VOCs gas leak detection method of claim 1, wherein: the step 1: acquiring infrared video data with and without leakage of VOCs and preprocessing the data, wherein the method specifically comprises the following subdivision steps S11-S12:
step S11: acquiring infrared video data with VOCs leakage and no leakage;
step S12: and carrying out preprocessing operations of random rotation, frame size normalization and scene segmentation on the infrared video data.
3. The utility model provides a self-adaptation robustness VOCs gas leak detection system based on extreme value theory which characterized in that: comprises the following units which are connected with each other,
the data acquisition and processing unit is used for acquiring data of VOCs leakage areas and non-leakage areas in the infrared video data to carry out preprocessing operation;
the one-dimensional network structure training unit is used for extracting one-dimensional time sequence characteristic data of pixel points with a certain length from the infrared video data and training a one-dimensional convolutional neural network classifier;
a parameter determination unit for sampling the space-time characteristics of a plurality of pixel points from the infrared video data for a plurality of times, using a one-dimensional convolution neural network classifier, leading the output value into an EVT algorithm in a Bayesian framework, and training the parameter alpha of the prior gamma distribution 0 And beta 0
The prediction unit is used for inputting relevant parameters, adjusting a threshold value through a self-adaptive algorithm and outputting a prediction result;
the one-dimensional network structure training unit comprises the following specific processing steps:
step S21: extracting one pixel from each 8X 8 or 16X 16 block of VOCs gas area of segmented scene video frame with VOCs leakage to form a plurality of pixel point one-dimensional time sequence leakage data (X) with length L L 1), L is the number of scene frames, where 1 represents that the data comes from the area where VOCs are leaking, and X is L =[x′ 1 x′ 2 ...x′ L ] T (ii) a Meanwhile, a plurality of pixel point one-dimensional time sequence normal data (X) with the same length are extracted from VOCs gas regions in segmentation scenes without VOCs leakage in the same way L 0), where 0 represents that the data is from a normal region;
step S22: firstly, extracting the obtained one-dimensional time sequence data X of the pixel points L Carrying out numerical value normalization to enable one-dimensional time sequence data X of pixel points L Each element satisfies 0 ≦ x' i 1, 2, L followed by x 'for each one-dimensional time series data element' i Carrying out zero equalization; then, the two types of data after the processing are respectively segmented, wherein 80% of the data are used as training data, and 20% of the data are used as verification data;
step S23: training a one-dimensional convolutional neural network classifier by using the processed training data, wherein the input of the one-dimensional convolutional neural network classifier is pixel point one-dimensional time sequence data X L The output is D (X) L ) Wherein D (X) L ) E (0, 1), stopping training when the classification accuracy of the classifier on the verification data set reaches more than 98%, thereby obtaining a one-dimensional convolution neural network classification model;
the parameter determination unit comprises the following specific processing steps:
step S31: randomly extracting one pixel from each 8X 8 or 16X 16 block of the dark part of the segmented scene video frame to be detected to obtain K pixel point one-dimensional time sequence data X with the length L L Sending the data into a one-dimensional convolutional neural network classifier to obtain an output D (X) L ) Wherein D (X) L )∈(0,1);
Step S32: mixing K X L =[x′ 1 x′ 2 ...x′ L ] T As data X, K D (X) L ) Constructing an EVT training algorithm data set T for the corresponding label sequence y;
step S33: picking out negative samples g ═ x from the dataset T i |y i 0, according to the right tail probability p u Searching an upper threshold u of the negative sample g, taking out a right tail t through the upper threshold, updating sufficient statistics n and s of data which are not marked as abnormal,
step S34: adjusting a parameter alpha of a prior gamma distribution 0 And beta 0 Is shown as
α 0 =1+w 0
Figure FDA0003749844660000061
Wherein, w 0 Is the weight assigned to the sample count of the training set;
the prediction unit comprises the following specific processing steps:
step S41: constructing a data set by using a one-dimensional convolution neural network classifier on a to-be-detected segmented scene video frame; according to the mode of the step S33, an upper threshold value of the data set is found, all samples of the right tail t1 are taken out, and a KolmogorovSmirnov test is carried out to find and eliminate the abnormity, wherein the KolmogorovSmirnov test method is that
Figure FDA0003749844660000071
Wherein
Figure FDA0003749844660000072
Figure FDA0003749844660000073
Figure FDA0003749844660000074
Is calculated to obtain D n Then, the largest sample is removed and Dn is calculated using the remaining samples; continuously circulating, and finally selecting D n The label value of the smallest sample is noted
Figure FDA0003749844660000079
Step S42: after removing the anomaly, selecting to
Figure FDA0003749844660000078
For the threshold, the right tail is extracted, the a priori estimate during training is used to calculate the a posteriori of the whole sequence, updated to α 1 And beta 1
Step S43: the posterior obtained by calculation in S42 is used as a prior, and a sample x is set j A window W as a center, denoted by
Figure FDA0003749844660000075
According to the right tail probability p u Searching the upper limit threshold u2 of the window, extracting the right tail part, and adjusting alpha, beta and
Figure FDA0003749844660000076
by calculating to obtain y j Is shown as
Figure FDA0003749844660000077
Wherein p is f Continuously and circularly obtaining the label values y of all samples for the target error rate j
4. A computer readable storage medium storing a computer program which, when executed by a processor, causes the processor to perform the steps of the extremum theory based adaptive robust VOCs gas leak detection method of claim 1 or 2.
CN202111013939.8A 2021-08-31 2021-08-31 Self-adaptive robustness VOCs gas leakage detection method, system and storage medium Active CN113780138B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111013939.8A CN113780138B (en) 2021-08-31 2021-08-31 Self-adaptive robustness VOCs gas leakage detection method, system and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111013939.8A CN113780138B (en) 2021-08-31 2021-08-31 Self-adaptive robustness VOCs gas leakage detection method, system and storage medium

Publications (2)

Publication Number Publication Date
CN113780138A CN113780138A (en) 2021-12-10
CN113780138B true CN113780138B (en) 2022-09-13

Family

ID=78840513

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111013939.8A Active CN113780138B (en) 2021-08-31 2021-08-31 Self-adaptive robustness VOCs gas leakage detection method, system and storage medium

Country Status (1)

Country Link
CN (1) CN113780138B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115751203B (en) * 2022-12-05 2025-06-06 南通中油燃气有限责任公司 A natural gas pipeline leakage monitoring system based on infrared thermal imager
CN117830731B (en) * 2024-01-02 2024-06-28 北京蓝耘科技股份有限公司 Multi-dimensional parallel scheduling method

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106599864B (en) * 2016-12-21 2020-01-07 中国科学院光电技术研究所 A deep face recognition method based on extreme value theory
CN110672278B (en) * 2019-10-16 2021-04-30 北京工业大学 Method for quantitatively measuring VOCs leakage of production device based on infrared imaging
CN111242404B (en) * 2019-11-12 2022-08-12 中国水利水电科学研究院 A method and system for extreme evaluation of flood events induced by heavy rainfall
CN113008470B (en) * 2020-07-22 2024-02-13 威盛电子股份有限公司 Gas leak detection device and gas leak detection method
CN112163624A (en) * 2020-09-30 2021-01-01 上海交通大学 Data anomaly judgment method and system based on deep learning and extreme value theory

Also Published As

Publication number Publication date
CN113780138A (en) 2021-12-10

Similar Documents

Publication Publication Date Title
CN108182454A (en) Safety check identifying system and its control method
CN113780138B (en) Self-adaptive robustness VOCs gas leakage detection method, system and storage medium
CN114581709B (en) Model training, method, device and medium for identifying objects in medical images
CN114092450B (en) Real-time image segmentation method, system and device based on gastroscopy video
CN116778482B (en) Embryo image blastomere target detection method, computer equipment and storage medium
CN111507998A (en) Depth cascade-based multi-scale excitation mechanism tunnel surface defect segmentation method
CN111191584B (en) Face recognition method and device
CN113205052A (en) Spore identification method, system and medium based on residual error deep network
CN110929635A (en) False face video detection method and system based on face cross-over ratio under trust mechanism
CN112949378A (en) Bacterial microscopic image segmentation method based on deep learning network
CN115410059A (en) Remote sensing image part supervision change detection method and device based on contrast loss
CN117314826A (en) Performance detection method of display screen
CN117934489A (en) Fundus hard exudate segmentation method based on residual error and pyramid segmentation attention
CN117456272A (en) A self-supervised anomaly detection method based on contrastive learning
CN118397367A (en) Tampering detection method based on convolution vision Mamba
CN114510715B (en) Method and device for testing functional safety of model, storage medium and equipment
CN118014048A (en) Low-illumination face detection model construction method, device and terminal
CN117541581A (en) Fine management system suitable for lithium battery production
CN117557872A (en) Unsupervised anomaly detection method and device for optimizing storage mode
CN111598184B (en) DenseNet-based image noise identification method and device
CN113255472A (en) Face quality evaluation method and system based on random embedding stability
CN113409281A (en) Image definition evaluation method and device based on depth model
CN112906707A (en) Semantic segmentation method and device for surface defect image and computer equipment
CN118608845B (en) A method and system for analyzing steganalysis images
Subha et al. Analysis of deep learning based optimization techniques for oral cancer detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20250324

Address after: 230000 no.5089 Wangjiang West Road, hi tech Zone, Hefei City, Anhui Province

Patentee after: INSTITUTE OF ADVANCED TECHNOLOGY University OF SCIENCE AND TECHNOLOGY OF CHINA

Country or region after: China

Patentee after: University of Science and Technology of China

Address before: 230088 5089 Wangjiang West Road, Hefei high tech Zone, Anhui

Patentee before: INSTITUTE OF ADVANCED TECHNOLOGY University OF SCIENCE AND TECHNOLOGY OF CHINA

Country or region before: China