CN113780138B - Self-adaptive robustness VOCs gas leakage detection method, system and storage medium - Google Patents
Self-adaptive robustness VOCs gas leakage detection method, system and storage medium Download PDFInfo
- Publication number
- CN113780138B CN113780138B CN202111013939.8A CN202111013939A CN113780138B CN 113780138 B CN113780138 B CN 113780138B CN 202111013939 A CN202111013939 A CN 202111013939A CN 113780138 B CN113780138 B CN 113780138B
- Authority
- CN
- China
- Prior art keywords
- data
- training
- dimensional
- neural network
- vocs
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 239000012855 volatile organic compound Substances 0.000 title claims abstract description 75
- 238000001514 detection method Methods 0.000 title claims abstract description 26
- 238000003860 storage Methods 0.000 title claims abstract description 9
- 238000012549 training Methods 0.000 claims abstract description 60
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 41
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 33
- 238000000034 method Methods 0.000 claims abstract description 28
- 238000009826 distribution Methods 0.000 claims abstract description 21
- 238000007781 pre-processing Methods 0.000 claims abstract description 13
- 230000003044 adaptive effect Effects 0.000 claims description 24
- 238000012545 processing Methods 0.000 claims description 22
- 238000004590 computer program Methods 0.000 claims description 11
- 238000013528 artificial neural network Methods 0.000 claims description 10
- 238000001276 Kolmogorov–Smirnov test Methods 0.000 claims description 9
- 238000010606 normalization Methods 0.000 claims description 9
- 238000012795 verification Methods 0.000 claims description 8
- 238000004364 calculation method Methods 0.000 claims description 7
- 238000013145 classification model Methods 0.000 claims description 6
- 230000011218 segmentation Effects 0.000 claims description 6
- 230000002159 abnormal effect Effects 0.000 claims description 4
- 238000005070 sampling Methods 0.000 claims description 4
- 238000000605 extraction Methods 0.000 claims 1
- 238000012216 screening Methods 0.000 abstract description 5
- 238000013459 approach Methods 0.000 abstract 1
- 239000000284 extract Substances 0.000 description 18
- 238000010586 diagram Methods 0.000 description 7
- 230000006870 function Effects 0.000 description 6
- 230000008569 process Effects 0.000 description 4
- 238000010200 validation analysis Methods 0.000 description 4
- 230000005856 abnormality Effects 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- 230000008676 import Effects 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 208000032170 Congenital Abnormalities Diseases 0.000 description 1
- 206010010356 Congenital anomaly Diseases 0.000 description 1
- 206010028980 Neoplasm Diseases 0.000 description 1
- CBENFWSGALASAD-UHFFFAOYSA-N Ozone Chemical compound [O-][O+]=O CBENFWSGALASAD-UHFFFAOYSA-N 0.000 description 1
- 238000010521 absorption reaction Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000007698 birth defect Effects 0.000 description 1
- 201000011510 cancer Diseases 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000005802 health problem Effects 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000001850 reproductive effect Effects 0.000 description 1
- 208000023504 respiratory system disease Diseases 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000010998 test method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Examining Or Testing Airtightness (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域technical field
本发明涉及环境检测领域中VOCs气体泄漏检测技术领域,具体涉及一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法、系统及存储介质。The invention relates to the technical field of VOCs gas leak detection in the field of environmental detection, in particular to an adaptive robust VOCs gas leak detection method, system and storage medium based on extreme value theory.
背景技术Background technique
近年来,随着石油化工行业的迅速发展,生产安全问题也越来越重要。如挥发性有机化合物(VOCs)的泄露会导致癌症、出生缺陷和生殖影响等人类健康问题。VOCs还有导致了臭氧的形成,臭氧是烟雾的主要来源,也是城市地区和炼油厂和化工厂附近地区呼吸系统疾病的主要原因之一,因此,对VOCs的检测与治理已成为当前空气处理问题的一个焦点。In recent years, with the rapid development of the petrochemical industry, the issue of production safety has become more and more important. The release of volatile organic compounds (VOCs) can lead to human health problems such as cancer, birth defects and reproductive effects. VOCs also lead to the formation of ozone, which is the main source of smog and one of the main causes of respiratory diseases in urban areas and areas near oil refineries and chemical plants. Therefore, the detection and treatment of VOCs has become a current air treatment problem. a focus.
考虑到VOCs气体对红外光的吸收,导致红外视频数据上VOCs泄漏区域的颜色较周围区域深(白热模式)。而红外视频数据会受到光照、气温、湿度和气候等因素的影响。成像条件复杂使得检测VOCs气体泄漏的可靠不高。综上,我们提出了一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法,抽取红外视频数据中的像素时空信息进行泄漏预筛查,使用贝叶斯框架内的极值理论(EVT)来优化调整筛查阈值,通过用指数分布(广义帕累托分布的特殊情况)逼近分数的概率密度函数的右尾部,并使用从训练数据中学习的伽马共轭先验,可以降低错误率的可变性并提高整体性能。旨在实现VOCs气体泄漏鲁棒性检测,从而适应各类成像条件的VOCs气体泄漏检测。Considering the absorption of infrared light by VOCs gas, the color of the VOCs leakage area on the infrared video data is darker than the surrounding area (white-hot mode). Infrared video data will be affected by factors such as light, temperature, humidity and climate. The complex imaging conditions make the detection of VOCs gas leaks less reliable. In summary, we propose an adaptive and robust VOCs gas leak detection method based on extreme value theory, extracting pixel spatiotemporal information in infrared video data for leak pre-screening, using extreme value theory in the Bayesian framework ( EVT) to optimally adjust the screening threshold, by approximating the right tail of the probability density function of the fraction with an exponential distribution (a special case of the generalized Pareto distribution), and using a gamma conjugate prior learned from the training data, which can be reduced Variability in error rates and improve overall performance. It aims to achieve robust detection of VOCs gas leakage, so as to adapt to VOCs gas leakage detection of various imaging conditions.
发明内容SUMMARY OF THE INVENTION
本发明提出的一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法,可以实现在成像条件复杂的情况下,可靠地进行VOCs泄漏检测。An adaptive robust VOCs gas leak detection method based on extreme value theory proposed by the invention can realize reliable VOCs leak detection under complex imaging conditions.
为实现上述目的,本发明采用了以下技术方案:To achieve the above object, the present invention has adopted the following technical solutions:
包括以下步骤,Include the following steps,
步骤1:获取红外视频数据中VOCs泄漏区和无泄漏区的数据进行预处理操作;Step 1: Obtain the data of the VOCs leakage area and the non-leakage area in the infrared video data for preprocessing;
步骤2:从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器;Step 2: Extract one-dimensional time series feature data of a certain length of pixels from the infrared video data, and train a one-dimensional convolutional neural network classifier;
步骤3:多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0;Step 3: Sample the spatiotemporal features of several pixels from the infrared video data multiple times, use a one-dimensional convolutional neural network classifier, import the output values into the EVT algorithm in the Bayesian framework, and train the parameters of the prior gamma distribution α 0 and β 0 ;
步骤4:输入相关参数,通过自适应算法调整阈值,输出预测结果;Step 4: Input the relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result;
其中,所述步骤2:从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器;具体包括如下细分步骤S21至S23:Wherein, the step 2: extracting a certain length of pixel point one-dimensional time series feature data from the infrared video data, and training a one-dimensional convolutional neural network classifier; specifically includes the following subdivision steps S21 to S23:
步骤S21:从具有VOCs泄漏的切分场景视频帧的VOCs气体区域每8*8或16*16的块中提取一个像素,形成若干具有长度L的像素点一维时序泄露数据(XL,1),L为场景帧数量,其中1代表此数据来自存在VOCs泄漏区域,且XL=[x′1 x′2 ... x′L]T;同时也从不具有VOCs泄漏的切分场景中VOCs气体区域以同样方式提取出若干具有相同长度的像素点一维时序正常数据(XL,0),其中0代表此数据来自正常区域;Step S21: Extract one pixel from every 8*8 or 16*16 block of the VOCs gas region of the segmented scene video frame with VOCs leakage to form one-dimensional time series leakage data of several pixels with length L (X L , 1 ), L is the number of scene frames, where 1 means that the data comes from an area with VOCs leakage, and XL = [x′ 1 x′ 2 ... x′ L ] T ; at the same time, there is never a segmented scene with no VOCs leakage In the VOCs gas area in the same way, several one-dimensional time series normal data (X L , 0) of pixels with the same length are extracted, where 0 means that this data comes from the normal area;
步骤S22:首先对提取得到的像素点一维时序数据XL进行数值归一化,使像素点一维时序数据XL每项元素满足0≤x′i≤255,i=1,2,...,L,随后对每一一维时序数据元素x′i进行零均值化;再分别对处理完成后的两类数据进行切分,80%作为训练数据,20%作为验证数据;Step S22: First, perform numerical normalization on the extracted one-dimensional time series data XL of pixels, so that each element of the one-dimensional time series data XL of pixels satisfies 0≤x′ i≤255 , i =1, 2, . .., L, and then zero-average each one-dimensional time series data element x′ i ; then divide the two types of data after processing, 80% as training data, and 20% as verification data;
步骤S23:使用处理完成的训练数据训练一维卷积神经网络分类器,一维卷积神经网络分类器的输入为像素点一维时序数据XL,输出为D(XL),其中D(XL)∈(0,1),当分类器在验证数据集上分类准确率达到98%以上后停止训练,从而得到一维卷积神经网络分类模型;Step S23: Use the processed training data to train a one-dimensional convolutional neural network classifier, the input of the one-dimensional convolutional neural network classifier is the one-dimensional time series data XL of pixels, and the output is D( XL ), where D( X L )∈(0, 1), stop training when the classification accuracy of the classifier on the validation data set reaches more than 98%, thus obtaining a one-dimensional convolutional neural network classification model;
所述步骤3:多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0,具体包括如下细分步骤S31至S34:Step 3: Sample the spatiotemporal features of several pixels from the infrared video data multiple times, use a one-dimensional convolutional neural network classifier, import the output value into the EVT algorithm in the Bayesian framework, and train the parameter α of the prior gamma distribution 0 and β 0 , specifically including the following subdivision steps S31 to S34:
步骤S31:从待测切分场景视频帧的暗部每8*8或16*16的块中随机提取一个像素,得到K个有长度L的像素点一维时序数据XL,送入一阶段一维卷积神经网络分类器,得到输出D(XL),其中D(XL)∈(0,1);Step S31: randomly extract a pixel from every 8*8 or 16*16 block of the dark part of the video frame of the segmented scene to be measured, obtain K one-dimensional time series data XL of pixels with length L , and send it into a stage-1 dimensional convolutional neural network classifier, and get the output D( XL ), where D( XL )∈(0,1);
步骤S32:将K个XL=[x′1 x′2 ... x′L]T作为数据x,K个D(XL)为对应的标签序列y,构造EVT训练算法数据集T;Step S32: Construct the EVT training algorithm data set T by using K pieces of XL =[x' 1 x' 2 ... x' L ] T as data x, and K pieces of D( XL ) as corresponding label sequences y;
步骤S33:从数据集T中挑选出负样本g={xi|yi=0},根据右尾部概率pu查找负样本g的上限阈值u,通过上限阈值取出右尾部t,更新未被标记为异常的数据的充分统计量n和s;Step S33: Select the negative sample g={x i |y i =0} from the data set T, find the upper threshold u of the negative sample g according to the right tail probability p u , take out the right tail t through the upper threshold, and update the Sufficient statistics n and s for data marked as anomalies;
步骤S34:调整先验伽马分布的参数α0和β0,表示为Step S34: Adjust the parameters α 0 and β 0 of the prior gamma distribution, expressed as
α0=1+w0 α 0 =1+w 0
其中,w0是分配给训练集的样本计数的权重;where w 0 is the weight assigned to the sample count of the training set;
所述步骤4:输入相关参数,通过自适应算法调整阈值,输出预测结果,具体包括如下细分步骤S41至S43:Described step 4: input relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result, which specifically includes the following subdivision steps S41 to S43:
步骤S41:将待测切分场景视频帧,使用一维卷积神经网络分类器,构造数据集;根据步骤S33的方式,找出数据集的上限阈值,并取出右尾部t1的所有样本执行KolmogorovSmirnov测试,以发现和消除异常,KolmogorovSmirnov测试方法为Step S41: Divide the video frame of the scene to be tested, and use a one-dimensional convolutional neural network classifier to construct a data set; according to the method of step S33, find out the upper limit threshold of the data set, and take out all the samples of the right tail t1 to execute KolmogorovSmirnov test to find and eliminate anomalies, the KolmogorovSmirnov test method is
其中in
计算得到Dn,然后,移除最大的样本,并使用剩余的样本计算Dn;不断循环,最后选择其中使得Dn最小的样本的标签值记为 Calculate D n , then remove the largest sample, and use the remaining samples to calculate Dn; keep looping, and finally select the label value of the sample that makes D n the smallest and record it as
步骤S42:在去除异常后,选择以为阈值,提取出右尾部,使用训练期间的先验估计来计算整个序列的后验,更新为α1和β1;Step S42: After removing the abnormality, select the is the threshold value, extract the right tail, use the prior estimation during training to calculate the posterior of the entire sequence, and update it to α 1 and β 1 ;
步骤S43:将S42中计算得到的后验作为先验,设置以样本xj为中心的窗口W,表示为根据右尾部概率pu查找窗口的上限阈值u2,提取右尾部,调整α、β和通过计算得到yj,表示为Step S43: Take the posterior calculated in S42 as the prior, and set the window W with the sample x j as the center, expressed as Find the upper threshold u2 of the window according to the right tail probability p u , extract the right tail, adjust α, β and y j is obtained by calculation, which is expressed as
其中pf为目标出错率,不断循环获得所有样本的标签值yj。where p f is the target error rate, and the label value y j of all samples is continuously obtained in a loop.
进一步的,上述步骤1:获取存在VOCs泄漏和无泄漏的红外视频数据并对进行数据预处理,具体包括如下细分步骤S11至S12:Further, the above step 1: obtaining infrared video data with and without leakage of VOCs and performing data preprocessing, which specifically includes the following subdivided steps S11 to S12:
步骤S11:获取存在VOCs泄漏和无泄漏的红外视频数据;Step S11: Acquire infrared video data with and without leakage of VOCs;
步骤S12:对红外视频数据进行随机旋转、帧尺寸归一化、场景切分这些预处理操作。Step S12: Perform preprocessing operations such as random rotation, frame size normalization, and scene segmentation on the infrared video data.
另一方面,本发明还公开一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测系统,包括以下单元,On the other hand, the present invention also discloses an adaptive and robust VOCs gas leak detection system based on extreme value theory, comprising the following units:
数据获取和处理单元,用于获取红外视频数据中VOCs泄漏区和无泄漏区的数据进行预处理操作;The data acquisition and processing unit is used to acquire the data of the VOCs leakage area and the non-leakage area in the infrared video data for preprocessing;
一维网络结构训练单元,用于从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器;The one-dimensional network structure training unit is used to extract one-dimensional time series feature data of a certain length of pixels from infrared video data, and train a one-dimensional convolutional neural network classifier;
参数确定单元,用于多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0;The parameter determination unit is used to sample the spatiotemporal features of several pixels from the infrared video data multiple times. It uses a one-dimensional convolutional neural network classifier, and the output value is imported into the EVT algorithm in the Bayesian framework to train the parameters of the prior gamma distribution. α 0 and β 0 ;
预测单元,用于输入相关参数,通过自适应算法调整阈值,输出预测结果;The prediction unit is used to input relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result;
其中,所述一维网络结构训练单元,具体处理步骤如下:The specific processing steps of the one-dimensional network structure training unit are as follows:
步骤S21:从具有VOCs泄漏的切分场景视频帧的VOCs气体区域每8*8或16*16的块中提取一个像素,形成若干具有长度L的像素点一维时序泄露数据(XL,1),L为场景帧数量,其中1代表此数据来自存在VOCs泄漏区域,且XL=[x′1 x′2 ... x′L]T;同时也从不具有VOCs泄漏的切分场景中VOCs气体区域以同样方式提取出若干具有相同长度的像素点一维时序正常数据(XL,0),其中0代表此数据来自正常区域;Step S21: Extract one pixel from every 8*8 or 16*16 block of the VOCs gas region of the segmented scene video frame with VOCs leakage to form one-dimensional time series leakage data of several pixels with length L (X L , 1 ), L is the number of scene frames, where 1 means that the data comes from an area with VOCs leakage, and XL = [x′ 1 x′ 2 ... x′ L ] T ; at the same time, there is never a segmented scene with no VOCs leakage In the VOCs gas area in the same way, several one-dimensional time series normal data (X L , 0) of pixels with the same length are extracted, where 0 means that this data comes from the normal area;
步骤S22:首先对提取得到的像素点一维时序数据XL进行数值归一化,使像素点一维时序数据XL每项元素满足0≤x′i≤255,i=1,2,...,L,随后对每一一维时序数据元素x′i进行零均值化;再分别对处理完成后的两类数据进行切分,80%作为训练数据,20%作为验证数据;Step S22: First, perform numerical normalization on the extracted one-dimensional time series data XL of pixels, so that each element of the one-dimensional time series data XL of pixels satisfies 0≤x′ i≤255 , i =1, 2, . .., L, and then zero-average each one-dimensional time series data element x′ i ; then divide the two types of data after processing, 80% as training data, and 20% as verification data;
步骤S23:使用处理完成的训练数据训练一维卷积神经网络分类器,一维卷积神经网络分类器的输入为像素点一维时序数据XL,输出为D(XL),其中D(XL)∈(0,1),当分类器在验证数据集上分类准确率达到98%以上后停止训练,从而得到一维卷积神经网络分类模型;Step S23: Use the processed training data to train a one-dimensional convolutional neural network classifier, the input of the one-dimensional convolutional neural network classifier is the one-dimensional time series data XL of pixels, and the output is D( XL ), where D( X L )∈(0, 1), stop training when the classification accuracy of the classifier on the validation data set reaches more than 98%, thus obtaining a one-dimensional convolutional neural network classification model;
所述参数确定单元具体处理步骤如下:The specific processing steps of the parameter determination unit are as follows:
步骤S31:从待测切分场景视频帧的暗部每8*8或16*16的块中随机提取一个像素,得到K个有长度L的像素点一维时序数据XL,送入一维卷积神经网络分类器,得到输出D(XL),其中D(XL)∈(0,1);Step S31: randomly extract a pixel from every 8*8 or 16*16 block of the dark part of the video frame of the segmented scene to be measured, obtain K one-dimensional time series data XL of pixels with length L , and send it into a one-dimensional volume Integrate the neural network classifier to get the output D( XL ), where D( XL )∈(0,1);
步骤S32:将K个XL=[x′1 x′2 ... x′L]T作为数据x,K个D(XL)为对应的标签序列y,构造EVT训练算法数据集T;Step S32: Construct the EVT training algorithm data set T by using K pieces of XL =[x' 1 x' 2 ... x' L ] T as data x, and K pieces of D( XL ) as corresponding label sequences y;
步骤S33:从数据集T中挑选出负样本g={xi|yi=0},根据右尾部概率pu查找负样本g的上限阈值u,通过上限阈值取出右尾部t,更新未被标记为异常的数据的充分统计量n和s,Step S33: Select the negative sample g={x i |y i =0} from the data set T, find the upper threshold u of the negative sample g according to the right tail probability p u , take out the right tail t through the upper threshold, and update the Sufficient statistics n and s for data marked as anomalies,
步骤S34:调整先验伽马分布的参数α0和β0,表示为Step S34: Adjust the parameters α 0 and β 0 of the prior gamma distribution, expressed as
α0=1+w0 α 0 =1+w 0
其中,w0是分配给训练集的样本计数的权重;where w 0 is the weight assigned to the sample count of the training set;
所述预测单元,具体处理步骤如下:The specific processing steps of the prediction unit are as follows:
步骤S41:将待测切分场景视频帧,使用一维卷积神经网络分类器,构造数据集;根据步骤S33的方式,找出数据集的上限阈值,并取出右尾部t1的所有样本执行KolmogorovSmirnov测试,以发现和消除异常,KolmogorovSmirnov测试方法为Step S41: Divide the video frame of the scene to be tested, and use a one-dimensional convolutional neural network classifier to construct a data set; according to the method of step S33, find out the upper limit threshold of the data set, and take out all the samples of the right tail t1 to execute KolmogorovSmirnov test to find and eliminate anomalies, the KolmogorovSmirnov test method is
其中in
计算得到Dn,然后,移除最大的样本,并使用剩余的样本计算Dn;不断循环,最后选择其中使得Dn最小的样本的标签值记为 Calculate D n , then remove the largest sample, and use the remaining samples to calculate Dn; keep looping, and finally select the label value of the sample that makes D n the smallest and record it as
步骤S42:在去除异常后,选择以为阈值,提取出右尾部,使用训练期间的先验估计来计算整个序列的后验,更新为α1和β1;Step S42: After removing the abnormality, select the is the threshold value, extract the right tail, use the prior estimation during training to calculate the posterior of the entire sequence, and update it to α 1 and β 1 ;
步骤S43:将S42中计算得到的后验作为先验,设置以样本xj为中心的窗口W,表示为根据右尾部概率pu查找窗口的上限阈值u2,提取右尾部,调整α、β和通过计算得到yj,表示为Step S43: Take the posterior calculated in S42 as the prior, and set the window W with the sample x j as the center, expressed as Find the upper threshold u2 of the window according to the right tail probability p u , extract the right tail, adjust α, β and y j is obtained by calculation, which is expressed as
其中pf为目标出错率,不断循环获得所有样本的标签值yj。where p f is the target error rate, and the label value y j of all samples is continuously obtained in a loop.
再一方面,本发明还公开一种计算机可读存储介质,存储有计算机程序,所述计算机程序被处理器执行时,使得所述处理器执行上述基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法的步骤。In another aspect, the present invention also discloses a computer-readable storage medium storing a computer program, which, when executed by a processor, causes the processor to execute the above-mentioned adaptive robustness VOCs gas based on extreme value theory Steps of a leak detection method.
由上述技术方案可知,本发明的基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法和系统,克服现有方法的不足,充分利用红外视频数据中VOCs气体区域像素点的时空特征使用卷积神经网络对红外视频图像进行预筛查,通过贝叶斯框架内的极值理论(EVT)来优化调整筛查阈值,通过用指数分布(广义帕累托分布的特殊情况)逼近分数的概率密度函数的右尾部,并使用从训练数据中学习的伽马共轭先验,可以降低错误率的可变性并提高整体性能。从而在实现VOCs泄漏鲁棒性检测。It can be seen from the above technical solutions that the adaptive and robust VOCs gas leak detection method and system based on the extreme value theory of the present invention overcomes the shortcomings of the existing methods, and makes full use of the spatiotemporal characteristics of the pixels in the VOCs gas region in the infrared video data. Infrared video images are pre-screened by an integrated neural network, and the screening threshold is optimally adjusted by extreme value theory (EVT) within a Bayesian framework, by approximating the probability of the score with an exponential distribution (a special case of the generalized Pareto distribution). The right tail of the density function, and using a gamma-conjugate prior learned from the training data, reduces error rate variability and improves overall performance. So as to realize the robust detection of VOCs leakage.
附图说明Description of drawings
图1为本发明方法总体网络模型示意图;1 is a schematic diagram of the overall network model of the method of the present invention;
图2是本发明的实验结果图。Figure 2 is a graph of the experimental results of the present invention.
具体实施方式Detailed ways
为使本发明实施例的目的、技术方案和优点更加清楚,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本发明一部分实施例,而不是全部的实施例。In order to make the purposes, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments These are some embodiments of the present invention, but not all embodiments.
如图1所示,本实施例所述的基于极值理论的自适应鲁棒性VOCs气体泄漏检测方法,包括以下步骤:As shown in FIG. 1 , the adaptive and robust VOCs gas leak detection method based on extreme value theory described in this embodiment includes the following steps:
步骤1:获取红外视频数据中VOCs泄漏区和无泄漏区的数据进行预处理操作;Step 1: Obtain the data of the VOCs leakage area and the non-leakage area in the infrared video data for preprocessing;
步骤2:从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器。Step 2: Extract one-dimensional time series feature data of a certain length of pixels from the infrared video data, and train a one-dimensional convolutional neural network classifier.
步骤3:多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0。Step 3: Sample the spatiotemporal features of several pixels from the infrared video data multiple times, use a one-dimensional convolutional neural network classifier, import the output values into the EVT algorithm in the Bayesian framework, and train the parameters of the prior gamma distribution α 0 and β 0 .
步骤4:输入相关参数,通过自适应算法调整阈值,输出预测结果。Step 4: Input the relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result.
以下具体说明:The following specific instructions:
进一步地,上述步骤1:获取存在VOCs泄漏和无泄漏的红外视频数据并对进行数据预处理。具体包括如下细分步骤S11至S12:Further, the above step 1: obtain infrared video data with and without leakage of VOCs and perform data preprocessing. Specifically, it includes the following subdivision steps S11 to S12:
S11:获取存在VOCs泄漏和无泄漏的红外视频数据。S11: Obtain infrared video data with and without leakage of VOCs.
S12:对红外视频数据进行随机旋转、帧尺寸归一化、场景切分等预处理操作。S12: Perform preprocessing operations such as random rotation, frame size normalization, and scene segmentation on the infrared video data.
进一步地,上述步骤2:从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器。具体包括如下细分步骤S21至S23:Further, in the above step 2: extracting one-dimensional time series feature data of pixels of a certain length from the infrared video data, and training a one-dimensional convolutional neural network classifier. Specifically, it includes the following subdivision steps S21 to S23:
S21:从具有VOCs泄漏的切分场景视频帧的暗部(VOCs气体区域)每8*8或16*16的块中提取一个像素,形成若干具有长度L(场景帧数量,本发明中选取L=160)的像素点一维时序泄露数据(XL,1),其中1代表此数据来自存在VOCs泄漏区域,且XL=[x′1 x′2 ... x′L]T;同时也从不具有VOCs泄漏的切分场景中暗部以同样方式提取出若干具有相同长度的像素点一维时序正常数据(XL,0),其中0代表此数据来自正常区域。S21: Extract one pixel from every 8*8 or 16*16 block of the dark part (VOCs gas region) of the segmented scene video frame with VOCs leakage to form a number of lengths L (the number of scene frames, in the present invention, L= 160) of the pixel point one-dimensional time series leakage data (X L , 1), where 1 represents that the data comes from the leakage area of VOCs, and XL =[x′ 1 x′ 2 ... x′ L ] T ; A number of one-dimensional time series normal data (X L , 0) of pixels with the same length are extracted from the dark part of the segmented scene without VOCs leakage in the same way, where 0 means that the data comes from the normal area.
S22:首先对提取得到的像素点一维时序数据XL进行数值归一化,使其每项元素满足0≤x′i≤255,i=1,2,...,L,随后对每一一维时序数据元素x′i进行零均值化。再分别对处理完成后的两类数据进行切分,80%作为训练数据,20%作为验证数据。S22: First, perform numerical normalization on the extracted one-dimensional time series data XL of pixels, so that each element satisfies 0≤x′ i ≤255, i=1, 2, . . . , L, and then for each element One-dimensional time series data elements x'i are zero-averaged. Then, the two types of data after processing are divided, 80% are used as training data, and 20% are used as verification data.
S23:使用处理完成的训练数据训练一维卷积神经网络分类器,一阶段分类器的输入为像素点一维时序数据XL,输出为得到输出D(XL),其中D(XL)∈(0,1),当分类器在验证数据集上分类准确率达到98%以上后停止训练。从而得到一阶段分类模型如Table1;S23: Use the processed training data to train a one-dimensional convolutional neural network classifier, the input of the one-stage classifier is the one-dimensional time series data XL of pixels, and the output is to obtain the output D( XL ), where D( XL ) ∈ (0, 1), stop training when the classifier achieves more than 98% classification accuracy on the validation dataset. Thus, a one-stage classification model such as Table1 is obtained;
Table 1Table 1
一阶段网络结构One-stage network structure
进一步地,上述步骤3:多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0。其中EVT训练算法过程如Table2;Further, the above-mentioned step 3: sampling the spatiotemporal features of several pixels from the infrared video data multiple times, using a one-dimensional convolutional neural network classifier, and importing the output value into the EVT algorithm in the Bayesian framework, training the prior gamma distribution. Parameters α 0 and β 0 . The EVT training algorithm process is shown in Table 2;
具体包括如下细分步骤S31至S34:Specifically, it includes the following subdivision steps S31 to S34:
S31:从待测切分场景视频帧的暗部每8*8或16*16的块中随机提取一个像素,得到K个有长度L(场景帧数量)的像素点一维时序数据XL,送入一阶段一维卷积神经网络,得到输出D(XL),其中D(XL)∈(0,1)。S31: Randomly extract a pixel from every 8*8 or 16*16 block of the dark part of the video frame of the segmented scene to be tested, obtain K one-dimensional time series data XL of pixels with length L (the number of scene frames), and send it to Enter the one-stage one-dimensional convolutional neural network to obtain the output D( XL ), where D( XL )∈(0, 1).
S32:将K个XL=[x′1 x′2 ... x′L]T作为数据x,K个D(XL)为对应的标签序列y,构造EVT训练算法数据集T。S32: Construct the EVT training algorithm data set T by taking K pieces of XL =[x' 1 x' 2 ... x' L ] T as data x, and K pieces of D( XL ) as corresponding label sequences y.
S33:从数据集T中挑选出负样本g={xi|yi=0},并将其代入公式{gi>u}=gpu,该公式表示为根据右尾部概率pu查找负样本g的上限阈值u,其中参数pu是右尾部的概率。通过上限阈值取出右尾部t,更新未被标记为异常的数据的充分统计量n和s,可表示为S33: Select the negative sample g={x i |y i =0} from the data set T, and substitute it into the formula { gi >u}=gp u , which is expressed as finding negative samples according to the right tail probability p u Upper threshold u for sample g, where parameter p u is the probability of the right tail. The right tail t is taken out by the upper threshold, and the sufficient statistics n and s of the data not marked as abnormal are updated, which can be expressed as
nj+1=nj+t nj+1 = nj +t
sj+1=sj+∑ts j+1 =s j +∑t
其中,n0和s0初始化为0。Among them, n 0 and s 0 are initialized to 0.
S34:调整先验伽马分布的参数α0和β0,可表示为S34: Adjust the parameters α 0 and β 0 of the prior gamma distribution, which can be expressed as
α0=1+w0 α 0 =1+w 0
其中,w0是分配给训练集的样本计数的权重。where w 0 is the weight assigned to the sample count of the training set.
Table 2Table 2
EVT训练算法细节EVT training algorithm details
进一步地,上述步骤4:输入相关参数,通过自适应算法调整阈值,输出预测结果,其中EVT自适应阈值算法细节如Table 3所示;Further, the above-mentioned step 4: input relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result, wherein the details of the EVT adaptive threshold algorithm are shown in Table 3;
具体包括如下细分步骤S41至S43:Specifically, it includes the following subdivision steps S41 to S43:
S41:根据步骤S33,找出上限阈值,取出尾部的所有样本执行一系列KolmogorovSmirnov(KS)测试,以发现和消除异常。即将待测切分场景视频帧,使用一维卷积神经网络分类器,构造数据集;根据步骤S33的方式,找出数据集的上限阈值,并取出右尾部t1的所有样本执行KS测试,KS测试方法为S41: According to step S33, find the upper threshold, take out all the samples in the tail and perform a series of Kolmogorov Smirnov (KS) tests to find and eliminate abnormalities. The video frame of the scene to be tested is divided, and a one-dimensional convolutional neural network classifier is used to construct a data set; according to the method of step S33, the upper threshold of the data set is found, and all samples in the right tail t1 are taken out to perform the KS test, KS The test method is
其中in
计算得到Dn或称之为Dn,1。然后,移除最大的样本,并使用剩余的样本计算Dn,2。不断迭代,直到得到最后选择使Dn,i最小的i的值记为S42:在去除异常后,选择以为阈值,根据步骤S33,提取出右尾部,使用训练期间的先验估计来计算整个序列的后验,更新为α1和β1。 Dn is calculated or referred to as Dn,1 . Then, the largest sample is removed, and Dn,2 is calculated using the remaining samples. Iterate continuously until you get Finally, select the value of i that minimizes D n, i and is recorded as S42: After removing the exception, select to As the threshold, according to step S33, the right tail is extracted, and the posterior of the entire sequence is calculated using the prior estimation during training, and updated to α 1 and β 1 .
S43:将S42中计算的到的后验作为先验。设置以样本为中心的窗口,其表示为根据{wi>u}=wpu查找上限阈值u2,根据S33,提取右尾部,调整α、β和通过计算得到yj,表示为S43: Use the posterior calculated in S42 as the prior. Sets the sample-centered window, which is represented as Find the upper threshold u2 according to { wi > u}=wp u , according to S33, extract the right tail, adjust α, β and y j is obtained by calculation, which is expressed as
其中pf为目标出错率。不断迭代可以获得调整后的分数yj,即实现自适应阈值。where p f is the target error rate. Continuous iteration can obtain an adjusted score y j , ie, an adaptive threshold is achieved.
Table 3Table 3
EVT自适应阈值算法细节EVT Adaptive Thresholding Algorithm Details
图2展示了红外视频中存在VOCs泄漏的一帧,该帧在一维卷积神经网络中未被识别为VOCs泄漏帧,而通过EVT阈值自适应算法,提取出右尾部计算yj,其中红框为泄漏的区域。从中可以看出本发明的方法可以有效检出VOCs气体泄漏情况,且在不同场景下,有EVT算法的模型总会比单一维CNN的检测性能高。Figure 2 shows a frame with VOCs leakage in the infrared video. This frame is not recognized as a VOCs leakage frame in the one-dimensional convolutional neural network. Through the EVT threshold adaptive algorithm, the right tail is extracted to calculate y j , where red The box is the leaked area. It can be seen from this that the method of the present invention can effectively detect VOCs gas leakage, and in different scenarios, the model with EVT algorithm will always have higher detection performance than single-dimensional CNN.
综上所述,本发明基于极值理论的自适应鲁棒VOCs气体泄漏检测的方法,使用一维卷积神经网络可以实现对VOCs气体泄漏快速预检测的同时降低计算量,使算法能够适应于性能受限的设备上;采用EVT阈值自适应的方式可以提高红外视频数据VOCs泄漏检测的鲁棒性,减少由光照、温度、气候等因素所造成错误识别的影响。To sum up, the method of the present invention based on the extreme value theory for adaptive and robust VOCs gas leak detection, using a one-dimensional convolutional neural network can achieve rapid pre-detection of VOCs gas leaks while reducing the amount of calculation, so that the algorithm can adapt to On devices with limited performance; adopting the EVT threshold adaptive method can improve the robustness of VOCs leakage detection in infrared video data, and reduce the impact of false identification caused by factors such as illumination, temperature, and climate.
另一方面,本发明还公开一种基于极值理论的自适应鲁棒性VOCs气体泄漏检测系统,包括以下单元,On the other hand, the present invention also discloses an adaptive and robust VOCs gas leak detection system based on extreme value theory, comprising the following units:
数据获取和处理单元,用于获取红外视频数据中VOCs泄漏区和无泄漏区的数据进行预处理操作;The data acquisition and processing unit is used to acquire the data of the VOCs leakage area and the non-leakage area in the infrared video data for preprocessing;
一维网络结构训练单元,用于从红外视频数据中提取一定长度像素点一维时序特征数据,训练一维卷积神经网络分类器;The one-dimensional network structure training unit is used to extract one-dimensional time series feature data of a certain length of pixels from infrared video data, and train a one-dimensional convolutional neural network classifier;
参数确定单元,用于多次从红外视频数据采样若干像素点的时空特征,使用一维卷积神经网络分类器,输出值导入贝叶斯框架内的EVT算法,训练先验伽马分布的参数α0和β0;The parameter determination unit is used to sample the spatiotemporal features of several pixels from the infrared video data multiple times. It uses a one-dimensional convolutional neural network classifier, and the output value is imported into the EVT algorithm in the Bayesian framework to train the parameters of the prior gamma distribution. α 0 and β 0 ;
预测单元,用于输入相关参数,通过自适应算法调整阈值,输出预测结果。The prediction unit is used to input relevant parameters, adjust the threshold through the adaptive algorithm, and output the prediction result.
进一步的,所述一维网络结构训练单元,具体处理步骤如下:Further, the specific processing steps of the one-dimensional network structure training unit are as follows:
S21:从具有VOCs泄漏的切分场景视频帧的VOCs气体区域每8*8或16*16的块中提取一个像素,形成若干具有长度L的像素点一维时序泄露数据(XL,1),L为场景帧数量,其中1代表此数据来自存在VOCs泄漏区域,且XL=[x′1 x′2 ... x′L]T;同时也从不具有VOCs泄漏的切分场景中VOCs气体区域以同样方式提取出若干具有相同长度的像素点一维时序正常数据(XL,0),其中0代表此数据来自正常区域;S21: Extract one pixel from every 8*8 or 16*16 block of the VOCs gas region of the segmented scene video frame with VOCs leakage to form one-dimensional time series leakage data of several pixels with length L (X L , 1) , L is the number of scene frames, where 1 means that the data comes from the area with VOCs leakage, and XL = [x′ 1 x′ 2 ... x′ L ] T ; at the same time, it also never has VOCs leakage in the segmentation scene The VOCs gas area extracts several one-dimensional time series normal data (X L , 0) of pixels with the same length in the same way, where 0 represents that the data comes from the normal area;
S22:首先对提取得到的像素点一维时序数据XL进行数值归一化,使像素点一维时序数据XL每项元素满足0≤x′i≤255,i=1,2,...,L,随后对每一一维时序数据元素x′i进行零均值化;再分别对处理完成后的两类数据进行切分,80%作为训练数据,20%作为验证数据;S22: First, perform numerical normalization on the extracted one-dimensional time series data XL of pixels, so that each element of the one-dimensional time series data XL of pixels satisfies 0≤x′ i≤255 , i =1, 2, .. , L, and then zero-average each one-dimensional time series data element x′ i ; then divide the two types of data after processing, 80% as training data, and 20% as verification data;
S23:使用处理完成的训练数据训练一维卷积神经网络分类器,一阶段分类器的输入为像素点一维时序数据XL,输出为D(XL),其中D(XL)∈(0,1),当分类器在验证数据集上分类准确率达到98%以上后停止训练,从而得到一阶段分类模型。S23: Use the processed training data to train a one-dimensional convolutional neural network classifier. The input of the one-stage classifier is the one-dimensional time series data XL of pixels, and the output is D( XL ), where D( XL )∈( 0, 1), when the classification accuracy rate of the classifier on the validation data set reaches more than 98%, the training is stopped, so as to obtain a one-stage classification model.
进一步的,所述参数确定单元具体处理步骤如下:Further, the specific processing steps of the parameter determination unit are as follows:
S31:从待测切分场景视频帧的暗部每8*8或16*16的块中随机提取一个像素,得到K个有长度L的像素点一维时序数据XL,送入一阶段一维卷积神经网络,得到输出D(XL),其中D(XL)∈(0,1);S31: Randomly extract one pixel from every 8*8 or 16*16 block in the dark part of the video frame of the segmented scene to be tested, obtain K one-dimensional time series data XL of pixels with length L , and send it into one-stage one-dimensional Convolutional Neural Network, get the output D( XL ), where D( XL )∈(0,1);
S32:将K个XL=[x′1 x′2 ... x′L]T作为数据x,K个D(XL)为对应的标签序列y,构造EVT训练算法数据集T;S32: Construct the EVT training algorithm dataset T by using K XL = [x' 1 x' 2 ... x' L ] T as data x, and K D(X L ) as corresponding label sequence y;
S33:从数据集T中挑选出负样本g={xi|yi=0},并将其代入公式{gi>u}=gpu查找上限阈值u,其中参数pu是右尾部的概率,通过上限阈值取出右尾部t,更新未被标记为异常的数据的充分统计量n和s,表示为S33: Select the negative sample g={x i |y i =0} from the data set T, and substitute it into the formula { gi >u}=gp u to find the upper threshold u, where the parameter p u is the right tail Probability, taking out the right tail t through the upper threshold, updating the sufficient statistics n and s of the data not marked as abnormal, expressed as
nj+1=nj+t nj+1 = nj +t
sj+1=sj+∑ts j+1 =s j +∑t
其中,n0和s0初始化为0;Among them, n 0 and s 0 are initialized to 0;
S34:调整先验伽马分布的参数α0和β0,表示为S34: Adjust the parameters α 0 and β 0 of the prior gamma distribution, expressed as
α0=1+w0 α 0 =1+w 0
其中,w0是分配给训练集的样本计数的权重。where w 0 is the weight assigned to the sample count of the training set.
进一步的,所述预测单元,具体处理步骤如下:Further, the specific processing steps of the prediction unit are as follows:
S41:根据步骤S33,找出上限阈值,取出尾部的所有样本即将待测切分场景视频帧,使用一维卷积神经网络,构造数据集;根据步骤S33的方式,找出数据集的上限阈值,并取出右尾部t1的所有样本执行一系列KolmogorovSmirnov(KS)测试,以发现和消除异常,KS为S41: According to step S33, find the upper limit threshold, take out all the samples at the tail, and then divide the video frame of the scene to be tested, and use a one-dimensional convolutional neural network to construct a data set; According to the method of step S33, find the upper limit threshold of the data set , and take out all samples of the right tail t1 to perform a series of KolmogorovSmirnov (KS) tests to find and eliminate anomalies, KS is
其中in
计算得到Dn或称之为Dn,1,然后,移除最大的样本,并使用剩余的样本计算Dn,2,不断迭代,直到得到最后选择使Dn,i最小的i的值记为 Calculate D n or call it D n, 1 , then remove the largest sample and use the remaining samples to calculate D n, 2 , and iterate until you get Finally, select the value of i that minimizes D n, i and is recorded as
S42:在去除异常后,选择以为阈值,提取出右尾部,使用训练期间的先验估计来计算整个序列的后验,更新为α1和β1;S42: After removing the exception, select to is the threshold value, extract the right tail, use the prior estimation during training to calculate the posterior of the entire sequence, and update it to α 1 and β 1 ;
S43:将S42中计算得到的后验作为先验,设置以样本xj为中心的窗口W,表示为根据右尾部概率pu查找窗口的上限阈值u2,提取右尾部,调整α、β和通过计算得到yj,表示为S43: Take the posterior calculated in S42 as the prior, and set the window W with the sample x j as the center, expressed as Find the upper threshold u2 of the window according to the right tail probability p u , extract the right tail, adjust α, β and y j is obtained by calculation, which is expressed as
其中pf为目标出错率,不断迭代获得调整后的分数yj,即实现自适应阈值。where p f is the target error rate, and iteratively obtains the adjusted score y j , that is, the adaptive threshold is realized.
本发明还公开一种计算机可读存储介质,存储有计算机程序,所述计算机程序被处理器执行时,使得所述处理器执行如上述方法的步骤。The present invention also discloses a computer-readable storage medium, which stores a computer program, and when the computer program is executed by a processor, causes the processor to execute the steps of the above method.
可理解的是,本发明实施例提供的系统与本发明实施例提供的方法相对应,相关内容的解释、举例和有益效果可以参考上述方法中的相应部分。It is understandable that the system provided by the embodiment of the present invention corresponds to the method provided by the embodiment of the present invention, and reference may be made to the corresponding part of the above-mentioned method for explanation, examples and beneficial effects of related content.
本领域内的技术人员应明白,本申请的实施例可提供为方法、系统、或计算机程序产品。因此,本申请可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。As will be appreciated by those skilled in the art, the embodiments of the present application may be provided as a method, a system, or a computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.
本申请是参照根据本申请实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the present application. It will be understood that each process and/or block in the flowchart illustrations and/or block diagrams, and combinations of processes and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to the processor of a general purpose computer, special purpose computer, embedded processor or other programmable data processing device to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing device produce Means for implementing the functions specified in a flow or flow of a flowchart and/or a block or blocks of a block diagram.
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions The apparatus implements the functions specified in the flow or flow of the flowcharts and/or the block or blocks of the block diagrams.
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded on a computer or other programmable data processing device to cause a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process such that The instructions provide steps for implementing the functions specified in the flow or blocks of the flowcharts and/or the block or blocks of the block diagrams.
以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。The above embodiments are only used to illustrate the technical solutions of the present invention, but not to limit them; although the present invention has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that: The recorded technical solutions are modified, or some technical features thereof are equivalently replaced; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of the embodiments of the present invention.
Claims (4)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111013939.8A CN113780138B (en) | 2021-08-31 | 2021-08-31 | Self-adaptive robustness VOCs gas leakage detection method, system and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202111013939.8A CN113780138B (en) | 2021-08-31 | 2021-08-31 | Self-adaptive robustness VOCs gas leakage detection method, system and storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113780138A CN113780138A (en) | 2021-12-10 |
CN113780138B true CN113780138B (en) | 2022-09-13 |
Family
ID=78840513
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202111013939.8A Active CN113780138B (en) | 2021-08-31 | 2021-08-31 | Self-adaptive robustness VOCs gas leakage detection method, system and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113780138B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115751203B (en) * | 2022-12-05 | 2025-06-06 | 南通中油燃气有限责任公司 | A natural gas pipeline leakage monitoring system based on infrared thermal imager |
CN117830731B (en) * | 2024-01-02 | 2024-06-28 | 北京蓝耘科技股份有限公司 | Multi-dimensional parallel scheduling method |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106599864B (en) * | 2016-12-21 | 2020-01-07 | 中国科学院光电技术研究所 | A deep face recognition method based on extreme value theory |
CN110672278B (en) * | 2019-10-16 | 2021-04-30 | 北京工业大学 | Method for quantitatively measuring VOCs leakage of production device based on infrared imaging |
CN111242404B (en) * | 2019-11-12 | 2022-08-12 | 中国水利水电科学研究院 | A method and system for extreme evaluation of flood events induced by heavy rainfall |
CN113008470B (en) * | 2020-07-22 | 2024-02-13 | 威盛电子股份有限公司 | Gas leak detection device and gas leak detection method |
CN112163624A (en) * | 2020-09-30 | 2021-01-01 | 上海交通大学 | Data anomaly judgment method and system based on deep learning and extreme value theory |
-
2021
- 2021-08-31 CN CN202111013939.8A patent/CN113780138B/en active Active
Also Published As
Publication number | Publication date |
---|---|
CN113780138A (en) | 2021-12-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108182454A (en) | Safety check identifying system and its control method | |
CN113780138B (en) | Self-adaptive robustness VOCs gas leakage detection method, system and storage medium | |
CN114581709B (en) | Model training, method, device and medium for identifying objects in medical images | |
CN114092450B (en) | Real-time image segmentation method, system and device based on gastroscopy video | |
CN116778482B (en) | Embryo image blastomere target detection method, computer equipment and storage medium | |
CN111507998A (en) | Depth cascade-based multi-scale excitation mechanism tunnel surface defect segmentation method | |
CN111191584B (en) | Face recognition method and device | |
CN113205052A (en) | Spore identification method, system and medium based on residual error deep network | |
CN110929635A (en) | False face video detection method and system based on face cross-over ratio under trust mechanism | |
CN112949378A (en) | Bacterial microscopic image segmentation method based on deep learning network | |
CN115410059A (en) | Remote sensing image part supervision change detection method and device based on contrast loss | |
CN117314826A (en) | Performance detection method of display screen | |
CN117934489A (en) | Fundus hard exudate segmentation method based on residual error and pyramid segmentation attention | |
CN117456272A (en) | A self-supervised anomaly detection method based on contrastive learning | |
CN118397367A (en) | Tampering detection method based on convolution vision Mamba | |
CN114510715B (en) | Method and device for testing functional safety of model, storage medium and equipment | |
CN118014048A (en) | Low-illumination face detection model construction method, device and terminal | |
CN117541581A (en) | Fine management system suitable for lithium battery production | |
CN117557872A (en) | Unsupervised anomaly detection method and device for optimizing storage mode | |
CN111598184B (en) | DenseNet-based image noise identification method and device | |
CN113255472A (en) | Face quality evaluation method and system based on random embedding stability | |
CN113409281A (en) | Image definition evaluation method and device based on depth model | |
CN112906707A (en) | Semantic segmentation method and device for surface defect image and computer equipment | |
CN118608845B (en) | A method and system for analyzing steganalysis images | |
Subha et al. | Analysis of deep learning based optimization techniques for oral cancer detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right | ||
TR01 | Transfer of patent right |
Effective date of registration: 20250324 Address after: 230000 no.5089 Wangjiang West Road, hi tech Zone, Hefei City, Anhui Province Patentee after: INSTITUTE OF ADVANCED TECHNOLOGY University OF SCIENCE AND TECHNOLOGY OF CHINA Country or region after: China Patentee after: University of Science and Technology of China Address before: 230088 5089 Wangjiang West Road, Hefei high tech Zone, Anhui Patentee before: INSTITUTE OF ADVANCED TECHNOLOGY University OF SCIENCE AND TECHNOLOGY OF CHINA Country or region before: China |