CN112394137A - Intelligent calibration method for monitoring environmental air quality - Google Patents

Intelligent calibration method for monitoring environmental air quality Download PDF

Info

Publication number
CN112394137A
CN112394137A CN201910747028.4A CN201910747028A CN112394137A CN 112394137 A CN112394137 A CN 112394137A CN 201910747028 A CN201910747028 A CN 201910747028A CN 112394137 A CN112394137 A CN 112394137A
Authority
CN
China
Prior art keywords
data
data set
model
unlabeled
monitoring
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910747028.4A
Other languages
Chinese (zh)
Inventor
祁柏林
王宁
张欣
魏景峰
刘闽
杜毅明
周晓磊
白雪
张镝
陈月
王兴刚
范秋枫
孟繁星
金继鑫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenyang Institute of Computing Technology of CAS
Original Assignee
Shenyang Institute of Computing Technology of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenyang Institute of Computing Technology of CAS filed Critical Shenyang Institute of Computing Technology of CAS
Priority to CN201910747028.4A priority Critical patent/CN112394137A/en
Publication of CN112394137A publication Critical patent/CN112394137A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N33/00Investigating or analysing materials by specific methods not covered by groups G01N1/00 - G01N31/00
    • G01N33/0004Gaseous mixtures, e.g. polluted air
    • G01N33/0006Calibrating gas analysers
    • G01N33/0008Details concerning storage of calibration data, e.g. in EEPROM
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Chemical & Material Sciences (AREA)
  • Combustion & Propulsion (AREA)
  • Food Science & Technology (AREA)
  • Medicinal Chemistry (AREA)
  • Analytical Chemistry (AREA)
  • Biochemistry (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention relates to an intelligent calibration method for monitoring environmental air quality. The method mainly aims at the problem of monitoring data deviation caused by low precision of a sensor of the current miniature monitoring instrument according to the gridding requirement of the current air environment monitoring, and calibrates the data monitored by the miniature monitoring instrument. And taking standard data of the national standard station as a target learning value. And placing the miniature monitoring instrument around a national standard station to train the learning rule under the same environment. The method can learn the rule close to the national standard station after training the micro monitoring instrument, and the accuracy of the monitoring data of the micro monitoring instrument can be improved by utilizing the new learning rule. SO mainly aiming at one of six air quality pollutants2The data are described as an example.

Description

Intelligent calibration method for monitoring environmental air quality
Technical Field
The invention relates to the field of air quality monitoring, in particular to an intelligent calibration method for monitoring the air quality of an environment.
Background
The quality status of the air environment is closely related to our life. Whether from media or other routes, the environment problem is more serious when the age of the people is the current time. This is a difficult task in view of the present. In order to realize a large number of standard stations released by the state for gridding supervision, the high-precision sensors have high cost, and monitoring point positions are not flexible enough in arrangement, so that the monitoring points are far away from the aim of realizing comprehensive supervision.
In the face of the emergence of the related policy of realizing gridding supervision, a plurality of miniature monitoring instruments which are convenient to distribute and have low cost emerge in the market. However, these monitoring devices have a common disadvantage: the physical characteristics of the sensors themselves cause some deviation in the monitored data. These data errors cause certain trouble to our normal environmental regulatory work.
Disclosure of Invention
In order to reduce the problem of data deviation caused by the physical characteristics of a common sensor, the intelligent calibration method for monitoring the air quality of the environment can improve the accuracy of a miniature monitor to a certain extent, and makes a certain reference value for solving the problem of data deviation of the miniature monitor caused by the physical characteristics of the sensor.
The technical scheme adopted by the invention for solving the technical problems is as follows: an intelligent calibration method for monitoring the quality of ambient air comprises the following steps:
data processing: integrating data transmitted into a database by a micro monitoring instrument and national standard data to obtain marked data and unmarked data; taking national standard data as a label, and duplicating a marked data set and an unmarked data set into two parts for collaborative training;
model training: respectively training the two marked data sets on an LSTM model to obtain two models with different parameters;
and (3) collaborative training: and applying the trained model to an unlabeled data set, selecting data with high confidence coefficient in the unlabeled data set, adding the data into the labeled data set of another trainer, and performing iterative training in such a way until all parameters of the model are stable.
The data processing adopts SO of one of six pollutants of air quality2The domestic standard data is used as a training target, and the data collected by the miniature monitoring instrument is used as input data.
The data processing comprises the following steps:
integrating data transmitted into a database by the micro monitoring instrument and the obtained national standard data according to the same time, wherein the national standard data corresponding to the data transmitted into the database by the micro monitoring instrument is existed at the same time, and the data transmitted into the database by the micro monitoring instrument at the time is marked data; the rest is unmarked data;
the marked data set is copied into two parts which are respectively used as a marked data set 1 and a marked data set 2; the unlabeled data set is also duplicated as unlabeled data set 1 and unlabeled data set 2, respectively.
The marked data is the data which is transmitted into the database by the miniature monitoring instrument and has corresponding national standard data at the same time, namely the data which accords with the national standard regulation; the non-marking data is the national standard data which does not correspond to the data of the micro monitoring instrument.
The model training adopts a long-term and short-term memory network model, and the neural network is divided into three layers.
The cooperative training comprises the following steps:
1) applying model 1 trained by labeled data set 1 to unlabeled data set 1, predicting by model, and obtaining data (x) from unlabeled data set 1μ1,yμ1) Confidence detection with the data of the marker dataset 1:
finding (x) in unlabeled dataset 1 in labeled dataset 1μ1,yμ1) The K-neighborhood values found form a set Z1, and y corresponds to the neighborhood value x in the labeled data set 1 and x in the unlabeled data set 1μ1Corresponding to yμ1Sorting the distance differences;
first unmarked dataset 1 (x)μ1,yμ1) Added to labeled dataset 1, ranked set Z1 as the test set, and model 2 trained using these results of the loss function output compared to the (x) results of the previous unlabeled dataset 1 not addedμ1,yμ1) Whether the loss function of (1) becomes small; decreasing indicates the piece of unmarked data (x)μ1,yμ1) The model 2 is optimized and used as an alternative for adding the unmarked data set 1; selecting a piece of data which can make the loss function reduce most from the alternatives of the unmarked data set 1 as the data with the highest confidence coefficient, and finally adding the piece of data into the marked data set 2 of the model 2; wherein xμ1In the unlabeled data set 1A piece of data xμ1,yμ1Is to pass the trained model 1 to a piece of data x in the unlabeled data set 1μ1The predicted value of (2);
applying model 2 trained by labeled data set 2 to unlabeled data set 2, predicting by model, according to (x) of unlabeled data set 2μ2,yμ2) Confidence detection with the data in the label data set 2:
finding (x) in unlabeled dataset 1 in labeled dataset 2μ2,yμ2) The K-neighborhood values found form a set Z2, and y corresponds to the neighborhood value x in the labeled data set 2 and x in the unlabeled data set 2μ2Corresponding to yμ2Sorting the distance differences;
first unmarked data set 2 (x)μ2,yμ2) Added to labeled dataset 2, ranked set Z2 as the test set, and model 1 trained using these results of the loss function output compared to the (x) results of the previous unlabeled dataset 2 not addedμ2,yμ2) Whether the loss function of (1) becomes small; decreasing indicates the piece of unmarked data (x)μ2,yμ2) The model 1 is optimized and used as an alternative for adding the marked data set 2; selecting a piece of data which can make the loss function reduce most from the alternatives of the unmarked data set 2 as the data with the highest confidence coefficient, and finally adding the piece of data into the marked data set 1 of the model 1; wherein xμ2Refers to a piece of data x in the unlabeled data set 2μ2,yμ2Is to pass the trained model 2 to a piece of data x in the unlabeled data set 2u2The predicted value of (2);
2) the above process is continuously and alternately repeated until the parameters of the last two trained models are closest.
The format of the marking data is (x, y); x represents the data collected by the micro monitoring instrument, namely the electric signal value, and y represents the national standard data.
The K-neighborhood value is used to calculate a prediction value y of the set of neighborhood valuesuA difference value with each neighboring value y;the K-proximity calculation formula is as follows;
d(xμ,xi)=||xμ,xi||2 (1)
in the formula: x is the number ofμRepresenting the data, i.e. electrical signal values, x, in an unmarked data setiIs xμIn the neighborhood of the marker dataset.
The invention has the following beneficial effects and advantages:
1. the data utilization rate is high. Through semi-supervised learning, both marked data and unmarked data can be fully utilized. Although we can now obtain a large amount of data, there is very little data that really meets the requirements that we can utilize. Unlabeled data may not work optimally in model training, but manual labeling presents significant difficulties. By using a semi-supervised learning method, the defect can be avoided, and the data can be fully utilized by people.
Multiple angles. The cooperative training is a multi-angle training method. It is known that viewing questions from different angles tends to be more comprehensive. It is also desirable to achieve such an effect here using cooperative training. The data set is divided into two parts which are the same, namely, the training of the model is carried out from different angles, so that the final fitting capability and the generalization capability of the model are better.
Drawings
FIG. 1 is a comparison of before and after data processing.
Fig. 2 is a flowchart of the overall algorithm.
FIG. 3 is a diagram of the structure of the LSTM.
Detailed Description
The present invention will be described in further detail with reference to examples.
As shown in fig. 2, an intelligent calibration method for monitoring ambient air quality includes the following steps:
step 1: and (6) data processing. The data transmitted into the database by the existing miniature monitoring instrument and the known national standard data are processed and then divided. And (3) taking national standard data as a label, copying the marked data set and the unmarked data set into two parts, and preparing for cooperative training in the early stage.
Step 2: and (5) training a model. Training the marked two data sets on an LSTM model to obtain two models with different parameters;
and step 3: and (5) detecting confidence. For data x in each unmarked datasetμPredicting y with trained modelsμThen find x in the tagged datasetμK proximity values that are close, and grouping the K proximity values into a set, the set being ordered by K-proximity distance. K-neighborhood distance is the predicted value y in calculating the set of neighborhood valuesuA difference value with each neighboring value y; the K-vicinity calculation formula is shown in formula (1). We will note the set of K neighbors sorted by K-neighbor distance as Z. We treat this set as a new test set, with each piece of prediction data (x)μ,yμ) And adding the new training set into the labeled data set to serve as a new training set, and training the corresponding model by using the new training set and the test set. Finding a (x) that minimizes the model loss functionμ,yμ) Of (x)μ,yμ) The data with the highest confidence.
And 4, step 4: and (5) performing cooperative training. According to the model trained in the step 2, two data sets can train out models with different parameters, then the trained model is applied to an unmarked data set, data with high confidence in the unmarked data set is selected and added into a marked data set of another trainer, and the iterative training is repeated in such a way until the marked data set is stable, that is, data which does not meet the confidence requirement in the unmarked data set needs to be added into the marked data set, which means that the model training is stable.
Data processing: SO used here is one of six air quality pollutants2The hour data of (2) and the monitoring data of a national standard station are used as training targets, and the data collected by a miniature monitoring instrument are used as input data to be trained. Because the obtained national standard data is limited, the data of a part of miniature monitoring instruments does not have the corresponding national standard data,we attribute such data as unlabeled data. Both the labeled and unlabeled data we duplicated in two to prepare for subsequent co-training.
The miniature monitoring instrument is small-scale air quality monitoring equipment, for example, small instruments erected above roads can be seen at each intersection in normal times. The national monitoring instrument needs to arrange a plurality of point positions, and has poor flexibility and high cost. Compared with the standard station put in China, the miniature monitoring instrument has low cost and high flexibility, but has the defect that the accuracy of the sensor is not high.
The model training method comprises the following steps: here we use the long short term memory network model (LSTM). Experiments show that the effect is optimal when the neural network is three layers. Two different rules are trained through two data sets.
The method for collaborative training comprises the following steps:
we apply the model trained with labeled data set 1 to unlabeled data set 1, and by model prediction, we predict (x) from unlabeled data 1μ1,yμ1) And carrying out confidence detection on the data in the labeled data 1, screening out data with high confidence, adding the data into a labeled data set 2 of another training model, and continuously and repeatedly training in a crossed manner until the last two labeled data sets are stable, namely, data which does not meet the confidence condition in the unlabeled data set 1, namely data which can reduce the loss function is not available, so that no new data is added into the labeled data set 2 to be stable, and each parameter of the model is stable (the parameter is not changed).
We apply the model trained with labeled data set 2 to unlabeled data set 2, and by model prediction, we predict (x) from unlabeled data 2μ2,yμ2) Carrying out confidence detection with the data in the labeled data 2, screening out the data with high confidence, adding the data into the labeled data set 1 of another training model, and repeating the training continuously and alternately until the last two labeled data sets are stable, namely, the data which does not meet the confidence condition in the unlabeled data set 2, namely the data which does not meet the confidence condition in the unlabeled data setThe existence of data enables the loss function to be reduced, so that the mark data set 1 has no new data to be added, and is stable, which also means that the parameters of the model are stable (the parameters are not changed).
Step 1: and (3) carrying out certain processing on the data, and pulling back the value with larger fluctuation to the upper limit, wherein the threshold value is set to be 2.3, namely when the data difference is greater than 2.3, the abnormal value is adjusted, and the difference value is controlled within the range of 2.3.
Step 2: we copy the processed labeled data and unlabeled data into two sets respectively, and model training is performed on the labeled data of the sets respectively, and the LSTM model is used here. Experimentally, the models for both datasets were adjusted to three layers.
And step 3: the trained models of the two labeled data sets are applied. The trained model of labeled dataset 1 is applied to unlabeled dataset 1, for data x in each unlabeled datasetμ1Predicting y with trained modelsμ1Then find x in the marked data setμ1The K neighbors that are close together are grouped into a set that is ordered by the K-close neighbors. K-proximity distance is calculated for the proximity value of x, predicted value yuThe difference between 1 and y. The K-vicinity calculation formula is shown in formula (1). At the same time, we also apply the trained model of labeled dataset 2 to unlabeled dataset 2, for data x in each unlabeled datasetμ2Predicting y with trained modelsμ2Then find x in the marked data setμ2The K neighboring values are similar and grouped into a set.
d(xμ,xi)=||xμ,xi||2
(1) In the formula: x is the number ofμIs the value of the electrical signal in the unmarked data set
xiIs xμAdjacent values in the labeled dataset.
And 4, step 4: according to step 3, we perform confidence detection. Firstly, distance sorting is carried out on the obtained K adjacent sets, and whether the obtained K-adjacent value sets and the original marked data sets reduce the cost brought by the loss function in the marked data sets or not is judged. The penalty function we use here is the root mean square error (mean squared error). If the loss function is decreased, we will find a piece of data in the set of K-neighborhood values that can maximize the decrease of the loss function, and add the piece of data with the highest confidence corresponding to the labeled data set 1 to the labeled data set 2, and similarly, add the piece of data with the highest confidence corresponding to the labeled data set 2 to the labeled data set 1. The two data sets are operated in a loop iteration mode until the two calibrated data sets can be kept stable or the maximum iteration times is reached, the algorithm is ended, and the currently updated optimal parameters are stored. Our trained model here is a three-layer, with the output dimension of the first layer being 22, the second layer being 22, the output being 22, and the third layer being 22, the output being one-dimensional. While the maximum number of iterations we set is 1000.
This is a comparison before and after processing the data, as shown in fig. 1. Because the variation amplitude of the monitoring data is stable, the difference between the data is ensured not to exceed 2.3 when the data is processed, and the threshold value is the optimal value after experimental verification.
After processing, we can implement our whole training according to the whole algorithm of fig. 2. According to FIG. 2, the labeled data are first copied into two copies, and then the LSTM model is trained respectively. The LSTM model has three control gates: forgetting gate, input gate and output gate. The LSTM model is shown in fig. 3, and their functional formula is as follows:
forget the door: f. oft=σ(Wf·[ht-1,xt]+bf) (2)
In the formula:
ft-indicating forgetting gate input
Sigma-an activation function, we use here sigmoid
Wf-weight matrix of forgetting gate
ht-1Output of last moment
xt-input of the current time
bf-forgetting gate bias terms
An input gate:
Figure BDA0002165886880000071
Ctstate unit of current time t input
ftOutput f of the forgetting gatet
Ct-1The state of the cell at the previous moment
it-input of the gate it
Figure BDA0002165886880000072
New cell state
it=σ(Wi·[ht-1,xt]+bi (4)
Wi-matrix weights of input gates
bi-biasing of the input gate
Ct=tanh(Wc·[ht-1,xt]+bc) (5)
Wc-matrix weights of output gates
bcBiasing of the output gate
tan h-output activation function
An output gate: o ist=σ(Wo[ht-1,xt]+bo) (6)
Ot-sigmoid gate output
Wo-calculating a weight matrix of cell states
bo-calculating the bias of the cell state
ht=Ot*tanh(Ct) (7)
We apply the trained models to unlabeled data 1, 2, respectively. The unlabeled data 1 is predicted by using an LSTM model trained by the labeled data 1, then data adjacent to K in the labeled data 1 is searched to form a K adjacent set, and then the distance is calculated according to the formula (1) and sorted according to the distance. And calculating the K neighbor data and the labeled data set 1 to retrain the model, and finding the data which can reduce the loss function of the model to the maximum in the K neighbor array if the loss function can be reduced, adding the data into the labeled data set 2, and repeating the iteration in the way until the model parameters are stable, and finishing the grating method.

Claims (8)

1. An intelligent calibration method for monitoring the quality of ambient air is characterized by comprising the following steps:
data processing: integrating data transmitted into a database by a micro monitoring instrument and national standard data to obtain marked data and unmarked data; taking national standard data as a label, and duplicating a marked data set and an unmarked data set into two parts for collaborative training;
model training: respectively training the two marked data sets on an LSTM model to obtain two models with different parameters;
and (3) collaborative training: and applying the trained model to an unlabeled data set, selecting data with high confidence coefficient in the unlabeled data set, adding the data into the labeled data set of another trainer, and performing iterative training in such a way until all parameters of the model are stable.
2. The intelligent calibration method for monitoring the air quality of the environment as claimed in claim 1, wherein the data processing employs SO of one of six pollutants of the air quality2The hour data of (2) takes national standard data as a training target, and data collected by a micro monitoring instrument is taken as input data.
3. The intelligent calibration method for monitoring the quality of the ambient air as claimed in claim 1, wherein the data processing comprises the following steps:
integrating data transmitted into a database by the micro monitoring instrument and the obtained national standard data according to the same time, wherein the national standard data corresponding to the data transmitted into the database by the micro monitoring instrument is existed at the same time, and the data transmitted into the database by the micro monitoring instrument at the time is marked data; the rest is unmarked data;
the marked data set is copied into two parts which are respectively used as a marked data set 1 and a marked data set 2; the unlabeled data set is also duplicated as unlabeled data set 1 and unlabeled data set 2, respectively.
4. The intelligent calibration method for monitoring the quality of the ambient air according to claim 1, wherein the labeled data is data which is transmitted into the database by the miniature monitoring instrument and has corresponding national standard data at the same time, namely data which meets the national standard regulation; the non-marking data is the national standard data which does not correspond to the data of the micro monitoring instrument.
5. The intelligent calibration method for monitoring the quality of the ambient air as claimed in claim 1, wherein the model training adopts a long-short term memory network model, and the neural network has three layers.
6. The intelligent calibration method for monitoring the quality of the ambient air as claimed in claim 1, wherein the cooperative training comprises the following steps:
1) applying model 1 trained by labeled data set 1 to unlabeled data set 1, predicting by model, and obtaining data (x) from unlabeled data set 1μ1,yμ1) Confidence detection with the data of the label data set 1:
finding (x) in unlabeled dataset 1 in labeled dataset 1μ1,yμ1) The K-neighborhood values found form a set Z1, and y corresponds to the neighborhood value x in the labeled data set 1 and x in the unlabeled data set 1μ1Corresponding to yμ1Sorting the distance differences;
first unmarked dataset 1 (x)μ1,yμ1) Added to the labeled dataset 1, the sorted set Z1 was used as the test set, and model 2 was trained using these to see the loss function result of the output compared to the previous (x) for the unlabeled dataset 1μ1,yμ1) Whether the loss function of (1) becomes small; the smaller the size, the unmarked data (x) isμ1,yμ1) The model 2 is optimized and used as an alternative for adding the unmarked data set 1; selecting a piece of data which can make the loss function reduce most from the alternatives of the unmarked data set 1 as the data with the highest confidence coefficient, and finally adding the piece of data into the marked data set 2 of the model 2; wherein xμ1Refers to a piece of data x in the unlabeled data set 1μ1,yμ1Is to pass the trained model 1 to a piece of data x in the unlabeled data set 1μ1The predicted value of (2);
applying model 2 trained by labeled data set 2 to unlabeled data set 2, predicting by model, according to (x) of unlabeled data set 2μ2,yμ2) Confidence detection with the data in the label data set 2:
finding (x) in unlabeled dataset 1 in labeled dataset 2μ2,yμ2) The K-neighborhood values found form a set Z2, and y corresponds to the neighborhood value x in the labeled data set 2 and x in the unlabeled data set 2μ2Corresponding to yμ2Sorting the distance differences;
first unmarked data set 2 (x)μ2,yμ2) Added to the labeled dataset 2, and the sorted set Z2 as the test set, using these to train model 1, see the loss function result of the output compared to the previous (x) for the unlabeled dataset 2μ2,yμ2) Whether the loss function of (1) becomes small; the smaller the size, the unmarked data (x) isμ2,yμ2) The model 1 is optimized and used as an alternative for adding the marked data set 2; selecting a piece of data which can make loss function reduce most from the alternatives of the unmarked data set 2 as the data with highest confidence degree, and finally using the piece of dataAdded to the labeled dataset 1 of model 1; wherein xμ2Refers to a piece of data x in the unlabeled data set 2μ2,yμ2Is to pass the trained model 2 to a piece of data x in the unlabeled data set 2μ2The predicted value of (2);
2) the above process is continuously and alternately repeated until the parameters of the last two trained models are closest.
7. The intelligent calibration method for monitoring the quality of the ambient air according to claim 5, wherein the format of the tag data is (x, y); x represents the data collected by the micro monitoring instrument, namely the electric signal value, and y represents the national standard data.
8. The intelligent calibration method for monitoring the quality of the ambient air according to claim 5, wherein the K-neighborhood value is used for calculating a neighborhood value set, and a predicted value yμA difference value with each neighboring value y; the K-proximity calculation formula is as follows;
d(xμ,xi)=||xμ,xi||2 (1)
in the formula: x is the number ofμRepresenting the data, i.e. electrical signal values, x, in an unmarked data setiIs xμIn the neighborhood of the marker dataset.
CN201910747028.4A 2019-08-14 2019-08-14 Intelligent calibration method for monitoring environmental air quality Pending CN112394137A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910747028.4A CN112394137A (en) 2019-08-14 2019-08-14 Intelligent calibration method for monitoring environmental air quality

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910747028.4A CN112394137A (en) 2019-08-14 2019-08-14 Intelligent calibration method for monitoring environmental air quality

Publications (1)

Publication Number Publication Date
CN112394137A true CN112394137A (en) 2021-02-23

Family

ID=74602708

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910747028.4A Pending CN112394137A (en) 2019-08-14 2019-08-14 Intelligent calibration method for monitoring environmental air quality

Country Status (1)

Country Link
CN (1) CN112394137A (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108469273A (en) * 2018-02-27 2018-08-31 济宁中科云天环保科技有限公司 High in the clouds data joint debugging calibration method based on machine learning algorithm
CN109558971A (en) * 2018-11-09 2019-04-02 河海大学 Intelligent landslide monitoring device and method based on LSTM shot and long term memory network

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108469273A (en) * 2018-02-27 2018-08-31 济宁中科云天环保科技有限公司 High in the clouds data joint debugging calibration method based on machine learning algorithm
CN109558971A (en) * 2018-11-09 2019-04-02 河海大学 Intelligent landslide monitoring device and method based on LSTM shot and long term memory network

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
孟想: "基于群智感知的空气质量传感器校准方法设计与实现", 《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》 *
蔡毅 等: "半监督集成学习综述", 《计算机化学》 *
邱云飞 等: "基于协同训练的意图分类优化方法", 《现代情报》 *

Similar Documents

Publication Publication Date Title
Siregar et al. Analysis of artificial neural network accuracy using backpropagation algorithm in predicting process (forecasting)
CN109142171B (en) Urban PM10 concentration prediction method based on feature expansion and fusing with neural network
Navone et al. Predicting Indian monsoon rainfall: a neural network approach
CN107798383B (en) Improved positioning method of nuclear extreme learning machine
CN110210625A (en) Modeling method, device, computer equipment and storage medium based on transfer learning
CN108334943A (en) The semi-supervised soft-measuring modeling method of industrial process based on Active Learning neural network model
WO2006120724A1 (en) Geographic information system using neural networks
CN105335619A (en) Collaborative optimization method applicable to parameter back analysis of high calculation cost numerical calculation model
CN108877947A (en) Depth sample learning method based on iteration mean cluster
Zeng et al. Short-term load forecasting of smart grid systems by combination of general regression neural network and least squares-support vector machine algorithm optimized by harmony search algorithm method
CN113393057A (en) Wheat yield integrated prediction method based on deep fusion machine learning model
Zhang et al. An improved genetic algorithm and neural network-based evaluation model of classroom teaching quality in colleges and universities
CN115017970A (en) Migration learning-based gas consumption behavior anomaly detection method and system
CN117350330A (en) Semi-supervised entity alignment method based on hybrid teaching
Lo Predicting software reliability with support vector machines
CN116611517A (en) Knowledge tracking method integrating graph embedding and attention
CN112394137A (en) Intelligent calibration method for monitoring environmental air quality
Chau Rainfall-runoff correlation with particle swarm optimization algorithm
CN116452854A (en) Adaptive image classification method based on width learning and random sensitivity
Khan et al. Forecasting renewable energy for environmental resilience through computational intelligence
CN115293249A (en) Power system typical scene probability prediction method based on dynamic time sequence prediction
CN114818945A (en) Small sample image classification method and device integrating category adaptive metric learning
Kavipriya et al. Adaptive Weight Deep Convolutional Neural Network (AWDCNN) Classifier for Predicting Student’s Performance in Job Placement Process
Wang et al. Prototype-based supervised contrastive learning method for noisy label correction in tire defect detection
Nagy Data-driven analysis of fractality and other characteristics of complex networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210223

RJ01 Rejection of invention patent application after publication