CN113627499A - Smoke intensity grade estimation method and device based on inspection station diesel vehicle tail gas image - Google Patents

Smoke intensity grade estimation method and device based on inspection station diesel vehicle tail gas image Download PDF

Info

Publication number
CN113627499A
CN113627499A CN202110859026.1A CN202110859026A CN113627499A CN 113627499 A CN113627499 A CN 113627499A CN 202110859026 A CN202110859026 A CN 202110859026A CN 113627499 A CN113627499 A CN 113627499A
Authority
CN
China
Prior art keywords
image
frame
network model
convolutional
residual
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110859026.1A
Other languages
Chinese (zh)
Other versions
CN113627499B (en
Inventor
康宇
庞现阳
张正
李兵兵
曹洋
夏秀山
许镇义
许楠钒
刘增林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN202110859026.1A priority Critical patent/CN113627499B/en
Publication of CN113627499A publication Critical patent/CN113627499A/en
Application granted granted Critical
Publication of CN113627499B publication Critical patent/CN113627499B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The invention discloses a smoke intensity grade estimation method and equipment based on a diesel vehicle tail gas image of an inspection station, which comprises the following steps of: s1: dividing videos collected by an inspection station into subblocks, sampling each subblock, and defining a residual frame according to an obtained RGB frame; s2: preprocessing the RGB image and the residual frame image; s3: constructing a space-time cross network model, training the model by utilizing a training data set and a verification data set, finely adjusting and optimizing the whole network based on a preset loss function and an optimizer until the network is converged, and storing the trained network model parameters; s4: and (4) carrying out grade estimation on the image to be detected by using the trained network model. The method firstly extracts the characteristics of the picture of the tail gas emitted by the diesel vehicle during the environmental annual inspection, effectively removes redundant information, calculates the amount directly related to the smoke concentration in the tail gas emitted by the vehicle through an algorithm, and has higher pertinence and higher feasibility.

Description

Smoke intensity grade estimation method and device based on inspection station diesel vehicle tail gas image
Technical Field
The invention relates to the technical field of environmental inspection, in particular to a smoke intensity level estimation method based on a diesel vehicle tail gas image of an inspection station.
Background
Along with the healthy and rapid development of economic development in China, the reserve of motor vehicles in China is increased. The diesel vehicle is one of the important components of motor vehicles in China, and the exhaust emission of the diesel vehicle causes serious atmospheric pollution. Therefore, in order to monitor and control the exhaust emission of various types of diesel vehicles more specifically, it is necessary to effectively classify the smoke emission level of the diesel vehicles.
At present, the motor vehicle exhaust monitoring in China is mainly divided into that arranged on a vehicle body: the system comprises a vehicle-mounted tail gas detection device (PEMS), a vehicle-mounted automatic diagnosis system (OBD) and a remote sensing monitoring instrument arranged at a fixed point; data monitored by the two methods are often subdivided into specific values of various pollutants, and integration of data of each part is needed to realize pollutant emission level estimation, which often brings heavy workload.
The grade estimation of the diesel vehicle exhaust emission is better realized through the exhaust, and the existing research has recently shown that the method for identifying the diesel vehicle exhaust emission by using a computer vision technology is a low-cost and convenient method. The space-time cross network provided by the invention constructs a double pyramid characteristic structure, and promotes mutual guidance of time domain characteristics and space domain characteristics to help the whole network learning, so that the smoke intensity level of the tail gas emission of the diesel vehicle can be accurately estimated, and the relevant departments can conveniently supervise and make policies.
Disclosure of Invention
The invention provides a smoke intensity grade estimation method and device based on a diesel vehicle tail gas image of an inspection station, which can solve the technical problem.
In order to achieve the purpose, the invention adopts the following technical scheme:
a smoke intensity grade estimation method based on a diesel vehicle tail gas image of an inspection station comprises the following steps:
s1: dividing videos collected by an inspection station into subblocks, sampling each subblock, and defining a residual frame according to an obtained RGB frame;
s2: preprocessing the RGB image and the residual frame image;
s3: constructing a space-time cross network model, training the model by utilizing a training data set and a verification data set, finely adjusting and optimizing the whole network based on a preset loss function and an optimizer until the network is converged, and storing the trained network model parameters;
s4: and (4) carrying out grade estimation on the image to be detected by using the trained network model.
Further, in S1, the operations of dividing the video collected by the checkpoint into subblocks, performing frame sampling on each subblock, and defining a residual frame according to the obtained RGB frame are as follows:
s11: dividing the monitoring video into N subblocks, extracting an RGB Frame from each subblock, and recording each Frame as a Framei,i<=N,
Figure BDA0003185075100000021
C, T, H, W are image channel number, time length, width and height;
s12: according to the RGB frame extracted at S11, the residual frame ResFrame is divided according to the following formula:
Figure BDA0003185075100000022
where alpha is a constant coefficient, beta is a set pixel value size to prevent residual frame size overflow,
Figure BDA0003185075100000023
c, T, H, W, which is the number of image channels, time length, width and height;
s13: the resulting image was displayed as 7: 2: the scale of 1 is randomly divided into a training set, a validation set, and a test set.
Further, the details of preprocessing the RGB image and the residual frame image in S2
The operation is as follows:
s21: simple scaling, random cropping, horizontal flipping, perspective transformation, region erasure and color dithering are performed on the RGB image and the residual frame image.
Further, a space-time cross network model is constructed in S3, the model is trained using a training data set and a verification data set, the whole network is fine-tuned and optimized based on a preset loss function and an optimizer until the network converges, and the specific operations of storing the trained network model parameters are as follows:
s31: the time domain and the space domain of the whole network model are consistent, and the main body adopts SE-ResNeXt-50 to extract space-time characteristics; the network structure comprises 4 convolutional residual blocks { res1, res2, res3, res4}, and the outputs of the blocks are {4,8,16,32} relative to stride of the input frame; the outside of the middle-included number is the repetition number of each residual block, on the basis of ResNeXt-50, the output identity mapping of the last layer of res1, res2 and res3 is selected at the residual stage of a time domain and a space domain to construct a double pyramid feature network, so that the time domain and the space domain are mutually guided, and feature maps output by res4 are fused and used for final classification head prediction; the feature fusion is the addition of corresponding elements of the spatial domain feature and the time domain feature, and the specific operation description is as follows:
Figure BDA0003185075100000031
wherein i is more than or equal to 1 and less than or equal to H, j is more than or equal to 1 and less than or equal to W, C is more than or equal to 1 and less than or equal to C and SF,
Figure BDA0003185075100000032
s32: the whole network architecture specifically comprises a 7 × 7 convolutional layer (conv1), a 3 × 3 pooling layer (pool1), three convolutional blocks (res1) comprising three convolutional layers, four convolutional blocks (res2) comprising three convolutional layers, 6 convolutional blocks (res3) comprising three convolutional layers, and three convolutional blocks (res4) comprising three convolutional layers, wherein the convolutional kernel size of the three convolutional layers contained in each convolutional block is 1 × 1, 3 × 3 and 1 × 1, and after module stacking, a feature fusion layer, an adaptive average pooling layer and a full connection layer are added, wherein the convolutional kernel size of the feature fusion layer and the adaptive average pooling layer is 1 × 1;
s33: the time domain flow part of the whole network model takes a residual error frame as input, the space domain part takes an RGB frame as input, a training set and a verification set are input into the network, Relu is selected as an activation function, a loss function is a cross entropy loss function, the batch size is set to be 3, the weight attenuation is set to be 0.0005, the momentum is set to be 0.9, the learning rate is 0.001, the weight of the model is optimized by using random gradient descent, and the iteration is stopped after the network converges, and the whole network model is stored.
Further, in S4, the trained network model is used, and the specific operation of performing the grade estimation on the image to be detected is as follows:
s41: and inputting the test set into a trained smoke degree estimation model in S3 to obtain a corresponding smoke degree grade.
On the other hand, the invention also discloses a smoke degree grade estimation system based on the inspection station diesel vehicle tail gas image, which comprises the following units:
the sub-block dividing unit is used for dividing the video collected by the inspection station into sub-blocks, sampling each sub-block and dividing a residual frame according to the obtained RGB frame;
the preprocessing unit is used for preprocessing the RGB image and the residual frame image;
the model construction unit is used for constructing a space-time cross network model, training the model by utilizing a training data set and a verification data set, finely adjusting and optimizing the whole network based on a preset loss function and an optimizer until the network is converged, and storing the trained network model parameters;
and the grade evaluation unit is used for carrying out grade estimation on the image to be detected by utilizing the trained network model.
In a second aspect, the invention also discloses a computer-readable storage medium, in which a computer program is stored, which, when executed by a processor, causes the processor to perform the steps of the method as described above.
In a third aspect, the present invention also discloses a computer device comprising a memory and a processor, the memory storing a computer program, which, when executed by the processor, causes the processor to perform the steps of the method as described above.
According to the technical scheme, the invention provides the smoke intensity level estimation method based on the diesel vehicle tail gas image of the detection station aiming at the problem that the environment-friendly detection of the vehicle detection station lacks an efficient supervision means.
In conclusion, the smoke degree grading of the diesel vehicle is estimated based on the black smoke discharged by the diesel vehicle at the monitoring station, and the method can highlight the fine motion characteristics of the black smoke by means of mutual guidance of a time domain and a space domain, so that the interference of trees and buildings can be effectively reduced. Compared with the traditional instrument measurement method, the method does not need expensive equipment, can greatly reduce the cost, has good real-time performance and high implementability, and is beneficial to popularization.
Drawings
FIG. 1 is a flow chart of the method of the present invention;
FIG. 2 is a diagram of a smoke level estimation network according to the present invention;
FIG. 3 is a double pyramid signature network architecture according to the present invention;
fig. 4, 4a, 4b, and 4c are graphs showing the recognition results of the model of the present invention for 3 diesel vehicle samples.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention.
As shown in fig. 1, the embodiment of the invention provides a smoke intensity level estimation method based on a diesel vehicle tail gas image of a detection station, aiming at the problem that environment-friendly detection of a vehicle detection station lacks an efficient supervision means.
The method comprises the following specific steps:
a smoke intensity grade estimation method based on a diesel vehicle tail gas image of an inspection station comprises the following steps:
s1: dividing videos collected by an inspection station into subblocks, sampling each subblock, and defining a residual frame according to an obtained RGB frame;
s2: preprocessing the RGB image and the residual frame image;
s3: constructing a space-time cross network model, training the model by utilizing a training data set and a verification data set, finely adjusting and optimizing the whole network based on a preset loss function and an optimizer until the network is converged, and storing the trained network model parameters;
s4: and (4) carrying out grade estimation on the image to be detected by using the trained network model.
The following are described separately:
in S1, the operations of dividing the video collected by the checkpoint into subblocks, performing frame sampling on each subblock, and defining a residual frame according to the obtained RGB frame are as follows:
s11: dividing a monitoring video into N subblocks (N is 8 in the invention), extracting an RGB Frame in each subblock, and recording each Frame as a Framei,i<=N,
Figure BDA0003185075100000061
C, T, H, W are the number of image channels, time length, width and height (in the present invention C is 3);
s12: according to the RGB frame extracted at S11, the residual frame ResFrame is divided according to the following formula:
Figure BDA0003185075100000062
where α is a constant coefficient, β is a set pixel value size to prevent the residual frame size from overflowing, (in the present invention β is 255),
Figure BDA0003185075100000063
c, T, H, W are the number of image channels, time length, width, and height (C is 3 in this embodiment).
S13: the resulting image was displayed as 7: 2: 1, randomly dividing the proportion into a training set, a verification set and a test set;
the specific operation of preprocessing the RGB image and the residual frame image in S2 is as follows:
s21: simple scaling, random cropping, horizontal flipping, perspective transformation, region erasure and color dithering are performed on the RGB image and the residual frame image.
In S3, a spatio-temporal crossing network model is constructed, a training dataset and a verification dataset are used to train the model, the whole network is fine-tuned and optimized based on a preset loss function and an optimizer until the network converges, and the specific operations of storing the trained network model parameters are as follows:
s31: the time domain and space domain structures of the whole network model are consistent, the main body adopts SE-ResNeXt-50 to extract space-time characteristics, and the basic structure of the network is shown in figure 2. The network contains 4 convolutional residual blocks { res1, res2, res3, res4} whose outputs are {4,8,16,32} relative to the stride of the input frame. The outside of the middle-included number is the repetition number of each residual block, on the basis of ResNeXt-50, the output identity mapping of the last layer of res1, res2 and res3 is selected at the residual stage of a time domain and a space domain to construct a double pyramid feature network, so that the time domain and the space domain are mutually guided, and feature maps output by res4 are fused and used for final classification head prediction; the feature fusion is the addition of corresponding elements of the spatial domain feature and the time domain feature, and the specific operation can be described as:
Figure BDA0003185075100000071
wherein i is more than or equal to 1 and less than or equal to H, j is more than or equal to 1 and less than or equal to W, C is more than or equal to 1 and less than or equal to C and SF,
Figure BDA0003185075100000072
s32: the whole network architecture specifically comprises a 7 × 7 convolutional layer (conv1), a 3 × 3 pooling layer (pool1), three convolutional blocks (res1) comprising three convolutional layers, four convolutional blocks (res2) comprising three convolutional layers, 6 convolutional blocks (res3) comprising three convolutional layers, and three convolutional blocks (res4) comprising three convolutional layers, wherein the convolutional kernel size of the three convolutional layers contained in each convolutional block is 1 × 1, 3 × 3, and 1 × 1, and after module stacking, a feature fusion layer, an adaptive average pooling layer, and a full connection layer are added, wherein the convolutional kernel size of the feature fusion layer and the adaptive average pooling layer is 1 × 1, and the whole structure is shown in fig. 3.
S33: the time domain flow part of the whole network model takes a residual error frame as input, the space domain part takes an RGB frame as input, a training set and a verification set are input into the network, Relu is selected as an activation function, a loss function is a cross entropy loss function, the batch size is set to be 3, the weight attenuation is set to be 0.0005, the momentum is set to be 0.9, the learning rate is 0.001, the weight of the model is optimized by using random gradient descent, and the iteration is stopped after the network converges, and the whole network model is stored.
In S4, using the trained network model, the specific operation of performing the grade estimation on the image to be detected is as follows:
s41: inputting the test set into the trained smoke intensity estimation model in S3 to obtain a corresponding smoke intensity grade, wherein the prediction effect is shown in FIG. 4, and the prediction result is consistent with the label.
S42: in order to further illustrate the performance advantage of the space-time cross fusion network provided by the invention, the STCNet of the method is compared with I3D-TSM, I3D-LSTM, I3D-NL and I3D-TC, namely RGB stream is taken as input, and a corresponding module is accessed after an I3D framework. Wherein TSM is a temporal shift module, LSTM is a long-term and short-term memory network module, NL is a non-local module, TC is a time sensing layer, an evaluation index is F Score (F-Score), and specific values of F scores of all models on a test set are as follows:
Model F-score
I3D-TSM 81.3%
I3D-LSTM 81.3%
I3D-NL 81.7%
I3D-TC 82.3%
STCNet 88.5%
as can be seen from the data in the table, the F-score value of the STCNet of the method of the embodiment of the invention is 88.5%, which is higher than that of other models, and the method has better performance.
In conclusion, the smoke degree grading of the diesel vehicle is estimated based on the black smoke discharged by the diesel vehicle at the monitoring station, and the method can highlight the fine motion characteristics of the black smoke by means of mutual guidance of a time domain and a space domain, so that the interference of trees and buildings can be effectively reduced. Compared with the traditional instrument measurement method, the method does not need expensive equipment, can greatly reduce the cost, has good real-time performance and high implementability, and is beneficial to popularization.
On the other hand, the invention also discloses a smoke degree grade estimation system based on the inspection station diesel vehicle tail gas image, which comprises the following units:
the sub-block dividing unit is used for dividing the video collected by the inspection station into sub-blocks, sampling each sub-block and dividing a residual frame according to the obtained RGB frame;
the preprocessing unit is used for preprocessing the RGB image and the residual frame image;
the model construction unit is used for constructing a space-time cross network model, training the model by utilizing a training data set and a verification data set, finely adjusting and optimizing the whole network based on a preset loss function and an optimizer until the network is converged, and storing the trained network model parameters;
and the grade evaluation unit is used for carrying out grade estimation on the image to be detected by utilizing the trained network model.
In a second aspect, the invention also discloses a computer-readable storage medium, in which a computer program is stored, which, when executed by a processor, causes the processor to perform the steps of the method as described above.
In a third aspect, the present invention also discloses a computer device comprising a memory and a processor, the memory storing a computer program, which, when executed by the processor, causes the processor to perform the steps of the method as described above.
It is understood that the system provided by the embodiment of the present invention corresponds to the method provided by the embodiment of the present invention, and the explanation, the example and the beneficial effects of the related contents can refer to the corresponding parts in the method.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.

Claims (8)

1. A smoke intensity grade estimation method based on a diesel vehicle tail gas image of an inspection station is characterized by comprising the following steps:
the method comprises the following steps:
s1: dividing videos collected by an inspection station into subblocks, sampling each subblock, and defining a residual frame according to an obtained RGB frame;
s2: preprocessing the RGB image and the residual frame image;
s3: constructing a space-time cross network model, training the model by utilizing a training data set and a verification data set, finely adjusting and optimizing the whole network based on a preset loss function and an optimizer until the network is converged, and storing the trained network model parameters;
s4: and (4) carrying out grade estimation on the image to be detected by using the trained network model.
2. The checkpoint diesel vehicle exhaust image-based smoke level estimation method as claimed in claim 1, wherein: in S1, the operations of dividing the video collected by the checkpoint into subblocks, performing frame sampling on each subblock, and defining a residual frame according to the obtained RGB frame are as follows:
s11: dividing the monitoring video into N subblocks, extracting an RGB Frame from each subblock, and recording each Frame as a Framei,i<=N,
Figure FDA0003185075090000011
C, T, H, W are image channel number, time length, width and height;
s12: according to the RGB frame extracted at S11, the residual frame ResFrame is divided according to the following formula:
Figure FDA0003185075090000012
where alpha is a constant coefficient, beta is a set pixel value size to prevent residual frame size overflow,
Figure FDA0003185075090000013
c, T, H, W, which is the number of image channels, time length, width and height;
s13: the resulting image was displayed as 7: 2: the scale of 1 is randomly divided into a training set, a validation set, and a test set.
3. The checkpoint diesel vehicle exhaust image-based smoke level estimation method as claimed in claim 2, wherein: the specific operation of preprocessing the RGB image and the residual frame image in S2 is as follows:
s21: simple scaling, random cropping, horizontal flipping, perspective transformation, region erasure and color dithering are performed on the RGB image and the residual frame image.
4. The checkpoint diesel vehicle exhaust image-based smoke level estimation method as claimed in claim 3, wherein: and (8) constructing a space-time cross network model in the S3, training the model by using a training data set and a verification data set, finely adjusting and optimizing the whole network based on a preset loss function and an optimizer until the network is converged, and storing the parameters of the trained network model as follows:
s31: the time domain and the space domain of the whole network model are consistent, and the main body adopts SE-ResNeXt-50 to extract space-time characteristics; the network structure comprises 4 convolutional residual blocks { res1, res2, res3, res4}, and the outputs of the blocks are {4,8,16,32} relative to stride of the input frame; the outside of the middle-included number is the repetition number of each residual block, on the basis of ResNeXt-50, the output identity mapping of the last layer of res1, res2 and res3 is selected at the residual stage of a time domain and a space domain to construct a double pyramid feature network, so that the time domain and the space domain are mutually guided, and feature maps output by res4 are fused and used for final classification head prediction; the feature fusion is the addition of corresponding elements of the spatial domain feature and the time domain feature, and the specific operation description is as follows:
Figure FDA0003185075090000021
wherein i is more than or equal to 1 and less than or equal to H, j is more than or equal to 1 and less than or equal to W, C is more than or equal to 1 and less than or equal to C,
Figure FDA0003185075090000022
s32: the whole network architecture specifically comprises a 7 × 7 convolutional layer (conv1), a 3 × 3 pooling layer (pool1), three convolutional blocks (res1) comprising three convolutional layers, four convolutional blocks (res2) comprising three convolutional layers, 6 convolutional blocks (res3) comprising three convolutional layers, and three convolutional blocks (res4) comprising three convolutional layers, wherein the convolutional kernel size of the three convolutional layers contained in each convolutional block is 1 × 1, 3 × 3 and 1 × 1, and after module stacking, a feature fusion layer, an adaptive average pooling layer and a full connection layer are added, wherein the convolutional kernel size of the feature fusion layer and the adaptive average pooling layer is 1 × 1;
s33: the time domain flow part of the whole network model takes a residual error frame as input, the space domain part takes an RGB frame as input, a training set and a verification set are input into the network, Relu is selected as an activation function, a loss function is a cross entropy loss function, the batch size is set to be 3, the weight attenuation is set to be 0.0005, the momentum is set to be 0.9, the learning rate is 0.001, the weight of the model is optimized by using random gradient descent, and the iteration is stopped after the network converges, and the whole network model is stored.
5. The checkpoint diesel vehicle exhaust image-based smoke level estimation method as claimed in claim 4, wherein: in S4, using the trained network model, the specific operation of performing the grade estimation on the image to be detected is as follows:
s41: and inputting the test set into a trained smoke degree estimation model in S3 to obtain a corresponding smoke degree grade.
6. The utility model provides a smoke intensity level estimation system based on checkpoint diesel vehicle tail gas image which characterized in that:
the method comprises the following units:
the sub-block dividing unit is used for dividing the video collected by the inspection station into sub-blocks, sampling each sub-block and dividing a residual frame according to the obtained RGB frame;
the preprocessing unit is used for preprocessing the RGB image and the residual frame image;
the model construction unit is used for constructing a space-time cross network model, training the model by utilizing a training data set and a verification data set, finely adjusting and optimizing the whole network based on a preset loss function and an optimizer until the network is converged, and storing the trained network model parameters;
and the grade evaluation unit is used for carrying out grade estimation on the image to be detected by utilizing the trained network model.
7. A computer-readable storage medium, storing a computer program which, when executed by a processor, causes the processor to carry out the steps of the method according to any one of claims 1 to 5.
8. A computer device comprising a memory and a processor, the memory storing a computer program that, when executed by the processor, causes the processor to perform the steps of the method according to any one of claims 1 to 5.
CN202110859026.1A 2021-07-28 2021-07-28 Smoke level estimation method and equipment based on diesel vehicle tail gas image of inspection station Active CN113627499B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110859026.1A CN113627499B (en) 2021-07-28 2021-07-28 Smoke level estimation method and equipment based on diesel vehicle tail gas image of inspection station

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110859026.1A CN113627499B (en) 2021-07-28 2021-07-28 Smoke level estimation method and equipment based on diesel vehicle tail gas image of inspection station

Publications (2)

Publication Number Publication Date
CN113627499A true CN113627499A (en) 2021-11-09
CN113627499B CN113627499B (en) 2024-04-02

Family

ID=78381346

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110859026.1A Active CN113627499B (en) 2021-07-28 2021-07-28 Smoke level estimation method and equipment based on diesel vehicle tail gas image of inspection station

Country Status (1)

Country Link
CN (1) CN113627499B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115798624A (en) * 2023-02-01 2023-03-14 中国汽车技术研究中心有限公司 Intelligent modeling method for diesel vehicle tail gas SCR reaction process

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2000030074A1 (en) * 1998-11-13 2000-05-25 Qualcomm Incorporated Low bit-rate coding of unvoiced segments of speech
CN106680281A (en) * 2016-12-31 2017-05-17 中国科学技术大学 Diesel vehicle tail gas smoke intensity detection method based on deep residual error learning network
WO2018171109A1 (en) * 2017-03-23 2018-09-27 北京大学深圳研究生院 Video action detection method based on convolutional neural network
CN110969577A (en) * 2019-11-29 2020-04-07 北京交通大学 Video super-resolution reconstruction method based on deep double attention network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2000030074A1 (en) * 1998-11-13 2000-05-25 Qualcomm Incorporated Low bit-rate coding of unvoiced segments of speech
CN106680281A (en) * 2016-12-31 2017-05-17 中国科学技术大学 Diesel vehicle tail gas smoke intensity detection method based on deep residual error learning network
WO2018171109A1 (en) * 2017-03-23 2018-09-27 北京大学深圳研究生院 Video action detection method based on convolutional neural network
CN110969577A (en) * 2019-11-29 2020-04-07 北京交通大学 Video super-resolution reconstruction method based on deep double attention network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
袁帅;秦贵和;晏婕;: "应用残差生成对抗网络的路况视频帧预测模型", 西安交通大学学报, no. 10 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115798624A (en) * 2023-02-01 2023-03-14 中国汽车技术研究中心有限公司 Intelligent modeling method for diesel vehicle tail gas SCR reaction process

Also Published As

Publication number Publication date
CN113627499B (en) 2024-04-02

Similar Documents

Publication Publication Date Title
CN110348445B (en) Instance segmentation method fusing void convolution and edge information
CN110826684B (en) Convolutional neural network compression method, convolutional neural network compression device, electronic device, and medium
CN111046821B (en) Video behavior recognition method and system and electronic equipment
CN110991311A (en) Target detection method based on dense connection deep network
CN113486726A (en) Rail transit obstacle detection method based on improved convolutional neural network
CN110532925B (en) Driver fatigue detection method based on space-time graph convolutional network
CN111428558A (en) Vehicle detection method based on improved YO L Ov3 method
CN106228528A (en) A kind of multi-focus image fusing method based on decision diagram Yu rarefaction representation
CN111126278A (en) Target detection model optimization and acceleration method for few-category scene
CN113569756B (en) Abnormal behavior detection and positioning method, system, terminal equipment and readable storage medium
CN110599459A (en) Underground pipe network risk assessment cloud system based on deep learning
CN110717863A (en) Single-image snow removing method based on generation countermeasure network
CN115965827A (en) Lightweight small target detection method and device integrating multi-scale features
CN113963333B (en) Traffic sign board detection method based on improved YOLOF model
CN113627499A (en) Smoke intensity grade estimation method and device based on inspection station diesel vehicle tail gas image
Wu et al. Single shot multibox detector for vehicles and pedestrians detection and classification
CN113850135A (en) Dynamic gesture recognition method and system based on time shift frame
CN116740495A (en) Training method and defect detection method for defect detection model of road and bridge tunnel
CN115588150A (en) Pet dog video target detection method and system based on improved YOLOv5-L
Ni et al. Enhanced knowledge distillation for face recognition
CN114782983A (en) Road scene pedestrian detection method based on improved feature pyramid and boundary loss
CN115620051A (en) Large-resolution image target detection method and system based on improved YOLOv5 model
CN115239733A (en) Crack detection method, crack detection device, terminal equipment and storage medium
CN114463732A (en) Scene text detection method and device based on knowledge distillation
CN114943834A (en) Full scene semantic segmentation method based on prototype queue learning under few-label samples

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant