CN110705376A - Abnormal behavior detection method based on generative countermeasure network - Google Patents

Abnormal behavior detection method based on generative countermeasure network Download PDF

Info

Publication number
CN110705376A
CN110705376A CN201910860850.1A CN201910860850A CN110705376A CN 110705376 A CN110705376 A CN 110705376A CN 201910860850 A CN201910860850 A CN 201910860850A CN 110705376 A CN110705376 A CN 110705376A
Authority
CN
China
Prior art keywords
frame
network
frames
training
predicted
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910860850.1A
Other languages
Chinese (zh)
Inventor
卢博文
郭文波
朱松豪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Post and Telecommunication University
Nanjing University of Posts and Telecommunications
Original Assignee
Nanjing Post and Telecommunication University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Post and Telecommunication University filed Critical Nanjing Post and Telecommunication University
Priority to CN201910860850.1A priority Critical patent/CN110705376A/en
Publication of CN110705376A publication Critical patent/CN110705376A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • G06V20/42Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items of sport video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting

Abstract

The invention discloses an abnormal behavior detection method based on a generative countermeasure network (GAN). The method comprises the following steps: first, a U-Net network is built and treated as a generator module in the GAN, where not only appearance (spatial) constraints are used, but also motion (temporal) constraints are introduced. Next, a patch discriminator (patch discriminator) is employed as the discriminator module in the GAN. The generator and discriminator are then trained alternately in confrontation until the discriminator cannot distinguish between the generated frames and the true frames. And finally, carrying out an abnormal behavior detection experiment through the trained GAN model. The experimental results on three publicly available abnormal detection data sets show that the method provided by the invention effectively improves the accuracy of abnormal behavior detection.

Description

Abnormal behavior detection method based on generative countermeasure network
Technical Field
The invention relates to the technical field of video analysis, in particular to an abnormal behavior detection method based on a generative countermeasure network, which relates to a method for detecting abnormal behaviors in a surveillance video scene.
Background
The traditional video monitoring mainly depends on artificially monitoring abnormal behaviors in a scene, which not only needs extremely high labor cost, but also is easy to generate visual fatigue, and even can cause some abnormal behaviors not to be observed in time. The abnormal behavior detection and analysis aims to automatically detect the abnormal behavior in the monitoring scene through algorithms such as video signal processing and machine learning, and therefore people are helped to take corresponding measures in time. Therefore, the method has very important significance and value in developing the abnormal behavior detection algorithm for the monitoring video.
Anomaly detection in video refers to identifying events that do not conform to expected behavior. Its application in video surveillance is an important task. However, since the abnormal events are not controlled in practical applications, it is almost impossible to collect and process various abnormal events by using a classification method, so that the task of detecting the abnormal events is very challenging at the same time.
All existing methods can be roughly divided into two categories, depending on the nature used:
i) artificial design features based on the study method. These methods typically represent each video with some artificially designed features including appearance and action. On the basis, a dictionary method for reconstructing a normal event by using a small reconstruction error is learned, so that an abnormal event corresponds to a larger reconstruction error. But since the dictionary is not trained on exceptional events and is often too complete, the expected results may not be guaranteed.
i i) deep learning based methods. These methods typically learn deep neural networks by means of an auto-encoder and use them to reconstruct normal events with relatively small errors compared to abnormal events. However, the deep neural network has certain fault tolerance, and a large reconstruction error does not necessarily occur for an abnormal event. It follows that almost all methods based on reconstruction of training data do not guarantee the detection of abnormal events.
Interestingly, although anomalies are defined as those events that are not expected, much of the existing work in computer vision is to solve the problem within the framework of reconstructing the training data. Current video frame prediction may be far from satisfactory. In recent years, with the advent of the generative countermeasure network (GAN), the performance of video prediction has improved dramatically. In this context, we do not reconstruct the training data for anomaly detection, but propose to identify an anomaly event by comparing it with its prediction, and introduce a method of anomaly detection based on video frame prediction. Specifically, given a video segment, we predict future frames based on its historical observations. We first train a predictor (generator) that can predict well future frames of normal training data. In the test phase, a frame may correspond to a normal event if it is consistent with its prediction. Otherwise, it may correspond to an exception event. Therefore, a good predictor is the key to the present technique.
Disclosure of Invention
The invention aims to overcome the defects of an abnormal behavior detection method in the existing monitoring video scene and provides an abnormal behavior detection method based on a generative countermeasure network. The invention constructs a generation type countermeasure network which takes U-Net as a generator and a patch discriminator as a discriminator, and carries out an abnormal behavior detection experiment through a trained GAN model. Compared with the most advanced method at present, the method further improves the accuracy rate of abnormal behavior detection and is more suitable for practical application.
The invention relates to an abnormal behavior detection method based on a generative countermeasure network, which comprises the following steps:
s1 selecting a frame I with t consecutive frames1,I2,…,ItDetecting a video to be detected;
s2, inputting the video to be detected into the trained generative confrontation network to obtain a prediction frame
Figure BDA0002199712370000021
S3 predicting the frame
Figure BDA0002199712370000026
Comparing with a ground truth value I, and judging whether the ground truth value I is normal or not;
the generative countermeasure network comprises a generator network
Figure BDA0002199712370000027
And discriminator network
Figure BDA0002199712370000022
Further, the generative confrontation network training process is as follows:
s2-1, acquiring data of each frame of a video to be detected, and preprocessing the data of each frame;
s2-2 building a prediction network based on U-Net and regarding the prediction network as a generator network
Figure BDA0002199712370000023
Employing a patch discriminator as a discriminator network
Figure BDA0002199712370000024
S2-3, introducing appearance constraint, and alternately optimizing the generator network by adopting a random gradient descent method
Figure BDA0002199712370000028
And discriminator network
Figure BDA0002199712370000025
Minimizing the intensity difference and gradient difference between the predicted frame and the ground truth value;
s2-4, introducing motion constraint, and predicting the difference value between the optical flow of the frame and the ground truth value;
s2-5 alternate pair generator network
Figure BDA0002199712370000029
And discriminator network
Figure BDA00021997123700000210
Performing antagonistic training
Figure BDA00021997123700000211
Training by using an objective function
Figure BDA00021997123700000212
A loss function is adopted;
s2-6, inputting the preprocessed data into the trained generative confrontation network, calculating the regular score of each frame, and judging whether the frame is normal or not according to a preset threshold value.
Further, in step S2-1, the data of each frame is preprocessed by normalizing the pixel intensities of all frames to [ -1,1], and the size of each frame is set to 256 × 256.
Further, S2-2, when a prediction network based on U-Net is built, the U-Net adds shortcuts between the high resolution layer and the second resolution layer.
The depth of the network has a great influence on the effect of the convolutional neural network, but simply increasing the depth of the network cannot simply improve the effect of the network, but may damage the effect of the model due to gradient divergence. One of the tricks to solve this problem is to introduce shortcuts (translations into shortcuts or shortcuts), i.e. connections between two non-adjacent layers.
Further, in step S2-3, the frame is predicted in the intensity spaceL between ground truth value I2Distance minimization, as in equation (1):
Figure BDA0002199712370000031
gradient loss function, as in equation (2):
wherein i, j represents a spatial index of the video frame, whereinRepresents the gradient of the predicted frame in the horizontal axis, | Ii,j-Ii-1,jI represents the gradient of the real frame in the direction of the horizontal axis,
Figure BDA00021997123700000311
represents the gradient of the predicted frame in the vertical axis direction, | Ii,j-Ii,j-1And | represents the gradient of the real frame in the direction of the vertical axis.
Further, in step S2-4, the difference between the optical flow of the predicted frame and the ground truth value is expressed by an optical flow loss function, as shown in formula (3),
Figure BDA0002199712370000033
wherein
Figure BDA0002199712370000034
Representing predicted t +1 frames, It+1Representing the real t +1 th frame, ItRepresenting the t-th frame. By means of I1, I2,…,ItRespectively representing successive t frames in a video to be detected, and inputting the t frames into a generator
Figure BDA0002199712370000035
The generator function isGenerating a predicted frame for a t +1 th frame
Figure BDA0002199712370000037
In the implementation, f is pre-trained on one integrated data set. In the detection process, the frame is predictedWith the real t +1 th frame It+1A comparison is made to detect anomalies.
Further, in step S2-5,
training
Figure BDA0002199712370000048
Is aimed att+1Classify to class1 and
Figure BDA0002199712370000041
classificationThrough class0, where 0 and 1 represent a false tag and a true tag, respectively; when training
Figure BDA0002199712370000049
At the same time, fix
Figure BDA00021997123700000410
Applying a Mean Square Error (MSE) loss function:
Figure BDA0002199712370000042
where i, j is the spatial patch index, LMSEAs a function of Mean Square Error (MSE), LMSEThe definition is as follows:
Figure BDA0002199712370000043
wherein the value range of Y is {0,1},
Figure BDA00021997123700000411
is in the value range of [0,1]];
When training
Figure BDA00021997123700000412
The following loss function is used:
Figure BDA0002199712370000044
training
Figure BDA00021997123700000413
Is to generate a quilt
Figure BDA00021997123700000414
Frames classified into class1, when trained
Figure BDA00021997123700000415
When the temperature of the water is higher than the set temperature,is fixed, and again a Mean Square Error (MSE) function is applied as follows:
Figure BDA0002199712370000045
when training
Figure BDA00021997123700000417
All these constraints on appearance, motion and antagonism training are then incorporated into the objective function, and the following objective functions are obtained:
Figure BDA0002199712370000046
in step S2-6, the peak snr for each frame of each test video is calculated as follows:
the peak signal-to-noise ratio of all frames in each test video is normalized to the range [0,1], and the canonical score for each frame is calculated using the following formula:
a threshold may be set to distinguish between regular or irregular frames, based on the fraction s (t) of a frame to predict whether it is normal or abnormal. And if the difference value between the regular score of the ground truth value and the regular score of the predicted frame is greater than the discrimination threshold, judging the video as the abnormal video.
The invention has the beneficial effects that: the method overcomes the defects of the existing abnormal behavior detection method in the monitoring video scene, constructs a generation type countermeasure network which takes U-Net as a generator and a patch discriminator as a discriminator, and carries out an abnormal behavior detection experiment through a trained GAN model. Compared with the most advanced method at present, the method further improves the accuracy rate of abnormal behavior detection and is more suitable for practical application.
Drawings
In order that the present invention may be more readily and clearly understood, there now follows a more particular description of the invention in terms of specific embodiments and reference to the accompanying drawings.
FIG. 1 is a flow chart of a method for detecting abnormal behavior according to the present disclosure;
FIG. 2 is a block diagram of a generator used in this patent;
FIG. 3 is a graph of optical flow generated with/without motion constraints;
FIG. 4 is a schematic diagram of the anomaly detection results of a video prediction network and an auto-encoder network;
fig. 5 is an assessment of abnormal behavior detection performance on three publicly available data sets based on the assessment index AUC.
Detailed Description
Mathematically, given a frame I with consecutive t frames1,I2,…,ItWe stack all these frames in order and use them to predict future frame It+1We useRepresenting our prediction. Let
Figure BDA0002199712370000053
Is close to It+1We minimize the distance between them in intensity and gradient. To protect temporal coherence between adjacent frames, we strengthen It+1And ItIn between, such that
Figure BDA0002199712370000054
To ItClose. Finally, the difference between the prediction of the future frame and itself determines whether the future frame is normal or abnormal. The structure of our network framework is shown in figure 1. Next, we will describe in detail all the components of the framework.
1. Building a generator network
The frame generation network or image generation network commonly used in the existing work generally comprises two modules: i) an encoder for extracting features by gradually reducing spatial resolution; ii) a decoder for gradually recovering the frame by increasing the spatial resolution. However, this solution faces the problem of gradient disappearance and the problem of unbalance of the information of the layers. To avoid this, U-Net is implemented by adding a fast path between the high resolution layer and the low resolution layer. This method suppresses the disappearance of the gradient, making the information symmetrical. In the implementation, U-Net is modified slightly to make future frame predictions. Specifically, the output resolution is kept constant for every two convolutional layers. Thus, there is no longer a need to crop and resize operations when adding shortcuts. Details of this network are shown in figure 2. The size of all convolution and deconvolution kernels is set to 3 x 3, and the maximum pooling layer size is set to 2 x 2.
2. Appearance constraint
In order to get the prediction close to its ground truth value, the method followed uses intensity and gradient differences. The intensity punishment guarantees the similarity of all pixels in the RGB space, and the gradient punishment can carry out sharpening processing on the generated image. In particular, in the intensity space, frames will be predicted
Figure BDA0002199712370000063
L between ground truth value I2The distance is minimized as follows:
further, the gradient loss defined previously is followed as follows:
where i, j represents the spatial index of the video frame.
3. Motion constraint
Previous work only considers the intensity difference and gradient difference when future frames are generated, and cannot guarantee that the predicted frames have correct motion. This is because, in a predicted frame, even if the pixel intensities of all pixels vary very little, it may lead to a complete difference in optical flow (which is a good motion estimator) even if it corresponds to a small prediction error in terms of gradient and intensity.
Previous work only considers the intensity difference and gradient difference when future frames are generated, and cannot guarantee that the predicted frames have correct motion. This is because, in a predicted frame, even if the pixel intensities of all pixels vary very little, it may lead to a complete difference in optical flow (which is a good motion estimator) even if it corresponds to a small prediction error in terms of gradient and intensity.
Therefore, it is necessary to ensure the correctness of the motion prediction. Especially in anomaly detection, coherence of motion is an important factor in evaluating normal events. Therefore, the present invention introduces a time penalty, defined as the difference between the optical flow of the predicted frame and the ground truth. However, the calculation of optical flow is not easy. Recently, a CNN-based optical flow estimation method has appeared, and the invention also adopts the flow to estimate the optical flow, and f represents the functional relationship in the flow, so the optical flow loss can be expressed as:
Figure BDA0002199712370000071
wherein
Figure BDA0002199712370000072
Representing predicted t +1 frames, It+1Representing the real t +1 th frame, ItRepresenting the t-th frame. By means of I1, I2,…,ItRespectively representing successive t frames in a video to be detected, and inputting the t frames into a generator
Figure BDA0002199712370000073
The generator function is
Figure BDA0002199712370000074
Generating a predicted frame for a t +1 th frame
Figure BDA0002199712370000075
In the implementation, f is pre-trained on one integrated data set. In the detection process, the frame is predicted
Figure BDA0002199712370000076
With the real t +1 th frame It+1A comparison is made to detect anomalies.
4. Counter training
Generative countermeasure networks (GANs) have demonstrated their usefulness for image and video generation. The present invention also utilizes variants of the GAN (least squares GAN) module to generate a more realistic framework. Typically the GAN comprises a network of discriminatorsAnd a generator network
Figure BDA0002199712370000077
Figure BDA00021997123700000715
Learning generation is difficult to be
Figure BDA00021997123700000716
Classified frames of
Figure BDA00021997123700000717
Is aimed at
Figure BDA00021997123700000718
The generated frames are distinguished. In practice, antagonism training is achieved by an alternative updating method. Furthermore, the prediction network based on U-Net is regarded as a generator network
Figure BDA0002199712370000078
For discriminator network
Figure BDA0002199712370000079
Using a patch discriminator (patch discriminator), i.e.
Figure BDA00021997123700000719
Each output scalar of (1) corresponds to a patch of the input image. The training program is as follows:
training
Figure BDA00021997123700000710
Training
Figure BDA00021997123700000720
Is aimed att+1Classify to class1 and
Figure BDA00021997123700000711
class to class0, where 0 and 1 represent a false label and a true label, respectively. When training
Figure BDA00021997123700000721
At the same time, fixApplying a Mean Square Error (MSE) loss function:
Figure BDA00021997123700000712
where i, j is the spatial patch index, LMSEIs a Mean Square Error (MSE) function. L isMSEThe definition is as follows:
wherein the value range of Y is {0,1},
Figure BDA00021997123700000723
is in the value range of [0,1]]。
TrainingTraining
Figure BDA0002199712370000087
Is to generate a quilt
Figure BDA0002199712370000088
Frames classified into class 1. When training
Figure BDA0002199712370000089
When the temperature of the water is higher than the set temperature,is fixed. Also, a Mean Square Error (MSE) function is applied as follows:
Figure BDA0002199712370000082
5. objective function
When training
Figure BDA00021997123700000811
All these constraints on appearance, motion and antagonism training are then combined into the objective function and the following objective function is obtained:
Figure BDA0002199712370000083
when trainingThe following loss function is used:
Figure BDA0002199712370000084
6. anomaly detection on a test set
Assuming normal events are well predicted, the predicted frames can be utilized
Figure BDA00021997123700000813
And its ground truth value I to make an anomaly prediction. MSE (mean square error) is a popular method to measure the quality of a predicted image by computing the euclidean distance between the predicted value and the ground truth of all pixels in the RGB color space. However, Mathieu shows that Peak Signal to Noise Ratio (PSNR) is a better image quality assessment method, and is expressed as follows:
Figure BDA0002199712370000085
the peak signal-to-noise ratio (PSNR) of the t-th frame is high, indicating that it is more likely to be normal. After calculating the PSNR of each frame of each test video, we normalized the PSNR of all frames in each test video to the range [0,1] and calculated the regular score (regular score) of each frame using the following formula:
therefore, it can be predicted whether it is normal or abnormal from the score s (t) of one frame. A threshold may be set to distinguish between regular or irregular frames.
The method specifically comprises the following steps:
1. preparation of Experimental data
The present invention evaluates the proposed method and the functionality of the different components on three publicly available anomaly detection datasets, including the CUHK Avenue dataset, the UCSD Peerreian dataset, and the ShanghaiTech dataset. We further verified the robustness of our method using a small dataset, that is, our method can correctly classify normal and abnormal events even if there is some uncertainty in the normal events.
2. Experimental details settings
To train the network, the pixel intensities of all frames are normalized to [ -1,1]The size of each frame is set to 256 × 256. I amThey set t 4 and use a random slice consisting of 5 ordered frames. And (3) performing parameter optimization by adopting a random gradient descent method based on Adam. The small batch size was 4. For gray scale datasets, the learning rates of the generator and discriminator are set to 0.0001 and 0.00001, respectively. While for the chrominance data sets (color scaled data sets), the learning rates of the generator and discriminator start from 0.0002 and 0.00002, respectively. Coefficient factor lambda for different data setsint,λgd,λopAnd λadvSlightly different. A simple method is to apply λint,λgd,λopAnd λadvSet to 1.0, 1.0, 2.0 and 0.05, respectively.
3. Antagonism training
During training, a network is generated
Figure BDA0002199712370000093
The aim of the method is to generate a real picture as much as possible to deceive a discrimination networkWhile
Figure BDA0002199712370000094
The aim is to identify the most probable
Figure BDA0002199712370000095
The generated false image and the real image. Thus, the training is divided and alternated
Figure BDA0002199712370000096
And
Figure BDA0002199712370000092
they form a dynamic "gaming process" and the final balance points, namely nash balance points, are considered as the fit state of the network.
4. Performance evaluation
In the literature of conventional anomaly detection, a common evaluation index is to calculate a receiver operating characteristic curve (ROC) by gradually changing a threshold value of a regularization score. The area under the curve (AUC) is then accumulated as a scalar for performance evaluation, with higher values giving better anomaly detection performance. In the present invention, the AUC at the frame level is used for performance evaluation. (there are three different levels of criteria for evaluating the performance of the abnormal behavior detection algorithm, from small to large, pixel level, frame level, behavior level.)
5. Compared with the prior method
In order to verify the effectiveness of the method, compared with the traditional method and the latest deep learning-based method, as shown in fig. 5, the results of performance evaluation of abnormal behavior detection on three publicly available data sets based on the evaluation index AUC are shown, and fig. 5 lists AUC performance comparison of different methods in a common data set. It can be seen that the method proposed by the present invention is about 3% -5% better than all existing methods in all datasets, which demonstrates the effectiveness of the method.
6. Effects of motion constraints
To assess the importance of motion constraints on video frame generation and anomaly detection, experiments were conducted in training by removing target constraints. Such a reference is then compared to the method of the present invention.
The motion constraints are evaluated using a flowsheet. In fig. 3, an optical flow diagram generated with/without motion constraint, a visualization of a prediction image and an optical flow on a Ped1 data set are shown, a red box represents an optical flow difference predicted by a model under the condition of motion constraint, and it can be seen that the optical flow generated with motion constraint is more consistent with ground truth, which shows that such motion constraint terms help our prediction network to capture motion information more accurately. In addition, the mean square error (7.51 and 8.26, respectively) between the optical flow graph generated with/without motion constraint and the ground truth is compared, which further illustrates the effectiveness of the motion constraint.
7. Comparing performance of predictive networks and autoencoder networks in anomaly detection
As shown in fig. 4, the video prediction network-based and the self-encoder network-based anomaly detection are also compared. For exceptions based on an autoencoder networkDetection, the most advanced anomaly detection performance was achieved using the latest Conv-AE. Due to the capability of deep neural networks, the method based on the self-encoder can well reconstruct the normal and abnormal frames of the test phase, and in order to evaluate the performance of the prediction network and the automatic encoder, the difference delta between the normal and abnormal scores is usedsGreater ΔsIndicating that the network is better able to distinguish between normal and abnormal patterns. The average scores for normal and abnormal frames in the Ped1, Ped2, and Avenue dataset test sets were first calculated. Then, I calculate the difference Δ of these two scoressTo measure our method and the ability of Conv-AE to distinguish between normal and abnormal frames. Greater difference ΔsCorresponding to a smaller false alarm rate and a higher detection rate. As can be seen from the results of FIG. 4, the method presented in the present application consistently outperforms the Conv-AE method in terms of the difference in scores between normal and abnormal events the solution of the method of the present invention can always reach a higher difference (Δ) than Conv-AEs) The effectiveness of the video prediction based anomaly detection method is verified.
The above description is only a preferred embodiment of the present invention, and is not intended to limit the present invention further, and all equivalent variations made by using the contents of the present specification and the drawings are within the protection scope of the present invention.

Claims (8)

1. An abnormal behavior detection method based on a generative countermeasure network comprises the following steps:
s1 selecting a frame I with t consecutive frames1,I2,…,ItDetecting a video to be detected;
s2, inputting the video to be detected into the trained generative confrontation network to obtain a prediction frame
S3 predicting the frame
Figure FDA0002199712360000012
Ground truth value It+1Comparing and judging whether the test sample is normal or not;
the generative countermeasure network comprises a generator network
Figure FDA0002199712360000013
And discriminator network
Figure FDA0002199712360000014
2. The method as claimed in claim 1, wherein the generative confrontation network training process comprises:
s2-1, acquiring data of each frame of a video to be detected, and preprocessing the data of each frame;
s2-2 building a prediction network based on U-Net and regarding the prediction network as a generator network
Figure FDA0002199712360000015
Employing a patch discriminator as a discriminator network
Figure FDA0002199712360000016
S2-3, introducing appearance constraint, and alternately optimizing the generator network by adopting a random gradient descent method
Figure FDA0002199712360000017
And discriminator networkMinimizing the intensity difference and gradient difference between the predicted frame and the ground truth value;
s2-4, introducing motion constraint, and predicting the difference value between the optical flow of the frame and the ground truth value;
s2-5 alternate pair generator network
Figure FDA0002199712360000019
And discriminator network
Figure FDA00021997123600000110
Performing antagonistic training
Figure FDA00021997123600000111
Training by using an objective function
Figure FDA00021997123600000112
A loss function is adopted;
s2-6, inputting the preprocessed data into the trained generative confrontation network, calculating the regular score of each frame, and judging whether the frame is normal or not according to a preset threshold value.
3. The method as claimed in claim 2, wherein in step S2-1, the data of each frame is preprocessed by normalizing the pixel intensity of all frames to [ -1,1], and the size of each frame is set to 256 × 256.
4. The abnormal behavior detection method based on the generative countermeasure network as claimed in claim 2, wherein, in step S2-2, when building a U-Net based prediction network, U-Net adds a shortcut between the high resolution layer and the second resolution layer.
5. The method for detecting abnormal behavior of generative-based countermeasure network as claimed in claim 2, wherein in step S2-3, the predicted frame is predicted in intensity space
Figure FDA00021997123600000113
L between ground truth value I2Distance minimization, as in equation (1):
Figure FDA00021997123600000114
gradient loss function, as in equation (2):
Figure FDA0002199712360000021
where i, j represent the spatial index of the video frame,
Figure FDA0002199712360000022
representing the gradient of the predicted frame in the direction of the horizontal axis,
Figure FDA0002199712360000023
representing the gradient of the predicted frame in the direction of the vertical axis.
6. The method for detecting abnormal behavior based on generative countermeasure network as claimed in claim 2, wherein in step S2-4, the difference between the optical flow of the predicted frame and the ground truth value is represented by optical flow loss function, as shown in formula (3),
Figure FDA0002199712360000024
wherein
Figure FDA0002199712360000025
Representing predicted t +1 frames, It+1Representing the real t +1 th frame, ItRepresenting the t-th frame. By means of I1,I2,…,ItRespectively representing successive t frames in a video to be detected, and inputting the t frames into a generator
Figure FDA0002199712360000026
The generator function is
Figure FDA0002199712360000027
Generating a predicted frame for a t +1 th frame
Figure FDA0002199712360000028
In the implementation, f is pre-trained on one integrated data set. In the detection process, the frame is predicted
Figure FDA0002199712360000029
With the real t +1 th frame It+1A comparison is made to detect anomalies.
7. The abnormal behavior detection method based on generative countermeasure network as claimed in claim 2, wherein in step S2-5,
training
Figure FDA00021997123600000210
Is aimed att+1Classify to class1 andclass0, where 0 and 1 represent false and true tags, respectively; when training
Figure FDA00021997123600000212
At the same time, fix
Figure FDA00021997123600000213
Applying a Mean Square Error (MSE) loss function:
Figure FDA00021997123600000214
where i, j is the spatial patch index, LMSEAs a function of Mean Square Error (MSE), LMSEThe definition is as follows:
Figure FDA00021997123600000215
wherein the value range of Y is {0,1},is in the value range of [0,1]];
When trainingThe following loss function is used:
Figure FDA0002199712360000031
trainingIs to generate a quilt
Figure FDA0002199712360000033
Frames classified into class1, when trained
Figure FDA0002199712360000034
When the temperature of the water is higher than the set temperature,
Figure FDA0002199712360000035
is fixed, and again a Mean Square Error (MSE) function is applied as follows:
when training
Figure FDA0002199712360000037
All these constraints on appearance, motion and antagonism training are then incorporated into the objective function, and the following objective functions are obtained:
8. the method for detecting abnormal behavior based on generative countermeasure network as claimed in claim 2, wherein in step S2-6, the peak snr of each frame of each test video is calculated as follows:
Figure FDA0002199712360000039
the peak signal-to-noise ratio of all frames in each test video is normalized to the range [0,1], and the canonical score for each frame is calculated using the following formula:
Figure FDA00021997123600000310
a threshold may be set to distinguish between regular or irregular frames, based on the fraction s (t) of a frame to predict whether it is normal or abnormal.
CN201910860850.1A 2019-09-11 2019-09-11 Abnormal behavior detection method based on generative countermeasure network Pending CN110705376A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910860850.1A CN110705376A (en) 2019-09-11 2019-09-11 Abnormal behavior detection method based on generative countermeasure network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910860850.1A CN110705376A (en) 2019-09-11 2019-09-11 Abnormal behavior detection method based on generative countermeasure network

Publications (1)

Publication Number Publication Date
CN110705376A true CN110705376A (en) 2020-01-17

Family

ID=69194461

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910860850.1A Pending CN110705376A (en) 2019-09-11 2019-09-11 Abnormal behavior detection method based on generative countermeasure network

Country Status (1)

Country Link
CN (1) CN110705376A (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111310915A (en) * 2020-01-21 2020-06-19 浙江工业大学 Data anomaly detection and defense method for reinforcement learning
CN111783582A (en) * 2020-06-22 2020-10-16 东南大学 Unsupervised monocular depth estimation algorithm based on deep learning
CN112036955A (en) * 2020-09-07 2020-12-04 贝壳技术有限公司 User identification method and device, computer readable storage medium and electronic equipment
CN112149596A (en) * 2020-09-29 2020-12-29 厦门理工学院 Abnormal behavior detection method, terminal device and storage medium
CN112364745A (en) * 2020-11-04 2021-02-12 北京瑞莱智慧科技有限公司 Method and device for generating countermeasure sample and electronic equipment
CN112418149A (en) * 2020-12-04 2021-02-26 清华大学深圳国际研究生院 Abnormal behavior detection method based on deep convolutional neural network
WO2021147055A1 (en) * 2020-01-22 2021-07-29 Beijing Didi Infinity Technology And Development Co., Ltd. Systems and methods for video anomaly detection using multi-scale image frame prediction network
WO2021147066A1 (en) * 2020-01-23 2021-07-29 Telefonaktiebolaget Lm Ericsson (Publ) Method and apparatus for image classification
CN113269104A (en) * 2021-05-28 2021-08-17 山东大学 Group abnormal behavior identification method, system, storage medium and equipment
CN113283849A (en) * 2021-07-26 2021-08-20 山东新北洋信息技术股份有限公司 Logistics abnormity intelligent detection method based on video context association
CN113313012A (en) * 2021-05-26 2021-08-27 北京航空航天大学 Dangerous driving behavior identification method based on convolution generation countermeasure network
CN113313037A (en) * 2021-06-02 2021-08-27 郑州大学 Method for detecting video abnormity of generation countermeasure network based on self-attention mechanism
CN113435432A (en) * 2021-08-27 2021-09-24 腾讯科技(深圳)有限公司 Video anomaly detection model training method, video anomaly detection method and device
CN113449679A (en) * 2021-07-14 2021-09-28 湖南长城科技信息有限公司 Method and device for identifying abnormal behaviors of human body
CN113837351A (en) * 2020-06-08 2021-12-24 爱思开海力士有限公司 New different detector
WO2022116322A1 (en) * 2020-12-02 2022-06-09 罗普特科技集团股份有限公司 Method and apparatus for generating anomaly detection model, and anomaly event detection method and apparatus
CN115049781A (en) * 2022-05-11 2022-09-13 西南石油大学 Shale digital core three-dimensional reconstruction method based on deep learning
CN115690665A (en) * 2023-01-03 2023-02-03 华东交通大学 Video anomaly detection method and device based on cross U-Net network
CN117411674A (en) * 2023-09-22 2024-01-16 南京中新赛克科技有限责任公司 Industrial Internet abnormal flow detection method and detection system based on generation and diffusion
CN117411674B (en) * 2023-09-22 2024-05-14 南京中新赛克科技有限责任公司 Industrial Internet abnormal flow detection method and detection system based on generation and diffusion

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
WEN LIU 等: "Future Frame Prediction for Anomaly Detection – A New Baseline", 《ARXIV:1712.09867V3 [CS.CV]》 *
胡正平 等: "视频监控系统异常目标检测与定位综述", 《燕山大学学报》 *

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111310915A (en) * 2020-01-21 2020-06-19 浙江工业大学 Data anomaly detection and defense method for reinforcement learning
CN111310915B (en) * 2020-01-21 2023-09-01 浙江工业大学 Data anomaly detection defense method oriented to reinforcement learning
WO2021147055A1 (en) * 2020-01-22 2021-07-29 Beijing Didi Infinity Technology And Development Co., Ltd. Systems and methods for video anomaly detection using multi-scale image frame prediction network
WO2021147066A1 (en) * 2020-01-23 2021-07-29 Telefonaktiebolaget Lm Ericsson (Publ) Method and apparatus for image classification
CN113837351B (en) * 2020-06-08 2024-04-23 爱思开海力士有限公司 New different detector
CN113837351A (en) * 2020-06-08 2021-12-24 爱思开海力士有限公司 New different detector
CN111783582A (en) * 2020-06-22 2020-10-16 东南大学 Unsupervised monocular depth estimation algorithm based on deep learning
CN112036955A (en) * 2020-09-07 2020-12-04 贝壳技术有限公司 User identification method and device, computer readable storage medium and electronic equipment
CN112036955B (en) * 2020-09-07 2021-09-24 贝壳找房(北京)科技有限公司 User identification method and device, computer readable storage medium and electronic equipment
CN112149596A (en) * 2020-09-29 2020-12-29 厦门理工学院 Abnormal behavior detection method, terminal device and storage medium
CN112364745A (en) * 2020-11-04 2021-02-12 北京瑞莱智慧科技有限公司 Method and device for generating countermeasure sample and electronic equipment
WO2022116322A1 (en) * 2020-12-02 2022-06-09 罗普特科技集团股份有限公司 Method and apparatus for generating anomaly detection model, and anomaly event detection method and apparatus
CN112418149A (en) * 2020-12-04 2021-02-26 清华大学深圳国际研究生院 Abnormal behavior detection method based on deep convolutional neural network
CN113313012A (en) * 2021-05-26 2021-08-27 北京航空航天大学 Dangerous driving behavior identification method based on convolution generation countermeasure network
CN113269104A (en) * 2021-05-28 2021-08-17 山东大学 Group abnormal behavior identification method, system, storage medium and equipment
CN113313037A (en) * 2021-06-02 2021-08-27 郑州大学 Method for detecting video abnormity of generation countermeasure network based on self-attention mechanism
CN113449679A (en) * 2021-07-14 2021-09-28 湖南长城科技信息有限公司 Method and device for identifying abnormal behaviors of human body
CN113283849A (en) * 2021-07-26 2021-08-20 山东新北洋信息技术股份有限公司 Logistics abnormity intelligent detection method based on video context association
CN113283849B (en) * 2021-07-26 2021-11-02 山东建筑大学 Logistics abnormity intelligent detection method based on video context association
CN113435432A (en) * 2021-08-27 2021-09-24 腾讯科技(深圳)有限公司 Video anomaly detection model training method, video anomaly detection method and device
CN115049781A (en) * 2022-05-11 2022-09-13 西南石油大学 Shale digital core three-dimensional reconstruction method based on deep learning
CN115690665A (en) * 2023-01-03 2023-02-03 华东交通大学 Video anomaly detection method and device based on cross U-Net network
CN117411674A (en) * 2023-09-22 2024-01-16 南京中新赛克科技有限责任公司 Industrial Internet abnormal flow detection method and detection system based on generation and diffusion
CN117411674B (en) * 2023-09-22 2024-05-14 南京中新赛克科技有限责任公司 Industrial Internet abnormal flow detection method and detection system based on generation and diffusion

Similar Documents

Publication Publication Date Title
CN110705376A (en) Abnormal behavior detection method based on generative countermeasure network
CN109360232B (en) Indoor scene layout estimation method and device based on condition generation countermeasure network
CN109870461B (en) Electronic components quality detection system
CN108549841A (en) A kind of recognition methods of the Falls Among Old People behavior based on deep learning
CN110321873A (en) Sensitization picture recognition methods and system based on deep learning convolutional neural networks
CN106127148A (en) A kind of escalator passenger's unusual checking algorithm based on machine vision
CN112597864B (en) Monitoring video anomaly detection method and device
CN108846826A (en) Object detecting method, device, image processing equipment and storage medium
CN113658115A (en) Image anomaly detection method for generating countermeasure network based on deep convolution
CN111627044A (en) Target tracking attack and defense method based on deep network
CN113569756B (en) Abnormal behavior detection and positioning method, system, terminal equipment and readable storage medium
CN113705490B (en) Anomaly detection method based on reconstruction and prediction
CN113313037A (en) Method for detecting video abnormity of generation countermeasure network based on self-attention mechanism
CN112990052A (en) Partially-shielded face recognition method and device based on face restoration
CN110807396A (en) Face changing video tampering detection method and system based on illumination direction consistency
CN113344475A (en) Transformer bushing defect identification method and system based on sequence modal decomposition
CN115171218A (en) Material sample feeding abnormal behavior recognition system based on image recognition technology
CN114067251B (en) Method for detecting anomaly of unsupervised monitoring video prediction frame
CN109815943B (en) Hazardous chemical storage stacking picture sample generation method and system
CN116823673B (en) High-speed elevator car passenger state visual perception method based on image processing
CN111401209B (en) Action recognition method based on deep learning
CN105844671B (en) A kind of fast background relief method under the conditions of change illumination
CN116994044A (en) Construction method of image anomaly detection model based on mask multi-mode generation countermeasure network
CN106375756A (en) Single object removing and tampering detection method for monitored video
CN116229347A (en) Crowd safety abnormal event identification method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200117