CN112149596A - Abnormal behavior detection method, terminal device and storage medium - Google Patents

Abnormal behavior detection method, terminal device and storage medium Download PDF

Info

Publication number
CN112149596A
CN112149596A CN202011049302.XA CN202011049302A CN112149596A CN 112149596 A CN112149596 A CN 112149596A CN 202011049302 A CN202011049302 A CN 202011049302A CN 112149596 A CN112149596 A CN 112149596A
Authority
CN
China
Prior art keywords
video
image
neural network
abnormal behavior
behavior detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN202011049302.XA
Other languages
Chinese (zh)
Inventor
苏鹭梅
陈鑫强
李天友
吴家俊
黄明勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xiamen University of Technology
Original Assignee
Xiamen University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xiamen University of Technology filed Critical Xiamen University of Technology
Priority to CN202011049302.XA priority Critical patent/CN112149596A/en
Publication of CN112149596A publication Critical patent/CN112149596A/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biophysics (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to an abnormal behavior detection method, a terminal device and a storage medium, wherein the method comprises the following steps: collecting a video image sequence formed by a plurality of groups of continuous t +1 frame video images, and forming a training set by all the video image sequences; constructing a U-Net neural network, inputting the U-Net neural network into a video image sequence consisting of 1 st to t th frame video images, outputting a reconstructed image reconstructed by the U-Net neural network, and training the U-Net neural network through a training set to minimize the difference between the reconstructed image and the real image of the acquired t +1 th frame; and continuously reconstructing images of continuous frames in the video through the trained U-Net neural network, and judging whether abnormal behaviors exist in the video or not through the relation between the difference between the reconstructed images and the corresponding real images and a threshold value. The invention autonomously learns the classification of behavior categories from the rules of a large amount of video data, and saves the workload of labeling labels.

Description

Abnormal behavior detection method, terminal device and storage medium
Technical Field
The present invention relates to the field of computer vision, and in particular, to an abnormal behavior detection method, a terminal device, and a storage medium.
Background
The abnormal behavior detection technology has a wide application prospect in a security system, the real-time monitoring of the abnormal behavior at present mostly depends on monitoring personnel to carry out manual inspection through a monitoring system, but due to the fact that the monitoring personnel are fatigued, the attention cannot be concentrated for a long time, and the like, the conditions of missed detection, false detection and the like of the abnormal behavior often occur. Therefore, a system capable of automatically identifying abnormal behavior in a video will improve the efficiency of fighting against illegal crimes.
The existing abnormal behavior detection methods are mainly divided into two types:
1) the abnormal behavior detection method is based on traditional manual feature extraction. The traditional manual feature extraction method comprises the following steps: 1. sampling a video and extracting characteristics; 2. encoding the features; 3. normalizing the encoded vector; 4. and training a classifier. However, the types of abnormal behaviors that can be detected by this method are few, and many abnormal behaviors are included in a real scene, so this method is not well suited for detecting complex scenes.
(2) Abnormal behavior detection method based on deep learning. Neural networks for abnormal behavior detection in deep learning mainly include Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs). The more classical approach is the dual-stream CNN (dual-stream CNN) proposed by simony et al for behavior recognition, which treats video as a sequence of images, where a spatial stream is used to compute the CNN features of an image frame, a temporal stream is used to compute the optical stream CNN features between several image frames, and finally both features are merged. Ji et al propose a 3D CNN (three-dimensional CNN) based method that adds time to a two-dimensional CNN so that a neural network can learn spatial and temporal information from a video. However, the above abnormal behavior detection method based on supervised learning only relies on a powerful behavior detection classifier and does not fully utilize prior knowledge. This method has a problem that the amount of calculation is large and the detection speed is affected. And creating labels for supervised learning requires a lot of work, the fewer labels created manually, the less data the algorithm can use for training.
Disclosure of Invention
In order to solve the above problems, the present invention provides an abnormal behavior detection method, a terminal device, and a storage medium.
The specific scheme is as follows:
an abnormal behavior detection method comprises the following steps:
s1: collecting a video image sequence formed by a plurality of groups of continuous t +1 frame video images, and forming a training set by all the video image sequences;
s2: constructing a U-Net neural network, inputting the U-Net neural network into a video image sequence consisting of 1 st to t th frame video images, outputting a reconstructed image reconstructed by the U-Net neural network, and training the U-Net neural network through a training set to minimize the difference between the reconstructed image and the real image of the acquired t +1 th frame; the training process of the U-Net neural network also comprises appearance constraint, motion constraint and information gain constraint;
s3: and continuously reconstructing images of continuous frames in the video through the trained U-Net neural network, and judging whether abnormal behaviors exist in the video or not through the relation between the difference between the reconstructed images and the corresponding real images and a threshold value.
Further, an auto-encoder formed by the U-Net neural network is divided into an encoder and a decoder, the encoder is responsible for extracting the features of the image and gradually reducing the spatial size of the pooling layer, and the decoder is responsible for reconstructing the image and restoring the details and the spatial size of the image.
Further, the appearance constraints include intensity constraints and gradient constraints; the intensity constraint is to compute the difference of all pixel values between the reconstructed image and the real image, and the gradient constraint is to compute the gradient between the reconstructed image and the real image.
Further, the motion constraint includes optical flow loss, which is used to calculate a difference between the optical flow of the reconstructed image and the optical flow of the real image.
Further, the information gain constraint is used for calculating a difference value between the information entropy of the reconstructed image and the information entropy of the real image.
Further, the difference is evaluated by the peak signal-to-noise ratio.
An abnormal behavior detection terminal device includes a processor, a memory, and a computer program stored in the memory and executable on the processor, and when the processor executes the computer program, the processor implements the steps of the method described above in the embodiment of the present invention.
A computer-readable storage medium, in which a computer program is stored, which, when being executed by a processor, carries out the steps of the method as described above for an embodiment of the invention.
The invention adopts the technical scheme and has the beneficial effects that:
1. compared with supervised learning which needs a large amount of manual labeling, the method adopts an unsupervised learning method based on the self-encoder, and the classifier independently learns the classification of behavior categories from the rules of a large amount of video data instead of learning from the existing experience of human beings, so that the workload of labeling labels is saved.
2. Compared with the supervised learning which does not fully utilize the prior knowledge, the unsupervised learning method based on the self-encoder is adopted, the appearance constraint, the motion constraint and the information gain constraint are added, and the prior knowledge is fully utilized.
3. Compared with the abnormal behavior detection method for training the manual feature extraction of the specific type of abnormal behaviors, the method adopts an unsupervised learning method based on the self-encoder, utilizes a large number of videos containing normal behaviors to train the self-encoder, when the normal behaviors appear in the videos, the difference between the reconstructed images generated by the self-encoder and the real images is small, and when the abnormal behaviors appear in the videos, the difference between the reconstructed images generated by the self-encoder and the real images is large, so that the abnormal behaviors in the complex scene can be judged.
Drawings
FIG. 1 is a flow chart of a method according to an embodiment of the present invention.
Fig. 2 is a schematic flow chart of the method in this embodiment.
FIG. 3 is a diagram showing the structure of the U-Net neural network in this embodiment.
Fig. 4 shows the abnormal behavior detection result of the test video in section 7 in the CUHK Avenue data set in this embodiment.
FIG. 5 shows the abnormal behavior detection results of segment 7 of the test video in the UCSD Ped1 data set in this embodiment.
FIG. 6 shows the abnormal behavior detection results of segment 4 of the test video in the UCSD Ped2 data set in this embodiment.
Detailed Description
To further illustrate the various embodiments, the invention provides the accompanying drawings. The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate embodiments of the invention and, together with the description, serve to explain the principles of the embodiments. Those skilled in the art will appreciate still other possible embodiments and advantages of the present invention with reference to these figures.
The invention will now be further described with reference to the accompanying drawings and detailed description.
The first embodiment is as follows:
an embodiment of the present invention provides an abnormal behavior detection method, as shown in fig. 1 and fig. 2, the method includes the following steps:
s1: and acquiring a video image sequence consisting of a plurality of groups of continuous t +1 frame video images, and forming a training set by all the video image sequences.
For the successive t +1 frames of video images to be randomly extracted from the video, the collected video images also need to be normalized, which is 256 × 256 pixels in this embodiment.
S2: and constructing a U-Net neural network, wherein the input of the U-Net neural network is a video image sequence consisting of 1 st to t th frame video images, outputting a reconstructed image reconstructed by the U-Net neural network, and training the U-Net neural network through a training set to ensure that the difference between the reconstructed image and the real image of the acquired t +1 th frame is minimum.
The structure diagram of the U-Net neural network in this embodiment is shown in fig. 3, the structure of the U-Net is similar to a Full Convolutional Network (FCN), the U-Net is named by its left-right symmetry, and its main structure is composed of a left half encoder (including 4 down-sampling processes) and a right half decoder (including 4 up-sampling processes). The encoder performs feature extraction on the image and gradually reduces the spatial size of the pooling layer, while the decoder is used to reconstruct the image and restore the details and spatial size of the image. There is usually a cascade of processes between the encoder and the decoder, and copying the features in the encoder to the corresponding positions of the decoder can help the decoder to better determine the details of the target. The intermediate cascade avoids the problems of gradient elimination and asymmetric information of each layer of image, so that the size of a reconstructed image is the same as that of an original image.
Appearance constraint, motion constraint and information gain constraint are added in the training process of the U-Net neural network, so that the reconstructed image is closer to a real image. The U-Net neural network inputs 1 st to t th frame video images I1,I2,...,ItReconstructing to generate a reconstructed image
Figure BDA0002709044130000061
The reconstructed image corresponds to the t +1 frame video image, and the result of network training is used for enabling the reconstructed image
Figure BDA0002709044130000062
The difference from the real image (t +1 th frame video image) is minimal. The added appearance constraint, the motion constraint and the information gain constraint are all used for ensuring that the reconstructed image generated by the U-Net neural network is similar to the real image as much as possible.
Three constraints are described separately below.
1) The appearance constraints include intensity constraints and gradient constraints, the intensity constraints utilizing l2Norm distance metric reconstructed image
Figure BDA0002709044130000063
Image of the real image IThe similarity of elements is calculated as follows:
Figure BDA0002709044130000064
the gradient constraint is used to calculate the gradient between the reconstructed image and the real image, and the calculation formula is as follows:
Figure BDA0002709044130000065
where i, j represents the width and height of a video frame.
2) The motion constraint comprises an optical flow constraint, and the calculation formula is as follows:
Figure BDA0002709044130000066
where f represents a function that calculates the optical flow of the image frame.
3) And (4) information gain constraint.
The information gain constraint calculates the difference between the information entropy of the reconstructed image and the information entropy of the real image. The concept of entropy first originates from the field of thermodynamics, and shannon introduces the concept of entropy into the information theory, where entropy is used to measure the degree of uncertainty of a system. The amount of information of an event is inversely proportional to the probability of the event occurring. Abnormal behavior can be defined as the occurrence of unexpected, non-normal events in the video, so abnormal behavior as a small probability event generally results in greater information gain. The calculation formula of the information entropy is as follows:
Figure BDA0002709044130000071
wherein, let the sample space H contain n elementary events w1,w2,...,w3,piRepresents wiProbability of occurrence, 0 ≦ pi≤1,
Figure BDA0002709044130000072
When the degree of disorder of the system is larger, the probability of occurrence of the event is smaller, and the information entropy is larger. Abnormal behavior is defined as an unexpected situation, so the entropy of the information can be used to detect abnormal behavior. That is, when the video frame is in normal behavior, the uncertainty of the event is small, and the information entropy is small; when abnormal behaviors exist in the video frame, the uncertainty of the event is large, and the information entropy is also large.
When the information entropy H of the reconstructed image is calculatedrAnd the information entropy H of the actual frameaThen, subtracting the information entropy of the real image from the information entropy of the reconstructed image, wherein the difference between the two is the information gain HmThe calculation formula of the information gain is as follows:
Hm=Hr-Ha
s3: and continuously reconstructing images of continuous frames in the video through the trained U-Net neural network, and judging whether abnormal behaviors exist in the video or not through the relation between the difference between the reconstructed images and the corresponding real images and a threshold value.
In practical application, for a certain monitoring video, acquiring continuous t frames of video images of the monitoring video, inputting the continuous t frames of video images into a trained U-Net neural network, judging whether the difference between a reconstructed image generated by the reconstruction of the U-Net neural network and a t +1 th frame of real image behind the continuous t frames of video images corresponding to the monitoring video is larger than a threshold value, and if so, judging that abnormal behaviors exist in the monitoring video.
The difference in this embodiment is evaluated by peak signal to noise ratio (PSNR), which is calculated as:
Figure BDA0002709044130000073
the peak signal-to-noise ratio PSNR is then normalized to [0, 1 ]]Can obtain an abnormality score ASThe calculation formula is as follows:
Figure BDA0002709044130000081
abnormal score ASIndicating the possibility of abnormal behaviors in the video sequence, the higher the abnormal score is, the higher the possibility of abnormal behaviors is, and the lower the abnormal score is, the lower the possibility of abnormal behaviors is, i.e. the higher the possibility of normal behaviors is. Therefore, a threshold value is set to judge whether abnormal behaviors occur in the video sequence, when the abnormal score is larger than the threshold value, the abnormal behaviors are considered to occur in the video frame, and when the abnormal score is smaller than the threshold value, the abnormal behaviors are not considered to occur in the video frame.
Simulation experiment
This example performed experiments on the CUHK Avenue dataset and the UCSD pedistrian dataset:
(1) the ucsdpeedestrian dataset contains 2 subsets (Ped1 and Ped2), each subset corresponding to a different scene. The scene of Ped1 is a pedestrian moving towards or away from the camera, and comprises 34 training video sequences and 36 test video sequences. Each video sequence has 200 frames and a resolution of 238 x 158 pixels. The scene of Ped2 is a pedestrian moving parallel to the direction of the camera, and comprises 16 training video sequences and 12 testing video sequences. Each video sequence has 180 frames and a resolution of 360 x 240 pixels. Abnormal events in a data set are caused by two situations: the running of vehicles on sidewalks (e.g., trucks, bicycles) and abnormal pedestrian movement patterns (e.g., practicing lawns, running).
The CUHK Avenue data set contains 16 training video sequences and 21 test video sequences, with 47 number of exceptional events, including throwing objects, wandering, running, etc. The resolution of each video frame is 360 x 640 pixels. Anomalous events in the data set include throwing objects, running, etc.
In this example, Area Under the Curve (AUC) is used as an evaluation index of frame level (frame-level). The AUC is used to evaluate whether the video frame is correctly classified, and a higher value of the AUC indicates better performance of abnormal behavior detection. Table 1 shows the results of AUC for different methods:
TABLE 1
Figure BDA0002709044130000091
As shown in fig. 4, the result of abnormal behavior detection for the test video of paragraph 7 in the CUHK Avenue dataset is shown. A child in jumping appears in the video, and the method can detect the abnormal behavior from the video.
As shown in fig. 5, is the result of abnormal behavior detection for segment 7 of the test video in the UCSD Ped1 data set. The abnormal behavior can be detected from the video by the method of the embodiment.
As shown in fig. 6, the result of abnormal behavior detection for segment 4 test video in the UCSD Ped2 data set is shown. The method can detect the abnormal behavior from the video.
Example two:
the invention further provides an abnormal behavior detection terminal device, which comprises a memory, a processor and a computer program which is stored in the memory and can run on the processor, wherein the processor executes the computer program to realize the steps of the method embodiment of the first embodiment of the invention.
Further, as an executable scheme, the abnormal behavior detection terminal device may be a desktop computer, a notebook, a palm computer, a cloud server, or other computing devices. The abnormal behavior detection terminal device may include, but is not limited to, a processor and a memory. It is understood by those skilled in the art that the above-mentioned structure of the abnormal behavior detection terminal device is only an example of the abnormal behavior detection terminal device, and does not constitute a limitation on the abnormal behavior detection terminal device, and may include more or less components than the above, or combine some components, or different components, for example, the abnormal behavior detection terminal device may further include an input/output device, a network access device, a bus, and the like, which is not limited in this embodiment of the present invention.
Further, as an executable solution, the Processor may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other Programmable logic device, a discrete Gate or transistor logic device, a discrete hardware component, and the like. The general-purpose processor may be a microprocessor or the processor may be any conventional processor, and the processor is a control center of the abnormal behavior detection terminal device, and various interfaces and lines are used to connect various parts of the entire abnormal behavior detection terminal device.
The memory may be configured to store the computer program and/or the module, and the processor may implement various functions of the abnormal behavior detection terminal device by executing or executing the computer program and/or the module stored in the memory and calling data stored in the memory. The memory can mainly comprise a program storage area and a data storage area, wherein the program storage area can store an operating system and an application program required by at least one function; the storage data area may store data created according to the use of the mobile phone, and the like. In addition, the memory may include high speed random access memory, and may also include non-volatile memory, such as a hard disk, a memory, a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), at least one magnetic disk storage device, a Flash memory device, or other volatile solid state storage device.
The invention also provides a computer-readable storage medium, in which a computer program is stored, which, when being executed by a processor, carries out the steps of the above-mentioned method of an embodiment of the invention.
The abnormal behavior detection terminal device integrated module/unit may be stored in a computer-readable storage medium if it is implemented in the form of a software functional unit and sold or used as an independent product. Based on such understanding, all or part of the flow of the method according to the embodiments of the present invention may also be implemented by a computer program, which may be stored in a computer-readable storage medium, and when the computer program is executed by a processor, the steps of the method embodiments may be implemented. Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, usb disk, removable hard disk, magnetic disk, optical disk, computer Memory, Read-Only Memory (ROM), Random Access Memory (RAM), software distribution medium, and the like.
While the invention has been particularly shown and described with reference to a preferred embodiment, it will be understood by those skilled in the art that various changes in form and detail may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (8)

1. An abnormal behavior detection method is characterized by comprising the following steps:
s1: collecting a video image sequence formed by a plurality of groups of continuous t +1 frame video images, and forming a training set by all the video image sequences;
s2: constructing a U-Net neural network, inputting the U-Net neural network into a video image sequence consisting of 1 st to t th frame video images, outputting a reconstructed image reconstructed by the U-Net neural network, and training the U-Net neural network through a training set to minimize the difference between the reconstructed image and the real image of the acquired t +1 th frame; the training process of the U-Net neural network also comprises appearance constraint, motion constraint and information gain constraint;
s3: and continuously reconstructing images of continuous frames in the video through the trained U-Net neural network, and judging whether abnormal behaviors exist in the video or not through the relation between the difference between the reconstructed images and the corresponding real images and a threshold value.
2. The abnormal behavior detection method according to claim 1, characterized in that: the self-encoder formed by the U-Net neural network is divided into an encoder and a decoder, wherein the encoder is responsible for extracting the features of the image and gradually reducing the spatial size of the pooling layer, and the decoder is responsible for reconstructing the image and recovering the details and the spatial size of the image.
3. The abnormal behavior detection method according to claim 1, characterized in that: appearance constraints include intensity constraints and gradient constraints; the intensity constraint is to compute the difference of all pixel values between the reconstructed image and the real image, and the gradient constraint is to compute the gradient between the reconstructed image and the real image.
4. The abnormal behavior detection method according to claim 1, characterized in that: the motion constraint includes optical flow losses, which are used to compute the difference between the optical flow of the reconstructed image and the optical flow of the real image.
5. The abnormal behavior detection method according to claim 1, characterized in that: the information gain constraint is used to calculate the difference between the information entropy of the reconstructed image and the information entropy of the real image.
6. The abnormal behavior detection method according to claim 1, characterized in that: the difference is evaluated by the peak signal-to-noise ratio.
7. An abnormal behavior detection terminal device, characterized in that: comprising a processor, a memory and a computer program stored in the memory and running on the processor, the processor implementing the steps of the method according to any one of claims 1 to 6 when executing the computer program.
8. A computer-readable storage medium storing a computer program, characterized in that: the computer program when executed by a processor implementing the steps of the method as claimed in any one of claims 1 to 6.
CN202011049302.XA 2020-09-29 2020-09-29 Abnormal behavior detection method, terminal device and storage medium Withdrawn CN112149596A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011049302.XA CN112149596A (en) 2020-09-29 2020-09-29 Abnormal behavior detection method, terminal device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011049302.XA CN112149596A (en) 2020-09-29 2020-09-29 Abnormal behavior detection method, terminal device and storage medium

Publications (1)

Publication Number Publication Date
CN112149596A true CN112149596A (en) 2020-12-29

Family

ID=73894984

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011049302.XA Withdrawn CN112149596A (en) 2020-09-29 2020-09-29 Abnormal behavior detection method, terminal device and storage medium

Country Status (1)

Country Link
CN (1) CN112149596A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113032780A (en) * 2021-03-01 2021-06-25 厦门服云信息科技有限公司 Webshell detection method based on image analysis, terminal device and storage medium
CN113516075A (en) * 2021-07-09 2021-10-19 厦门恩格节能科技有限公司 Method, terminal equipment and medium for weighing by utilizing variable frequency signals of garbage crane
CN116450880A (en) * 2023-05-11 2023-07-18 湖南承希科技有限公司 Intelligent processing method for vehicle-mounted video of semantic detection

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102831447A (en) * 2012-08-30 2012-12-19 北京理工大学 Method for identifying multi-class facial expressions at high precision
CN102968779A (en) * 2012-12-03 2013-03-13 北方民族大学 Non-subsample Contourlet domain type MRI (Magnetic Resonance Imaging) image enhancing method based on FCM (Fuzzy C-means) cluster
CN108898173A (en) * 2018-06-25 2018-11-27 重庆知遨科技有限公司 A kind of the electrocardiogram Medical image fusion and classification method of multiple dimensioned multiple features
US20190228313A1 (en) * 2018-01-23 2019-07-25 Insurance Services Office, Inc. Computer Vision Systems and Methods for Unsupervised Representation Learning by Sorting Sequences
CN110705376A (en) * 2019-09-11 2020-01-17 南京邮电大学 Abnormal behavior detection method based on generative countermeasure network
CN111292323A (en) * 2020-03-16 2020-06-16 清华大学深圳国际研究生院 Partial-reference image quality evaluation method, control device, and computer-readable storage medium
CN111401239A (en) * 2020-03-16 2020-07-10 科大讯飞(苏州)科技有限公司 Video analysis method, device, system, equipment and storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102831447A (en) * 2012-08-30 2012-12-19 北京理工大学 Method for identifying multi-class facial expressions at high precision
CN102968779A (en) * 2012-12-03 2013-03-13 北方民族大学 Non-subsample Contourlet domain type MRI (Magnetic Resonance Imaging) image enhancing method based on FCM (Fuzzy C-means) cluster
US20190228313A1 (en) * 2018-01-23 2019-07-25 Insurance Services Office, Inc. Computer Vision Systems and Methods for Unsupervised Representation Learning by Sorting Sequences
CN108898173A (en) * 2018-06-25 2018-11-27 重庆知遨科技有限公司 A kind of the electrocardiogram Medical image fusion and classification method of multiple dimensioned multiple features
CN110705376A (en) * 2019-09-11 2020-01-17 南京邮电大学 Abnormal behavior detection method based on generative countermeasure network
CN111292323A (en) * 2020-03-16 2020-06-16 清华大学深圳国际研究生院 Partial-reference image quality evaluation method, control device, and computer-readable storage medium
CN111401239A (en) * 2020-03-16 2020-07-10 科大讯飞(苏州)科技有限公司 Video analysis method, device, system, equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
丛润民 等: "视频显著性检测研究进展" *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113032780A (en) * 2021-03-01 2021-06-25 厦门服云信息科技有限公司 Webshell detection method based on image analysis, terminal device and storage medium
CN113516075A (en) * 2021-07-09 2021-10-19 厦门恩格节能科技有限公司 Method, terminal equipment and medium for weighing by utilizing variable frequency signals of garbage crane
CN116450880A (en) * 2023-05-11 2023-07-18 湖南承希科技有限公司 Intelligent processing method for vehicle-mounted video of semantic detection
CN116450880B (en) * 2023-05-11 2023-09-01 湖南承希科技有限公司 Intelligent processing method for vehicle-mounted video of semantic detection

Similar Documents

Publication Publication Date Title
Ravanbakhsh et al. Training adversarial discriminators for cross-channel abnormal event detection in crowds
CN108805015B (en) Crowd abnormity detection method for weighted convolution self-coding long-short term memory network
Zhao et al. Spatio-temporal autoencoder for video anomaly detection
Kim et al. Fully deep blind image quality predictor
Han et al. Density-based multifeature background subtraction with support vector machine
US8811663B2 (en) Object detection in crowded scenes
CN112329685A (en) Method for detecting crowd abnormal behaviors through fusion type convolutional neural network
CN112149596A (en) Abnormal behavior detection method, terminal device and storage medium
Wang et al. A cognitive memory-augmented network for visual anomaly detection
Nasrollahi et al. Deep learning based super-resolution for improved action recognition
Chen et al. A vision based traffic accident detection method using extreme learning machine
CN111860414A (en) Method for detecting Deepfake video based on multi-feature fusion
CN111259919B (en) Video classification method, device and equipment and storage medium
US12106541B2 (en) Systems and methods for contrastive pretraining with video tracking supervision
CN113569756B (en) Abnormal behavior detection and positioning method, system, terminal equipment and readable storage medium
Luo et al. Traffic analytics with low-frame-rate videos
CN111079539A (en) Video abnormal behavior detection method based on abnormal tracking
Hu et al. Parallel spatial-temporal convolutional neural networks for anomaly detection and location in crowded scenes
Gnanavel et al. Abnormal event detection in crowded video scenes
Prawiro et al. Abnormal event detection in surveillance videos using two-stream decoder
Lamba et al. A texture based mani-fold approach for crowd density estimation using Gaussian Markov Random Field
Schneider et al. Unsupervised anomaly detection from time-of-flight depth images
Yang et al. Video anomaly detection for surveillance based on effective frame area
Panigrahi et al. A ResNet-101 deep learning framework induced transfer learning strategy for moving object detection
Tsai et al. Joint detection, re-identification, and LSTM in multi-object tracking

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication
WW01 Invention patent application withdrawn after publication

Application publication date: 20201229