CN107274432B - A kind of intelligent video monitoring method - Google Patents
A kind of intelligent video monitoring method Download PDFInfo
- Publication number
- CN107274432B CN107274432B CN201710434834.7A CN201710434834A CN107274432B CN 107274432 B CN107274432 B CN 107274432B CN 201710434834 A CN201710434834 A CN 201710434834A CN 107274432 B CN107274432 B CN 107274432B
- Authority
- CN
- China
- Prior art keywords
- network
- input
- encoder
- training
- sample
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/254—Analysis of motion involving subtraction of images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20024—Filtering details
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The present invention relates to the common scene intelligent video monitoring methods that a kind of view-based access control model conspicuousness and depth encode certainly, it include: that single frames decomposition is carried out to video, motion information is extracted using vision significance, then the light stream of consecutive frame moving object is calculated, detection process later is divided into two processes of training and test, in training, using the light stream of training sample as the input from coding, entire autoencoder network is trained by minimizing loss function, in test phase, respectively using trained and test sample light stream as input, extract the encoder in trained autoencoder network, the feature of input is extracted by dimensionality reduction, then the result after dimensionality reduction is visualized, the visualization range of training sample is indicated using suprasphere, in input test sample, it is visualized using same method, if the visual result of sample falls into suprasphere range It is interior, then determine that sample is normal;Conversely, falling in except suprasphere range, determines that sample is abnormal, be achieved in the intelligent monitoring of video.
Description
Technical field
The present invention relates to image processing techniques, public from what is encoded more particularly to a kind of view-based access control model conspicuousness and depth
Scene intelligent video monitoring method.
Background technique
In recent years, monitoring device was used in all trades and professions, and the common scenes such as modern airport, station, hospital are covered into
Thousand monitoring devices up to ten thousand depend merely on Security Personnel's analysis, filter out the normal behaviour under normal scene since video data is numerous,
Note abnormalities behavior in time, is a very big workload, and increasing with analysis quantity, the attention and work effect of personnel
Rate can generate apparent decline, in order to free people from a large amount of analysis and understanding, study a kind of intelligent video monitoring side
Method is of great significance.
Intelligent monitor system relates generally to the content of three parts: the extraction of motion information in video, i.e., in extraction video
Moving target, since monitoring system is fixed, so this part is mainly the motion information for extracting foreground target in video;Row
The extraction being characterized, a major challenge in intelligent monitor system will have the characteristics that uniqueness, robustness, extract feature;It is abnormal
Behavioral value is divided into rule-based detection, such as detects whether target violates certain predefined rules, and based on statistics
The mode of behavior is found in detection in a large amount of sample, use pattern knows method for distinguishing and model carries out abnormal behaviour identification.
Existing second of technology multidigit, use pattern knows method for distinguishing to be identified, but this method precision compares deep learning
Method precision it is low, therefore service precision of the present invention is higher carries out abnormal row based on depth autoencoder network in deep learning
For identification.
Summary of the invention
In view of this, the main purpose of the present invention is to provide a kind of detection accuracy, high, strong robustness view-based access control model is aobvious
Work property and depth substantially increase detection accuracy from the common scene intelligent video monitoring method of coding, meanwhile, it can cope with a variety of
Abnormal behaviour identification under scene, robustness are very strong.
In order to achieve the above object, technical solution proposed by the present invention are as follows: a kind of view-based access control model conspicuousness and depth are self-editing
The common scene intelligent video monitoring method of code realizes that steps are as follows:
Video under step 1, reading common scene, resolves into single frame for video, is then based on difference of Gaussian assemblage zone
Bandpass filter calculates the vision significance figure of each frame, extracts motion information with this;
Step 2, on the basis of every frame Saliency maps, the light stream of consecutive frame is calculated, to extract the movement of foreground target
Information obtains motion feature;
Step 3 includes two processes of training and test, in the training process, calculating training in the algorithm of anomalous identification
The vision significance figure of sample simultaneously extracts motion feature, and obtained Optical-flow Feature is converted to column vector as depth and encodes net certainly
The input of network loses letter by minimizing using the dimensionality reduction of encoder in depth autoencoder network and the restructuration of decoder
Number rebuilds input, training depth autoencoder network;
Step 4 rebuilds input by minimizing loss function, after training depth autoencoder network, extracts trained depth
The encoder section for spending autoencoder network is used as the network in test process, calculates separately out showing for training sample and test sample
After work property figure and motion feature, using the Optical-flow Feature of each sample as the input of encoder in depth autoencoder network, pass through
The dimensionality reduction of the encoder network operates, and the low-dimensional feature that can most represent input is extracted with low-dimensional vector;
Step 5, in three-dimensional coordinate during visual testing encoder network as a result, indicating it with a suprasphere
Distribution after middle training sample dimensionality reduction;
Step 6, for the anomalous identification of input test sample, if the visual range of test sample falls into the model of suprasphere
In enclosing, then determine the test sample for normal sequence;Conversely, falling in other than suprasphere range, then determine that the test sample is different
Chang Xulie is achieved in the identification of abnormal behaviour, the intelligent monitoring of video under common scene.
The method of vision significance figure is as follows in the step 1:
Step i) is for a frame image, the significance of each point in image is defined as:
S (x, y)=| | Iμ-Iwhc(x,y)||
Wherein, IμThe mean value of input picture each pixel color in Lab space, Iwhc(x, y) is to carry out height to image
After this is fuzzy, for each pixel in the value of Lab space, S (x, y) indicates the significance of each pixel, for both Euclidean away from
From;
Step ii) Gaussian Blur, two-dimensional gauss of distribution function are carried out to image first are as follows:
Wherein, x and y respectively corresponds the transverse and longitudinal coordinate of 8 points around central point, and σ is the variance of gauss of distribution function, G (x,
It y) is the fog-level of each pixel;
For color image, in R, convolution operation is done using Gaussian kernel and original image respectively in tri- channels G, B, will be each
The result in channel merges, as the image after Gaussian Blur, respectively by after Gaussian Blur image and original image be transformed into Lab sky
Between
Step iii) calculate Gaussian Blur after each pixel of image Lab value IwhcThe each pixel of (x, y) and original image
In the mean value I of Lab space colorμ, the Euclidean distance of the two is calculated to get the vision significance figure of original image is arrived.
The step 3 uses the detailed process for training depth autoencoder network are as follows:
Only include normal sample in step i) training sample, in the training process, calculates training sample consecutive frame image
Optical-flow Feature is converted to column vector by Optical-flow Feature, is one to connect entirely from coding as the input of depth autoencoder network
Mode, allow output as closely as possible equal to input structure be input layer-hidden layer-output layer network, whole network is by left-half
Encoder and right half part decoder composition, encoder is used for Data Dimensionality Reduction, and the characteristic information of input can most be represented by extracting;
Decoder rebuilds being originally inputted for whole network using the output of encoder as the input of decoder with error as small as possible,
Depth autoencoder network is increased in encoder network and decoder network several hidden on the basis of autoencoder network
Layer;
Step ii) with light stream be input X={ x1,x2...xn, the activation primitive of network uses ReLU function f (x)=max
(0, x), wherein x is the input of activation primitive, i.e., independent variable, f (x) are the dependent variables of activation primitive, and the first half of network is
The output of encoder network are as follows: Z=f (wX+b), wherein w is the weight of encoder network, and b is the biasing of encoder network, Z
For the output of encoder network, i.e. Z is after X dimensionality reduction as a result, the characteristic information of X can be represented;The latter half of network decodes
The output of device are as follows: Y=f (w'Z+b'), wherein w' is the weight of decoder network, and b' is the biasing of decoder network, i.e. Y is X
Reconstruction, entire encoder network indicates with formulating are as follows: Y=f (w'(f (wX+b))+b').
Step iii) loss function use mean square error: MSE=| | X-Y | |2=| | X-f (w'(f (wX+b))+b') | |2,
And loss function is minimized to rebuild input, it is exactly the training process by depth autoencoder network, keeps mean square error minimum, this
When output be input reconstruction.
The step 4 extracts the encoder section of trained depth autoencoder network as the network mistake in test process
Journey are as follows:
Step i) is firstly, the pretreatment of image is similar with training process, with the Optical-flow Feature of training sample and test sample
Input of the column vector of conversion as network;
Step ii) it is different from network used in training process, the training that training process obtains is extracted in test process
Network of the encoder as test process in good depth autoencoder network, is acted on using the dimensionality reduction of encoder network, will be defeated
Enter 3 neurons of boil down to, by the characteristics of encoder it is found that these three neurons can comprising input all information.
In conclusion the common scene intelligent video that a kind of view-based access control model conspicuousness of the present invention and depth encode certainly
Monitoring method, comprising: single frames decomposition is carried out to the video under common scene, it is aobvious using vision in decompositing the video frame come
Work property extracts motion information, then the light stream of calculating consecutive frame moving object, the size and Orientation including movement velocity, later
Detection process is divided into two processes of training and test, in training, the input of coding certainly using the light stream of training sample as depth,
Entire depth autoencoder network is trained by minimizing loss function, in test phase, respectively with trained and test sample
Light stream extracts the encoder in trained depth autoencoder network as input, is mentioned by the dimensionality reduction effect of encoder network
The feature for taking input, according to the characteristic of encoder network, the feature after dimensionality reduction can represent all information of input, then visualize
It is after dimensionality reduction as a result, using suprasphere indicate training sample visualization range use same side in input test sample
Method visualization, if the visual result of sample is fallen within the scope of suprasphere, determines that sample is normal;Conversely, falling in suprasphere model
Except enclosing, determines that sample is abnormal, be achieved in the intelligent monitoring of video.
The advantages of the present invention over the prior art are that:
(1) present invention tentatively extracts motion information using vision significance and optical flow method based on the identification of abnormal behaviour,
Then feature is extracted using the depth self-encoding encoder in deep learning, is trained and detects, since depth self-encoding encoder can be with
It minimizes loss function and rebuilds input, the dimensionality reduction effect of encoder can extract the low-dimensional feature that can indicate input information, so mentioning
The feature taken has very strong robustness, and just because of the robustness of feature, it can efficiently carry out very much the knowledge of abnormal behaviour
Not, arithmetic accuracy is improved.Due to using suprasphere to indicate normal range, when carrying out abnormal differentiation, it is only necessary to which judgement can
Range depending on changing result, so judging that speed is fast.
(2) present invention have the characteristics that detection accuracy height, strong robustness, can be widely applied to Community Safety protection, hospital,
The safeguard protection of the common scenes such as bank.By using the depth autoencoder network in optical flow method and deep learning, extracting can table
Show the low-dimensional feature of object all information, judgement is accurate, strong robustness, due to using suprasphere to indicate normal range, into
When the differentiation of row exception, it is only necessary to the range of visualization result is judged, so judging that speed is fast.
Detailed description of the invention
Fig. 1 is implementation flow chart of the present invention.
Specific embodiment
To make the object, technical solutions and advantages of the present invention clearer, right below in conjunction with the accompanying drawings and the specific embodiments
The present invention is described in further detail.
The common scene intelligent video monitoring method that a kind of view-based access control model conspicuousness and depth of the present invention encode certainly,
Include: that single frames decomposition is carried out to video under common scene, in decompositing the video frame come, is extracted and moved using vision significance
Information, then calculates the light stream of consecutive frame moving object, the size and Orientation including movement velocity, and detection process later is divided into
Two processes of training and test, using the light stream of training sample as depth from the input of coding, are damaged in training by minimizing
It loses function and trains entire depth autoencoder network, in test phase, respectively using trained and test sample light stream as inputting,
The encoder in trained depth autoencoder network is extracted, the feature of input is extracted by the dimensionality reduction effect of encoder network,
According to the characteristic of encoder network, the feature after dimensionality reduction can represent all information of input, then visualize dimensionality reduction after as a result,
Indicate that the visualization range of training sample is visualized in input test sample using same method using suprasphere, if sample
This visual result is fallen within the scope of suprasphere, then determines that sample is normal;Conversely, falling in except suprasphere range, sample is determined
This exception is achieved in the intelligent monitoring of video.
As shown in Figure 1, the present invention is implemented as follows step:
Video under step 1), reading common scene, resolves into single frame for video, is then based on difference of Gaussian assemblage zone
Bandpass filter calculates the vision significance figure of each frame, extracts motion information with this;
Step 2), on the basis of every frame Saliency maps, the light stream of consecutive frame is calculated, to extract the movement of foreground target
Information obtains motion feature;
Step 3) includes two processes of training and test, in the training process, calculating training in the algorithm of anomalous identification
The vision significance figure of sample simultaneously extracts motion feature, and the Optical-flow Feature of obtained every frame image is converted to column vector as deep
The input of degree autoencoder network passes through minimum using the reconstruction of the dimensionality reduction and decoder of encoder in depth autoencoder network
Loss function rebuilds input, training depth autoencoder network;
Step 4) rebuilds input by minimizing loss function, after training depth autoencoder network, extracts trained depth
The encoder section for spending autoencoder network is used as the network in test process, calculates separately out showing for training sample and test sample
After work property figure and motion feature, using the Optical-flow Feature of each picture frame sample as the input of encoder in depth autoencoder network,
It is operated by the dimensionality reduction of the encoder network, the low-dimensional feature that can most represent input is extracted with low-dimensional vector;
Step 5), in three-dimensional coordinate during visual testing encoder network as a result, being indicated with a suprasphere
The wherein distribution after training sample dimensionality reduction;
Step 6), for the anomalous identification of input test sample, if the visual range of test sample falls into suprasphere
In range, then determine the test sample for normal sequence;Conversely, falling in other than suprasphere range, then determine that the test sample is
Unusual sequences are achieved in the identification of abnormal behaviour, the intelligent monitoring of video under common scene.
The calculation method of vision significance figure is as follows in the step 1):
Step i) is for a frame image, the significance of each point in image is defined as:
S (x, y)=| | Iμ-Iwhc(x,y)||
Wherein, IμThe mean value of input picture each pixel color in Lab space, Iwhc(x, y) is to carry out height to image
After this is fuzzy, for each pixel in the value of Lab space, S (x, y) indicates the significance of each pixel, for both Euclidean away from
From;
Step ii) Gaussian Blur, two-dimensional gauss of distribution function are carried out to image first are as follows:
Wherein, x and y respectively corresponds the transverse and longitudinal coordinate of 8 points around central point, and σ is the variance of gauss of distribution function, G (x,
It y) is the fog-level of each pixel;
In R, convolution operation is done using Gaussian kernel and original image respectively in tri- channels G, B, and the result in each channel is closed
And the image as after Gaussian Blur, respectively by after Gaussian Blur image and original image be transformed into Lab space.
Step iii) calculate Gaussian Blur after each pixel of image Lab value IwhcThe each pixel of (x, y) and original image
In the mean value I of Lab space colorμ, the Euclidean distance of the two is calculated to get the vision significance figure of original image is arrived.
Step 3) the training depth autoencoder network principle is as follows:
Only include normal sample in step i) training sample, in the training process, calculates training sample consecutive frame image
Optical-flow Feature is converted to column vector by Optical-flow Feature, is one to connect entirely from coding as the input of depth autoencoder network
Mode allows output as closely as possible equal to the structure of input to be input layer-hidden layer-output layer network, whole network by encoder and
Decoder composition, encoder are used for Data Dimensionality Reduction, extract the characteristic information that can most represent input;Decoder is with mistake as small as possible
Difference rebuilds being originally inputted for whole network using the output of encoder as the input of decoder, and depth autoencoder network is certainly
On the basis of coding network, several hidden layers are increased in encoder network and decoder network;
Step ii) with light stream be input X={ x1,x2...xn, the activation primitive of network uses ReLU function f (x)=max
(0, x), wherein x is the input of activation primitive, i.e., independent variable, f (x) are the dependent variables of activation primitive, and the first half of network is
The output of encoder network are as follows: Z=f (wX+b), wherein w is the weight of encoder network, and b is the biasing of encoder network, Z
For the output of encoder network, i.e. Z is after X dimensionality reduction as a result, the characteristic information of X can be represented;The latter half of network decodes
The output of device are as follows: Y=f (w'Z+b'), wherein w' is the weight of decoder network, and b' is the biasing of decoder network, i.e. Y is X
Reconstruction, entire encoder network indicates with formulating are as follows: Y=f (w'(f (wX+b))+b').
Step iii) loss function use mean square error: MSE=| | X-Y | |2=| | X-f (w'(f (wX+b))+b') | |2,
And loss function is minimized to rebuild input, it is exactly the training process by depth autoencoder network, keeps mean square error minimum, this
When output be input reconstruction.
The step 4) extracts the encoder section of trained depth autoencoder network as the network in test process
Detailed process are as follows:
Step i) is firstly, the pretreatment of image is similar with training process, with the Optical-flow Feature of training sample and test sample
Input of the column vector of conversion as network;
Step ii) it is different from network used in training process, the training that training process obtains is extracted in test process
Network of the encoder as test process in good depth autoencoder network, is acted on using the dimensionality reduction of encoder network, will be defeated
Enter 3 neurons of boil down to, 3 neurons can include all information of input.
In conclusion the above is merely preferred embodiments of the present invention, being not intended to limit the scope of the present invention.
All within the spirits and principles of the present invention, any modification, equivalent replacement, improvement and so on should be included in of the invention
Within protection scope.
Claims (4)
1. the common scene intelligent video monitoring method that a kind of view-based access control model conspicuousness and depth encode certainly, which is characterized in that real
It is existing that steps are as follows:
Video under step 1, reading common scene, resolves into single frame for video, is then based on difference of Gaussian combination band logical filter
Wave device calculates the vision significance figure of each frame, extracts motion information with this;
Step 2, on the basis of every frame Saliency maps, calculate the light stream of consecutive frame, to extract the motion information of foreground target,
Obtain motion feature;
Step 3 includes two processes of training and test, in the training process, calculating training sample in the algorithm of anomalous identification
Vision significance figure and extract motion feature, obtained Optical-flow Feature is converted into column vector as depth autoencoder network
Input, using the dimensionality reduction of encoder in depth autoencoder network and the restructuration of decoder, by minimizing loss function weight
Build input, training depth autoencoder network;
Step 4 passes through minimum loss function reconstruction input, after training depth autoencoder network, extracts trained depth certainly
The encoder section of coding network calculates separately out the conspicuousness of training sample and test sample as the network in test process
After figure and motion feature, using the Optical-flow Feature of each sample as the input of encoder in autoencoder network, by described self-editing
The dimensionality reduction operation of encoder, the low-dimensional feature that can most represent input is extracted with low-dimensional vector in code network;
Step 5, in three-dimensional coordinate during visual testing encoder network as a result, indicating wherein to instruct with a suprasphere
Distribution after practicing sample dimensionality reduction;
Step 6, for the anomalous identification of input test sample, if the visual range of test sample falls into the range of suprasphere
It is interior, then determine the test sample for normal sequence;Conversely, falling in other than suprasphere range, then determine the test sample for exception
Sequence is achieved in the identification of abnormal behaviour, the intelligent monitoring of video under common scene.
2. the common scene intelligent video monitoring that a kind of view-based access control model conspicuousness according to claim 1 and depth encode certainly
Method, it is characterised in that: in the step 1, the method for calculating the vision significance figure of each frame is as follows:
Step i) is for a frame image, the significance of each point in image are as follows:
S (x, y)=| | Iμ-Iwhc(x,y)||
Wherein, IμThe mean value of input picture each pixel color in Lab space, Iwhc(x, y) is to carry out Gaussian mode to image
After paste, each pixel is in the value of Lab space, and S (x, y) indicates the significance of each pixel, for the Euclidean distance of the two;
Step ii) Gaussian Blur, two-dimensional gauss of distribution function are carried out to image first are as follows:
Wherein, x and y respectively corresponds the transverse and longitudinal coordinate of 8 points around central point, and σ is the variance of gauss of distribution function, and G (x, y) is
The fog-level of each pixel;
In R, convolution operation is done using Gaussian kernel and original image respectively in tri- channels G, B, the result in each channel is merged, i.e.,
For the image after Gaussian Blur;Respectively by after Gaussian Blur image and original image be transformed into Lab space;
Step iii) calculate Gaussian Blur after each pixel of image Lab value IwhcThe each pixel of (x, y) and original image exists
The mean value I of Lab space colorμ, the Euclidean distance of the two is calculated to get the vision significance figure of original image is arrived.
3. the common scene intelligent video monitoring that a kind of view-based access control model conspicuousness according to claim 1 and depth encode certainly
Method, it is characterised in that: in the step 3, the specific implementation process of training depth autoencoder network are as follows:
Only include normal sample in step i) training sample, in the training process, calculates the light stream of training sample consecutive frame image
Optical-flow Feature is converted to column vector by feature, is one with full connection side from coding as the input of depth autoencoder network
Formula, allow output as closely as possible equal to input structure be input layer-hidden layer-output layer network, whole network is by left-half
The decoder of encoder and right half part composition, encoder are used for Data Dimensionality Reduction, extract the characteristic information that can most represent input;Solution
Code device rebuilds being originally inputted for whole network using the output of encoder as the input of decoder with error as small as possible, deep
Degree autoencoder network is on the basis of autoencoder network, and it is hidden in encoder network and decoder network to increase several
Layer;
Step ii) with light stream be input X={ x1,x2...xn, the activation primitive of network using ReLU function f (x)=max (0,
X), wherein x is the input of activation primitive, i.e. independent variable, and f (x) is the dependent variable of activation primitive, and the first half of network is compiled
The output of code device network are as follows: Z=f (wX+b), wherein w is the weight of encoder network, and b is the biasing of encoder network, and Z is
The output of encoder network, i.e. Z are after X dimensionality reduction as a result, the characteristic information of X can be represented;Latter half, that is, decoder of network
Output are as follows: Y=f (w'Z+b'), wherein w' be decoder network weight, b' be decoder network biasing, i.e. Y is X
It rebuilds, entire encoder network is indicated with formulating are as follows: Y=f (w'(f (wX+b))+b');
Step iii) loss function use mean square error: MSE=| | X-Y | |2=| | X-f (w'(f (wX+b))+b') | |2, and most
Smallization loss function inputs to rebuild, and is exactly the training process by depth autoencoder network, keeps mean square error minimum, at this time
Output is the reconstruction of input.
4. the common scene intelligent video monitoring that a kind of view-based access control model conspicuousness according to claim 1 and depth encode certainly
Method, it is characterised in that: in the step 4, the encoder section for extracting trained depth autoencoder network, which is used as, was tested
The detailed process of network in journey are as follows:
Step i) is firstly, the pretreatment of image is similar with training process, with training sample and the conversion of the Optical-flow Feature of test sample
Input of the column vector as network;
Step ii) it is different from network used in training process, extraction training process obtains trained in test process
Network of the encoder as test process in depth autoencoder network is acted on using the dimensionality reduction of encoder network, input is pressed
3 neurons are condensed to, 3 neurons can include all information of input.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710434834.7A CN107274432B (en) | 2017-06-10 | 2017-06-10 | A kind of intelligent video monitoring method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710434834.7A CN107274432B (en) | 2017-06-10 | 2017-06-10 | A kind of intelligent video monitoring method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107274432A CN107274432A (en) | 2017-10-20 |
CN107274432B true CN107274432B (en) | 2019-07-26 |
Family
ID=60066526
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710434834.7A Active CN107274432B (en) | 2017-06-10 | 2017-06-10 | A kind of intelligent video monitoring method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107274432B (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107833208B (en) * | 2017-10-27 | 2021-07-13 | 哈尔滨工业大学 | Hyperspectral anomaly detection method based on dynamic weight depth self-encoding |
CN108830882B (en) * | 2018-05-25 | 2022-05-17 | 中国科学技术大学 | Video abnormal behavior real-time detection method |
CN110009866A (en) * | 2019-04-03 | 2019-07-12 | 天津中航亿达科技有限公司 | A kind of method of video detection temperature anomaly |
CN111107107B (en) * | 2019-12-31 | 2022-03-29 | 奇安信科技集团股份有限公司 | Network behavior detection method and device, computer equipment and storage medium |
US11748629B2 (en) * | 2020-01-21 | 2023-09-05 | Moxa Inc. | Device and method of handling anomaly detection |
CN113222926B (en) * | 2021-05-06 | 2023-04-18 | 西安电子科技大学 | Zipper abnormity detection method based on depth support vector data description model |
CN113592390A (en) * | 2021-07-12 | 2021-11-02 | 嘉兴恒创电力集团有限公司博创物资分公司 | Warehousing digital twin method and system based on multi-sensor fusion |
CN115714731B (en) * | 2022-09-27 | 2023-06-27 | 中国人民解放军63921部队 | Deep space measurement and control link abnormality detection method based on deep learning self-encoder |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106778921A (en) * | 2017-02-15 | 2017-05-31 | 张烜 | Personnel based on deep learning encoding model recognition methods again |
-
2017
- 2017-06-10 CN CN201710434834.7A patent/CN107274432B/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106778921A (en) * | 2017-02-15 | 2017-05-31 | 张烜 | Personnel based on deep learning encoding model recognition methods again |
Non-Patent Citations (4)
Title |
---|
"Dynamic Background Learning through Deep Auto-encoder Networks";Pei Xu etc.;《Proceedings of the 22nd ACM international conference on Multimedia》;20141107;第107-116页 |
"Optical flow in the presence of spatially-varying motion blur";Travis Portz etc.;《IEEE Conference on Computer Vision and Pattern Recognition》;20120726;第1752-1759页 |
"基于视觉的目标检测与跟踪综述";尹宏鹏等;《自动化学报》;20161031;第42卷(第10期);第1467-1481页 |
"视觉显著性检测方法及其应用研究";凌南平;《万方数据知识服务平台》;20141106;第4.1-4.3节 |
Also Published As
Publication number | Publication date |
---|---|
CN107274432A (en) | 2017-10-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107274432B (en) | A kind of intelligent video monitoring method | |
CN108846365B (en) | Detection method and device for fighting behavior in video, storage medium and processor | |
CN107330364B (en) | A kind of people counting method and system based on cGAN network | |
CN104361327B (en) | A kind of pedestrian detection method and system | |
CN106960202B (en) | Smiling face identification method based on visible light and infrared image fusion | |
CN108805002B (en) | Monitoring video abnormal event detection method based on deep learning and dynamic clustering | |
CN105956560B (en) | A kind of model recognizing method based on the multiple dimensioned depth convolution feature of pondization | |
CN103839065B (en) | Extraction method for dynamic crowd gathering characteristics | |
Mancas et al. | Abnormal motion selection in crowds using bottom-up saliency | |
CN106910204B (en) | A kind of method and system to the automatic Tracking Recognition of sea ship | |
CN104732236B (en) | A kind of crowd's abnormal behaviour intelligent detecting method based on layered shaping | |
CN110176016B (en) | Virtual fitting method based on human body contour segmentation and skeleton recognition | |
He et al. | A robust method for wheatear detection using UAV in natural scenes | |
CN107330360A (en) | A kind of pedestrian's clothing colour recognition, pedestrian retrieval method and device | |
CN104298974A (en) | Human body behavior recognition method based on depth video sequence | |
CN109740609A (en) | A kind of gauge detection method and device | |
CN113312973B (en) | Gesture recognition key point feature extraction method and system | |
CN103902989A (en) | Human body motion video recognition method based on non-negative matrix factorization | |
CN106971158A (en) | A kind of pedestrian detection method based on CoLBP symbiosis feature Yu GSS features | |
CN110490055A (en) | A kind of Weakly supervised Activity recognition localization method and device recoded based on three | |
Yu et al. | Robust median filtering forensics by CNN-based multiple residuals learning | |
CN109117774A (en) | A kind of multi-angle video method for detecting abnormality based on sparse coding | |
CN110084201A (en) | A kind of human motion recognition method of convolutional neural networks based on specific objective tracking under monitoring scene | |
CN105631405B (en) | Traffic video intelligent recognition background modeling method based on Multilevel Block | |
CN113837154B (en) | Open set filtering system and method based on multitask assistance |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |