CN107274432B - A kind of intelligent video monitoring method - Google Patents

A kind of intelligent video monitoring method Download PDF

Info

Publication number
CN107274432B
CN107274432B CN201710434834.7A CN201710434834A CN107274432B CN 107274432 B CN107274432 B CN 107274432B CN 201710434834 A CN201710434834 A CN 201710434834A CN 107274432 B CN107274432 B CN 107274432B
Authority
CN
China
Prior art keywords
network
input
encoder
training
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710434834.7A
Other languages
Chinese (zh)
Other versions
CN107274432A (en
Inventor
王田
乔美娜
陈阳
陶飞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Original Assignee
Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University filed Critical Beihang University
Priority to CN201710434834.7A priority Critical patent/CN107274432B/en
Publication of CN107274432A publication Critical patent/CN107274432A/en
Application granted granted Critical
Publication of CN107274432B publication Critical patent/CN107274432B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/254Analysis of motion involving subtraction of images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20024Filtering details
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The present invention relates to the common scene intelligent video monitoring methods that a kind of view-based access control model conspicuousness and depth encode certainly, it include: that single frames decomposition is carried out to video, motion information is extracted using vision significance, then the light stream of consecutive frame moving object is calculated, detection process later is divided into two processes of training and test, in training, using the light stream of training sample as the input from coding, entire autoencoder network is trained by minimizing loss function, in test phase, respectively using trained and test sample light stream as input, extract the encoder in trained autoencoder network, the feature of input is extracted by dimensionality reduction, then the result after dimensionality reduction is visualized, the visualization range of training sample is indicated using suprasphere, in input test sample, it is visualized using same method, if the visual result of sample falls into suprasphere range It is interior, then determine that sample is normal;Conversely, falling in except suprasphere range, determines that sample is abnormal, be achieved in the intelligent monitoring of video.

Description

A kind of intelligent video monitoring method
Technical field
The present invention relates to image processing techniques, public from what is encoded more particularly to a kind of view-based access control model conspicuousness and depth Scene intelligent video monitoring method.
Background technique
In recent years, monitoring device was used in all trades and professions, and the common scenes such as modern airport, station, hospital are covered into Thousand monitoring devices up to ten thousand depend merely on Security Personnel's analysis, filter out the normal behaviour under normal scene since video data is numerous, Note abnormalities behavior in time, is a very big workload, and increasing with analysis quantity, the attention and work effect of personnel Rate can generate apparent decline, in order to free people from a large amount of analysis and understanding, study a kind of intelligent video monitoring side Method is of great significance.
Intelligent monitor system relates generally to the content of three parts: the extraction of motion information in video, i.e., in extraction video Moving target, since monitoring system is fixed, so this part is mainly the motion information for extracting foreground target in video;Row The extraction being characterized, a major challenge in intelligent monitor system will have the characteristics that uniqueness, robustness, extract feature;It is abnormal Behavioral value is divided into rule-based detection, such as detects whether target violates certain predefined rules, and based on statistics The mode of behavior is found in detection in a large amount of sample, use pattern knows method for distinguishing and model carries out abnormal behaviour identification. Existing second of technology multidigit, use pattern knows method for distinguishing to be identified, but this method precision compares deep learning Method precision it is low, therefore service precision of the present invention is higher carries out abnormal row based on depth autoencoder network in deep learning For identification.
Summary of the invention
In view of this, the main purpose of the present invention is to provide a kind of detection accuracy, high, strong robustness view-based access control model is aobvious Work property and depth substantially increase detection accuracy from the common scene intelligent video monitoring method of coding, meanwhile, it can cope with a variety of Abnormal behaviour identification under scene, robustness are very strong.
In order to achieve the above object, technical solution proposed by the present invention are as follows: a kind of view-based access control model conspicuousness and depth are self-editing The common scene intelligent video monitoring method of code realizes that steps are as follows:
Video under step 1, reading common scene, resolves into single frame for video, is then based on difference of Gaussian assemblage zone Bandpass filter calculates the vision significance figure of each frame, extracts motion information with this;
Step 2, on the basis of every frame Saliency maps, the light stream of consecutive frame is calculated, to extract the movement of foreground target Information obtains motion feature;
Step 3 includes two processes of training and test, in the training process, calculating training in the algorithm of anomalous identification The vision significance figure of sample simultaneously extracts motion feature, and obtained Optical-flow Feature is converted to column vector as depth and encodes net certainly The input of network loses letter by minimizing using the dimensionality reduction of encoder in depth autoencoder network and the restructuration of decoder Number rebuilds input, training depth autoencoder network;
Step 4 rebuilds input by minimizing loss function, after training depth autoencoder network, extracts trained depth The encoder section for spending autoencoder network is used as the network in test process, calculates separately out showing for training sample and test sample After work property figure and motion feature, using the Optical-flow Feature of each sample as the input of encoder in depth autoencoder network, pass through The dimensionality reduction of the encoder network operates, and the low-dimensional feature that can most represent input is extracted with low-dimensional vector;
Step 5, in three-dimensional coordinate during visual testing encoder network as a result, indicating it with a suprasphere Distribution after middle training sample dimensionality reduction;
Step 6, for the anomalous identification of input test sample, if the visual range of test sample falls into the model of suprasphere In enclosing, then determine the test sample for normal sequence;Conversely, falling in other than suprasphere range, then determine that the test sample is different Chang Xulie is achieved in the identification of abnormal behaviour, the intelligent monitoring of video under common scene.
The method of vision significance figure is as follows in the step 1:
Step i) is for a frame image, the significance of each point in image is defined as:
S (x, y)=| | Iμ-Iwhc(x,y)||
Wherein, IμThe mean value of input picture each pixel color in Lab space, Iwhc(x, y) is to carry out height to image After this is fuzzy, for each pixel in the value of Lab space, S (x, y) indicates the significance of each pixel, for both Euclidean away from From;
Step ii) Gaussian Blur, two-dimensional gauss of distribution function are carried out to image first are as follows:
Wherein, x and y respectively corresponds the transverse and longitudinal coordinate of 8 points around central point, and σ is the variance of gauss of distribution function, G (x, It y) is the fog-level of each pixel;
For color image, in R, convolution operation is done using Gaussian kernel and original image respectively in tri- channels G, B, will be each The result in channel merges, as the image after Gaussian Blur, respectively by after Gaussian Blur image and original image be transformed into Lab sky Between
Step iii) calculate Gaussian Blur after each pixel of image Lab value IwhcThe each pixel of (x, y) and original image In the mean value I of Lab space colorμ, the Euclidean distance of the two is calculated to get the vision significance figure of original image is arrived.
The step 3 uses the detailed process for training depth autoencoder network are as follows:
Only include normal sample in step i) training sample, in the training process, calculates training sample consecutive frame image Optical-flow Feature is converted to column vector by Optical-flow Feature, is one to connect entirely from coding as the input of depth autoencoder network Mode, allow output as closely as possible equal to input structure be input layer-hidden layer-output layer network, whole network is by left-half Encoder and right half part decoder composition, encoder is used for Data Dimensionality Reduction, and the characteristic information of input can most be represented by extracting; Decoder rebuilds being originally inputted for whole network using the output of encoder as the input of decoder with error as small as possible, Depth autoencoder network is increased in encoder network and decoder network several hidden on the basis of autoencoder network Layer;
Step ii) with light stream be input X={ x1,x2...xn, the activation primitive of network uses ReLU function f (x)=max (0, x), wherein x is the input of activation primitive, i.e., independent variable, f (x) are the dependent variables of activation primitive, and the first half of network is The output of encoder network are as follows: Z=f (wX+b), wherein w is the weight of encoder network, and b is the biasing of encoder network, Z For the output of encoder network, i.e. Z is after X dimensionality reduction as a result, the characteristic information of X can be represented;The latter half of network decodes The output of device are as follows: Y=f (w'Z+b'), wherein w' is the weight of decoder network, and b' is the biasing of decoder network, i.e. Y is X Reconstruction, entire encoder network indicates with formulating are as follows: Y=f (w'(f (wX+b))+b').
Step iii) loss function use mean square error: MSE=| | X-Y | |2=| | X-f (w'(f (wX+b))+b') | |2, And loss function is minimized to rebuild input, it is exactly the training process by depth autoencoder network, keeps mean square error minimum, this When output be input reconstruction.
The step 4 extracts the encoder section of trained depth autoencoder network as the network mistake in test process Journey are as follows:
Step i) is firstly, the pretreatment of image is similar with training process, with the Optical-flow Feature of training sample and test sample Input of the column vector of conversion as network;
Step ii) it is different from network used in training process, the training that training process obtains is extracted in test process Network of the encoder as test process in good depth autoencoder network, is acted on using the dimensionality reduction of encoder network, will be defeated Enter 3 neurons of boil down to, by the characteristics of encoder it is found that these three neurons can comprising input all information.
In conclusion the common scene intelligent video that a kind of view-based access control model conspicuousness of the present invention and depth encode certainly Monitoring method, comprising: single frames decomposition is carried out to the video under common scene, it is aobvious using vision in decompositing the video frame come Work property extracts motion information, then the light stream of calculating consecutive frame moving object, the size and Orientation including movement velocity, later Detection process is divided into two processes of training and test, in training, the input of coding certainly using the light stream of training sample as depth, Entire depth autoencoder network is trained by minimizing loss function, in test phase, respectively with trained and test sample Light stream extracts the encoder in trained depth autoencoder network as input, is mentioned by the dimensionality reduction effect of encoder network The feature for taking input, according to the characteristic of encoder network, the feature after dimensionality reduction can represent all information of input, then visualize It is after dimensionality reduction as a result, using suprasphere indicate training sample visualization range use same side in input test sample Method visualization, if the visual result of sample is fallen within the scope of suprasphere, determines that sample is normal;Conversely, falling in suprasphere model Except enclosing, determines that sample is abnormal, be achieved in the intelligent monitoring of video.
The advantages of the present invention over the prior art are that:
(1) present invention tentatively extracts motion information using vision significance and optical flow method based on the identification of abnormal behaviour, Then feature is extracted using the depth self-encoding encoder in deep learning, is trained and detects, since depth self-encoding encoder can be with It minimizes loss function and rebuilds input, the dimensionality reduction effect of encoder can extract the low-dimensional feature that can indicate input information, so mentioning The feature taken has very strong robustness, and just because of the robustness of feature, it can efficiently carry out very much the knowledge of abnormal behaviour Not, arithmetic accuracy is improved.Due to using suprasphere to indicate normal range, when carrying out abnormal differentiation, it is only necessary to which judgement can Range depending on changing result, so judging that speed is fast.
(2) present invention have the characteristics that detection accuracy height, strong robustness, can be widely applied to Community Safety protection, hospital, The safeguard protection of the common scenes such as bank.By using the depth autoencoder network in optical flow method and deep learning, extracting can table Show the low-dimensional feature of object all information, judgement is accurate, strong robustness, due to using suprasphere to indicate normal range, into When the differentiation of row exception, it is only necessary to the range of visualization result is judged, so judging that speed is fast.
Detailed description of the invention
Fig. 1 is implementation flow chart of the present invention.
Specific embodiment
To make the object, technical solutions and advantages of the present invention clearer, right below in conjunction with the accompanying drawings and the specific embodiments The present invention is described in further detail.
The common scene intelligent video monitoring method that a kind of view-based access control model conspicuousness and depth of the present invention encode certainly, Include: that single frames decomposition is carried out to video under common scene, in decompositing the video frame come, is extracted and moved using vision significance Information, then calculates the light stream of consecutive frame moving object, the size and Orientation including movement velocity, and detection process later is divided into Two processes of training and test, using the light stream of training sample as depth from the input of coding, are damaged in training by minimizing It loses function and trains entire depth autoencoder network, in test phase, respectively using trained and test sample light stream as inputting, The encoder in trained depth autoencoder network is extracted, the feature of input is extracted by the dimensionality reduction effect of encoder network, According to the characteristic of encoder network, the feature after dimensionality reduction can represent all information of input, then visualize dimensionality reduction after as a result, Indicate that the visualization range of training sample is visualized in input test sample using same method using suprasphere, if sample This visual result is fallen within the scope of suprasphere, then determines that sample is normal;Conversely, falling in except suprasphere range, sample is determined This exception is achieved in the intelligent monitoring of video.
As shown in Figure 1, the present invention is implemented as follows step:
Video under step 1), reading common scene, resolves into single frame for video, is then based on difference of Gaussian assemblage zone Bandpass filter calculates the vision significance figure of each frame, extracts motion information with this;
Step 2), on the basis of every frame Saliency maps, the light stream of consecutive frame is calculated, to extract the movement of foreground target Information obtains motion feature;
Step 3) includes two processes of training and test, in the training process, calculating training in the algorithm of anomalous identification The vision significance figure of sample simultaneously extracts motion feature, and the Optical-flow Feature of obtained every frame image is converted to column vector as deep The input of degree autoencoder network passes through minimum using the reconstruction of the dimensionality reduction and decoder of encoder in depth autoencoder network Loss function rebuilds input, training depth autoencoder network;
Step 4) rebuilds input by minimizing loss function, after training depth autoencoder network, extracts trained depth The encoder section for spending autoencoder network is used as the network in test process, calculates separately out showing for training sample and test sample After work property figure and motion feature, using the Optical-flow Feature of each picture frame sample as the input of encoder in depth autoencoder network, It is operated by the dimensionality reduction of the encoder network, the low-dimensional feature that can most represent input is extracted with low-dimensional vector;
Step 5), in three-dimensional coordinate during visual testing encoder network as a result, being indicated with a suprasphere The wherein distribution after training sample dimensionality reduction;
Step 6), for the anomalous identification of input test sample, if the visual range of test sample falls into suprasphere In range, then determine the test sample for normal sequence;Conversely, falling in other than suprasphere range, then determine that the test sample is Unusual sequences are achieved in the identification of abnormal behaviour, the intelligent monitoring of video under common scene.
The calculation method of vision significance figure is as follows in the step 1):
Step i) is for a frame image, the significance of each point in image is defined as:
S (x, y)=| | Iμ-Iwhc(x,y)||
Wherein, IμThe mean value of input picture each pixel color in Lab space, Iwhc(x, y) is to carry out height to image After this is fuzzy, for each pixel in the value of Lab space, S (x, y) indicates the significance of each pixel, for both Euclidean away from From;
Step ii) Gaussian Blur, two-dimensional gauss of distribution function are carried out to image first are as follows:
Wherein, x and y respectively corresponds the transverse and longitudinal coordinate of 8 points around central point, and σ is the variance of gauss of distribution function, G (x, It y) is the fog-level of each pixel;
In R, convolution operation is done using Gaussian kernel and original image respectively in tri- channels G, B, and the result in each channel is closed And the image as after Gaussian Blur, respectively by after Gaussian Blur image and original image be transformed into Lab space.
Step iii) calculate Gaussian Blur after each pixel of image Lab value IwhcThe each pixel of (x, y) and original image In the mean value I of Lab space colorμ, the Euclidean distance of the two is calculated to get the vision significance figure of original image is arrived.
Step 3) the training depth autoencoder network principle is as follows:
Only include normal sample in step i) training sample, in the training process, calculates training sample consecutive frame image Optical-flow Feature is converted to column vector by Optical-flow Feature, is one to connect entirely from coding as the input of depth autoencoder network Mode allows output as closely as possible equal to the structure of input to be input layer-hidden layer-output layer network, whole network by encoder and Decoder composition, encoder are used for Data Dimensionality Reduction, extract the characteristic information that can most represent input;Decoder is with mistake as small as possible Difference rebuilds being originally inputted for whole network using the output of encoder as the input of decoder, and depth autoencoder network is certainly On the basis of coding network, several hidden layers are increased in encoder network and decoder network;
Step ii) with light stream be input X={ x1,x2...xn, the activation primitive of network uses ReLU function f (x)=max (0, x), wherein x is the input of activation primitive, i.e., independent variable, f (x) are the dependent variables of activation primitive, and the first half of network is The output of encoder network are as follows: Z=f (wX+b), wherein w is the weight of encoder network, and b is the biasing of encoder network, Z For the output of encoder network, i.e. Z is after X dimensionality reduction as a result, the characteristic information of X can be represented;The latter half of network decodes The output of device are as follows: Y=f (w'Z+b'), wherein w' is the weight of decoder network, and b' is the biasing of decoder network, i.e. Y is X Reconstruction, entire encoder network indicates with formulating are as follows: Y=f (w'(f (wX+b))+b').
Step iii) loss function use mean square error: MSE=| | X-Y | |2=| | X-f (w'(f (wX+b))+b') | |2, And loss function is minimized to rebuild input, it is exactly the training process by depth autoencoder network, keeps mean square error minimum, this When output be input reconstruction.
The step 4) extracts the encoder section of trained depth autoencoder network as the network in test process Detailed process are as follows:
Step i) is firstly, the pretreatment of image is similar with training process, with the Optical-flow Feature of training sample and test sample Input of the column vector of conversion as network;
Step ii) it is different from network used in training process, the training that training process obtains is extracted in test process Network of the encoder as test process in good depth autoencoder network, is acted on using the dimensionality reduction of encoder network, will be defeated Enter 3 neurons of boil down to, 3 neurons can include all information of input.
In conclusion the above is merely preferred embodiments of the present invention, being not intended to limit the scope of the present invention. All within the spirits and principles of the present invention, any modification, equivalent replacement, improvement and so on should be included in of the invention Within protection scope.

Claims (4)

1. the common scene intelligent video monitoring method that a kind of view-based access control model conspicuousness and depth encode certainly, which is characterized in that real It is existing that steps are as follows:
Video under step 1, reading common scene, resolves into single frame for video, is then based on difference of Gaussian combination band logical filter Wave device calculates the vision significance figure of each frame, extracts motion information with this;
Step 2, on the basis of every frame Saliency maps, calculate the light stream of consecutive frame, to extract the motion information of foreground target, Obtain motion feature;
Step 3 includes two processes of training and test, in the training process, calculating training sample in the algorithm of anomalous identification Vision significance figure and extract motion feature, obtained Optical-flow Feature is converted into column vector as depth autoencoder network Input, using the dimensionality reduction of encoder in depth autoencoder network and the restructuration of decoder, by minimizing loss function weight Build input, training depth autoencoder network;
Step 4 passes through minimum loss function reconstruction input, after training depth autoencoder network, extracts trained depth certainly The encoder section of coding network calculates separately out the conspicuousness of training sample and test sample as the network in test process After figure and motion feature, using the Optical-flow Feature of each sample as the input of encoder in autoencoder network, by described self-editing The dimensionality reduction operation of encoder, the low-dimensional feature that can most represent input is extracted with low-dimensional vector in code network;
Step 5, in three-dimensional coordinate during visual testing encoder network as a result, indicating wherein to instruct with a suprasphere Distribution after practicing sample dimensionality reduction;
Step 6, for the anomalous identification of input test sample, if the visual range of test sample falls into the range of suprasphere It is interior, then determine the test sample for normal sequence;Conversely, falling in other than suprasphere range, then determine the test sample for exception Sequence is achieved in the identification of abnormal behaviour, the intelligent monitoring of video under common scene.
2. the common scene intelligent video monitoring that a kind of view-based access control model conspicuousness according to claim 1 and depth encode certainly Method, it is characterised in that: in the step 1, the method for calculating the vision significance figure of each frame is as follows:
Step i) is for a frame image, the significance of each point in image are as follows:
S (x, y)=| | Iμ-Iwhc(x,y)||
Wherein, IμThe mean value of input picture each pixel color in Lab space, Iwhc(x, y) is to carry out Gaussian mode to image After paste, each pixel is in the value of Lab space, and S (x, y) indicates the significance of each pixel, for the Euclidean distance of the two;
Step ii) Gaussian Blur, two-dimensional gauss of distribution function are carried out to image first are as follows:
Wherein, x and y respectively corresponds the transverse and longitudinal coordinate of 8 points around central point, and σ is the variance of gauss of distribution function, and G (x, y) is The fog-level of each pixel;
In R, convolution operation is done using Gaussian kernel and original image respectively in tri- channels G, B, the result in each channel is merged, i.e., For the image after Gaussian Blur;Respectively by after Gaussian Blur image and original image be transformed into Lab space;
Step iii) calculate Gaussian Blur after each pixel of image Lab value IwhcThe each pixel of (x, y) and original image exists The mean value I of Lab space colorμ, the Euclidean distance of the two is calculated to get the vision significance figure of original image is arrived.
3. the common scene intelligent video monitoring that a kind of view-based access control model conspicuousness according to claim 1 and depth encode certainly Method, it is characterised in that: in the step 3, the specific implementation process of training depth autoencoder network are as follows:
Only include normal sample in step i) training sample, in the training process, calculates the light stream of training sample consecutive frame image Optical-flow Feature is converted to column vector by feature, is one with full connection side from coding as the input of depth autoencoder network Formula, allow output as closely as possible equal to input structure be input layer-hidden layer-output layer network, whole network is by left-half The decoder of encoder and right half part composition, encoder are used for Data Dimensionality Reduction, extract the characteristic information that can most represent input;Solution Code device rebuilds being originally inputted for whole network using the output of encoder as the input of decoder with error as small as possible, deep Degree autoencoder network is on the basis of autoencoder network, and it is hidden in encoder network and decoder network to increase several Layer;
Step ii) with light stream be input X={ x1,x2...xn, the activation primitive of network using ReLU function f (x)=max (0, X), wherein x is the input of activation primitive, i.e. independent variable, and f (x) is the dependent variable of activation primitive, and the first half of network is compiled The output of code device network are as follows: Z=f (wX+b), wherein w is the weight of encoder network, and b is the biasing of encoder network, and Z is The output of encoder network, i.e. Z are after X dimensionality reduction as a result, the characteristic information of X can be represented;Latter half, that is, decoder of network Output are as follows: Y=f (w'Z+b'), wherein w' be decoder network weight, b' be decoder network biasing, i.e. Y is X It rebuilds, entire encoder network is indicated with formulating are as follows: Y=f (w'(f (wX+b))+b');
Step iii) loss function use mean square error: MSE=| | X-Y | |2=| | X-f (w'(f (wX+b))+b') | |2, and most Smallization loss function inputs to rebuild, and is exactly the training process by depth autoencoder network, keeps mean square error minimum, at this time Output is the reconstruction of input.
4. the common scene intelligent video monitoring that a kind of view-based access control model conspicuousness according to claim 1 and depth encode certainly Method, it is characterised in that: in the step 4, the encoder section for extracting trained depth autoencoder network, which is used as, was tested The detailed process of network in journey are as follows:
Step i) is firstly, the pretreatment of image is similar with training process, with training sample and the conversion of the Optical-flow Feature of test sample Input of the column vector as network;
Step ii) it is different from network used in training process, extraction training process obtains trained in test process Network of the encoder as test process in depth autoencoder network is acted on using the dimensionality reduction of encoder network, input is pressed 3 neurons are condensed to, 3 neurons can include all information of input.
CN201710434834.7A 2017-06-10 2017-06-10 A kind of intelligent video monitoring method Active CN107274432B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710434834.7A CN107274432B (en) 2017-06-10 2017-06-10 A kind of intelligent video monitoring method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710434834.7A CN107274432B (en) 2017-06-10 2017-06-10 A kind of intelligent video monitoring method

Publications (2)

Publication Number Publication Date
CN107274432A CN107274432A (en) 2017-10-20
CN107274432B true CN107274432B (en) 2019-07-26

Family

ID=60066526

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710434834.7A Active CN107274432B (en) 2017-06-10 2017-06-10 A kind of intelligent video monitoring method

Country Status (1)

Country Link
CN (1) CN107274432B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107833208B (en) * 2017-10-27 2021-07-13 哈尔滨工业大学 Hyperspectral anomaly detection method based on dynamic weight depth self-encoding
CN108830882B (en) * 2018-05-25 2022-05-17 中国科学技术大学 Video abnormal behavior real-time detection method
CN110009866A (en) * 2019-04-03 2019-07-12 天津中航亿达科技有限公司 A kind of method of video detection temperature anomaly
CN111107107B (en) * 2019-12-31 2022-03-29 奇安信科技集团股份有限公司 Network behavior detection method and device, computer equipment and storage medium
US11748629B2 (en) * 2020-01-21 2023-09-05 Moxa Inc. Device and method of handling anomaly detection
CN113222926B (en) * 2021-05-06 2023-04-18 西安电子科技大学 Zipper abnormity detection method based on depth support vector data description model
CN113592390A (en) * 2021-07-12 2021-11-02 嘉兴恒创电力集团有限公司博创物资分公司 Warehousing digital twin method and system based on multi-sensor fusion
CN115714731B (en) * 2022-09-27 2023-06-27 中国人民解放军63921部队 Deep space measurement and control link abnormality detection method based on deep learning self-encoder

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106778921A (en) * 2017-02-15 2017-05-31 张烜 Personnel based on deep learning encoding model recognition methods again

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106778921A (en) * 2017-02-15 2017-05-31 张烜 Personnel based on deep learning encoding model recognition methods again

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
"Dynamic Background Learning through Deep Auto-encoder Networks";Pei Xu etc.;《Proceedings of the 22nd ACM international conference on Multimedia》;20141107;第107-116页
"Optical flow in the presence of spatially-varying motion blur";Travis Portz etc.;《IEEE Conference on Computer Vision and Pattern Recognition》;20120726;第1752-1759页
"基于视觉的目标检测与跟踪综述";尹宏鹏等;《自动化学报》;20161031;第42卷(第10期);第1467-1481页
"视觉显著性检测方法及其应用研究";凌南平;《万方数据知识服务平台》;20141106;第4.1-4.3节

Also Published As

Publication number Publication date
CN107274432A (en) 2017-10-20

Similar Documents

Publication Publication Date Title
CN107274432B (en) A kind of intelligent video monitoring method
CN108846365B (en) Detection method and device for fighting behavior in video, storage medium and processor
CN107330364B (en) A kind of people counting method and system based on cGAN network
CN104361327B (en) A kind of pedestrian detection method and system
CN106960202B (en) Smiling face identification method based on visible light and infrared image fusion
CN108805002B (en) Monitoring video abnormal event detection method based on deep learning and dynamic clustering
CN105956560B (en) A kind of model recognizing method based on the multiple dimensioned depth convolution feature of pondization
CN103839065B (en) Extraction method for dynamic crowd gathering characteristics
Mancas et al. Abnormal motion selection in crowds using bottom-up saliency
CN106910204B (en) A kind of method and system to the automatic Tracking Recognition of sea ship
CN104732236B (en) A kind of crowd's abnormal behaviour intelligent detecting method based on layered shaping
CN110176016B (en) Virtual fitting method based on human body contour segmentation and skeleton recognition
He et al. A robust method for wheatear detection using UAV in natural scenes
CN107330360A (en) A kind of pedestrian's clothing colour recognition, pedestrian retrieval method and device
CN104298974A (en) Human body behavior recognition method based on depth video sequence
CN109740609A (en) A kind of gauge detection method and device
CN113312973B (en) Gesture recognition key point feature extraction method and system
CN103902989A (en) Human body motion video recognition method based on non-negative matrix factorization
CN106971158A (en) A kind of pedestrian detection method based on CoLBP symbiosis feature Yu GSS features
CN110490055A (en) A kind of Weakly supervised Activity recognition localization method and device recoded based on three
Yu et al. Robust median filtering forensics by CNN-based multiple residuals learning
CN109117774A (en) A kind of multi-angle video method for detecting abnormality based on sparse coding
CN110084201A (en) A kind of human motion recognition method of convolutional neural networks based on specific objective tracking under monitoring scene
CN105631405B (en) Traffic video intelligent recognition background modeling method based on Multilevel Block
CN113837154B (en) Open set filtering system and method based on multitask assistance

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant