CN104166841A - Rapid detection identification method for specified pedestrian or vehicle in video monitoring network - Google Patents

Rapid detection identification method for specified pedestrian or vehicle in video monitoring network Download PDF

Info

Publication number
CN104166841A
CN104166841A CN201410356465.0A CN201410356465A CN104166841A CN 104166841 A CN104166841 A CN 104166841A CN 201410356465 A CN201410356465 A CN 201410356465A CN 104166841 A CN104166841 A CN 104166841A
Authority
CN
China
Prior art keywords
image
pixel
histogram
edge
pedestrian
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410356465.0A
Other languages
Chinese (zh)
Other versions
CN104166841B (en
Inventor
于慧敏
谢奕
郑伟伟
汪东旭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201410356465.0A priority Critical patent/CN104166841B/en
Publication of CN104166841A publication Critical patent/CN104166841A/en
Application granted granted Critical
Publication of CN104166841B publication Critical patent/CN104166841B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention relates to a rapid detection identification method for a specified pedestrian or vehicle in a video monitoring network. According to a target template provided by a user, accurate and rapid detection identification is carried out on a specified target in a monitoring network video. In the method, firstly, normalized scale change is carried out on a target template image and a plurality of kinds of mixed characteristics of the template image are calculated; a mixed Gaussian model is adopted to carry out background modeling on each monitoring video, and then parallel processing is carried out and area filtering and morphological postprocessing are used to extract motion foregrounds; normalized scale change is carried out on all motion foregrounds which meet a preliminary screening condition and a plurality of mixed characteristics are calculated and then weighted similarity distances between the motion foregrounds and the target template are calculated and first N objects, which are smaller than a judgment threshold and have smallest weighted similarity distances with the target template image, are returned and used as detection identification results. On the basis of ensuring the detection identification accuracy, the method significantly improves the processing speed of an algorithm through methods of parallel processing and frame skipping and the like.

Description

In a kind of video surveillance network, specify the fast detecting recognition methods of pedestrian or vehicle
Technical field
The present invention relates to technical field of video image processing, be specially appointment pedestrian in a kind of video surveillance network or the fast detecting recognition methods of vehicle.
Background technology
In field of intelligent monitoring, the appointment pedestrian in monitor video network or vehicle are carried out to detection and Identification and can help public security organ to determine fast time of occurrence and the position of suspect or suspected vehicles, accelerate case investigation efficiency.But owing to existing in practical application scene, illumination condition changes, exists between different monitoring camera parameter differences, moving object (pedestrian, vehicle etc.) exists attitude change and mutually block, and pedestrian or the vehicle size in monitor video is conventionally less, such as the detailed information such as face or car plate can not be guaranteed, therefore how in multi-cam monitor network fast and accurately detection and Identification to go out to specify pedestrian or vehicle be focus and the difficult point of computer vision field research always.
Oreifej is published in technology " Human identity recognition in aerial images " (pedestrian's identification in Aerial Images) article of delivering on 709 to 716 pages of " IEEE Comptuer Society Conference on Computer Vision and Pattern Recognition " (computer society of international IEEE computer vision and pattern-recognition meeting) technology collection and has proposed a kind of taking photo by plane across the pedestrian's recognition methods under video complex background based on vote by ballot in 2010, first this technology utilizes the pedestrian region in HoG (Histograms of Oriented Gradient gradient orientation histogram) feature detection video, form the pedestrian's set occurring in video sequence, be called candidate's set, candidate is voted by characteristic similarity for target pedestrian's characteristic set of training by pre-prepd, the highest pedestrian of gained vote is suspicious object.This algorithm preferably resolves the target pedestrian's identification problem under the complex environments such as image resolution ratio is lower, pedestrian's attitude is changeable.But the method be because will be in video image multiple dimensioned scanning repeatedly detect pedestrian region one by one, and need multiple To Template images to gather as voter, therefore Shortcomings all in speed and practical feasibility.
Because monitor video has fixed viewpoint conventionally, the method of cutting apart sport foreground by background modeling can effectively be implemented, than utilizing HoG feature, video image is carried out to multiple dimensioned scanning differentiation, these class methods are cut apart fast and efficiently the detection of moving object under fixed viewpoint scene, but the method will be no longer applicable in the time that camera existence is mobile.
Reducing training link is also the difficult point in target detection identification, although some detection recognizer recognition accuracies based on artificial neural network and support vector machine are high at present, but calculated amount is larger, and need to provide in advance a large amount of To Template images for training, the field using value less in To Template information such as criminal investigations is not high.
Said method respectively has relative merits, how for application-specific scene, said method to be carried out to effective combination, and current research is also insufficient.This impels finds a kind of rational method framework that is applicable to criminal investigation work in monitor network, and in ensuring method accuracy rate, raising method speed, reaches real-time processing.
Summary of the invention
The invention provides the fast detecting recognition methods of specifying pedestrian or vehicle in a kind of video surveillance network, slower to solve prior art detection speed, and mostly need the problem of training in advance under the enterprising line of training set.
For achieving the above object, the present invention adopts the technical scheme of following meter: in a kind of video surveillance network, specify the fast detecting recognition methods of pedestrian or vehicle, comprise the following steps:
Step 1: the mode of first demarcating by rectangle frame specifies picture or the video center of pedestrian or vehicle to get identification target to be detected from comprising;
Step 2: selected To Template image is normalized to change of scale, and calculates the multiple composite character of To Template image;
Step 3: offer independently thread for the monitor video that each camera is taken, to each camera parallel processing in monitor network, utilize gauss hybrid models to carry out respectively background modeling to the residing monitoring scene of each camera;
Step 4: utilize area filtering and morphology aftertreatment to extract the sport foreground target in each camera; Carry out the too small moving target of filtering area by setting moving target area threshold, to filter out other moving object that does not belong to pedestrian or vehicle in scene;
Step 5: each monitor video frame-skipping is extracted to sport foreground, all moving object that meets rough handling requirement in video surveillance network is normalized change of scale and calculates multiple composite character, be weighted similarity measurement with To Template image afterwards;
Step 6: according to the similarity size of template image, the information of the detected moving object in each monitor video in monitor network is dynamically put into the result vector V of designated length result, when after all monitor videos of traversal, return to this result vector as final detection recognition result.
Further, the multiple composite character described in step 2 is main color characteristic, block margin direction histogram feature, piecemeal HSV histogram feature, HoG feature and invariable rotary LBP feature of equal value.Its specific implementation process is:
Step 2.1: wherein the generative process of main color characteristic is:
1. the To Template image after Normalized Scale conversion is transformed into HSV space from rgb space, only extracts the tone component of image in HSV space;
2. the span of tone component is divided into 8 intervals, the tone component of To Template image is become to 8 dimension H histogram of component chist according to these 8 interval projections, wherein i of chist interval value chist iobtained by following formula:
chist i = Σ ( x , y ) ∈ Rect δ i ( h x , y )
Wherein h x,ythe tone component value that refers to image (x, y) coordinate position pixel, Rect refers to the foreground area that To Template image or sport foreground detect, δ i(h x,y) be defined as follows:
δ i ( h x , y ) = 1 , h x , y ∈ [ 45 * i , 45 * ( i + 1 ) ) 0 , otherwise
3. calculate respectively 8 average color tone pitch c between tone zone i, and 8 dimension histograms are normalized, obtain from 8 dimension normalization histograms, choose p i3 maximum dimensions, use vector v dcpreserve the corresponding c of this 3 dimension iand p iform the main color characteristic of this target.Therefore main color characteristic v dcformed altogether corresponding tone value and the percentage v of 3 color components of accounting maximum in target image by 6 dimensions dc=[c 1, p 1, c 2, p 2, c 3, p 3];
Step 2.2: wherein the generative process of block margin direction histogram is:
1. first the target image after Normalized Scale conversion is converted into gray level image by RGB coloured image, and is divided into 4 × 4 totally 16;
2. utilize Sobel horizontal edge detection operator and Sobel vertical edge to detect operator gray level image is carried out to filtering, obtain horizontal edge intensity and the vertical edge intensity of each pixel, and utilize them to obtain the edge direction e of each pixel x,yand edge strength | e x,y|;
3. pixel is divided into five types according to edge direction and intensity: horizontal edge, vertical edge, 45 ° of edges, 135 ° of edges and directionless edge; Utilize four kinds of edges above to build edge orientation histogram;
4. set directionless edge threshold T e, in each piecemeal, all edge strengths are greater than to T epixel carry out statistics with histogram according to edge direction, generate 4 dimension edge orientation histograms; 4 × 4 piecemeal symbiosis become 64 dimension edge orientation histograms, and this 64 dimensional vector is saved as to local edge direction feature ehist l; Again the dimension of 4 in 16 piecemeals edge orientation histogram is added up, form the overall edge direction characteristic ehist of 4 dimension of view picture target image g; They are merged into 68 dimensional vectors, save as block margin direction histogram feature v ed;
Step 2.3: wherein the histogrammic generative process of piecemeal HSV is:
1. To Template image or sport foreground image after Normalized Scale conversion are projected to HSV color space by rgb color space, and by image from top to bottom level be divided into 4;
2. calculate respectively 4 HSV histograms in piecemeal, wherein tone component is divided into 16 intervals, and saturation degree component and luminance component are respectively divided into 4 intervals, and each piecemeal finally obtains 24 dimension histogram features; 4 24 dimension histograms are spliced to form to 96 of entire image and tie up piecemeal HSV histograms;
Step 2.4: wherein the generative process of HoG feature is:
1. the To Template image after Normalized Scale conversion is converted into gray level image, and gray level image is carried out to Gamma correction, reduce the impact on characteristic extraction procedure of local shade and illumination variation;
2. use [1,0,1] gradient operator to do convolution algorithm to image, obtain the horizontal gradient G of each pixel x(x, y), then use [1,0 ,-1] tgradient operator is done convolution algorithm to image, obtains the VG (vertical gradient) G of each pixel y(x, y); And utilize horizontal gradient and VG (vertical gradient) compute gradient amplitude and gradient direction:
G ( x , y ) = G x ( x , y ) 2 + G y ( x , y ) 2
dir ( x , y ) = tan - 1 ( G y ( x , y ) G x ( x , y ) )
3. image is divided into the cell unit of multiple 8 × 8 pixels, gradient direction is on average divided into 9 intervals by span, according to the gradient direction of each pixel in cell unit, generate 9 dimension histogram of gradients;
4. set every 4 adjacent piecemeals of cell unit composition, each piecemeal has 16 × 16 pixels, and a point interblock can be overlapped; By 9 dimension histogram of gradients series connection of 4 cell unit in each piecemeal, obtain the piecemeal description vectors of 36 dimensions;
5. set 64 × 64 window, window is vertically slided from top to bottom, slip is spaced apart 8 pixels, the description vectors of all piecemeals that window is comprised is connected, form the description vectors of whole window, again the description vectors of all windows is connected, finally obtain HoG proper vector v hog;
Step 2.5: wherein the generative process of invariable rotary LBP feature of equal value is:
1. To Template image or sport foreground image after Normalized Scale conversion are converted into gray level image, for each pixel in figure, relatively itself and the gray-scale value size of 8 neighbor pixels around, neighbours put gray-scale value and are greater than this pixel and are set to 1, otherwise are set to 0; In the direction of the clock 8 numeric strings are unified into one 82 system numbers since 12 o ' clock positions again, computing formula is as follows:
LBP P , R = Σ y = 0 P - 1 s ( g y - g c ) 2 y
s ( x ) = 1 , x &GreaterEqual; 0 0 , x < 0
Wherein, P is neighbours' number, and R is radius size, g yand g cbe respectively the gray-scale value of neighbor pixel point and the gray-scale value of center pixel;
2. each 8 the 2 system numbers that obtain are joined end to end, form an annular, this annular is turned clockwise 7 times, can obtain altogether 8 groups 82 system numbers, choose 8 group of 2 system and count 2 system numbers of intermediate value minimum, this number is the invariable rotary LBP value that pixel is corresponding;
3. the invariable rotary LBP value 2. being obtained by step is divided into two classes, wherein no more than 2 times of 0-1 conversion times be classified as a class, be called consistent LBP operator; Remaining 2 system number is all classified as another kind of; By above-mentioned classification, consistent invariable rotary LBP operator has 9 kinds, and non-uniform LBP operator has a kind;
4. image is divided into the fritter of multiple 16 × 16 pixels, calculates pixel LBP value in each fritter, form the invariable rotary LBP histogram of equal value of one 10 dimension; The LBP histogram of all fritters is connected, finally obtain the invariable rotary LBP feature v of equal value of 840 dimensions lBP;
Further, the foreground target in each camera is extracted in utilize area filtering and morphology aftertreatment described in step 4, and its specific implementation comprises following sub-step:
Step 4.1: mixed Gaussian background modeling and thresholding binary image after treatment are carried out to an area filtering;
Step 4.2: binary image is carried out to 3 expansive workings by 3 × 3 templates;
Step 4.3: use again 3 × 3 templates to carry out 1 corrosion operation to binary image;
Step 4.4: binary image is finally carried out to an area filtering again.
Area filtering described in step 4 refers to the fritter that binary image is divided into multiple 4 × 4 pixels, if the number of foreground pixel point is less than or equal at 3 o'clock in certain fritter, all pixels in this fritter are defined as to background pixel point, otherwise retain the foreground pixel point in this fritter;
Passing through described in step 4 set moving target area threshold and carrys out the too small moving target of filtering area, its specific implementation process is: find out and process all connected regions in rear binary image, calculate the number of each connected region foreground pixel point, set area threshold Th cfor 1/400 of video frame images area, if the area of connected region is greater than Th cretain connected region, and return to the minimum rectangle frame that comprises this connected region as sport foreground region; If connected region area is less than or equal to Th cthis connected region is made as to background area.
Further, described in step 5, each monitor video frame-skipping is extracted to sport foreground, its specific implementation process is: F is counted in the frame-skipping of setting according to user, again the sport foreground object extracting in step 4 is weighted to similarity measurement with To Template image one by one every F frame;
Weighting similarity measurement described in step 5, its specific implementation process is: the main color characteristic distance B of first calculating respectively sport foreground object and To Template image dc, block margin direction histogram characteristic distance D ed, piecemeal HSV histogram feature distance B hsv, HoG characteristic distance D hogwith invariable rotary LBP characteristic distance D of equal value lBP, then the distance of five kinds of features is normalized to weighting similarity measurement, obtain weighting similarity distance D all:
Wherein α, beta, gamma, λ, for normalization weighting coefficient, the impact for the each feature of balance on global similarity tolerance, tests to adjust according to specific environment in advance.
Further, the result vector V that the information of the detected moving object in each monitor video in monitor network is dynamically put into designated length described in step 6 result, its specific implementation comprises following sub-step:
Step 6.1: create a vectorial V with designated length N result, wherein N be in the monitor video network that need to return with specify pedestrian or the immediate object number of car modal image;
Step 6.2: when there being new sport foreground to be detected in monitor video network, first judge V resultthe object number of middle preservation;
Step 6.3: if V resultthe object number of middle preservation is less than N, as the weighting similarity distance D of object and To Template image allmeet predefined similarity threshold value Th alltime, by the camera of living in numbering of this object, there is frame number, occur that the information such as subject image after coordinate position and the Normalized Scale conversion of rectangle frame in this frame are retained in vectorial V resultin;
Step 6.4: if V resultthe object number of middle preservation has reached N, and works as the weighting similarity distance D of object and To Template image allmeet predefined similarity threshold value Th alltime, the weighting similarity distance D obtaining with this object allwith V resultin the weighting similarity distance that obtains apart from maximum object with To Template compare, if substitute V with this object resultin with To Template apart from maximum object, and by V resultthe N of a middle preservation object is according to resequencing from small to large with To Template similarity distance;
Step 6.5: after the monitor video that in traversal monitor network, all cameras are taken, the V returning resultin vector in store whole monitor network with specify pedestrian or the immediate N of a car modal image object, and comprise they separately residing camera numbering, the frame number occurring, there is the object image information after the coordinate position of rectangle frame in this frame and Normalized Scale convert.
Principle of the present invention is, pedestrian in monitor video network or vehicle must exist the process of motion just can enter and leave monitoring range in certain time period, and monitoring camera all has fixing visual angle, therefore if to specifying pedestrian or vehicle to carry out fast detecting identification, only need carry out respectively background modeling to each camera in network, extract the foreground object moving in each camera, after preliminary screening, each camera is offered separately to thread, carry out many characteristic weighings with the moving object after template image and the primary election of appointment pedestrian or vehicle and merge similarity measurements, again the similarity result in each camera is gathered to sequence, can obtain top n moving object the most similar to specifying pedestrian or vehicle in whole monitor network, as the result of fast detecting identification, because the sport foreground comprising in adjacent several frames in monitor video does not generally have too large variation, therefore the present invention can carry out frame-skipping processing in the process of background modeling and similarity measurement, carry out sport foreground similarity measurement one time every F frame, with the arithmetic speed of this further quickening system.
The present invention has the following advantages and good effect:
1) compared with prior art, the present invention has utilized monitoring camera visual angle to fix with pedestrian or vehicle must there is this two dot characteristics of motion process, adopt the method for Analysis on Prospect tentatively to determine target detection scope, compare multiple dimensioned searching algorithm, can effectively dwindle the target zone that detects identification, improve the efficiency that detects identification;
2) the present invention fully utilizes various features, considers to specify pedestrian or vehicle and each sport foreground in many-sided similarity degree such as color, shape, textures simultaneously, and multi-aspect information is weighted to fusion;
3) the present invention only needs simple target template as input, does not need to provide the database of detected target for training, has greatly increased the method feasibility in actual applications;
4) the present invention has introduced multi-cam parallel processing and frame-skipping processing scheduling algorithm, specifies the detection of pedestrian or vehicle to identify, the operation time that can effectively reduce again method thereby the present invention had both accurately been completed in complex scene robust.
Brief description of the drawings
Fig. 1 is overall flow schematic diagram of the present invention;
Fig. 2 (a) is monitoring scene schematic diagram of the present invention;
Fig. 2 (b) is mixed Gaussian background modeling result figure of the present invention;
Fig. 2 (c) is morphology aftertreatment result figure of the present invention;
Fig. 2 (d) is cut apart net result figure for sport foreground of the present invention;
Fig. 3 is intended target template image schematic diagram of the present invention;
Fig. 4 is intended target fast detecting recognition result figure of the present invention;
Embodiment
Below in conjunction with specific embodiment, further set forth this invention.The present embodiment is implemented under taking technical solution of the present invention as prerequisite, provided detailed embodiment and concrete operating process, but protection scope of the present invention is not limited to following embodiment.
Embodiment
The selected software environment of the present invention: Windows7.0 operating system, VS2010 development platform, C/C++ development language, OpenCV2.3.1 kit.
The present embodiment is processed certain monitor network.In this monitor network, video scene is comparatively complicated, has a large amount of moving object, and has mutually and block between moving object, between each camera, also due to parameter and position relationship, exists illumination variation and aberration.
Ask for an interview Fig. 1, the technical solution adopted in the present invention is: the appointment pedestrian in a kind of video surveillance network or the fast detecting recognition methods of vehicle, comprise the following steps:
Step 1: the mode of first demarcating by rectangle frame specifies picture or the video center of pedestrian or vehicle to get identification target to be detected from comprising;
User is provided and is comprised picture or the monitor video data of specifying pedestrian or vehicle by man-machine interface, the method of utilizing rectangle frame to demarcate, determine pedestrian or the residing position of vehicle from this picture or video frame images center fetching, the region that rectangle frame is comprised is divided into separately piece image.
Step 2: selected To Template image is normalized to change of scale, and calculates the multiple composite character of To Template image;
Be 64 × 128 pixel sizes by To Template image normalization change of scale, calculate respectively main color characteristic, block margin direction histogram feature, piecemeal HSV histogram feature, HoG feature and the invariable rotary LBP feature of equal value of To Template image, appointment pedestrian or vehicle are expressed from many-sides such as color, shape, textures.
Step 2.1: wherein the generative process of main color characteristic is:
1. the To Template image after Normalized Scale conversion is transformed into HSV space from rgb space, only extracts the tone component of image in HSV space; Main color characteristic only adopts tone component to calculate to be equivalent to the color composition of only considering target, thereby makes illumination variation and the saturation difference of feature to different cameras insensitive, with the robustness of this Enhanced feature.
2. the span of tone component is divided into 8 intervals, the tone component of To Template image is become to 8 dimension H histogram of component chist according to these 8 interval projections, wherein i of chist interval value chist iobtained by following formula:
chist i = &Sigma; ( x , y ) &Element; Rect &delta; i ( h x , y )
Wherein h x,ythe tone component value that refers to image (x, y) coordinate position pixel, Rect refers to the foreground area that To Template image or sport foreground detect.δ i(h x,y) be defined as follows:
&delta; i ( h x , y ) = 1 , h x , y &Element; [ 45 * i , 45 * ( i + 1 ) ) 0 , otherwise
3. calculate respectively 8 average color tone pitch c between tone zone i, and 8 dimension histograms are normalized, obtain from 8 dimension normalization histograms, choose p i3 maximum dimensions, use vector v dcpreserve the corresponding c of this 3 dimension iand p iform the main color characteristic of this target.Therefore main color characteristic v dcformed altogether corresponding tone value and the percentage v of 3 color components of accounting maximum in target image by 6 dimensions dc=[c 1, p 1, c 2, p 2, c 3, p 3]
Given two main color feature vectors, the distance B between them dcbe defined as follows:
D dc ( v dc , v dc &prime; ) = &Sigma; i = 1 3 p i 2 + &Sigma; j = 1 3 p j &prime; 2 - &Sigma; i = 1 3 &Sigma; j = 1 3 2 a i , j p i p j &prime;
A in above formula i,jbe used for characterizing color component c iand c jsimilarity degree, it defines by following formula:
a i , j = 1 - d i , j d max , i f d i , j &le; T d 0 , ohterwise
T in this example d=25, d max=50.
Step 2.2: wherein the generative process of block margin direction histogram is:
1. first the target image after Normalized Scale conversion is converted into gray level image by RGB coloured image, and is divided into 4 × 4 totally 16.
2. utilize Sobel horizontal edge detection operator and Sobel vertical edge to detect operator gray level image is carried out to filtering, obtain horizontal edge intensity and the vertical edge intensity of each pixel, and utilize them to obtain the edge direction e of each pixel x,yand edge strength | e x,y|.
3. conventional main color characteristic is divided into five types by pixel according to edge direction and intensity: horizontal edge, vertical edge, 45 ° of edges, 135 ° of edges and directionless edge; Wherein directionless edge refers to that edge strength is less than the pixel of a certain specific threshold; Because directionless edge pixel point accounts for significant proportion conventionally in entire image, therefore only consider in the present invention that four kinds of edges above form edge orientation histogram.
4. set directionless edge threshold T e, T in this example e=100.In each piecemeal, all edge strengths are greater than to T epixel carry out statistics with histogram according to edge direction, generate 4 dimension edge orientation histograms.4 × 4 piecemeal symbiosis become 64 dimension edge orientation histograms, and this 64 dimensional vector is saved as to local edge direction feature ehist l; Again the dimension of 4 in 16 piecemeals edge orientation histogram is added up, form the overall edge direction characteristic ehist of 4 dimension of view picture target image g; They are merged into 68 dimensional vectors, save as block margin direction histogram feature v ed.
Given two block margin direction histogram features, the distance B between them edbe defined as:
D ed ( v ed , v ed &prime; ) = &Sigma; i = 1 64 | ehist i l - ehist i &prime; l | + 5 &times; &Sigma; i = 1 4 | ehist i g - ehist i &prime; g |
Than local component, the present invention has given larger weight to global component.
Step 2.3: wherein the histogrammic generative process of piecemeal HSV is:
1. To Template image or sport foreground image after Normalized Scale conversion are projected to HSV color space by rgb color space, and by image from top to bottom level be divided into 4.
2. calculate respectively 4 HSV histograms in piecemeal, wherein tone component is divided into 16 intervals, and saturation degree component and luminance component are respectively divided into 4 intervals, and therefore each piecemeal finally obtains 24 dimension histogram features.4 24 dimension histograms are spliced to form to 96 of entire image and tie up piecemeal HSV histograms.
In order to adapt to the application scenarios of vehicle and pedestrian detection identification, be in To Template or sport foreground image, to include some background element topmost and bottom, and center section is mainly the important information such as vehicle body or pedestrian's clothes, this example has been given larger weight to middle two-layer piecemeal in the time of the distance of measuring between two piecemeal HSV histogram features.
Given two piecemeal HSV histogram features, distance B between them hsvbe defined as follows:
D hsv(v hsv,v′ hsv)=0.8·d(hist 1,hist 1′)+1.2·d(hist 2,hist′ 2)+1.0·d(hist 3,hist 3′)+0.8·d(hist 4,hist′ 4)
Wherein d (hist, hist ') is defined as card side's distance in this example:
d ( hist , hist &prime; ) = &Sigma; i = 1 24 ( hist i - hist i &prime; ) 2 hist i + hist i &prime;
Step 2.4: wherein the generative process of HoG feature is:
1. the To Template image after Normalized Scale conversion is converted into gray level image, and gray level image is carried out to Gamma correction, reduce the impact on characteristic extraction procedure of local shade and illumination variation.
2. use [1,0,1] gradient operator to do convolution algorithm to image, obtain the horizontal gradient G of each pixel x(x, y), then use [1,0 ,-1] tgradient operator is done convolution algorithm to image, obtains the VG (vertical gradient) G of each pixel y(x, y); And utilize horizontal gradient and VG (vertical gradient) compute gradient amplitude and gradient direction:
G ( x , y ) = G x ( x , y ) 2 + G y ( x , y ) 2
dir ( x , y ) = tan - 1 ( G y ( x , y ) G x ( x , y ) )
3. image is divided into the cell unit of multiple 8 × 8 pixels, every width image will be divided into 128 cell unit in this example, gradient direction is on average divided into 9 intervals by span, according to the gradient direction of each pixel in cell unit, generates 9 dimension histogram of gradients.
4. set every 4 adjacent piecemeals of cell unit composition, each piecemeal has 16 × 16 pixels, and a point interblock can be overlapped; By 9 dimension histogram of gradients series connection of 4 cell unit in each piecemeal, obtain the piecemeal description vectors of 36 dimensions.
5. set 64 × 64 window, window is vertically slided from top to bottom, slip is spaced apart 8 pixels, the description vectors of all piecemeals that window is comprised is connected, form the description vectors of whole window, again the description vectors of all windows is connected, finally obtain HoG proper vector v hog; The HoG feature generating in the present embodiment has 15876 dimensions, and the distance metric between Hog proper vector adopts card side's distance:
D hog ( v hog , v hog &prime; ) = &Sigma; i = 1 15876 ( v hog ( i ) - v hog &prime; ( i ) ) 2 v hog ( i ) + v hog &prime; ( i )
Step 2.5: wherein the generative process of invariable rotary LBP feature of equal value is:
1. To Template image or sport foreground image after Normalized Scale conversion are converted into gray level image, for each pixel in figure, relatively itself and the gray-scale value size of 8 neighbor pixels around, neighbours put gray-scale value and are greater than this pixel and are set to 1, otherwise be set to 0, more in the direction of the clock 8 numeric strings be unified into one 82 system numbers since 12 o ' clock positions.Computing formula is as follows:
LBP P , R = &Sigma; y = 0 P - 1 s ( g y - g c ) 2 y
s ( x ) = 1 , x &GreaterEqual; 0 0 , x < 0
Wherein, P is neighbours' number, and in this example, being made as 8, R is radius size, is made as 1, g in this example yand g cbe respectively the gray-scale value of neighbor pixel point and the gray-scale value of center pixel.
2. each 8 the 2 system numbers that obtain are joined end to end, form an annular, this annular is turned clockwise 7 times, can obtain altogether 8 groups 82 system numbers, choose 8 group of 2 system and count 2 system numbers of intermediate value minimum, this number is the invariable rotary LBP value that pixel is corresponding; By this step, 2 possible systems are exported by 2 8be reduced to 36 kinds, and the LBP value obtaining has robustness to the rotation of image.
3. the invariable rotary LBP value 2. being obtained by step is divided into two classes, wherein no more than 2 times of 0-1 conversion times be classified as a class, be called consistent LBP operator; Remaining 2 system number is all classified as another kind of; By above-mentioned classification, consistent invariable rotary LBP operator has 9 kinds, and non-uniform LBP operator has a kind.
4. image is divided into the fritter of multiple 16 × 16 pixels, calculates pixel LBP value in each fritter, form the invariable rotary LBP histogram of equal value of one 10 dimension; The LBP histogram of all fritters is connected, finally obtain the invariable rotary LBP feature v of equal value of 840 dimensions lBP; Distance B between invariable rotary LBP feature of equal value lBPthe same card side's distance that adopts is measured:
D LBP ( v LBP , v LBP &prime; ) = &Sigma; i = 1 840 ( v LBP ( i ) - v LBP &prime; ( i ) ) 2 v LBP ( i ) + v LBP &prime; ( i )
Step 3: offer independently thread for the monitor video that each camera is taken, to each camera parallel processing in monitor network, utilize gauss hybrid models to carry out respectively background modeling to the residing monitoring scene of each camera;
Each camera in monitor network has fixing video scene, and in video scene, the gray-scale value of each pixel can be described with mixed Gauss model
P ( X t ) = &Sigma; k = 1 K w k t &CenterDot; &eta; ( X t , &mu; k t , &Sigma; k t )
Wherein, η is Gaussian probability-density function, and K is the number upper limit of Gaussian function in mixed Gauss model, and K is set as 5 in this example, with be respectively weight, average and the variance of k Gauss model in the t frame moment.
In the present invention, the mixed Gaussian background modeling process of each camera realizes parallel computation by offering separately thread.Utilize gauss hybrid models to carry out the concrete steps of background modeling as follows:
1. carry out the mixed Gauss model of the each pixel of initialization with the gray-scale value of the each pixel of monitor video the first frame.Now mixed Gauss model only has a Gaussian function to be initialised, and its average is the gray-scale value of current pixel, and variance is set to fixed value σ 2=30, Gauss's weights are made as 0.05.
2. in the time reading in a new two field picture, by Gaussian function weights order from large to small check each Gaussian function whether therewith the gray-scale value of pixel match.The difference that is grey scale pixel value and this Gaussian function average is no more than Th d=2.5 σ=13.69.If find the Gaussian function of coupling, jump to step 3..If this gray scale is not mated with any one Gaussian function, 1. reinitialize a new Gaussian function according to step.In the time there is the Gaussian function of no initializtion in mixture model, directly carry out initialization with new Gaussian function; In the time that K the Gaussian function of setting all used, replace the Gaussian function of weights minimum in current mixture model with this new Gaussian function.
3. when determining after the Gaussian function that current pixel gray-scale value is corresponding, need to upgrade weights, average and the variance of each Gaussian function having used in mixture model.The modeling of background and renewal need regular hour accumulation, specify this time length of window L=200.When the frame number reading in when video is less than 200, Gaussian function more new formula is:
w ^ k N + 1 = w ^ k N + 1 N + 1 ( p ^ ( &omega; k | X N + 1 ) - w ^ k N )
&mu; ^ k N + 1 = &mu; ^ k N + p ^ ( &omega; k | X N + 1 ) &Sigma; i = 1 N + 1 p ^ ( &omega; k | X i ) ( X N + 1 - &mu; ^ k N )
&Sigma; ^ k N + 1 = &Sigma; ^ k N + p ^ ( &omega; k | X N + 1 ) &Sigma; i = 1 N + 1 p ^ ( &omega; k | X i ) ( ( X N + 1 - &mu; ^ k N ) ( X N + 1 - &mu; ^ k N ) T - &Sigma; ^ k N )
Wherein, N is frame number, ω kfor recording the sequence number of k Gaussian function in weights descending sort. for two-valued function, work as ω kwhen identical with the Gaussian function sequence number of mating, its value is 1, otherwise is 0.
When after frame number overtime length of window L, more new formula is:
w ^ k N + 1 = w ^ k N + 1 L ( p ^ ( &omega; k | X N + 1 ) - w ^ k N )
&mu; ^ k N + 1 = &mu; ^ k N + 1 L ( p ^ ( &omega; k | X N + 1 ) X N + 1 w ^ k N + 1 - &mu; ^ k N )
&Sigma; ^ k N + 1 = &Sigma; ^ k N + 1 L ( p ^ ( &omega; k | X N + 1 ) ( X N + 1 - &mu; ^ k N ) ( X N + 1 - &mu; ^ k N ) T w ^ k N + 1 - &Sigma; ^ k N )
After renewal, then the weights of each Gaussian function in mixed Gauss model are normalized.
4. the descending sequence of weight after each Gaussian function being upgraded according to it, determines that weight addition is greater than Th w=0.7 front B Gaussian function is the Gaussian function of describing background.If the Gaussian function mating with current pixel comes front B position, judge that this pixel is background pixel, otherwise, be foreground pixel.
Step 4: utilize area filtering and morphology aftertreatment to extract the sport foreground target in each camera; Carry out the too small moving target of filtering area by setting moving target area threshold, to filter out other moving object that does not belong to pedestrian or vehicle in scene;
The binary image that mixed Gauss model is obtained carries out area filtering and morphology processing can be removed noise and hole, and its specific implementation comprises following sub-step:
Step 4.1: mixed Gaussian background modeling and thresholding binary image after treatment are carried out to an area filtering;
Step 4.2: binary image is carried out to 3 expansive workings by 3 × 3 templates;
Step 4.3: use again 3 × 3 templates to carry out 1 corrosion operation to binary image;
Step 4.4: binary image is finally carried out to an area filtering again.
Wherein area filtering refers to the fritter that binary image is divided into multiple 4 × 4 pixels, if the number of foreground pixel point is less than or equal at 3 o'clock in certain fritter, all pixels in this fritter are defined as to background pixel point, otherwise retain the foreground pixel point in this fritter;
Passing through described in step 4 set moving target area threshold and carrys out the too small moving target of filtering area, its specific implementation process is: find out and process all connected regions in rear binary image, calculate the number of each connected region foreground pixel point, set area threshold Th cfor 1/400 of video frame images area, if the area of connected region is greater than Th cretain connected region, and return to the minimum rectangle frame that comprises this connected region as sport foreground region; If connected region area is less than or equal to Th cthis connected region is made as to background area.
Step 5: each monitor video frame-skipping is extracted to sport foreground, all moving object that meets rough handling requirement in video surveillance network is normalized change of scale and calculates multiple composite character, be weighted similarity measurement with To Template image afterwards;
Because each second of the video sequence in monitor video network comprises 25 frames, therefore pedestrian or vehicle all can repeat in adjacent several frames, in order to accelerate algorithm speed, the present invention introduces frame-skipping algorithm, F is counted in the frame-skipping of setting according to user, again the sport foreground object extracting in step 4 is weighted to similarity measurement with To Template image one by one every F frame.In this example, set F=3.
Sport foreground image is normalized to change of scale, be converted into 64 × 128 pixel images, and calculate respectively main color characteristic, block margin direction histogram feature, piecemeal HSV histogram feature, HoG feature and the invariable rotary LBP feature of equal value of sport foreground according to the method in step 2.
For described weighting similarity measurement, its specific implementation process is: first utilize respectively the definition of distance metric between each feature to calculate the main color characteristic distance B of sport foreground object and To Template image dc, block margin direction histogram characteristic distance D ed, piecemeal HSV histogram feature distance B hsv, HoG characteristic distance D hogwith invariable rotary LBP characteristic distance D of equal value lBP, then the distance of five kinds of features is normalized to weighting similarity measurement, obtain weighting similarity distance D all:
Wherein α, beta, gamma, λ, for normalization weighting coefficient, the impact for the each feature of balance on global similarity tolerance, tests to adjust according to specific environment in advance.Their value is respectively 10,1 in this example, and 5,1.2 and 1.
Step 6: according to the similarity size of template image, the information of the detected moving object in each monitor video in monitor network is dynamically put into the result vector V of designated length result, when after all monitor videos of traversal, return to this result vector as final detection recognition result.
Further, the result vector V that the information of the detected moving object in each monitor video in monitor network is dynamically put into designated length described in step 6 result, its specific implementation comprises following sub-step:
Step 6.1: create a vectorial V with designated length N result, wherein N be in the monitor video network that need to return with specify pedestrian or the immediate object number of car modal image;
Step 6.2: when there being new sport foreground to be detected in monitor video network, first judge V resultthe object number of middle preservation;
Step 6.3: if V resultthe object number of middle preservation is less than N, as the weighting similarity distance D of object and To Template image allmeet predefined similarity threshold value Th alltime, by the camera of living in numbering of this object, there is frame number, occur that the information such as subject image after coordinate position and the Normalized Scale conversion of rectangle frame in this frame are retained in vectorial V resultin;
Step 6.4: if V resultthe object number of middle preservation has reached N, and works as the weighting similarity distance D of object and To Template image allmeet predefined similarity threshold value Th alltime, the weighting similarity distance D obtaining with this object allwith V resultin the weighting similarity distance that obtains apart from maximum object with To Template compare, if substitute V with this object resultin with To Template apart from maximum object, and by V resultthe N of a middle preservation object is according to resequencing from small to large with To Template similarity distance;
Step 6.5: after the monitor video that in traversal monitor network, all cameras are taken, the V returning resultin vector in store whole monitor network with specify pedestrian or the immediate N of a car modal image object, and comprise they separately residing camera numbering, the frame number occurring, there is the object image information after the coordinate position of rectangle frame in this frame and Normalized Scale convert.
N=4 in this example, and Th all=50.
Implementation result
According to above-mentioned steps, the video in monitor network is specified to fast detecting and the identification of pedestrian or vehicle.Fig. 2 has provided the result figure that sport foreground is cut apart each stage, can see that from Fig. 2 (c) sport foreground region is estimated comparatively accurately by mixed Gaussian background modeling, area filtering and morphology aftertreatment; In Fig. 2 (d), by by connected region filtering too small area, net result has accurately returned to the residing rectangle frame of all sport foreground objects in scene.
Fig. 3 is that in this example, certain specifies pedestrian's template image schematic diagram, and Fig. 4 the present invention is directed to the retrieval result figure that this template image returns.Can find out, 4 return results is all the correct coupling to To Template image, even if change has occurred for personage's attitude and background of living in, the present invention still can find correct result from complex scene.
In this example, all experiments all realize on common PC computing machine, COMPUTER PARAMETER: central processing unit Interl (R) Core (TM) i3-2100 CPU@3.10GHz, internal memory 4GB.The processing speed of monitor video is relevant with the dense degree of moving object in scene, the processing speed of average every frame is 29ms in this example, add frame-skipping operation and parallel processing, can carry out fast detecting identification to the appointment pedestrian in monitor network or vehicle in real time.

Claims (5)

1. a fast detecting recognition methods for the appointment pedestrian in video surveillance network or vehicle, is characterized in that, comprises the steps:
Step 1: the mode of first demarcating by rectangle frame specifies picture or the video center of pedestrian or vehicle to get identification target to be detected from comprising;
Step 2: selected To Template image is normalized to change of scale, and generates the multiple composite character of To Template image;
Step 3: offer independently thread for the monitor video that each camera is taken, to each camera parallel processing in monitor network, utilize gauss hybrid models to carry out respectively background modeling to the residing monitoring scene of each camera;
Step 4: utilize area filtering and morphology aftertreatment to extract the sport foreground target in each camera; Carry out the too small moving target of filtering area by setting moving target area threshold, to filter out other moving object that does not belong to pedestrian or vehicle in scene;
Step 5: each monitor video frame-skipping is extracted to sport foreground, all moving object that meets rough handling requirement in video surveillance network is normalized change of scale and calculates multiple composite character, be weighted similarity measurement with To Template image afterwards;
Step 6: according to the similarity size of template image, the information of the detected moving object in each monitor video in monitor network is dynamically put into the result vector V of designated length result, when after all monitor videos of traversal, return to this result vector as final detection recognition result.
2. the fast detecting recognition methods of the appointment pedestrian in video surveillance network according to claim 1 or vehicle, it is characterized in that: the multiple composite character described in step 2 is main color characteristic, block margin direction histogram feature, piecemeal HSV histogram feature, HoG feature and invariable rotary LBP feature of equal value, and its concrete generative process is:
Step 2.1: wherein the generative process of main color characteristic is:
1. the To Template image after Normalized Scale conversion is transformed into HSV space from rgb space, only extracts the tone component of image in HSV space;
2. the span of tone component is divided into 8 intervals, the tone component of To Template image is become to 8 dimension H histogram of component chist according to these 8 interval projections, wherein i of chist interval value chist iobtained by following formula:
Wherein h x,ythe tone component value that refers to image (x, y) coordinate position pixel, Rect refers to the foreground area that To Template image or sport foreground detect, δ i(h x,y) be defined as follows:
3. calculate respectively 8 average color tone pitch c between tone zone i, and 8 dimension histograms are normalized, obtain from 8 dimension normalization histograms, choose p i3 maximum dimensions, use vector v dcpreserve the corresponding c of this 3 dimension iand p iform the main color characteristic of this target.Therefore main color characteristic v dcformed altogether corresponding tone value and the percentage v of 3 color components of accounting maximum in target image by 6 dimensions dc=[c 1, p 1, c 2, p 2, c 3, p 3];
Step 2.2: wherein the generative process of block margin direction histogram is:
1. first the target image after Normalized Scale conversion is converted into gray level image by RGB coloured image, and is divided into 4 × 4 totally 16;
2. utilize Sobel horizontal edge detection operator and Sobel vertical edge to detect operator gray level image is carried out to filtering, obtain horizontal edge intensity and the vertical edge intensity of each pixel, and utilize them to obtain the edge direction e of each pixel x,yand edge strength | e x,y|;
3. pixel is divided into five types according to edge direction and intensity: horizontal edge, vertical edge, 45 ° of edges, 135 ° of edges and directionless edge; Utilize four kinds of edges above to build edge orientation histogram;
4. set directionless edge threshold T e, in each piecemeal, all edge strengths are greater than to T epixel carry out statistics with histogram according to edge direction, generate 4 dimension edge orientation histograms; 4 × 4 piecemeal symbiosis become 64 dimension edge orientation histograms, and this 64 dimensional vector is saved as to local edge direction feature ehist l; Again the dimension of 4 in 16 piecemeals edge orientation histogram is added up, form the overall edge direction characteristic ehist of 4 dimension of view picture target image g; They are merged into 68 dimensional vectors, save as block margin direction histogram feature v ed;
Step 2.3: wherein the histogrammic generative process of piecemeal HSV is:
1. To Template image or sport foreground image after Normalized Scale conversion are projected to HSV color space by rgb color space, and by image from top to bottom level be divided into 4;
2. calculate respectively 4 HSV histograms in piecemeal, wherein tone component is divided into 16 intervals, and saturation degree component and luminance component are respectively divided into 4 intervals, and each piecemeal finally obtains 24 dimension histogram features; 4 24 dimension histograms are spliced to form to 96 of entire image and tie up piecemeal HSV histograms;
Step 2.4: wherein the generative process of HoG feature is:
1. the To Template image after Normalized Scale conversion is converted into gray level image, and gray level image is carried out to Gamma correction, reduce the impact on characteristic extraction procedure of local shade and illumination variation;
2. use [1,0,1] gradient operator to do convolution algorithm to image, obtain the horizontal gradient G of each pixel x(x, y), then use [1,0 ,-1] tgradient operator is done convolution algorithm to image, obtains the VG (vertical gradient) G of each pixel y(x, y); And utilize horizontal gradient and VG (vertical gradient) compute gradient amplitude and gradient direction:
3. image is divided into the cell unit of multiple 8 × 8 pixels, gradient direction is on average divided into 9 intervals by span, according to the gradient direction of each pixel in cell unit, generate 9 dimension histogram of gradients;
4. set every 4 adjacent piecemeals of cell unit composition, each piecemeal has 16 × 16 pixels, and a point interblock can be overlapped; By 9 dimension histogram of gradients series connection of 4 cell unit in each piecemeal, obtain the piecemeal description vectors of 36 dimensions;
5. set 64 × 64 window, window is vertically slided from top to bottom, slip is spaced apart 8 pixels, the description vectors of all piecemeals that window is comprised is connected, form the description vectors of whole window, again the description vectors of all windows is connected, finally obtain HoG proper vector v hog;
Step 2.5: wherein the generative process of invariable rotary LBP feature of equal value is:
1. To Template image or sport foreground image after Normalized Scale conversion are converted into gray level image, for each pixel in figure, relatively itself and the gray-scale value size of 8 neighbor pixels around, neighbours put gray-scale value and are greater than this pixel and are set to 1, otherwise are set to 0; In the direction of the clock 8 numeric strings are unified into one 82 system numbers since 12 o ' clock positions again, computing formula is as follows:
Wherein, P is neighbours' number, and R is radius size, g yand g cbe respectively the gray-scale value of neighbor pixel point and the gray-scale value of center pixel;
2. each 8 the 2 system numbers that obtain are joined end to end, form an annular, this annular is turned clockwise 7 times, can obtain altogether 8 groups 82 system numbers, choose 8 group of 2 system and count 2 system numbers of intermediate value minimum, this number is the invariable rotary LBP value that pixel is corresponding;
3. the invariable rotary LBP value 2. being obtained by step is divided into two classes, wherein no more than 2 times of 0-1 conversion times be classified as a class, be called consistent LBP operator; Remaining 2 system number is all classified as another kind of; By above-mentioned classification, consistent invariable rotary LBP operator has 9 kinds, and non-uniform LBP operator has a kind;
4. image is divided into the fritter of multiple 16 × 16 pixels, calculates pixel LBP value in each fritter, form the invariable rotary LBP histogram of equal value of one 10 dimension; The LBP histogram of all fritters is connected, finally obtain the invariable rotary LBP feature v of equal value of 840 dimensions lBP.
3. the fast detecting recognition methods of the appointment pedestrian in video surveillance network according to claim 1 or vehicle, it is characterized in that: the foreground target in each camera is extracted in utilize area filtering and morphology aftertreatment described in step 4, and its specific implementation comprises following sub-step:
Step 4.1: mixed Gaussian background modeling and thresholding binary image after treatment are carried out to an area filtering;
Step 4.2: binary image is carried out to 3 expansive workings by 3 × 3 templates;
Step 4.3: use again 3 × 3 templates to carry out 1 corrosion operation to binary image;
Step 4.4: binary image is finally carried out to an area filtering again;
Area filtering described in step 4 refers to the fritter that binary image is divided into multiple 4 × 4 pixels, if the number of foreground pixel point is less than or equal at 3 o'clock in certain fritter, all pixels in this fritter are defined as to background pixel point, otherwise retain the foreground pixel point in this fritter;
Passing through described in step 4 set moving target area threshold and carrys out the too small moving target of filtering area, its specific implementation process is: find out and process all connected regions in rear binary image, calculate the number of each connected region foreground pixel point, set area threshold Th cfor 1/400 of video frame images area, if the area of connected region is greater than Th cretain connected region, and return to the minimum rectangle frame that comprises this connected region as sport foreground region; If connected region area is less than or equal to Th cthis connected region is made as to background area.
4. the fast detecting recognition methods of the appointment pedestrian in video surveillance network according to claim 1 or vehicle, it is characterized in that: described in step 5, each monitor video frame-skipping is extracted to sport foreground, its specific implementation process is: F is counted in the frame-skipping of setting according to user, again the sport foreground object extracting in step 4 is weighted to similarity measurement with To Template image one by one every F frame;
Weighting similarity measurement described in step 5, its specific implementation process is: the main color characteristic distance B of first calculating respectively sport foreground object and To Template image dc, block margin direction histogram characteristic distance D ed, piecemeal HSV histogram feature distance B hsv, HoG characteristic distance D hogwith invariable rotary LBP characteristic distance D of equal value lBP, then the distance of five kinds of features is normalized to weighting similarity measurement, obtain weighting similarity distance D all:
Wherein α, beta, gamma, λ, for normalization weighting coefficient, the impact for the each feature of balance on global similarity tolerance, tests to adjust according to specific environment in advance.
5. the fast detecting recognition methods of the appointment pedestrian in video surveillance network according to claim 1 or vehicle, is characterized in that: the result vector V that the information of the detected moving object in each monitor video in monitor network is dynamically put into designated length described in step 6 result, its specific implementation comprises following sub-step:
Step 6.1: create a vectorial V with designated length N result, wherein N be in the monitor video network that need to return with specify pedestrian or the immediate object number of car modal image;
Step 6.2: when there being new sport foreground to be detected in monitor video network, first judge V resultthe object number of middle preservation;
Step 6.3: if V resultthe object number of middle preservation is less than N, as the weighting similarity distance D of object and To Template image allmeet predefined similarity threshold value Th alltime, by the camera of living in numbering of this object, there is frame number, occur that the information such as subject image after coordinate position and the Normalized Scale conversion of rectangle frame in this frame are retained in vectorial V resultin;
Step 6.4: if V resultthe object number of middle preservation has reached N, and works as the weighting similarity distance D of object and To Template image allmeet predefined similarity threshold value Th alltime, the weighting similarity distance D obtaining with this object allwith V resultin the weighting similarity distance that obtains apart from maximum object with To Template compare, if substitute V with this object resultin with To Template apart from maximum object, and by V resultthe N of a middle preservation object is according to resequencing from small to large with To Template similarity distance;
Step 6.5: after the monitor video that in traversal monitor network, all cameras are taken, the V returning resultin vector in store whole monitor network with specify pedestrian or the immediate N of a car modal image object, and comprise they separately residing camera numbering, the frame number occurring, there is the object image information after the coordinate position of rectangle frame in this frame and Normalized Scale convert.
CN201410356465.0A 2014-07-24 2014-07-24 The quick detection recognition methods of pedestrian or vehicle is specified in a kind of video surveillance network Active CN104166841B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410356465.0A CN104166841B (en) 2014-07-24 2014-07-24 The quick detection recognition methods of pedestrian or vehicle is specified in a kind of video surveillance network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410356465.0A CN104166841B (en) 2014-07-24 2014-07-24 The quick detection recognition methods of pedestrian or vehicle is specified in a kind of video surveillance network

Publications (2)

Publication Number Publication Date
CN104166841A true CN104166841A (en) 2014-11-26
CN104166841B CN104166841B (en) 2017-06-23

Family

ID=51910645

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410356465.0A Active CN104166841B (en) 2014-07-24 2014-07-24 The quick detection recognition methods of pedestrian or vehicle is specified in a kind of video surveillance network

Country Status (1)

Country Link
CN (1) CN104166841B (en)

Cited By (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104463151A (en) * 2015-01-05 2015-03-25 北京邮电大学 Rapid target matching method based on fusion of main color histograms and spatial position information
CN104796664A (en) * 2015-03-26 2015-07-22 成都市斯达鑫辉视讯科技有限公司 Video monitoring device
CN104866616A (en) * 2015-06-07 2015-08-26 中科院成都信息技术股份有限公司 Method for searching monitor video target
CN105005773A (en) * 2015-07-24 2015-10-28 成都市高博汇科信息科技有限公司 Pedestrian detection method with integration of time domain information and spatial domain information
CN105447457A (en) * 2015-11-15 2016-03-30 安徽清新互联信息科技有限公司 License plate character identification method based on adaptive characteristic
CN105447448A (en) * 2015-11-12 2016-03-30 江南大学 Bayonet vehicle positioning method based on Gaussian color distance
CN105869148A (en) * 2016-03-24 2016-08-17 北京小米移动软件有限公司 Target detection method and device
CN105957093A (en) * 2016-06-07 2016-09-21 浙江树人大学 ATM retention detection method of texture discrimination optimization HOG operator
CN106295466A (en) * 2015-05-18 2017-01-04 佳能株式会社 Image processing method and device
CN106504264A (en) * 2016-10-27 2017-03-15 锐捷网络股份有限公司 Video foreground image extraction method and device
CN106559645A (en) * 2015-09-25 2017-04-05 杭州海康威视数字技术股份有限公司 Based on the monitoring method of video camera, system and device
CN106845359A (en) * 2016-12-26 2017-06-13 四川农业大学 Tunnel portal driving prompt apparatus and method based on infrared emission
CN107111756A (en) * 2015-01-14 2017-08-29 萨罗尼科斯贸易与服务人有限公司 Apparatus and method for the entity in detection image
CN107239780A (en) * 2017-04-29 2017-10-10 安徽慧视金瞳科技有限公司 A kind of image matching method of multiple features fusion
CN107292892A (en) * 2017-06-23 2017-10-24 北京奇艺世纪科技有限公司 The dividing method and device of video frame images
CN107578045A (en) * 2017-09-19 2018-01-12 北京工业大学 A kind of Underwater targets recognition based on machine vision
CN107645633A (en) * 2016-07-21 2018-01-30 三菱电机大楼技术服务株式会社 Image processing apparatus
CN107704797A (en) * 2017-08-08 2018-02-16 深圳市安软慧视科技有限公司 Real-time detection method and system and equipment based on pedestrian in security protection video and vehicle
CN107895376A (en) * 2017-12-11 2018-04-10 福州大学 Based on the solar panel recognition methods for improving Canny operators and contour area threshold value
CN107918767A (en) * 2017-11-27 2018-04-17 北京旷视科技有限公司 Object detection method, device, electronic equipment and computer-readable medium
CN108921083A (en) * 2018-06-28 2018-11-30 浙江工业大学 Illegal flowing street pedlar recognition methods based on deep learning target detection
CN109214996A (en) * 2018-08-29 2019-01-15 深圳市元征科技股份有限公司 A kind of image processing method and device
CN109389031A (en) * 2018-08-27 2019-02-26 浙江大丰实业股份有限公司 Performance personnel's automatic positioning mechanism
CN109635720A (en) * 2018-12-10 2019-04-16 江南大学 The illegal road occupying real-time detection method actively monitored based on video
CN110232314A (en) * 2019-04-28 2019-09-13 广东工业大学 A kind of image pedestrian's detection method based on improved Hog feature combination neural network
CN110263788A (en) * 2019-07-17 2019-09-20 广西盖德科技有限公司 Vehicle pass-through method for quickly identifying and system
CN110378341A (en) * 2019-07-24 2019-10-25 西南交通大学 A kind of binocular vision pedestrian distance detection method
CN110490253A (en) * 2019-08-19 2019-11-22 浙江工业大学 A kind of image similarity calculation method based on spatial color EMD
CN110516649A (en) * 2019-09-02 2019-11-29 南京微小宝信息技术有限公司 Alumnus's authentication method and system based on recognition of face
CN110517183A (en) * 2019-07-26 2019-11-29 电子科技大学 A kind of high-speed low-power-consumption image processor based on retinal mechanisms
CN110529788A (en) * 2018-10-30 2019-12-03 永康市道可道科技有限公司 The explosion-proof pendent lamp of automatic projection formula
CN110765926A (en) * 2019-10-21 2020-02-07 杭州易现先进科技有限公司 Drawing book identification method and device, electronic equipment and storage medium
CN110956200A (en) * 2019-11-05 2020-04-03 哈尔滨工程大学 Tire pattern similarity detection method
CN111062971A (en) * 2019-12-13 2020-04-24 深圳龙岗智能视听研究院 Cross-camera mud head vehicle tracking method based on deep learning multi-mode
CN111179302A (en) * 2019-12-27 2020-05-19 普联国际有限公司 Moving target detection method and device, storage medium and terminal equipment
CN111461242A (en) * 2020-04-08 2020-07-28 北京航天新风机械设备有限责任公司 Multi-material rapid comparison and matching method for production line
CN112149581A (en) * 2020-09-25 2020-12-29 北京嘀嘀无限科技发展有限公司 Working condition monitoring method, system, device, electronic equipment and readable storage medium
CN112232153A (en) * 2020-09-30 2021-01-15 广东职业技术学院 Method and system for acquiring track of target person
TWI716788B (en) * 2018-11-09 2021-01-21 大陸商北京嘀嘀無限科技發展有限公司 System and method for detecting in-vehicle conflicts
CN113255705A (en) * 2021-07-15 2021-08-13 广东省大湾区集成电路与系统应用研究院 Intelligent identification device and identification method for protective color animals
CN113807328A (en) * 2021-11-18 2021-12-17 济南和普威视光电技术有限公司 Target detection method, device and medium based on algorithm fusion
CN110110111B (en) * 2018-02-02 2021-12-31 兴业数字金融服务(上海)股份有限公司 Method and device for monitoring screen
CN117079219A (en) * 2023-10-08 2023-11-17 山东车拖车网络科技有限公司 Vehicle running condition monitoring method and device applied to trailer service

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100142809A1 (en) * 2008-12-08 2010-06-10 Electronics And Telecommunications Research Institute Method for detecting multi moving objects in high resolution image sequences and system thereof
CN102012213A (en) * 2010-08-31 2011-04-13 吉林大学 Method for measuring foreground height through single image
CN102592454A (en) * 2012-02-29 2012-07-18 北京航空航天大学 Intersection vehicle movement parameter measuring method based on detection of vehicle side face and road intersection line
CN102708367A (en) * 2012-03-11 2012-10-03 无锡派图半导体设备有限公司 Image identification method based on target contour features
CN103049751A (en) * 2013-01-24 2013-04-17 苏州大学 Improved weighting region matching high-altitude video pedestrian recognizing method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100142809A1 (en) * 2008-12-08 2010-06-10 Electronics And Telecommunications Research Institute Method for detecting multi moving objects in high resolution image sequences and system thereof
CN102012213A (en) * 2010-08-31 2011-04-13 吉林大学 Method for measuring foreground height through single image
CN102592454A (en) * 2012-02-29 2012-07-18 北京航空航天大学 Intersection vehicle movement parameter measuring method based on detection of vehicle side face and road intersection line
CN102708367A (en) * 2012-03-11 2012-10-03 无锡派图半导体设备有限公司 Image identification method based on target contour features
CN103049751A (en) * 2013-01-24 2013-04-17 苏州大学 Improved weighting region matching high-altitude video pedestrian recognizing method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
芮挺 等: "基于高斯混合建模的多尺度HOG行人头肩特征检测", 《山东科技大学学报(自然科学版)》 *

Cited By (61)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104463151A (en) * 2015-01-05 2015-03-25 北京邮电大学 Rapid target matching method based on fusion of main color histograms and spatial position information
CN107111756A (en) * 2015-01-14 2017-08-29 萨罗尼科斯贸易与服务人有限公司 Apparatus and method for the entity in detection image
CN104796664A (en) * 2015-03-26 2015-07-22 成都市斯达鑫辉视讯科技有限公司 Video monitoring device
CN106295466A (en) * 2015-05-18 2017-01-04 佳能株式会社 Image processing method and device
CN104866616A (en) * 2015-06-07 2015-08-26 中科院成都信息技术股份有限公司 Method for searching monitor video target
CN104866616B (en) * 2015-06-07 2019-01-22 中科院成都信息技术股份有限公司 Monitor video Target Searching Method
CN105005773A (en) * 2015-07-24 2015-10-28 成都市高博汇科信息科技有限公司 Pedestrian detection method with integration of time domain information and spatial domain information
CN106559645A (en) * 2015-09-25 2017-04-05 杭州海康威视数字技术股份有限公司 Based on the monitoring method of video camera, system and device
CN106559645B (en) * 2015-09-25 2020-01-17 杭州海康威视数字技术股份有限公司 Monitoring method, system and device based on camera
CN105447448A (en) * 2015-11-12 2016-03-30 江南大学 Bayonet vehicle positioning method based on Gaussian color distance
CN105447448B (en) * 2015-11-12 2019-01-11 江南大学 A kind of bayonet vehicle localization method based on Gauss color distance
CN105447457A (en) * 2015-11-15 2016-03-30 安徽清新互联信息科技有限公司 License plate character identification method based on adaptive characteristic
CN105869148A (en) * 2016-03-24 2016-08-17 北京小米移动软件有限公司 Target detection method and device
CN105957093B (en) * 2016-06-07 2018-12-25 浙江树人大学 A kind of texture distinguishes the ATM delay object detecting method of optimization HOG operator
CN105957093A (en) * 2016-06-07 2016-09-21 浙江树人大学 ATM retention detection method of texture discrimination optimization HOG operator
CN107645633A (en) * 2016-07-21 2018-01-30 三菱电机大楼技术服务株式会社 Image processing apparatus
CN106504264A (en) * 2016-10-27 2017-03-15 锐捷网络股份有限公司 Video foreground image extraction method and device
CN106504264B (en) * 2016-10-27 2019-09-20 锐捷网络股份有限公司 Video foreground image extraction method and device
CN106845359A (en) * 2016-12-26 2017-06-13 四川农业大学 Tunnel portal driving prompt apparatus and method based on infrared emission
CN107239780A (en) * 2017-04-29 2017-10-10 安徽慧视金瞳科技有限公司 A kind of image matching method of multiple features fusion
CN107292892A (en) * 2017-06-23 2017-10-24 北京奇艺世纪科技有限公司 The dividing method and device of video frame images
CN107292892B (en) * 2017-06-23 2020-04-21 北京奇艺世纪科技有限公司 Video frame image segmentation method and device
CN107704797A (en) * 2017-08-08 2018-02-16 深圳市安软慧视科技有限公司 Real-time detection method and system and equipment based on pedestrian in security protection video and vehicle
CN107704797B (en) * 2017-08-08 2020-06-23 深圳市安软慧视科技有限公司 Real-time detection method, system and equipment based on pedestrians and vehicles in security video
CN107578045A (en) * 2017-09-19 2018-01-12 北京工业大学 A kind of Underwater targets recognition based on machine vision
CN107918767B (en) * 2017-11-27 2019-11-08 北京旷视科技有限公司 Object detection method, device, electronic equipment and computer-readable medium
CN107918767A (en) * 2017-11-27 2018-04-17 北京旷视科技有限公司 Object detection method, device, electronic equipment and computer-readable medium
CN107895376A (en) * 2017-12-11 2018-04-10 福州大学 Based on the solar panel recognition methods for improving Canny operators and contour area threshold value
CN110110111B (en) * 2018-02-02 2021-12-31 兴业数字金融服务(上海)股份有限公司 Method and device for monitoring screen
CN108921083A (en) * 2018-06-28 2018-11-30 浙江工业大学 Illegal flowing street pedlar recognition methods based on deep learning target detection
CN109389031A (en) * 2018-08-27 2019-02-26 浙江大丰实业股份有限公司 Performance personnel's automatic positioning mechanism
CN109389031B (en) * 2018-08-27 2021-12-03 浙江大丰实业股份有限公司 Automatic positioning mechanism for performance personnel
CN109214996B (en) * 2018-08-29 2021-11-12 深圳市元征科技股份有限公司 Image processing method and device
CN109214996A (en) * 2018-08-29 2019-01-15 深圳市元征科技股份有限公司 A kind of image processing method and device
CN110529788A (en) * 2018-10-30 2019-12-03 永康市道可道科技有限公司 The explosion-proof pendent lamp of automatic projection formula
US10922564B2 (en) 2018-11-09 2021-02-16 Beijing Didi Infinity Technology And Development Co., Ltd. System and method for detecting in-vehicle conflicts
US11615545B2 (en) 2018-11-09 2023-03-28 Bejing Didi Infinity Technology And Development Co., Ltd. System and method for detecting in-vehicle conflicts
TWI716788B (en) * 2018-11-09 2021-01-21 大陸商北京嘀嘀無限科技發展有限公司 System and method for detecting in-vehicle conflicts
CN109635720A (en) * 2018-12-10 2019-04-16 江南大学 The illegal road occupying real-time detection method actively monitored based on video
CN110232314A (en) * 2019-04-28 2019-09-13 广东工业大学 A kind of image pedestrian's detection method based on improved Hog feature combination neural network
CN110263788B (en) * 2019-07-17 2021-07-30 广西盖德科技有限公司 Method and system for quickly identifying vehicle passing
CN110263788A (en) * 2019-07-17 2019-09-20 广西盖德科技有限公司 Vehicle pass-through method for quickly identifying and system
CN110378341A (en) * 2019-07-24 2019-10-25 西南交通大学 A kind of binocular vision pedestrian distance detection method
CN110517183A (en) * 2019-07-26 2019-11-29 电子科技大学 A kind of high-speed low-power-consumption image processor based on retinal mechanisms
CN110490253A (en) * 2019-08-19 2019-11-22 浙江工业大学 A kind of image similarity calculation method based on spatial color EMD
CN110516649A (en) * 2019-09-02 2019-11-29 南京微小宝信息技术有限公司 Alumnus's authentication method and system based on recognition of face
CN110516649B (en) * 2019-09-02 2023-08-22 南京微小宝信息技术有限公司 Face recognition-based alumni authentication method and system
CN110765926A (en) * 2019-10-21 2020-02-07 杭州易现先进科技有限公司 Drawing book identification method and device, electronic equipment and storage medium
CN110956200A (en) * 2019-11-05 2020-04-03 哈尔滨工程大学 Tire pattern similarity detection method
CN111062971B (en) * 2019-12-13 2023-09-19 深圳龙岗智能视听研究院 Deep learning multi-mode-based mud head vehicle tracking method crossing cameras
CN111062971A (en) * 2019-12-13 2020-04-24 深圳龙岗智能视听研究院 Cross-camera mud head vehicle tracking method based on deep learning multi-mode
CN111179302A (en) * 2019-12-27 2020-05-19 普联国际有限公司 Moving target detection method and device, storage medium and terminal equipment
CN111179302B (en) * 2019-12-27 2023-03-24 普联国际有限公司 Moving target detection method and device, storage medium and terminal equipment
CN111461242A (en) * 2020-04-08 2020-07-28 北京航天新风机械设备有限责任公司 Multi-material rapid comparison and matching method for production line
CN112149581A (en) * 2020-09-25 2020-12-29 北京嘀嘀无限科技发展有限公司 Working condition monitoring method, system, device, electronic equipment and readable storage medium
CN112232153A (en) * 2020-09-30 2021-01-15 广东职业技术学院 Method and system for acquiring track of target person
CN113255705B (en) * 2021-07-15 2021-11-05 广东省大湾区集成电路与系统应用研究院 Intelligent identification device for protective color animal
CN113255705A (en) * 2021-07-15 2021-08-13 广东省大湾区集成电路与系统应用研究院 Intelligent identification device and identification method for protective color animals
CN113807328A (en) * 2021-11-18 2021-12-17 济南和普威视光电技术有限公司 Target detection method, device and medium based on algorithm fusion
CN117079219A (en) * 2023-10-08 2023-11-17 山东车拖车网络科技有限公司 Vehicle running condition monitoring method and device applied to trailer service
CN117079219B (en) * 2023-10-08 2024-01-09 山东车拖车网络科技有限公司 Vehicle running condition monitoring method and device applied to trailer service

Also Published As

Publication number Publication date
CN104166841B (en) 2017-06-23

Similar Documents

Publication Publication Date Title
CN104166841A (en) Rapid detection identification method for specified pedestrian or vehicle in video monitoring network
Li et al. Automatic pavement crack detection by multi-scale image fusion
CN108875600A (en) A kind of information of vehicles detection and tracking method, apparatus and computer storage medium based on YOLO
CN102542289B (en) Pedestrian volume statistical method based on plurality of Gaussian counting models
CN105404886B (en) Characteristic model generation method and characteristic model generating means
Siriborvornratanakul An automatic road distress visual inspection system using an onboard in‐car camera
CN102867188B (en) Method for detecting seat state in meeting place based on cascade structure
CN104978567B (en) Vehicle checking method based on scene classification
CN105574550A (en) Vehicle identification method and device
CN106682586A (en) Method for real-time lane line detection based on vision under complex lighting conditions
CN103310194A (en) Method for detecting head and shoulders of pedestrian in video based on overhead pixel gradient direction
CN102521565A (en) Garment identification method and system for low-resolution video
CN104036284A (en) Adaboost algorithm based multi-scale pedestrian detection method
CN104517095B (en) A kind of number of people dividing method based on depth image
CN112560675B (en) Bird visual target detection method combining YOLO and rotation-fusion strategy
CN101142584A (en) Method for facial features detection
CN107730515A (en) Panoramic picture conspicuousness detection method with eye movement model is increased based on region
CN109711322A (en) A kind of people&#39;s vehicle separation method based on RFCN
CN105426903A (en) Cloud determination method and system for remote sensing satellite images
CN103577875A (en) CAD (computer-aided design) people counting method based on FAST (features from accelerated segment test)
CN102915433A (en) Character combination-based license plate positioning and identifying method
US8094971B2 (en) Method and system for automatically determining the orientation of a digital image
CN110287798B (en) Vector network pedestrian detection method based on feature modularization and context fusion
CN109784171A (en) Car damage identification method for screening images, device, readable storage medium storing program for executing and server
CN114049572A (en) Detection method for identifying small target

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant