CN105868734A - Power transmission line large-scale construction vehicle recognition method based on BOW image representation model - Google Patents

Power transmission line large-scale construction vehicle recognition method based on BOW image representation model Download PDF

Info

Publication number
CN105868734A
CN105868734A CN201610258089.0A CN201610258089A CN105868734A CN 105868734 A CN105868734 A CN 105868734A CN 201610258089 A CN201610258089 A CN 201610258089A CN 105868734 A CN105868734 A CN 105868734A
Authority
CN
China
Prior art keywords
image
value
pixel
dtri
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610258089.0A
Other languages
Chinese (zh)
Inventor
程伟华
袁杰
刘刚
赵琳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
State Grid Jiangsu Electric Power Co Ltd
Jiangsu Electric Power Information Technology Co Ltd
Original Assignee
State Grid Jiangsu Electric Power Co Ltd
Jiangsu Electric Power Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by State Grid Jiangsu Electric Power Co Ltd, Jiangsu Electric Power Information Technology Co Ltd filed Critical State Grid Jiangsu Electric Power Co Ltd
Priority to CN201610258089.0A priority Critical patent/CN105868734A/en
Publication of CN105868734A publication Critical patent/CN105868734A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/584Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights

Abstract

The invention discloses a power transmission line large-scale construction breakage-prevention large-scale construction vehicle recognition method based on a BOW image representation model. The method comprises the steps that median filtering is performed on a current image frame obtained by a camera; a Gaussian mixture modeling method is used for background modeling and foreground recognition; an intra-large-region texture-based method is used for eliminating shadows of a moving object; BOW model characteristics of each foreground region are extracted; the extracted characteristics are fed into a multi-classification SVM learned in advance for vehicle type recognition. According to the power transmission line large-scale construction breakage-prevention large-scale construction vehicle recognition method, the defect that other color-based methods can only detect vehicles of a certain specific color can be overcome, types of various large-scale construction vehicles can be detected, and the accuracy is high.

Description

Transmission line of electricity Large Construction vehicle identification method based on BOW characterization image model
Technical field
The invention belongs to transmission line of electricity external force damage prevention field, relate to a kind of transmission line of electricity Large Construction vehicle identification method based on BOW characterization image model.
Background technology
Whether the algorithm that on-line monitoring device for transmission line based on intelligent video analysis is used is identified being divided into two classes to detection target by it: a class only detects intrusion target and carries out unifying identifier, testing result is submitted to user and manually differentiates decision-making;The another kind of intrusion target that first detects, then uses the technology such as artificial intelligence, machine learning, pattern recognition that intrusion target is identified classification, the result of point good class is submitted to user.First kind method is normally only used for the situation that foreground moving object is few, and when prospect moving object is more, real effective target can be flooded by its a large amount of invalid targets detected.And technology is required higher by Equations of The Second Kind method.
First kind algorithm difficult point is that moving target detects.Moving object detection is divided can be divided three classes by principle: optical flow method, frame differential method and background subtraction method.Optical flow method is based on the dividing method scape relief method detecting the estimation of optical flow field.Optical flow method is that its amount of calculation is huge, and algorithm is the most sufficiently complex, poor real based on the dividing method detecting the estimation of optical flow field, to transmission line of electricity big machinery intrusion detection inapplicable.Frame differential method uses the Differential Detection between image sequence to determine that moving target, algorithm the most easily realize, but when target travel slowly or speed quickly time, the target of extraction can not be close to its true shape.Background subtraction method is the moving object detection algorithm of current main-stream, and it is that video image and background reference image are carried out difference, and foreground point is i.e. moving target.Its processing speed is fast, very strong to the adaptability of different scenes, can relatively accurately extract the true shape of target.
First Equations of The Second Kind algorithm is also required to carry out moving object detection, and then intrusion target is identified classification, and its difficult point essentially consists in target recognition, i.e. judges whether moving target is the Large Construction vehicle such as crane, cement pump truck.Document [realization of the intelligent early-warning function that anti-big machinery external force is destroyed in transmission line of electricity] uses background subtraction method to carry out background detection, and uses color as the characteristics of image of big machinery, it is modeled.Carry out arm detection according to arm region area, eccentricity and compactness, use HOUGH transformation calculations arm to stretch angle, send alarm according to the distance between transmission line of electricity.Document [Intelligent Measurement of moving target and identification in transmission line of electricity monitoring system] uses feature based on color to identify big machinery in foreground target, then location crane wheel, use area, eccentricity and 3 indexs of compactness as the input of grader characteristic vector, identify arm, and and then hazard recognition behavior.Method in this two documents all uses color characteristic to carry out target recognition, it is believed that the color of crane is yellow.In actual applications, when the crane of non-yellow occur, its effect is the most very poor.
Summary of the invention
For the problem overcoming prior art to exist, it is an object of the invention to provide a kind of transmission line of electricity Large Construction vehicle identification method based on BOW characterization image model, the method can detect the classification of various Large Construction vehicle, owing to using the feature insensitive to color, it can overcome the shortcoming that can only detect certain color vehicle specific that other method based on color exists, and accuracy is high.
The purpose of the present invention is achieved through the following technical solutions:
A kind of transmission line of electricity Large Construction vehicle identification method based on BOW characterization image model, comprises the following steps:
(1) current image frame obtaining photographic head carries out medium filtering;
(2) method of Gaussian modeling is used to carry out background modeling and prospect identification;
(3) in using big region, method based on texture eliminates the shade of moving target;
(4) the BOW aspect of model of each foreground area is extracted;
(5) the many classification SVM support vector machine feature feeding of extraction succeeded in school in advance carry out vehicle cab recognition.
Further, the described current image frame obtaining photographic head carries out median filtering step and is:
First image is become RGB color from YUV color space conversion, converts thereof into gray-scale map further, then for from the second row to row second from the bottom, from each pixel f of secondary series to row second from the bottom, (x, y), by 8 pixel f (x-1 around it and it, y-1), and f (x-1, y), f (x-1, y+1), f (x, y-1), f (x, y+1), f (x+1, y-1), f (x+1, y), f (x+1, y+1) is arranged from small to large by gray value, take the value of that number middle as f (x, value y).
The described method using Gaussian modeling carries out the step of background modeling and prospect identification:
(1) background modeling:
Setting mixed Gauss model to be made up of 5 Gaussian functions, first in model, the average of each Gaussian function, variance, weights are both configured to 0, i.e. initialization model matrix parameter.
30 frames in video are used to be used for training mixed Gauss model.For each pixel, set up its mixed Gauss model.When first pixel is come, fixing initial mean value, variance, and each weights are set for it and are disposed as 0.2.
During non-first frame training, when tail pixel value, with the average ratio of the most existing Gaussian function relatively, if the value of this pixel and the equal value difference of its model are in the variance of 3 times, then it is assumed that this point belongs to this Gaussian function.Now it is updated by equation below:
π ^ m ← π ^ m + α ( o m ( t ) - π ^ m ) ,
Wherein, It is 1.
When the difference of this pixel point value and average is not in its variance of 3 times, takeIt is 0.
When after the frame number 30 arriving training, carry out the different pixels point adaptive selection of mixed Gauss model number.First divided by variance, each Gauss is sorted from big to small with weights, then choose B, foremost Gauss so that it is meet, wherein
B = arg min b ( Σ m = 1 b π ^ m > ( 1 - c f ) ) ,
Then accumulation weight is normalized to 1.
(2) prospect identification: for the dynamic picture frame of complexity, superposition to multiple Gauss distribution of the different weights that each pixel is learnt by upper step models, when the value of current pixel point and the mean bias of modeling are in 2.5 σ, then it is assumed that this pixel belongs to background, otherwise belongs to prospect.
In described big region, the step of method based on texture elimination moving target shade is:
(41) use the preselected shadows pixels of following formula (x, y):
SP value is to represent when 1 that current pixel is shadows pixels, is to represent when 0 not to be.In above formula, α=0.8, β=0.94.It(x, y) V is t picture frame (x, y) value of the V component of position, Bt(x, y) V is t background image (x, y) value of the V component of position using Gaussian modeling to obtain;It(x, y) H is t picture frame (x, y) value of the H component of position, Bt(x, y) H is t moment background image (x, y) value of the H component of position using Gaussian modeling to obtain;It(x, y) S is t picture frame (x, y) value of the S component of position, Bt(x, y) S is background image (x, y) value of the S component of position using Gaussian modeling t to obtain.FPMTLIt it is foreground image areas flag bit.
(42) all connected domains of extraction are concentrated from preselected shadows pixels.
(43) for each connected domain, calculate each pixel p=(x, y) gradient direction at place and gradient-norm:
| ▿ p | = ▿ x 2 + ▿ y 2
θ p = a r c t a n 2 ( ▿ y / ▿ x )
WhereinFor vertical gradient, i.e. | f (x, y)-f (x+1, y) |,For horizontal direction gradient, i.e. | f (x, y)-f (x, y+1) |, (x, is y) that image is at point (x, y) gray value at place to f, f (x, y+1) is the image gray value at point (x, y+1) place;Wherein θpIn the range of [-π, π], the most only retain | θp| value more than a certain threshold taumThose pixels.
(44) for each pixel p remaining=(x, y), the gradient direction calculating present image and background image is poor
Δθ p = a r c c o s ▿ x F ▿ x B + ▿ y F ▿ y B { ( ▿ x F 2 + ▿ y F 2 ) ( ▿ x B 2 + ▿ y B 2 ) } 1 2
Represent in present image at pixel p the Grad of horizontal direction,Represent the Grad of horizontal direction at pixel p in background image,Represent the Grad of vertical direction at pixel p in present image,Represent the Grad of vertical direction at p in background image.
(45) present image and the gradient direction dependency of background image are estimated with following formula
c = { Σ p = 1 n H ( τ a - Δθ p ) } / n
Wherein n is the selected pixels number of candidate shadow region, and H () is unit progression function, and i.e. if differential seat angle is less than or equal to threshold tau a, then the value of function is 1, is otherwise 0.C is the ratio of pixel count similar with gradient direction in background image in present image in region.As c compares threshold taucGreatly, then it is assumed that candidate region is shadow region.
The step of the described BOW aspect of model extracting each foreground area is:
(51) local shape factor
Divide image with the uniform grid (grid) that the length of side is 8 pixels, the block of 4 grid protocol extracts SIFT and describes son.Each piece is the rectangle of 16 × 16 pixels, comprises 4 × 4 buckets, and each bucket is 4 × 4 pixels.Calculating the gradient information in 8 directions in each bucket, the most each piece is characterized by 4 × 4 × 8=128 dimensional vector.The step-length that block moves is 8 pixels.
(52) study visual dictionary
The Dense SIFT local feature using KMEANS method to go out all image zooming-out in training storehouse clusters.Each cluster centre is i.e. a visual vocabulary in dictionary, and all visual vocabularies form a visual dictionary.
(53) characteristic quantification coding and spatial pyramid
After obtaining visual dictionary, then each feature of image can be mapped on certain word in dictionary, then each visual word occurrence number on this image in statistics dictionary, thus obtains the histogram vectors of the phenogram picture of one 300 dimension.
Use multiple dimensioned method of partition, image is divided into three layers of pyramid of 1 × 1,2 × 2,4 × 4 space separatings, adds up the feature of each sub-block respectively, finally the merging features of all pieces is got up, form complete feature.So feature of piece image is 6300 dimensions.
(54) train classification models predicting
Image in all training sets is used above 1)-3) method in step extracts feature, uses SVM support vector machine to be trained the characteristic vector of training set, it is thus achieved that the disaggregated model of Large Construction vehicle;Kernel function form such as following formula:
k ( x , y ) = Σ k = 1 π m i n ( x k , y k )
The step carrying out vehicle cab recognition in the described many classification SVM support vector machine feature feeding of extraction succeeded in school in advance is:
For the foreground area after each removal shade detected, use the Dense SIFT feature of aforementioned extraction image, and carry out feature after quantization is quantified according to visual dictionary, the feature in region is obtained again through pyramid coding, send in the SVM classifier trained and classify, as it is categorized as crane, excavator or cement pump truck, then it is assumed that be to have potential outer broken possible Large Construction vehicle, otherwise it is assumed that not outer broken probability.
Compared with prior art, the present invention identifies transmission line of electricity Large Construction vehicle based on BOW characterization image model, use the feature insensitive to color, overcome the shortcoming that can only detect certain color vehicle specific that in prior art, method based on color exists, can detect the classification of various Large Construction vehicles, accuracy is high.
Accompanying drawing explanation
Fig. 1 is Large Construction vehicle identification method flow chart based on Computer Vision;
Detailed description of the invention
In order to be better understood from technical scheme, below in conjunction with accompanying drawing 1, the invention will be further described.
Fig. 1 describes a kind of transmission line of electricity external force damage prevention Large Construction vehicle identification method based on Computer Vision, comprises the following steps:
(1) current image frame obtaining photographic head carries out medium filtering;Step is as follows:
First image is become RGB color from YUV color space conversion, converts thereof into gray-scale map further, then for from the second row to row second from the bottom, from each pixel f of secondary series to row second from the bottom, (x, y), by 8 pixel f (x-1 around it and it, y-1), and f (x-1, y), f (x-1, y+1), f (x, y-1), f (x, y+1), f (x+1, y-1), f (x+1, y), f (x+1, y+1) is arranged from small to large by gray value, take the value of that number middle as f (x, value y).
(2) method of Gaussian modeling is used to carry out background modeling and prospect identification;Step is as follows:
1) background modeling:
Setting mixed Gauss model to be made up of 5 Gaussian functions, first in model, the average of each Gaussian function, variance, weights are both configured to 0, i.e. initialization model matrix parameter.
30 frames in video are used to be used for training mixed Gauss model.For each pixel, set up its mixed Gauss model.When first pixel is come, fixing initial mean value, variance, and each weights are set for it and are disposed as 0.2.
During non-first frame training, when tail pixel value, with the average ratio of the most existing Gaussian function relatively, if the value of this pixel and the equal value difference of its model are in the variance of 3 times, then it is assumed that this point belongs to this Gaussian function.Now it is updated by equation below:
π ^ m ← π ^ m + α ( o m ( t ) - π ^ m ) ,
Wherein, It is 1.
When the difference of this pixel point value and average is not in its variance of 3 times, takeIt is 0.
When after the frame number 30 arriving training, carry out the different pixels point adaptive selection of mixed Gauss model number.
First divided by variance, each Gauss is sorted from big to small with weights, then choose B, foremost Gauss so that it is meet, wherein
B = arg min b ( Σ m = 1 b π ^ m > ( 1 - c f ) ) ,
Then accumulation weight is normalized to 1.
2) prospect identification: for the dynamic picture frame of complexity, superposition to multiple Gauss distribution of the different weights that each pixel is learnt by upper step models, when the value of current pixel point and the mean bias of modeling are in 2.5 σ, then it is assumed that this pixel belongs to background, otherwise belongs to prospect.
(3) in using big region, method based on texture eliminates the shade of moving target;Step is as follows:
41) use the preselected shadows pixels of following formula (x, y):
SP value is to represent when 1 that current pixel is shadows pixels, is to represent when 0 not to be.In above formula, α=0.8, β=0.94.It(x, y) V is t picture frame (x, y) value of the V component of position, Bt(x, y) V is t background image (x, y) value of the V component of position using Gaussian modeling to obtain;It(x, y) H is t picture frame (x, y) value of the H component of position, Bt(x, y) H is t background image (x, y) value of the H component of position using Gaussian modeling to obtain;It(x, y) S is t picture frame (x, y) value of the S component of position, Bt(x, y) S is background image (x, y) value of the S component of position using Gaussian modeling t to obtain.FPMTLIt it is foreground image areas flag bit.
42) all connected domains of extraction are concentrated from preselected shadows pixels.
43) for each connected domain, calculate each pixel p=(x, y) gradient direction at place and gradient-norm:
| ▿ p | = ▿ x 2 + ▿ y 2
θ p = a r c t a n 2 ( ▿ y / ▿ x )
WhereinFor vertical gradient, i.e. | f (x, y)-f (x+1, y) |,For horizontal direction gradient, i.e. | f (x, y)-f (x, y+1) |, (x, is y) that image is at point (x, y) gray value at place to f, f (x, y+1) is the image gray value at point (x, y+1) place;Wherein θpIn the range of [-π, π], the most only retain | θp| value more than a certain threshold taumThose pixels.
44) for each pixel p remaining=(x, y), the gradient direction calculating present image and background image is poor
Δθ p = a r c c o s ▿ x F ▿ x B + ▿ y F ▿ y B { ( ▿ x F 2 + ▿ y F 2 ) ( ▿ x B 2 + ▿ y B 2 ) } 1 2
Represent in present image at pixel p the Grad of horizontal direction,Represent the Grad of horizontal direction at pixel p in background image,Represent the Grad of vertical direction at pixel p in present image,Represent the Grad of vertical direction at p in background image.
45) present image and the gradient direction dependency of background image are estimated with following formula
c = { Σ p = 1 n H ( τ a - Δθ p ) } / n
Wherein n is the selected pixels number of candidate shadow region, and H () is unit progression function, and i.e. if differential seat angle is less than or equal to threshold tau a, then the value of function is 1, is otherwise 0.C is the ratio of pixel count similar with gradient direction in background image in present image in region.As c is bigger than threshold tau c, then it is assumed that candidate region is shadow region.
(4) the BOW aspect of model of each foreground area is extracted;Step is as follows:
51) local shape factor
Divide image with the uniform grid (grid) that the length of side is 8 pixels, the block of 4 grid protocol extracts SIFT and describes son.Each piece is the rectangle of 16 × 16 pixels, comprises 4 × 4 buckets, and each bucket is 4 × 4 pixels.Calculating the gradient information in 8 directions in each bucket, the most each piece is characterized by 4 × 4 × 8=128 dimensional vector.The step-length that block moves is 8 pixels.
52) study visual dictionary
The Dense SIFT local feature using KMEANS method to go out all image zooming-out in training storehouse clusters.Each cluster centre is i.e. a visual vocabulary in dictionary, and all visual vocabularies form a visual dictionary.
53) characteristic quantification coding and spatial pyramid
After obtaining visual dictionary, then each feature of image can be mapped on certain word in dictionary, then each visual word occurrence number on this image in statistics dictionary, thus obtains the histogram vectors of the phenogram picture of one 300 dimension.
Use multiple dimensioned method of partition, image is divided into three layers of pyramid of 1 × 1,2 × 2,4 × 4 space separatings, adds up the feature of each sub-block respectively, finally the merging features of all pieces is got up, form complete feature.So feature of piece image is 6300 dimensions.
54) train classification models predicting
Image in all training sets is used above 1)-3) method in step extracts feature, uses SVM support vector machine to be trained the characteristic vector of training set, it is thus achieved that the disaggregated model of Large Construction vehicle;Kernel function form such as following formula:
k ( x , y ) = Σ k = 1 π m i n ( x k , y k )
(5) the many classification SVM support vector machine feature feeding of extraction succeeded in school in advance carry out vehicle cab recognition.Step is:
For the foreground area after each removal shade detected, use the Dense SIFT feature of aforementioned extraction image, and carry out feature after quantization is quantified according to visual dictionary, the feature in region is obtained again through pyramid coding, send in the SVM classifier trained and classify, as it is categorized as crane, excavator or cement pump truck, then it is assumed that be to have potential outer broken possible Large Construction vehicle, otherwise it is assumed that not outer broken probability.
Embodiment
The image obtained for monitoring camera, is handled as follows:
(1) current image frame obtaining monitoring camera carries out medium filtering.
(2) use the method for Gaussian modeling to carry out background modeling and prospect identification, then image is carried out shadow Detection elimination.
(3) the BOW aspect of model of each foreground area is extracted.
(4) the many classification SVM support vector machine feature feeding of extraction succeeded in school in advance carry out vehicle cab recognition, result marks all Large Construction vehicles.

Claims (6)

1. a transmission line of electricity Large Construction vehicle identification method based on BOW characterization image model, it is special Levy and be to comprise the following steps:
(1) current image frame obtaining photographic head carries out medium filtering;
(2) method of Gaussian modeling is used to carry out background modeling and prospect identification;
(3) in using big region, method based on texture eliminates the shade of moving target;
(4) the BOW aspect of model of each foreground area is extracted;
(5) the many classification SVM support vector machine feature feeding of extraction succeeded in school in advance carry out vehicle Identify.
Transmission line of electricity Large Construction vehicle based on BOW characterization image model the most according to claim 1 Recognition methods, it is characterised in that: the described current image frame obtaining photographic head carries out median filtering step and is:
First image is become RGB color from YUV color space conversion, convert thereof into ash further Degree figure, then for from the second row to row second from the bottom, from each pixel of secondary series to row second from the bottom (x, y), by its gray value f (x, the gray value of 8 pixels y) and around it (x-1, y), f (x-1, y+1), f (x, y-1), f (x, y+1), (x+1, y), f (x+1, y+1) is by ash for f (x+1, y-1), f for f (x-1, y-1), f Angle value arranges from small to large, takes the value of that number middle as pixel (x, gray value f (x, value y) y).
Transmission line of electricity Large Construction vehicle based on BOW characterization image model the most according to claim 1 Recognition methods, it is characterised in that: the described method using Gaussian modeling carries out background modeling and prospect is known Other step is:
(1) background modeling:
Set mixed Gauss model to be made up of 5 Gaussian functions, first the average of each Gaussian function in model, Variance, weights are both configured to 0, i.e. initialization model matrix parameter;
30 frames in video are used to be used for training mixed Gauss model;For each pixel, set up it and mix Close Gauss model, when first pixel is come, fixing initial mean value, variance, and each weights are set for it It is disposed as 0.2;
During non-first frame training, when tail pixel value, with the average of the most existing Gaussian function Relatively, if the value of this pixel and the equal value difference of its model are in the variance of 3 times, then it is assumed that this point belongs to this height This function, is now updated by equation below:
π ^ m ← π ^ m + α ( o m ( t ) - π ^ m ) ,
Wherein,It is 1;
When the difference of this pixel point value and average is not in its variance of 3 times, takeIt is 0;
When after the frame number 30 arriving training, carry out the different pixels point adaptive selection of mixed Gauss model number; First divided by variance, each Gauss is sorted from big to small with weights, then chooses B, foremost Gauss, It is made to meet, wherein
B = arg min b ( Σ m = 1 b π ^ m > ( 1 - c f ) ) ,
Then accumulation weight is normalized to 1;
2) prospect identification: for the dynamic picture frame of complexity, to each pixel by the different power of upper step study The superposition of multiple Gauss distribution of value models, when the value of current pixel point and the mean bias of modeling are at 2.5 σ Time interior, then it is assumed that this pixel belongs to background, otherwise belong to prospect.
Transmission line of electricity Large Construction vehicle based on BOW characterization image model the most according to claim 1 Recognition methods, it is characterised in that: in described big region, method based on texture eliminates the step of moving target shade Suddenly it is:
(41) use the preselected shadows pixels of following formula (x, y):
SP value is to represent when 1 that current pixel is shadows pixels, is to represent when 0 not to be.In above formula, α=0.8, β=0.94; It(x, y) V is t picture frame (x, y) value of the V component of position, Bt(x, y) V is to use mixed Gaussian T background image (x, y) value of the V component of position that modeling obtains;It(x, y) H is t image Frame (x, y) value of the H component of position, Bt(x, y) H is the t background using Gaussian modeling to obtain Image (x, y) value of the H component of position;It(x, y) S is that (x, y) S of position divides t picture frame The value of amount, Bt(x, y) S is t background image (x, y) S of position using Gaussian modeling to obtain The value of component.FPMTLIt it is foreground image areas flag bit;
(42) all connected domains of extraction are concentrated from preselected shadows pixels;
(43) for each connected domain, calculate each pixel p=(x, y) gradient direction at place and gradient-norm:
| ▿ p | = ▿ x 2 + ▿ y 2
θ p = arctan 2 ( ▿ y / ▿ x )
WhereinFor horizontal direction gradient, i.e. | f (x, y)-f (x, y+1) |,For vertical gradient, i.e. | f (x, y)-f (x+1, y) |, (x is y) that at point, (x, y) gray value at place, f (x, y+1) is that image is at point to image to f The gray value at (x, y+1) place;Wherein θpIn the range of [-π, π], the most only retain | θp| value be more than A certain threshold taumThose pixels;
(44) for each pixel p remaining=(x, y), the gradient direction calculating present image and background image is poor
Δθ p = arccos ▿ x F ▿ x B + ▿ y F ▿ y B { ( ▿ x F 2 + ▿ y F 2 ) ( ▿ x B 2 + ▿ y B 2 ) } 1 2
Represent in present image at pixel p the Grad of horizontal direction,Represent in background image The Grad of horizontal direction at pixel p,Represent the Grad of vertical direction at pixel p in present image,Represent the Grad of vertical direction at p in background image;
(45) present image and the gradient direction dependency of background image are estimated with following formula
c = { Σ p = 1 n H ( τ a - Δθ p ) } / n
Wherein n is the selected pixels number of candidate shadow region, and H () is unit progression function, i.e. Such as differential seat angle Δ θpLess than or equal to threshold tau a, then the value of function is 1, is otherwise 0;C is currently in region The ratio of pixel count similar with gradient direction in background image in image, as c is bigger than threshold tau c, then it is assumed that Candidate region is shadow region.
Transmission line of electricity Large Construction vehicle based on BOW characterization image model the most according to claim 1 Recognition methods, it is characterised in that: the step of the described BOW aspect of model extracting each foreground area is:
(51) local shape factor
Divide image with the uniform grid that the length of side is 8 pixels, the block block of 4 grid protocol extracts SIFT Son is described;Each piece is the rectangle of 16 × 16 pixels, comprises 4 × 4 buckets, and each bucket is 4 × 4 pixels; Calculating the gradient information in 8 directions in each bucket, the most each piece is characterized by 4 × 4 × 8=128 dimensional vector; The step-length that block moves is 8 pixels;
(52) study visual dictionary
The Dense SIFT local feature using KMEANS method to go out all image zooming-out in training storehouse is carried out Cluster, each cluster centre is i.e. a visual vocabulary in dictionary, and all visual vocabularies form a visual word Allusion quotation;
(53) characteristic quantification coding and spatial pyramid
After obtaining visual dictionary, then each feature of image can be mapped on certain word in dictionary, then unites Meter dictionary in each visual word occurrence number on this image, thus obtain one 300 dimension phenogram as Histogram vectors;
Use multiple dimensioned method of partition, image is divided into three layers of gold of 1 × 1,2 × 2,4 × 4 space separatings Word tower, adds up the feature of each sub-block respectively, is finally got up by the merging features of all pieces, form complete spy Levy.So feature of piece image is 6300 dimensions;
(54) train classification models predicting
Image in all training sets is used above 51) to 53) method in step extracts feature, uses SVM The characteristic vector of training set is trained by support vector machine, it is thus achieved that the disaggregated model of Large Construction vehicle;Core letter Number form formula such as following formula:
k ( x , y ) = Σ k = 1 π m i n ( x k , y k ) .
Transmission line of electricity Large Construction vehicle based on BOW characterization image model the most according to claim 1 Recognition methods, it is characterised in that: described props up many classification SVM that the feature feeding of extraction succeeds in school in advance Hold and vector machine carries out the step of vehicle cab recognition be:
For the foreground area after each removal shade detected, use the Dense of aforementioned extraction image SIFT feature, and carry out feature after quantization is quantified according to visual dictionary, then obtain district through pyramid coding The feature in territory, sends in the SVM classifier trained and classifies, as it is categorized as crane, excavator Or cement pump truck, then it is assumed that it is to have potential outer broken possible Large Construction vehicle, otherwise it is assumed that outer broken possible Property.
CN201610258089.0A 2016-04-22 2016-04-22 Power transmission line large-scale construction vehicle recognition method based on BOW image representation model Pending CN105868734A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610258089.0A CN105868734A (en) 2016-04-22 2016-04-22 Power transmission line large-scale construction vehicle recognition method based on BOW image representation model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610258089.0A CN105868734A (en) 2016-04-22 2016-04-22 Power transmission line large-scale construction vehicle recognition method based on BOW image representation model

Publications (1)

Publication Number Publication Date
CN105868734A true CN105868734A (en) 2016-08-17

Family

ID=56633983

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610258089.0A Pending CN105868734A (en) 2016-04-22 2016-04-22 Power transmission line large-scale construction vehicle recognition method based on BOW image representation model

Country Status (1)

Country Link
CN (1) CN105868734A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106529578A (en) * 2016-10-20 2017-03-22 中山大学 Vehicle brand model fine identification method and system based on depth learning
CN106650622A (en) * 2016-11-18 2017-05-10 华南农业大学 Excavator working state identification method based on hybrid LBF shape regression model
CN107220943A (en) * 2017-04-02 2017-09-29 南京大学 The ship shadow removal method of integration region texture gradient
CN110148117A (en) * 2019-04-22 2019-08-20 南方电网科学研究院有限责任公司 Power equipments defect recognition methods, device and storage medium based on electric power image
CN110349172A (en) * 2019-06-28 2019-10-18 华南理工大学 Transmission line of electricity external force damage prevention method for early warning based on image procossing and binocular solid ranging
CN110458090A (en) * 2019-08-08 2019-11-15 成都睿云物联科技有限公司 Working state of excavator detection method, device, equipment and storage medium
CN110688935A (en) * 2019-09-24 2020-01-14 南京慧视领航信息技术有限公司 Single-lane vehicle detection method based on rapid search
CN111368742A (en) * 2020-03-05 2020-07-03 江苏警官学院 Double-yellow traffic marking reconstruction identification method and system based on video analysis
CN111724426A (en) * 2019-03-19 2020-09-29 杭州海康威视数字技术股份有限公司 Background modeling method and camera for background modeling

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102147861A (en) * 2011-05-17 2011-08-10 北京邮电大学 Moving target detection method for carrying out Bayes judgment based on color-texture dual characteristic vectors
CN103136537A (en) * 2012-12-12 2013-06-05 惠州学院 Vehicle type identification method based on support vector machine
CN105225482A (en) * 2015-09-02 2016-01-06 上海大学 Based on vehicle detecting system and the method for binocular stereo vision
CN105354533A (en) * 2015-09-28 2016-02-24 江南大学 Bag-of-word model based vehicle type identification method for unlicensed vehicle at gate

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102147861A (en) * 2011-05-17 2011-08-10 北京邮电大学 Moving target detection method for carrying out Bayes judgment based on color-texture dual characteristic vectors
CN103136537A (en) * 2012-12-12 2013-06-05 惠州学院 Vehicle type identification method based on support vector machine
CN105225482A (en) * 2015-09-02 2016-01-06 上海大学 Based on vehicle detecting system and the method for binocular stereo vision
CN105354533A (en) * 2015-09-28 2016-02-24 江南大学 Bag-of-word model based vehicle type identification method for unlicensed vehicle at gate

Non-Patent Citations (7)

* Cited by examiner, † Cited by third party
Title
JIANXIN WU: "Histogram Intersection Kernel Learning for Multimedia Applications", 《THE 2010 PACIFIC-RIM CONFERENCE ON MULTIMEDIA》 *
TORNADOMEET: "前景检测算法", 《博客园,HTTP://WWW.CNBLOGS.COM/TORNADOMEET/ARCHIVE/2012/06/02/2531565.HTML》 *
孙立峰等编著: "《和谐人机环境2010》", 29 February 2012, 清华大学出版社 *
宓超等编著: "《装卸机器视觉及其应用》", 31 January 2016, 上海科学技术出版社 *
曾泽兴等: "基于多特征融合的运动阴影去除算法研究", 《电视技术》 *
罗铁镇: "基于混合高斯模型的运动检测及阴影消除算法研究", 《万方数据知识服务平台》 *
肖进胜等: "一种基于历史背景的混合高斯背景建模算法", 《湖南大学学报(自然科学版)》 *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106529578A (en) * 2016-10-20 2017-03-22 中山大学 Vehicle brand model fine identification method and system based on depth learning
CN106650622A (en) * 2016-11-18 2017-05-10 华南农业大学 Excavator working state identification method based on hybrid LBF shape regression model
CN106650622B (en) * 2016-11-18 2019-06-25 华南农业大学 Working state of excavator recognition methods based on mixing LBF shape regression model
CN107220943A (en) * 2017-04-02 2017-09-29 南京大学 The ship shadow removal method of integration region texture gradient
CN111724426A (en) * 2019-03-19 2020-09-29 杭州海康威视数字技术股份有限公司 Background modeling method and camera for background modeling
CN111724426B (en) * 2019-03-19 2023-08-04 杭州海康威视数字技术股份有限公司 Background modeling method and camera for background modeling
CN110148117A (en) * 2019-04-22 2019-08-20 南方电网科学研究院有限责任公司 Power equipments defect recognition methods, device and storage medium based on electric power image
CN110349172A (en) * 2019-06-28 2019-10-18 华南理工大学 Transmission line of electricity external force damage prevention method for early warning based on image procossing and binocular solid ranging
CN110349172B (en) * 2019-06-28 2022-12-16 华南理工大学 Power transmission line external damage prevention early warning method based on image processing and binocular stereo ranging
CN110458090A (en) * 2019-08-08 2019-11-15 成都睿云物联科技有限公司 Working state of excavator detection method, device, equipment and storage medium
CN110688935A (en) * 2019-09-24 2020-01-14 南京慧视领航信息技术有限公司 Single-lane vehicle detection method based on rapid search
CN111368742A (en) * 2020-03-05 2020-07-03 江苏警官学院 Double-yellow traffic marking reconstruction identification method and system based on video analysis

Similar Documents

Publication Publication Date Title
CN105868734A (en) Power transmission line large-scale construction vehicle recognition method based on BOW image representation model
CN109118479B (en) Capsule network-based insulator defect identification and positioning device and method
CN110163187B (en) F-RCNN-based remote traffic sign detection and identification method
Siriborvornratanakul An automatic road distress visual inspection system using an onboard in-car camera
CN105894701B (en) The identification alarm method of transmission line of electricity external force damage prevention Large Construction vehicle
CN107491762B (en) A kind of pedestrian detection method
CN111695514B (en) Vehicle detection method in foggy days based on deep learning
CN110543837A (en) visible light airport airplane detection method based on potential target point
CN108009518A (en) A kind of stratification traffic mark recognition methods based on quick two points of convolutional neural networks
CN107273832B (en) License plate recognition method and system based on integral channel characteristics and convolutional neural network
CN109918971B (en) Method and device for detecting number of people in monitoring video
CN104715244A (en) Multi-viewing-angle face detection method based on skin color segmentation and machine learning
CN106682641A (en) Pedestrian identification method based on image with FHOG- LBPH feature
Abbas et al. A comprehensive review of vehicle detection using computer vision
CN105069447A (en) Facial expression identification method
Wali et al. Shape matching and color segmentation based traffic sign detection system
Mammeri et al. North-American speed limit sign detection and recognition for smart cars
Supreeth et al. An approach towards efficient detection and recognition of traffic signs in videos using neural networks
CN105354547A (en) Pedestrian detection method in combination of texture and color features
CN115376108A (en) Obstacle detection method and device in complex weather
CN111931577A (en) Intelligent inspection method for specific foreign matters of power grid line
CN106778504A (en) A kind of pedestrian detection method
CN111339823A (en) Threshing and sunning ground detection method based on machine vision and back projection algorithm
Oluchi et al. Development of a Nigeria vehicle license plate detection system
Karaimer et al. Detection and classification of vehicles from omnidirectional videos using temporal average of silhouettes

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20160817