CN106815576A - Target tracking method based on consecutive hours sky confidence map and semi-supervised extreme learning machine - Google Patents

Target tracking method based on consecutive hours sky confidence map and semi-supervised extreme learning machine Download PDF

Info

Publication number
CN106815576A
CN106815576A CN201710047829.0A CN201710047829A CN106815576A CN 106815576 A CN106815576 A CN 106815576A CN 201710047829 A CN201710047829 A CN 201710047829A CN 106815576 A CN106815576 A CN 106815576A
Authority
CN
China
Prior art keywords
target
sample
trail
followed
semi
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710047829.0A
Other languages
Chinese (zh)
Other versions
CN106815576B (en
Inventor
年睿
邱书琦
常瑞杰
肖玫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ocean University of China
Original Assignee
Ocean University of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ocean University of China filed Critical Ocean University of China
Priority to CN201710047829.0A priority Critical patent/CN106815576B/en
Publication of CN106815576A publication Critical patent/CN106815576A/en
Application granted granted Critical
Publication of CN106815576B publication Critical patent/CN106815576B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of target tracking method based on consecutive hours sky confidence map and semi-supervised extreme learning machine, the method is in time continuous in view of video frame image, target location to be followed the trail of will not also undergo mutation simultaneously, other video frame image is also spatially continuous, spatial continuity is embodied in target and target ambient background has certain particular kind of relationship, when great changes will take place for the outward appearance of target, this relation can help distinguish between target and background region to be followed the trail of.The present invention is directed to deformation and occlusion issue, fully take into account the information that real goal can be provided, fully excavate the distribution similarity for having exemplar and unlabeled exemplars, improve the precision followed the trail of, propose a kind of semi-supervised method for tracing based on extreme learning machine for excavating and having exemplar and unlabeled exemplars distribution similarity, above two method is combined in a tracking framework for coupling, the present invention realizes that a kind of robustness is good, high robust tracking.

Description

Target tracking method based on consecutive hours sky confidence map and semi-supervised extreme learning machine
Technical field
Target tracking method the present invention relates to be based on consecutive hours sky confidence map and semi-supervised extreme learning machine, belongs to intelligence Information processing and target tracking technical field.
Background technology
Target following is indispensable link in most vision systems.(such as video in specific scene application The fields such as monitoring), automation, the quick, target tracking of high robust attract attention.Video monitoring, Vehicle Detection, intelligence machine The aspects such as people, sea floor object detection tracking have broad application prospects.
Target tracking is an extremely important part in computer vision field, and moving objects in video track algorithm is logical The information of the video image for analyzing each frame in sequence of video images to be followed the trail of is crossed, data mining is carried out in video, learn mesh Mark behavior simultaneously carries out substantial amounts of motion capture, and a series for the treatment of is carried out to information, obtains and marks tracked target and exist Corresponding position in video image.The complexity for blocking deformation, background between object, the change of illumination light and shade, real-time and strong Strong property difference etc. is tracing process problem demanding prompt solution.Classical method for tracing such as Meanshift, particle filter etc. is depended on and regarded The abundant degree of contained target information in frequency, in actual sequence of video images, the information that target can be provided is fairly limited, Cause the tracking target that can not stablize, such as there is deformation to block in scene, these classic algorithms are even more helpless.
That is subject matter present in prior art:(1) real-time and strong during being followed the trail of in video scene to be followed the trail of Strong property is poor, and target time space position information is deficient, the unobvious problem of target signature;(2) when scene has shelter and target to be followed the trail of In the case of deforming upon, especially occur that whole target is blocked and target to be followed the trail of occurs the situation of huge deformation, The problem that the target followed the trail of can be caused to lose.
The content of the invention
It is an object of the invention to provide a kind of target tracking based on consecutive hours sky confidence map and semi-supervised extreme learning machine Method, to make up the deficiencies in the prior art.
The present invention is in time continuous in view of video frame image, and the continuity of time is embodied in neighbouring interframe and waits to chase after Track object variations will not be very big, while target location to be followed the trail of will not also undergo mutation;At the same time video frame image is in space On be also continuous, spatial continuity is embodied in target and target ambient background and there is certain particular kind of relationship, when the outward appearance of target When great changes will take place, this relation can help distinguish between target and background region to be followed the trail of, and propose to utilize the vacant letter of consecutive hours The method for tracing that graphics is practised overcomes real-time and robustness is poor, target time space position information scarcity and target signature are obvious etc. Problem.For deformation and occlusion issue, the information that real goal can be provided is fully taken into account, fully excavating has exemplar With the distribution similarity of unlabeled exemplars, the precision followed the trail of is improved, it is proposed that one kind is excavated exemplar and unlabeled exemplars The semi-supervised method for tracing based on extreme learning machine of distribution similarity, above two method is combined in a tracking for coupling In framework, realize that a kind of robustness is good, high robust tracking.
To reach above-mentioned purpose, concrete technical scheme that the present invention takes is through the following steps that realize:
Step one, it is specific treat tracing monitoring scene in gather n frames target video A={ I to be followed the trail of1,…,Ii,…In, Wherein IiThe i-th frame sequence of video images to be followed the trail of is represented, video to be followed the trail of is pre-processed using image filtering denoising, contrast enhancing SequenceReduce noise and protrusion interested treats trace regions;
Step 2, in t frames sequence of video images I to be followed the trail oftMiddle use rectangular window chooses target O to be followed the trail of, it is determined that Target's center position o*, O represents fresh target presence in the scene, and o represents new target location, defines a two dimension mesh to be followed the trail of Target confidence map MODEL Ct(o);It is by target area expansion twice formation local background region representation to be followed the trail of It is interior Intensity locations feature w (k) at coordinate position k are extracted, intensity locations feature set is constituted The brightness of image at I (k) denotation coordinations position k,Denotation coordination o*Neighborhood;Set up the priori of t frames target to be followed the trail of Model P (w (k) | O), t frame space-time models are gone out in prediction on such basis
Step 3, in target's center position to be followed the trail of region overlap sampling, obtain N1Individual region unit image is used as just Sample and N2Individual region unit image extracts positive and negative sample data feature x as negative samplej, the class label for remembering positive sample is 1, is born The class label of sample is 0, yj∈{1,0};Foundation has mark sample setWith without mark sample set XuGroup Into training sample set X={ Xs,Xu}={ (xj,yj), j=1 ..., N1+N2
Step 4, the training sample set X obtained with step 3 train semi-supervised extreme learning machine network model;
Step 5, in It+1In, the t frame space-time models tried to achieve using step 2Model modification is carried out, is calculated The space-time model of t+1 framesUsing the t+1 frame space-time models tried to achieveConvolution It+1Obtain the space-time confidence map of fresh target Ct+1O (), maximizes Ct+1O () determines the target location o in t+1 frames;
Step 6, judge whether target is blocked, if target is not blocked, into step 5, conversely, into step 7;
Step 7, in It+1In, by ItIn the o that has tried to achieve*It is target location, in target location o*Region, with target Region rectangle window size overlap sampling, obtains N number of region unit image as candidate target, extracts candidate target data characteristicsSet up target image block test sample collection to be followed the trail ofBy test sample collection input step four The semi-supervised extreme learning machine network for completing is trained, t+1 frames test output T is obtained, online semi-supervised extreme learning machine is maximized Maximum classification response position, obtains target location o in t+1 frames;
Step 8, carry out online semi-supervised extreme learning machine network model to maximum classification response result and update threshold value sentencing It is fixed, if online semi-supervised extreme learning machine model need not update, into step 5, otherwise enter step 9;
Step 9, there is labeled data collection by what step 3 was obtainedThe test specimens obtained with step 7 This collectionAs without labeled data collection Xu=Xt+1, carry out step 4, the semi-supervised limit of re -training Habit machine network model;
Circulating repetition above-mentioned steps, until follow the trail of completing whole video sequence.
Further, the step 3:In target's center position o to be followed the trail of*Region, with target area rectangular window Size overlap sampling, j-th sampled point be to the Euclidean distance of target's center positionWhenWhen, sampling obtains N1It is individual Region unit image as positive sample, whenWhen, sampling obtains N2Individual region unit image is used as negative sample, r1、r2And r3 It is respectively sample radius;Extract positive and negative sample data feature xj, target image block training sample set to be followed the trail of is set up, (N is collected altogether1 +N2) individual target image block is used as training sample set X={ (xj,yj), j=1 ..., N1+N2, remembering the class label of positive sample is 1, the class label of negative sample is 0, yj∈{1,0};The sample order that training sample is concentrated is upset and reset, is taken before coming most A certain proportion of sample (usual ratio is relatively low) conduct in face has marked sample set Xs, take remaining sample (usual ratio is higher) work Not mark sample set Xu, and X={ Xs,Xu}。
The step 4:Input weights are set using random fashion and hidden layer is biased, if representing hidden layer knot with (a, b) Input weight a and threshold value b that point is obtained, training sample are have labeled data collectionWithout labeled data collectionWherein XsAnd XuRepresent input sample, YsIt is and XsCorresponding output sample;The mapping function of hidden layer is G X (), mapping function form can be expressed as G (x)=1/ (1+e-x), output weight is represented with β, h (xi)=[G (a1,b1, xi),…,G(aL,bL,xi)]s×mI-th hidden layer output matrix is represented, the nodes of hidden layer are m, eiRepresent i-th input The learning error (residual error) of node
The object function of semi-supervised extreme learning machine is:
fi=h (xi) β, i=1 ..., s+u
Wherein ciPunishment parameter is represented, λ represents balance parameter, and L is the Tula obtained by label data and without label data This operation result of pula, F is the output matrix of network, and Tr is mark computing;
Represent that semi-supervised extreme learning machine object function is with matrix form:
WhereinIt is that preceding s rows are equal to Ys, the rear null output label sample of u rows, C is that preceding s diagonal entry is CiIt is remaining It is zero diagonal matrix;
Local derviation is asked to obtain β above formula:
It is zero to make local derviation, and solution obtains output weight beta and is:
When there is label data to be more than hidden layer nodal point number
When there is label data to be less than hidden layer nodal point number
Wherein, HTIt is the transposed matrix of matrix H.
The method whether target is blocked of judging described in step 6 is the result to confidence mapCarry out occlusion threshold th1Judge, ifWhen, show that blocking occurs in target, th1The critical value that expression is blocked, can become according to the difference of scene Change, when this algorithm is applied to different scenes, artificially adjust th1Value, can fluctuate, when target quilt in certain scope under normal circumstances When blocking,Can decline rapidly, the value after will fall rapidly upon is defined as th1Value, judges whether target is blocked with this.
It is to maximum that whether the online semi-supervised extreme learning machine network model of judgement described in step 8 needs the method for updating Classification response result TmaxIt is updated threshold value th2Judge, if Tmax>th2When, show online semi-supervised extreme learning machine network mould Type is not required to update, to update threshold value th2Judge whether network model needs to update.
Beneficial effects of the present invention:The vacant letter graphics of consecutive hours is practised method for tracing and semi-supervised extreme learning machine by the present invention Method for tracing is combined, and solves in tracing process that real-time and robustness are poor, target time space position information is deficient, target signature Not obvious and deformation blocks to cause to follow the trail of the problem that target is lost.The present invention is particular by calculating consecutive hours sky confidence map Carry out occlusion threshold judgement, obtained it is a kind of judge target whether enter occlusion area method, effectively solve target screening The decision problem of gear, and carry out semi-supervised limit study by calculating semi-supervised extreme learning machine network output maximum response Machine network model update threshold determination, obtained it is a kind of judge the network model whether need renewal method, efficiently solve net The problem of network model generalization difference.Present invention greatly enhances the precision followed the trail of, realize that a kind of robustness is good, high robust Tracing process.
Brief description of the drawings
Fig. 1 is overall trace flow schematic diagram of the invention.
Fig. 2 is the zone marker figure of target to be followed the trail of in specific embodiment.
Fig. 3 is based on consecutive hours sky confidence map target tracking method block diagram in specific embodiment.
Fig. 4 is the basic framework figure of semi-supervised extreme learning machine network.
Fig. 5 be specific embodiment in be based on semi-supervised extreme learning machine target tracking method block diagram.
To follow the trail of effect example under circumstance of occlusion in specific embodiment, figure (a) is that have interesting target to be followed the trail of to Fig. 6 Frame of video, (b), (c), (d), (e), (f) are respectively the frame of video being tracked to the interesting target after (a) frame.
Specific embodiment
To make the purpose of the present invention, embodiment and advantage relatively sharp, below in conjunction with the accompanying drawings and by specific embodiment To further illustrate the present invention.
Particular flow sheet of the invention is as shown in Figure 1.
It is specific using one section of corridor monitor video caviar (384*288 pixels, 25 frame per second) of classics in the present embodiment As video to be followed the trail of.
Step one, pre-process video sequence to be followed the trail of using image filtering denoising, contrast enhancing, reduction noise and protrusion It is interested to treat trace regions;Specifically include following steps:
Step 1-1, the corridor monitor video caviar to one section of classics are defined as A, carry out sub-frame processing 200 frames of acquisition and treat Follow the trail of sequence of video images, i.e. A={ I1,…,Ii,…I200, wherein IiRepresent that corridor monitoring video the i-th frames of caviar are waited to chase after Track video image;
Step 1-2, to the 200 frame video image sequenceIt is filtered denoising, the enhanced pretreatment of contrast.
Step 2, in t=1 frames sequence of video images I to be followed the trail ofT=1In choose target O to be followed the trail of, determine target's center Position o*, O represents fresh target presence in the scene, and o represents new target location, defines a confidence for two dimension target to be followed the trail of Graph model Ct(o);The prior model P (w (k) | O) of t frames target to be followed the trail of is set up, t frame space-time models are gone out in prediction on such basisAs schemed Shown in 3;Specifically include following steps:
Step 2-1, in IT=1In choose interesting target O to be followed the trail of, target area rectangle using rectangular window W by user The width of window is w and height is h, and o represents new target location, and target area is expanded into twice forms local background's region representation ForAs shown in Figure 2;Intensity locations feature w (k) at coordinate position k are extracted in local background region, intensity locations are constituted Feature setThe brightness of image at I (k) denotation coordinations position k,Denotation coordination o* Neighborhood;
Step 2-2, tracing problem is converted into calculates interesting target position confidence map problem to be followed the trail of:
Wherein CtO () represents the confidence graph model of t frames, represent new target location o and old target location o*Relative position And direction, new target location is nearer apart from old target location, and the value of the confidence is bigger;Represent space-time mould Type, the relative position and direction of description fresh target o and local background area coordinate points k, P (w (k) | O) prior model is represented, retouch State old target locationIntensity and relative position direction with local background area coordinate points k, simulate interesting target O to be followed the trail of Rudimentary profile information;
Step 2-3, calculating t=1 frames IT=1Confidence mapSimultaneously maximum is obtained to put Letter value
Step 2-4, the prior model for calculating t=1 framesWhereinIt is chi Degree parameter;
Step 2-5, by calculating t=1 frames IT=1Confidence map MODEL CtO () and prior model P (w (k), O) calculate the The space-time model of the interesting target of t=1 frames
Wherein F represents FFT, F-1Represent fast fourier inverse transformation.
Step 3, in target's center position to be followed the trail of region overlap sampling, obtain N1Individual region unit image is used as just Sample and N2Individual region unit image extracts positive and negative sample data feature x as negative samplej, setting up has mark sample setWith without mark sample set XuComposition training sample set X={ (xj,yj), j=1 ..., N1+N2, such as Fig. 5 It is shown, specifically include following steps:
Step 3-1, in center o*Region, with target area rectangular window size overlap sampling, j-th sampling Euclidean distance of the point to target's center position beWhenWhen, sampling obtains 45 region unit images as positive sample, WhenWhen, sampling obtains 31 region unit images as negative sample, sample radius r1、r2And r3It is respectively to set (the unit of parameter 5,10 and 20:Pixel);
Step 3-2, the positive and negative sample data feature x of extractionj, target image block training sample set to be followed the trail of is set up, collect altogether 76 target image blocks are used as training sample set X={ (xj,yj), j=1 ..., 76, the class label for remembering positive sample is 1, is born The class label of sample is 0, yj∈{1,0};
Step 3-3, by training sample concentrate sample order upset and reset, take 50 that come foremost sample make To have marked sample set Xs, remaining 26 samples are taken as not marking sample set Xu, and X={ Xs,Xu}。
Step 4, the training sample set X obtained with step 3 train online semi-supervised extreme learning machine network model, specifically Comprise the following steps:
Step 4-1, semi-supervised extreme learning machine are a kind of BP network models of single hidden layer as shown in figure 4, whole Network model is divided into three layers, including:Input layer, hidden layer and output layer, input weights and hidden layer are set using random fashion Biasing, independently of training sample, algorithm structure simple computation efficiency high, if representing the input that hidden layer node is obtained with (a, b) Weight a and threshold value b, training sample is have labeled data collectionWithout labeled data collectionWherein XsAnd XuRepresent input sample, YsIt is and XsCorresponding output sample;The mapping function of hidden layer is G (x), mapping function form G (x)=1/ (1+e can be expressed as-x), output weight is represented with β, h (xi)=[G (a1,b1,xi),…,G(a2000,b2000, xi)]50×2000I-th hidden layer output matrix is represented, the nodes of hidden layer are 2000, eiRepresent i-th of input node Practise error (residual error);
Step 4-2, the object function of the semi-supervised extreme learning machine that need to be trained are:
fi=h (xi) β, i=1 ..., s+u
Wherein ciPunishment parameter is represented, λ represents balance parameter, and L is the Tula obtained by label data and without label data This operation result of pula, F is the output matrix of network, TrIt is mark computing;
Step 4-3, represent that semi-supervised extreme learning machine object function is with matrix form:
WhereinIt is that preceding 50 row is equal to Ys, the null output label sample of 26 rows afterwards.C is that preceding 50 diagonal entries are CiRemaining as zero diagonal matrix;
Step 4-4, local derviation is asked to obtain β above formula:
Step 4-5, local derviation is made to be zero, solution obtains output weight beta and is:
When there is label data to be more than hidden layer nodal point number
When there is label data to be less than hidden layer nodal point number
Wherein, HTIt is the transposed matrix of matrix H, so far semi-supervised extreme learning machine network model training is finished.
Step 5, in t+1 frames, the t frame space-time models tried to achieve using step 2Model modification is carried out, is calculated Obtain the space-time model of t+1 framesUsing the t+1 frame space-time models tried to achieveConvolved image It+1Obtain fresh target when Empty confidence map Ct+1O (), maximizes the confidence map C for trying to achievet+1O () determines the target location o in t+1 frames, as shown in figure 3, tool Body is comprised the following steps:
Step 5-1, in It+1In, with o*The local background region of twice target sizes is taken for target location Intensity locations feature is extracted in the region, intensity locations feature set is constituted
Step 5-2, the space-time model of t frames interesting target to be followed the trail of update:
Wherein ρ is learning rate,It is interesting target space-time model to be followed the trail of that t frames are calculated,In frequency domain It is expressed as:
WhereinIt isTime domain Fourier transform.Time domain filtering FwIt is expressed as:
Fw=ρ/(ejw-(1-ρ))
Wherein j is imaginary unit;
Step 5-3, calculating t+1 frames interesting target confidence map to be followed the trail of:
Step 5-4, in t+1 frames interesting target position o be maximize t+1 frames confidence map:
O=arg maxCt+1(o)
Maximum the value of the confidence is
Step 6, judge whether target is blocked, specifically include following steps:
Step 6-1, the maximum the value of the confidence obtained to step 5-4Carry out occlusion threshold th1Judge, ifWhen, Show that blocking occurs in target, judge whether target is blocked with this.th1The critical value that expression is blocked, can be according to the difference of scene And change, when this algorithm is applied to different scenes, artificially adjust th1Value, can fluctuate in certain scope under normal circumstances, work as mesh When mark is blocked,Can decline rapidly, the value after will fall rapidly upon is defined as th1Value;Defined in this concrete scheme
If step 6-2,When, show that blocking does not occur in target, step 5-1 is carried out, otherwise carry out step 7- 1。
Step 7, in t+1 frames, tracked overlap sampling at target's center position in t frames, extract candidate target Data characteristics, sets up target image block test sample to be followed the trail of, and test sample is input into the above-mentioned online semi-supervised pole for training Be prediction new target location by maximum classification response position in test sample in limit learning machine, as shown in figure 5, specifically include with Lower step:
Step 7-1, to t+1 frame video images, with o*It is target location, in center o*Region, with target area Domain rectangular window size overlap sampling, j-th sampled point to o*Euclidean distance beWhenWhen, sampling obtains 232 Region unit image is test data as candidate target, and the sample data of extraction is characterized asAnd remember that test set isSample radius r1The parameter of setting is 20 (units:Pixel);
Step 7-2, test are output as:
T=H*β
Wherein β is the output weight that t frames are calculated, H*It is the hidden layer output matrix of test,
Step 7-3, in t+1 frames, interesting target position o to be followed the trail of be maximize the semi-supervised extreme learning machine of t+1 frames Maximum classification response position:
O=arg max T
Maximum classification response value is Tmax
Step 8, carry out online semi-supervised extreme learning machine network model to maximum classification response result and update threshold value sentencing It is fixed, comprise the following steps that:
Step 8-1, to maximum classification response result TmaxCarry out the renewal threshold value th of semi-supervised extreme learning machine2Judge, if Tmax>th2When, show that online semi-supervised extreme learning machine network model is not required to update, judge that the online semi-supervised limit learns with this Whether machine model needs to update, th2The critical value for updating is represented, th defined in this concrete scheme2=0.
If step 8-2, Tmax>When 0, show that online semi-supervised extreme learning machine network model is not required to update, carry out step 5- 1, otherwise carry out step 9.
The online semi-supervised extreme learning machine network model of step 9, re -training, as shown in figure 5, specific as follows:By step What 3-3 was obtained has labeled data collectionThe test set obtained with step 7-1 As without labeled data collection Xu=Xt+1, carry out the online semi-supervised extreme learning machine network model of step 4-1 re -trainings.
Circulating repetition above-mentioned steps, until completing the tracking for entirely treating tracing monitoring video sequence.
Tracing monitoring video is treated to above-mentioned, respectively with particle filter, Meanshift and the inventive method traceability The comparing of energy, the results are shown in Table 1, it can be seen that no matter the inventive method center bias contribution or deviation mean square deviation result Particle filter and Meanshift methods are superior to, robustness and robustness to target tracking is realized.
Table 1 is contrasted for display particle filter, Meanshift and the inventive method tracking performance
Particle filter Meanshift The inventive method
Center deviation 75.4796 22.9740 10.1834
Deviation mean square deviation 47.8903 12.2607 7.9702
Fig. 6 is tracking effect example under circumstance of occlusion in specific embodiment, it can be seen that serious twice having successively gone through Circumstance of occlusion under, can accurately still follow the trail of target, further demonstrate the robustness and robustness of the inventive method.
The above is the preferred embodiment of the present invention, it should be pointed out that:For those skilled in the art For, under the premise without departing from the principles of the invention, some improvements and modifications can also be made, these improvements and modifications should also be regarded It is protection scope of the present invention.

Claims (5)

1. it is a kind of based on consecutive hours sky confidence map and semi-supervised extreme learning machine target tracking method, it is characterised in that including Following steps:
Step one, it is specific treat tracing monitoring scene in gather n frames target video A={ I to be followed the trail of1,…,Ii,…In, wherein Ii The i-th frame sequence of video images to be followed the trail of is represented, video sequence to be followed the trail of is pre-processed using image filtering denoising, contrast enhancingReduce noise and protrusion interested treats trace regions;
Step 2, in t frames sequence of video images I to be followed the trail oftMiddle use rectangular window chooses target O to be followed the trail of, in determining target Heart position o*, O represent fresh target presence in the scene, and o represents new target location, one two dimension target O's to be followed the trail of of definition Confidence map MODEL Ct(o);It is by target area expansion twice formation local background region representation to be followed the trail of Interior extraction Intensity locations feature w (k) at coordinate position k, constitute intensity locations feature setI(k) The brightness of image at the k of denotation coordination position,The neighborhood of denotation coordination o*.Set up the prior model of t frames target to be followed the trail of P (w (k) | O), t frame space-time models are gone out in prediction on such basis
Step 3, in target's center position to be followed the trail of region overlap sampling, obtain N1Individual region unit image as positive sample and N2Individual region unit image extracts positive and negative sample data feature x as negative samplej, the class label for remembering positive sample is 1, negative sample Class label is 0, yj∈{1,0};Foundation has mark sample setWith without mark sample set XuComposition training Sample set X={ Xs,Xu}={ (xj,yj), j=1 ..., N1+N2
Step 4, the training sample set X obtained with step 3 train semi-supervised extreme learning machine network model;
Step 5, in It+1In, the t frame space-time models tried to achieve using step 2Model modification is carried out, t+1 is calculated The space-time model of frameUsing the t+1 frame space-time models tried to achieveConvolution It+1Obtain the space-time confidence map C of fresh targett+1 O (), maximizes Ct+1O () determines the target location o in t+1 frames;
Step 6, judge whether target is blocked, if target is not blocked, into step 5, conversely, into step 7;
Step 7, in It+1In, by ItIn the o* that has tried to achieve be target location, in target location o* regions, with target area Rectangular window size overlap sampling, obtains N number of region unit image as candidate target, extracts candidate target data characteristicsBuild Found target image block test sample collection to be followed the trail ofTest sample collection input step four has been trained Into semi-supervised extreme learning machine network, obtain t+1 frames test output T, maximize the maximum classification of semi-supervised extreme learning machine and ring Position is answered, target location o in t+1 frames is obtained;
Step 8, semi-supervised extreme learning machine network model is carried out to maximum classification response result update threshold determination, if half supervises Superintending and directing extreme learning machine model need not update, and into step 5, otherwise enter step 9;
Step 9, there is labeled data collection by what step 3 was obtainedThe test sample collection obtained with step 7As without labeled data collection Xu=Xt+1, carry out step 4, the semi-supervised extreme learning machine of re -training Network model;
Circulating repetition above-mentioned steps, until follow the trail of completing whole video sequence.
2. target tracking method as claimed in claim 1, it is characterised in that the step 3 is specially:In target to be followed the trail of Center o*Region, with target area rectangular window size overlap sampling, j-th sampled point to target's center position Euclidean distance isWhenWhen, sampling obtains N1Individual region unit image as positive sample, whenWhen, sampling Obtain N2Individual region unit image is used as negative sample, r1、r2And r3It is respectively sample radius.Extract positive and negative sample data feature xj, build Target image block training sample set to be followed the trail of is found, (N is collected altogether1+N2) individual target image block is used as training sample set X={ (xj, yj), j=1 ..., N1+N2, the class label for remembering positive sample is 1, and the class label of negative sample is 0, yj∈{1,0};Will training Sample order in sample set is upset and is reset, and takes and comes a certain proportion of sample of foremost as having marked sample set Xs, Remaining sample is taken as not marking sample set Xu, and X={ Xs,Xu}。
3. target tracking method as claimed in claim 1, it is characterised in that the step 4 is specially:Using random fashion Input weights and hidden layer biasing are set, if representing input weight a and threshold value b that hidden layer node is obtained with (a, b), sample is trained This is have labeled data collectionWithout labeled data collectionWherein XsAnd XuRepresent input sample, YsIt is and XsCorresponding output sample;The mapping function of hidden layer is G (x), and mapping function form can be expressed as G (x)=1/ (1 +e-x), output weight is represented with β, h (xi)=[G (a1,b1,xi),…,G(aL,bL,xi)]1×mRepresent i-th hidden layer output square Battle array, the nodes of hidden layer are m, eiRepresent i-th learning error of input node;
The object function of semi-supervised extreme learning machine is:
fi=h (xi) β, i=1 ..., s+u
Wherein ciPunishment parameter is represented, λ represents balance parameter, and L is the figure Laplce obtained by label data and without label data Operation result, F is the output matrix of network, and Tr is mark computing;
Represent that semi-supervised extreme learning machine object function is with matrix form:
WhereinIt is that preceding s rows are equal to Ys, the rear null output label sample of u rows, C is that preceding s diagonal entry is CiRemaining as zero Diagonal matrix;
Local derviation is asked to obtain β above formula:
It is zero to make local derviation, and solution obtains output weight beta and is:
When there is label data to be more than hidden layer nodal point number
When there is label data to be less than hidden layer nodal point number
Wherein, HTIt is the transposed matrix of matrix H.
4. target tracking method as claimed in claim 1, it is characterised in that judge whether target is blocked described in step 6 Method be result to confidence mapCarry out occlusion threshold th1Judge, ifWhen, show that blocking occurs in target, th1 The critical value that expression is blocked, can change according to the difference of scene, when this algorithm is applied to different scenes, artificially adjust th1 Value, can fluctuate in certain scope under normal circumstances, when target is blocked,Can decline rapidly, the value after will fall rapidly upon It is defined as th1Value, judges whether target is blocked with this.
5. target tracking method as claimed in claim 1, it is characterised in that the semi-supervised limit study of judgement described in step 8 It is to maximum classification response result T that whether machine network model needs the method for updatingmaxIt is updated threshold value th2Judge, if Tmax>th2 When, show that semi-supervised extreme learning machine network model is not required to update, to update threshold value th2Judge whether network model needs to update.
CN201710047829.0A 2017-01-20 2017-01-20 Target tracking method based on continuous space-time confidence map and semi-supervised extreme learning machine Active CN106815576B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710047829.0A CN106815576B (en) 2017-01-20 2017-01-20 Target tracking method based on continuous space-time confidence map and semi-supervised extreme learning machine

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710047829.0A CN106815576B (en) 2017-01-20 2017-01-20 Target tracking method based on continuous space-time confidence map and semi-supervised extreme learning machine

Publications (2)

Publication Number Publication Date
CN106815576A true CN106815576A (en) 2017-06-09
CN106815576B CN106815576B (en) 2020-07-07

Family

ID=59111417

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710047829.0A Active CN106815576B (en) 2017-01-20 2017-01-20 Target tracking method based on continuous space-time confidence map and semi-supervised extreme learning machine

Country Status (1)

Country Link
CN (1) CN106815576B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107423762A (en) * 2017-07-26 2017-12-01 江南大学 Semi-supervised fingerprinting localization algorithm based on manifold regularization
CN109255321A (en) * 2018-09-03 2019-01-22 电子科技大学 A kind of visual pursuit classifier construction method of combination history and instant messages
CN110211104A (en) * 2019-05-23 2019-09-06 复旦大学 A kind of image analysis method and system for pulmonary masses computer aided detection
CN110675382A (en) * 2019-09-24 2020-01-10 中南大学 Aluminum electrolysis superheat degree identification method based on CNN-LapseLM
CN113378673A (en) * 2021-05-31 2021-09-10 中国科学技术大学 Semi-supervised electroencephalogram signal classification method based on consistency regularization
CN113408984A (en) * 2021-06-21 2021-09-17 北京思路智园科技有限公司 Hazardous chemical substance transportation tracking system and method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8254633B1 (en) * 2009-04-21 2012-08-28 Videomining Corporation Method and system for finding correspondence between face camera views and behavior camera views
CN102663453A (en) * 2012-05-03 2012-09-12 西安电子科技大学 Human motion tracking method based on second generation Bandlet transform and top-speed learning machine
CN103942749A (en) * 2014-02-24 2014-07-23 西安电子科技大学 Hyperspectral ground feature classification method based on modified cluster hypothesis and semi-supervised extreme learning machine
CN104992453A (en) * 2015-07-14 2015-10-21 国家电网公司 Target tracking method under complicated background based on extreme learning machine
CN106296734A (en) * 2016-08-05 2017-01-04 合肥工业大学 Based on extreme learning machine and the target tracking algorism of boosting Multiple Kernel Learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8254633B1 (en) * 2009-04-21 2012-08-28 Videomining Corporation Method and system for finding correspondence between face camera views and behavior camera views
CN102663453A (en) * 2012-05-03 2012-09-12 西安电子科技大学 Human motion tracking method based on second generation Bandlet transform and top-speed learning machine
CN103942749A (en) * 2014-02-24 2014-07-23 西安电子科技大学 Hyperspectral ground feature classification method based on modified cluster hypothesis and semi-supervised extreme learning machine
CN104992453A (en) * 2015-07-14 2015-10-21 国家电网公司 Target tracking method under complicated background based on extreme learning machine
CN106296734A (en) * 2016-08-05 2017-01-04 合肥工业大学 Based on extreme learning machine and the target tracking algorism of boosting Multiple Kernel Learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
何传阳等: "基于智能监控的中小人群异常行为检测", 《计算机应用》 *
朱博等: "基于极限学习机(ELM)的视线落点估计方法", 《东北大学学报(自然科学版)》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107423762A (en) * 2017-07-26 2017-12-01 江南大学 Semi-supervised fingerprinting localization algorithm based on manifold regularization
CN109255321A (en) * 2018-09-03 2019-01-22 电子科技大学 A kind of visual pursuit classifier construction method of combination history and instant messages
CN109255321B (en) * 2018-09-03 2021-12-10 电子科技大学 Visual tracking classifier construction method combining history and instant information
CN110211104A (en) * 2019-05-23 2019-09-06 复旦大学 A kind of image analysis method and system for pulmonary masses computer aided detection
CN110211104B (en) * 2019-05-23 2023-01-06 复旦大学 Image analysis method and system for computer-aided detection of lung mass
CN110675382A (en) * 2019-09-24 2020-01-10 中南大学 Aluminum electrolysis superheat degree identification method based on CNN-LapseLM
CN113378673A (en) * 2021-05-31 2021-09-10 中国科学技术大学 Semi-supervised electroencephalogram signal classification method based on consistency regularization
CN113378673B (en) * 2021-05-31 2022-09-06 中国科学技术大学 Semi-supervised electroencephalogram signal classification method based on consistency regularization
CN113408984A (en) * 2021-06-21 2021-09-17 北京思路智园科技有限公司 Hazardous chemical substance transportation tracking system and method
CN113408984B (en) * 2021-06-21 2024-03-22 北京思路智园科技有限公司 Dangerous chemical transportation tracking system and method

Also Published As

Publication number Publication date
CN106815576B (en) 2020-07-07

Similar Documents

Publication Publication Date Title
CN106815576A (en) Target tracking method based on consecutive hours sky confidence map and semi-supervised extreme learning machine
CN107203781B (en) End-to-end weak supervision target detection method based on significance guidance
CN106778604B (en) Pedestrian re-identification method based on matching convolutional neural network
CN103871029B (en) A kind of image enhaucament and dividing method
CN106997597B (en) It is a kind of based on have supervision conspicuousness detection method for tracking target
CN109543606A (en) A kind of face identification method that attention mechanism is added
CN107506722A (en) One kind is based on depth sparse convolution neutral net face emotion identification method
CN102256065B (en) Automatic video condensing method based on video monitoring network
CN102722714B (en) Artificial neural network expanding type learning method based on target tracking
CN108053419A (en) Inhibited and the jamproof multiscale target tracking of prospect based on background
CN103942557B (en) A kind of underground coal mine image pre-processing method
CN106650690A (en) Night vision image scene identification method based on deep convolution-deconvolution neural network
CN106530246A (en) Image dehazing method and system based on dark channel and non-local prior
CN107341480A (en) A kind of crowd massing detection method of modified PCCNN neural network models
CN107025420A (en) The method and apparatus of Human bodys' response in video
CN110110649A (en) Alternative method for detecting human face based on directional velocity
CN104751466B (en) A kind of changing object tracking and its system based on conspicuousness
CN107833239B (en) Optimization matching target tracking method based on weighting model constraint
CN108009493A (en) Face anti-fraud recognition methods based on action enhancing
CN105825233B (en) A kind of pedestrian detection method based on on-line study random fern classifier
CN107833241A (en) To real-time vision object detection method of the ambient lighting change with robustness
CN106611421A (en) SAR image segmentation method based on feature learning and sketch line constraint
CN107895379A (en) The innovatory algorithm of foreground extraction in a kind of video monitoring
CN113139501B (en) Pedestrian multi-attribute identification method combining local area detection and multi-level feature grabbing
CN104021544A (en) Greenhouse vegetable disease surveillance video key frame extracting method and extracting system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant