CN105187801A - Condensed video generation system and method - Google Patents

Condensed video generation system and method Download PDF

Info

Publication number
CN105187801A
CN105187801A CN201510600278.7A CN201510600278A CN105187801A CN 105187801 A CN105187801 A CN 105187801A CN 201510600278 A CN201510600278 A CN 201510600278A CN 105187801 A CN105187801 A CN 105187801A
Authority
CN
China
Prior art keywords
module
background
video
sigma
moving objects
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510600278.7A
Other languages
Chinese (zh)
Other versions
CN105187801B (en
Inventor
蔡晓东
陈超村
王丽娟
吕璐
刘馨婷
宋宗涛
王迪
甘凯今
杨超
赵勤鲁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
GUILIN TOPINTELLIGENT COMMUNICATION TECHNOLOGY Co Ltd
Original Assignee
GUILIN TOPINTELLIGENT COMMUNICATION TECHNOLOGY Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by GUILIN TOPINTELLIGENT COMMUNICATION TECHNOLOGY Co Ltd filed Critical GUILIN TOPINTELLIGENT COMMUNICATION TECHNOLOGY Co Ltd
Priority to CN201510600278.7A priority Critical patent/CN105187801B/en
Publication of CN105187801A publication Critical patent/CN105187801A/en
Application granted granted Critical
Publication of CN105187801B publication Critical patent/CN105187801B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention relates to a condensed video generation system and method. The system comprises a detection recognition tracking and extracting module, an optimizing and combining module, a video frame background module, a trajectory background pasting module, and a video condensing module. The detection recognition tracking and extracting module detects and recognizes the moving objects within a moving targeted area, tracks the detected and recognized moving objects and extracts the trajectories of the detected and recognized moving objects. The optimizing and combining module optimizes and combines the extracted moving trajectories of the moving objects. The video frame background module dynamically updates the video background of the moving targeted area. The trajectory background pasting module combines the moving objected with the updated video background in a matched pasting manner according to the combined and optimized moving trajectories in order to generate a condensed video frame. The video condensing module generates a condensed video according to the condensed video frame. In comparison with the prior art, the condensed video formed by the invention is shorter than an original video and most of information concerning the moving objects are still kept in the condensed video so that people can conveniently browse and the system and method are high in processing efficiency.

Description

A kind of generation system of summarized radio and method
Technical field
The present invention relates to field of video image processing, particularly relate to a kind of generation system and method for summarized radio.
Background technology
Along with the develop rapidly of monitoring technique and network technology, high-definition network monitoring camera is widely used in industry-by-industry.These high-definition camera all weather operations, create the monitor video data of magnanimity, but, because the limitation of monitoring human resources and time resource usually causes some massive video data from not processed.How long monitor video of browsing complete fast has become the current problem demanding prompt solution of monitoring trade; The important means that summarized radio solves exactly " massive video data process ".But, the traditional summarized radio generation method based on key frame adopt frame sampling cannot each object of complete representation movement locus and result in the loss of a large amount of useful video information.
Summary of the invention
Technical problem to be solved by this invention is to provide a kind of summarized radio shorter than original video, remains the information of most of Moving Objects in original video, is convenient to fast browsing, and the generation system of the high summarized radio for the treatment of effeciency and method.
The technical scheme that the present invention solves the problems of the technologies described above is as follows: a kind of generation system of summarized radio, comprises and detects recognition and tracking extraction module, Combinatorial Optimization module, frame of video background module, track background stickup module, summarized radio module;
Described detection recognition and tracking extraction module, for detecting the Moving Objects identified in motion target area, follows the tracks of and movement locus extraction detecting the Moving Objects identified;
Described Combinatorial Optimization module, for carrying out Combinatorial Optimization to the movement locus extracted;
Described frame of video background module, for dynamically updating the video background in motion target area;
Described track background stickup module, for according to the movement locus after Combinatorial Optimization by Moving Objects with upgrade after video background carry out mate pasting and merge, generation summarized radio frame;
Described summarized radio module, for generating summarized radio according to summarized radio frame.
The invention has the beneficial effects as follows: by detecting the running of recognition and tracking extraction module, Combinatorial Optimization module, frame of video background module, track background stickup module and summarized radio module coordination, generate the summarized radio shorter than original video, remain the information of most of Moving Objects in original video, be convenient to fast browsing, and treatment effeciency is high.
On the basis of technique scheme, the present invention can also do following improvement.
Further, described detection recognition and tracking extraction module identifies the Moving Objects in motion target area by setting up rank adaptive model; Described rank adaptive model is with the HOG (HistogramofGradient of cart headstock and dolly vehicle body image, finger direction histogram of gradients) feature is as Expressive Features, and in conjunction with Linear SVM (SupportVectorMachine, referring to SVMs) full features training sorting technique carries out the detection identification of vehicle vehicle, simultaneously, target vehicle region was extracted roughly by GMM (GaussianMixtureModel refers to gauss hybrid models) before HOG feature extraction.
The beneficial effect of above-mentioned further scheme is adopted to be: to establish rank adaptive model, ensure that, on the Demand Base that people and Che separately browse, the accuracy of detection and speed have all had larger lifting, and taken into account the robustness of reply scene changes, reduce operand, improve precision.
Further, the full features training sorting technique of described SVM is specially:
Can assign to say for linear two classes, hyperplane, i.e. a linear classifier be found, all training samples correctly can be classified, namely meet:
y i [ w T x i + b ] > 0 , ∀ i = 1 , ... , n
Wherein
Yi represents the class label giving input data, and xi represents i-th input HOG characteristic vector;
Due to the nonuniqueness of hyperplane, solve optimal hyperlane by the method for quadratic programming;
w * = Σ i = 1 n α i y i x i b * = y i - w * T x i
Wherein, α irepresent fixing Lagrange multiplier, w *represent the normal vector of optimal hyperlane, b *represent the side-play amount of optimal hyperlane.
Further, described detection recognition and tracking extraction module carries out moving object tracking by multiple-camera method for tracking target, and described multiple-camera method for tracking target is specially:
p x 1 c a m A / t ( x , y ) = H c a m B - c a m A × F o o t A c a m B / t ( x , y ) p x 2 c a m A / t ( x , y ) = H c a m B - c a m A × F o o t B c a m B / t ( x , y )
p x 1 c a m A / t ( x , y ) ∈ A r e a ( x ) p x 2 c a m A / t ( x , y ) ∈ A r e a ( x )
Wherein represent t, the coordinate position of minimum point coordinate in camB of target A, H camB-camAthe affine projection matrix of camB to camA scene, represent the coordinate of any point in t camA, Area (x) is ground regional extent shared by target.
The beneficial effect of above-mentioned further scheme is adopted to be: multiple-camera method for tracking target solves target and blocks or be separated the failed problem of rear tracking, has higher robustness; Multiple situations is there is in the same target of situation eliminating space-time confusion in a summarized radio picture.
Further, the video background of described frame of video background module dynamically updates and is specially: vote to video background at each Moving Objects track of a frame of video, selects the background of all Moving Objects tracks under original video environment.
Further, the video background after renewal and movement locus agglomerate background are carried out the pixel fusion algorithm process of Gaussian Profile by described track background stickup module.
Further, the image pixel blending algorithm of described Gaussian Profile is specially:
F(m,n)=ω 1P MV(m,n)+ω 2P BG(m,n)
In above formula, P mV(m, n) and P bG(m, n) is respectively the pixel value of the image BG of each Moving Objects MV in summarized radio frame and frame background of making a summary; M and n is respectively line number and the row number of pixel in image, and m=1,2 ..., M, n=1,2 ..., N; w 1, w 2for weight coefficient, and w 1+ w 2=1;
Corr ( B G , M V ) = Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] [ P B G ( m , n ) - P B G ‾ ] Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] 2 Σ m = 1 M Σ n = 1 N [ P B G ( m , n ) - P B G ‾ ] 2
In formula, Corr is the coefficient correlation of MV and BG two width source images, for the pixel average of source images MV, for the pixel average of source images BG;
ω 1 = 1 2 ( 1 - | C o r r | ) ω 2 = 1 - ω 1
G ( x , y ) = 1 2 πσ 2 e - ( x - x c ) 2 + ( y - y c ) 2 2 σ 2
Wherein, σ is the standard deviation of normal distribution, (x c, y c) for Gaussian function distribution barycenter.
The beneficial effect of above-mentioned further scheme is adopted to be: the image pixel blending algorithm of Gaussian Profile can eliminate splicing vestige, realizes the seamless fusion of the movement locus of summarized radio background and tracking; Retain the information of original image, eliminate deceptive information, obtain reliable and stable syncretizing effect.
A generation method for summarized radio, comprises the following steps:
Step S1. detects the Moving Objects identified in motion target area, follows the tracks of and movement locus extraction detecting the Moving Objects identified;
Step S2. carries out Combinatorial Optimization to the movement locus extracted; Video background in motion target area is dynamically updated;
Step S3. according to the movement locus after Combinatorial Optimization by Moving Objects with upgrade after video background carry out mate pasting and merge, generation summarized radio frame;
Step S4. generates summarized radio according to summarized radio frame.
Further, the specific implementation of the Moving Objects identified in motion target area is detected in described step S1: by setting up rank adaptive model, the Moving Objects in motion target area is identified; Described rank adaptive model is using the HOG feature of cart headstock and dolly vehicle body image as Expressive Features, and the detection identification of vehicle vehicle is carried out in conjunction with the full features training sorting technique of Linear SVM, meanwhile, before HOG feature extraction, target vehicle region is extracted roughly by GMM.
Further, the specific implementation of described step S3: the pixel fusion algorithm process of the video background after renewal and movement locus agglomerate background being carried out Gaussian Profile; The image pixel blending algorithm of described Gaussian Profile is specially:
F(m,n)=ω 1P MV(m,n)+ω 2P BG(m,n)
In above formula, P mV(m, n) and P bG(m, n) is respectively the pixel value of the image BG of each Moving Objects MV in summarized radio frame and frame background of making a summary; M and n is respectively line number and the row number of pixel in image, and m=1,2 ..., M, n=1,2 ..., N; w 1, w 2for weight coefficient, and w 1+ w 2=1;
Corr ( B G , M V ) = Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] [ P B G ( m , n ) - P B G ‾ ] Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] 2 Σ m = 1 M Σ n = 1 N [ P B G ( m , n ) - P B G ‾ ] 2
In formula, Corr is the coefficient correlation of MV and BG two width source images, for the pixel average of source images MV, for the pixel average of source images BG;
ω 1 = 1 2 ( 1 - | C o r r | ) ω 2 = 1 - ω 1
G ( x , y ) = 1 2 πσ 2 e - ( x - x c ) 2 + ( y - y c ) 2 2 σ 2
Wherein, σ is the standard deviation of normal distribution, (x c, y c) for Gaussian function distribution barycenter.
The invention has the beneficial effects as follows: the inventive method proposes a kind of target detection recognition methods based on rank adaptive model, and propose a kind of multiple-camera method for tracking target based on event analysis, in the process that monitoring summarized radio generates, by voting in the best background of all Moving Objects tracks under original video environment that can reflect and participate in merging, carry out dynamically updating of summarized radio background; Meanwhile, adopt rectangle Gauss weight allocation strategy, well achieve the seamless pixel fusion effect of summarized radio frame background and Moving Objects track.
Accompanying drawing explanation
Fig. 1 is the module frame chart of the generation system of a kind of summarized radio of the present invention;
Fig. 2 is the flow chart of the generation method of a kind of summarized radio of the present invention;
Fig. 3 is the algorithm of target detection schematic diagram based on rank adaptive model;
Fig. 4 is that two video cameras carry out target tracking algorism flow chart;
Fig. 5 is sdi video target profile;
Fig. 6 is sdi video goal programming distribution map.
In accompanying drawing, the list of parts representated by each label is as follows:
1, recognition and tracking extraction module is detected, 2, Combinatorial Optimization module, 3, frame of video background module, 4, track background stickup module, 5, summarized radio module.
Embodiment
Be described principle of the present invention and feature below in conjunction with accompanying drawing, example, only for explaining the present invention, is not intended to limit scope of the present invention.
As shown in Figure 1, a kind of generation system of summarized radio, comprises and detects recognition and tracking extraction module 1, Combinatorial Optimization module 2, frame of video background module 3, track background stickup module 4, summarized radio module 5;
Described detection recognition and tracking extraction module 1, for detecting the Moving Objects identified in motion target area, follows the tracks of and movement locus extraction detecting the Moving Objects identified;
Described Combinatorial Optimization module 2, for carrying out Combinatorial Optimization to the movement locus extracted;
Described frame of video background module 3, for dynamically updating the video background in motion target area;
Described track background stickup module 4, for according to the movement locus after Combinatorial Optimization by Moving Objects with upgrade after video background carry out mate pasting and merge, generation summarized radio frame;
Described summarized radio module 5, for generating summarized radio according to summarized radio frame.
Preferably, described detection recognition and tracking extraction module 1 identifies the Moving Objects in motion target area by setting up rank adaptive model; Described rank adaptive model is using the HOG feature of cart headstock and dolly vehicle body image as Expressive Features, and the detection identification of vehicle vehicle is carried out in conjunction with the full features training sorting technique of Linear SVM, meanwhile, before HOG feature extraction, target vehicle region is extracted roughly by GMM.
Preferably, the full features training sorting technique of described SVM is specially:
Can assign to say for linear two classes, hyperplane, i.e. a linear classifier be found, all training samples correctly can be classified, namely meet:
y i [ w T x i + b ] > 0 , ∀ i = 1 , ... , n
Wherein
Yi represents the class label giving input data, and xi represents i-th input HOG characteristic vector;
Due to the nonuniqueness of hyperplane, solve optimal hyperlane by the method for quadratic programming;
w * = Σ i = 1 n α i y i x i b * = y i - w * T x i
Wherein, α irepresent fixing Lagrange multiplier, w *represent the normal vector of optimal hyperlane, b *represent the side-play amount of optimal hyperlane.
Preferably, described detection recognition and tracking extraction module 1 carries out moving object tracking by multiple-camera method for tracking target, and described multiple-camera method for tracking target is specially:
p x 1 c a m A / t ( x , y ) = H c a m B - c a m A × F o o t A c a m B / t ( x , y ) p x 2 c a m A / t ( x , y ) = H c a m B - c a m A × F o o t B c a m B / t ( x , y )
p x 1 c a m A / t ( x , y ) ∈ A r e a ( x ) p x 2 c a m A / t ( x , y ) ∈ A r e a ( x )
Wherein represent t, the coordinate position of minimum point coordinate in camB of target A, H camB-camAthe affine projection matrix of camB to camA scene, represent the coordinate of any point in t camA, Area (x) is ground regional extent shared by target.
Preferably, the video background of described frame of video background module 3 dynamically updates and is specially: vote to video background at each Moving Objects track of a frame of video, selects the background of all Moving Objects tracks under original video environment.
Preferably, the video background after renewal and movement locus agglomerate background are carried out the pixel fusion algorithm process of Gaussian Profile by described track background stickup module 4.
Preferably, the image pixel blending algorithm of described Gaussian Profile is specially:
F(m,n)=ω 1P MV(m,n)+ω 2P BG(m,n)
In above formula, P mV(m, n) and P bG(m, n) is respectively the pixel value of the image BG of each Moving Objects MV in summarized radio frame and frame background of making a summary; M and n is respectively line number and the row number of pixel in image, and m=1,2 ..., M, n=1,2 ..., N; w 1, w 2for weight coefficient, and w 1+ w 2=1;
Corr ( B G , M V ) = Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] [ P B G ( m , n ) - P B G ‾ ] Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] 2 Σ m = 1 M Σ n = 1 N [ P B G ( m , n ) - P B G ‾ ] 2
In formula, Corr is the coefficient correlation of MV and BG two width source images, for the pixel average of source images MV, for the pixel average of source images BG;
ω 1 = 1 2 ( 1 - | C o r r | ) ω 2 = 1 - ω 1
G ( x , y ) = 1 2 πσ 2 e - ( x - x c ) 2 + ( y - y c ) 2 2 σ 2
Wherein, σ is the standard deviation of normal distribution, (x c, y c) for Gaussian function distribution barycenter.
As shown in Figure 2, a kind of generation method of summarized radio, comprises the following steps:
Step S1. detects the Moving Objects identified in motion target area, follows the tracks of and movement locus extraction detecting the Moving Objects identified;
Step S2. carries out Combinatorial Optimization to the movement locus extracted; Video background in motion target area is dynamically updated;
Step S3. according to the movement locus after Combinatorial Optimization by Moving Objects with upgrade after video background carry out mate pasting and merge, generation summarized radio frame;
Step S4. generates summarized radio according to summarized radio frame.
Preferably, the specific implementation of the Moving Objects identified in motion target area is detected in described step S1: by setting up rank adaptive model, the Moving Objects in motion target area is identified; Described rank adaptive model is using the HOG feature of cart headstock and dolly vehicle body image as Expressive Features, and the detection identification of vehicle vehicle is carried out in conjunction with the full features training sorting technique of Linear SVM, meanwhile, before HOG feature extraction, target vehicle region is extracted roughly by GMM.
Preferably, the specific implementation of described step S3: the pixel fusion algorithm process of the video background after renewal and movement locus agglomerate background being carried out Gaussian Profile; The image pixel blending algorithm of described Gaussian Profile is specially:
F(m,n)=ω 1P MV(m,n)+ω 2P BG(m,n)
In above formula, P mV(m, n) and P bG(m, n) is respectively the pixel value of the image BG of each Moving Objects MV in summarized radio frame and frame background of making a summary; M and n is respectively line number and the row number of pixel in image, and m=1,2 ..., M, n=1,2 ..., N; w 1, w 2for weight coefficient, and w 1+ w 2=1;
Corr ( B G , M V ) = Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] [ P B G ( m , n ) - P B G ‾ ] Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] 2 Σ m = 1 M Σ n = 1 N [ P B G ( m , n ) - P B G ‾ ] 2
In formula, Corr is the coefficient correlation of MV and BG two width source images, for the pixel average of source images MV, for the pixel average of source images BG;
ω 1 = 1 2 ( 1 - | C o r r | ) ω 2 = 1 - ω 1
G ( x , y ) = 1 2 πσ 2 e - ( x - x c ) 2 + ( y - y c ) 2 2 σ 2
Wherein, σ is the standard deviation of normal distribution, (x c, y c) for Gaussian function distribution barycenter.
Embodiment 2:
The track of Moving Objects be according to early stage moving target detection identification and track and extract obtain, wherein, moving object detection cognitive phase is the target detection identification done after Gaussian Background rebuilds extraction motion target area, a grader will identify a kind of object, and then be the track combination optimization by the simulated annealing improved, make the whole track energy of all objects minimum, finally will obtain pedestrian's summary that people and Che separate and vehicle is made a summary.
Traditional monitoring summarized radio, when carrying out the choosing of summarized radio background, selects a Gauss model set up original video often.But in the monitor video scene of reality, the background of original video is change, such as, have a static vehicle at current video frame, resting is known from experience by as background, along with the movement of time shaft, this static vehicle there occurs motion, and it is no longer just background.If choose be initial stationary vehicle time background, the background of this original video namely extracted, when make an abstract frame background, cannot reflect real video monitoring scene like this.Based on addressing this problem, propose a kind of strategy that can dynamically update the background of summary frame herein, the background making each summary frame can be the background meeting the movement locus of object in original video.
When object trajectory and background are directly pasted, the background of following the tracks of object agglomerate and the summary frame obtained is not mated, and has and significantly splices vestige.Based on addressing this problem, devising a kind of Trace Formation strategy based on Gaussian Profile herein, reaching the seamless spliced effect of Moving Objects and frame of video background.
As shown in Figure 3 in the algorithm of target detection schematic diagram of rank adaptive model, the HOG feature that have selected cart headstock and dolly vehicle body image is as Expressive Features and carry out the detection identification of vehicle vehicle in conjunction with the full features training sorting technique of Linear SVM, simultaneously, target vehicle region was extracted roughly by GMM (GaussianMixtureModel) before HOG feature extraction, thus establish rank adaptive model, reduce operand, improve precision.
Target tracking algorism flow chart as shown in Figure 4, the method to be applicable in multiple-camera video camera between two and to carry out the situation of multiple target tracking.
Sdi video target profile as shown in Figure 5, transverse axis represents the initial time of this target, and the longitudinal axis represents each target.Very dense at summarized radio zone line, and all can there is a sparse region in beginning and end, obviously do not meet and be uniformly distributed, the present invention adopts the Kd-Trees method based on cluster, balance the moving target number of each frame in summarized radio, avoid target too concentrated in a certain frame, to such an extent as to be difficult to target not to be fused to overlappingly in summarized radio.Adopt the method, can also fall and the moving target of starting point in adjacent area may be gathered a class, be allowed to condition in summarized radio and in chronological sequence sequentially occur in video streaming.Like the scene dispatched a car in a track, preceding vehicle is first left, after vehicle can not have when overlaid pixel with preceding vehicle and dispatch a car again.So just can avoid not the conflicting in summarized radio of moving target, overlapping or intersect, effectively reducing the probability of conflict.
Sdi video goal programming distribution map as shown in Figure 6, transverse axis represents the initial time of this target, and the longitudinal axis represents each target.With Fig. 5 unlike, when total line number is certain, there is multiple target every provisional capital.Can ensure that the number of targets that each frame occurs is consistent like this, ensure that moving target being uniformly distributed in each frame, make full use of the space of each frame, effectively shorten the length of summarized radio.
Table 1
Conventional method as shown in table 1 and the contrast of to improve one's methods herein, when track number, primary power are identical with termination energy condition, the simulated annealing method convergence rate improved herein is higher than traditional simulated annealing.This is owing to first having carried out clustering planning analysis herein, and the initial distribution making track is enough even, has then carried out again that energy based on simulated annealing is minimum to be solved.Initial distribution enough evenly greatly speeded the convergence rate of simulated annealing.
Do not lose problem for the fast browsing of monitor video and most of action message, in conjunction with the feature of monitor video itself, propose a kind of summarized radio generation technique framework based on object trajectory, and achieve the key technology in this technological frame.
The foregoing is only preferred embodiment of the present invention, not in order to limit the present invention, within the spirit and principles in the present invention all, any amendment done, equivalent replacement, improvement etc., all should be included within protection scope of the present invention.

Claims (10)

1. a generation system for summarized radio, is characterized in that: comprise and detect recognition and tracking extraction module (1), Combinatorial Optimization module (2), frame of video background module (3), track background stickup module (4), summarized radio module (5);
Described detection recognition and tracking extraction module (1), for detecting the Moving Objects identified in motion target area, follows the tracks of and movement locus extraction detecting the Moving Objects identified;
Described Combinatorial Optimization module (2), for carrying out Combinatorial Optimization to the movement locus extracted;
Described frame of video background module (3), for dynamically updating the video background in motion target area;
Described track background stickup module (4), for according to the movement locus after Combinatorial Optimization by Moving Objects with upgrade after video background carry out mate pasting and merge, generation summarized radio frame;
Described summarized radio module (5), for generating summarized radio according to summarized radio frame.
2. the generation system of a kind of summarized radio according to claim 1, is characterized in that: described detection recognition and tracking extraction module (1) identifies the Moving Objects in motion target area by setting up rank adaptive model; Described rank adaptive model is using the HOG feature of cart headstock and dolly vehicle body image as Expressive Features, and the detection identification of vehicle vehicle is carried out in conjunction with the full features training sorting technique of Linear SVM, meanwhile, before HOG feature extraction, target vehicle region is extracted roughly by GMM.
3. the generation system of a kind of summarized radio according to claim 2, it is characterized in that: described SVM full features training sorting technique is specially: can assign to say for linear two classes, find hyperplane, i.e. a linear classifier, all training samples correctly can be classified, namely meet:
y i [ w T x i + b ] > 0 , ∀ i = 1 , ... , n
Wherein
Yi represents the class label giving input data, and xi represents i-th input HOG characteristic vector;
Due to the nonuniqueness of hyperplane, solve optimal hyperlane by the method for quadratic programming;
w * = Σ i = 1 n α i y i x i b * = y i - w * T x i
Wherein, α irepresent fixing Lagrange multiplier, w *represent the normal vector of optimal hyperlane, b *represent the side-play amount of optimal hyperlane.
4. the generation system of a kind of summarized radio according to claim 1, it is characterized in that: described detection recognition and tracking extraction module (1) carries out moving object tracking by multiple-camera method for tracking target, and described multiple-camera method for tracking target is specially:
p x 1 c a m A / t ( x , y ) = H c a m B - c a m A × F o o t A c a m B / t ( x , y ) p x 2 c a m A / t ( x , y ) = H c a m B - c a m A × F o o t B c a m B / t ( x , y )
p x 1 c a m A / t ( x , y ) ∈ A r e a ( x ) p x 2 c a m A / t ( x , y ) ∈ A r e a ( x )
Wherein represent the coordinate position of minimum point coordinate in camB of t target A, H camB-camAthe affine projection matrix of camB to camA scene, represent the coordinate of any point in t camA, Area (x) is ground regional extent shared by target.
5. the generation system of a kind of summarized radio according to claim 1, it is characterized in that: the video background of described frame of video background module (3) dynamically updates and is specially: at each Moving Objects track of a frame of video, video background is voted, select the background of all Moving Objects tracks under original video environment.
6. a kind of generation system of summarized radio according to any one of claim 1 to 5, is characterized in that: the video background after upgrading and movement locus agglomerate background are carried out the pixel fusion algorithm process of Gaussian Profile by described track background stickup module (4).
7. the generation system of a kind of summarized radio according to claim 6, is characterized in that: the image pixel blending algorithm of described Gaussian Profile is specially:
F(m,n)=ω 1P MV(m,n)+ω 2P BG(m,n)
In above formula, P mV(m, n) and P bG(m, n) is respectively the pixel value of the image BG of each Moving Objects MV in summarized radio frame and frame background of making a summary; M and n is respectively line number and the row number of pixel in image, and m=1,2 ..., M, n=1,2 ..., N; w 1, w 2for weight coefficient, and w 1+ w 2=1;
Corr ( B G , M V ) = Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] [ P B G ( m , n ) - P B G ‾ ] Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] 2 Σ m = 1 M Σ n = 1 N [ P B G ( m , n ) - P B G ‾ ] 2
In formula, Corr is the coefficient correlation of MV and BG two width source images, the pixel average of source images MV, for the pixel average of source images BG;
ω 1 = 1 2 ( 1 - | C o r r | ) ω 2 = 1 - ω 1
G ( x , y ) = 1 2 πσ 2 e - ( x - x c ) 2 + ( y - y c ) 2 2 σ 2
Wherein, σ is the standard deviation of normal distribution, (x c, y c) for Gaussian function distribution barycenter.
8. a generation method for summarized radio, is characterized in that, comprise the following steps:
Step S1. detects the Moving Objects identified in motion target area, follows the tracks of and movement locus extraction detecting the Moving Objects identified;
Step S2. carries out Combinatorial Optimization to the movement locus extracted; Video background in motion target area is dynamically updated;
Step S3. according to the movement locus after Combinatorial Optimization by Moving Objects with upgrade after video background carry out mate pasting and merge, generation summarized radio frame;
Step S4. generates summarized radio according to summarized radio frame.
9. the generation method of a kind of summarized radio according to claim 8, it is characterized in that, detecting the specific implementation of the Moving Objects identified in motion target area in described step S1: by setting up rank adaptive model, the Moving Objects in motion target area is identified; Described rank adaptive model is using the HOG feature of cart headstock and dolly vehicle body image as Expressive Features, and the detection identification of vehicle vehicle is carried out in conjunction with the full features training sorting technique of Linear SVM, meanwhile, before HOG feature extraction, target vehicle region is extracted roughly by GMM.
10. the generation method of a kind of summarized radio according to claim 8 or claim 9, is characterized in that, the specific implementation of described step S3: the pixel fusion algorithm process of the video background after upgrading and movement locus agglomerate background being carried out Gaussian Profile; The image pixel blending algorithm of described Gaussian Profile is specially:
F(m,n)=ω 1P MV(m,n)+ω 2P BG(m,n)
In above formula, P mV(m, n) and P bG(m, n) is respectively the pixel value of the image BG of each Moving Objects MV in summarized radio frame and frame background of making a summary; M and n is respectively line number and the row number of pixel in image, and m=1,2 ..., M, n=1,2 ..., N; w 1, w 2for weight coefficient, and w 1+ w 2=1;
Corr ( B G , M V ) = Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] [ P B G ( m , n ) - P B G ‾ ] Σ m = 1 M Σ n = 1 N [ P M V ( m , n ) - P M V ‾ ] 2 Σ m = 1 M Σ n = 1 N [ P B G ( m , n ) - P B G ‾ ] 2
In formula, Corr is the coefficient correlation of MV and BG two width source images, for the pixel average of source images MV, for the pixel average of source images BG;
ω 1 = 1 2 ( 1 - | C o r r | ) ω 2 = 1 - ω 1
G ( x , y ) = 1 2 πσ 2 e - ( x - x c ) 2 + ( y - y c ) 2 2 σ 2
Wherein, σ is the standard deviation of normal distribution, (x c, y c) for Gaussian function distribution barycenter.
CN201510600278.7A 2015-09-17 2015-09-17 System and method for generating abstract video Active CN105187801B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510600278.7A CN105187801B (en) 2015-09-17 2015-09-17 System and method for generating abstract video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510600278.7A CN105187801B (en) 2015-09-17 2015-09-17 System and method for generating abstract video

Publications (2)

Publication Number Publication Date
CN105187801A true CN105187801A (en) 2015-12-23
CN105187801B CN105187801B (en) 2021-07-27

Family

ID=54909623

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510600278.7A Active CN105187801B (en) 2015-09-17 2015-09-17 System and method for generating abstract video

Country Status (1)

Country Link
CN (1) CN105187801B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106101487A (en) * 2016-07-04 2016-11-09 石家庄铁道大学 Video spatiotemporal motion track extraction method
CN108364006A (en) * 2018-01-17 2018-08-03 超凡影像科技股份有限公司 Medical Images Classification device and its construction method based on multi-mode deep learning
CN110166851A (en) * 2018-08-21 2019-08-23 腾讯科技(深圳)有限公司 A kind of video abstraction generating method, device and storage medium
US20210390713A1 (en) * 2020-06-12 2021-12-16 Beijing Didi Infinity Technology And Development Co., Ltd. Systems and methods for performing motion transfer using a learning model
CN117726821A (en) * 2024-02-05 2024-03-19 武汉理工大学 Medical behavior identification method for region shielding in medical video

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080269924A1 (en) * 2007-04-30 2008-10-30 Huang Chen-Hsiu Method of summarizing sports video and apparatus thereof
US20120293686A1 (en) * 2011-05-18 2012-11-22 Keith Stoll Karn Video summary including a feature of interest
CN103079117A (en) * 2012-12-30 2013-05-01 信帧电子技术(北京)有限公司 Video abstract generation method and video abstract generation device
CN103413330A (en) * 2013-08-30 2013-11-27 中国科学院自动化研究所 Method for reliably generating video abstraction in complex scene

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080269924A1 (en) * 2007-04-30 2008-10-30 Huang Chen-Hsiu Method of summarizing sports video and apparatus thereof
US20120293686A1 (en) * 2011-05-18 2012-11-22 Keith Stoll Karn Video summary including a feature of interest
CN103079117A (en) * 2012-12-30 2013-05-01 信帧电子技术(北京)有限公司 Video abstract generation method and video abstract generation device
CN103413330A (en) * 2013-08-30 2013-11-27 中国科学院自动化研究所 Method for reliably generating video abstraction in complex scene

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
唐铭谦: "基于对象的监控视频摘要算法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
王雪锋,叶飞: "一种基于HOG与SVM的监控视频车辆识别方法", 《微型机及应用》 *
蔡晓东,吴迪等: "基于事件分析的多摄像机鲁棒跟踪算法", 《电视技术》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106101487A (en) * 2016-07-04 2016-11-09 石家庄铁道大学 Video spatiotemporal motion track extraction method
CN106101487B (en) * 2016-07-04 2018-05-08 石家庄铁道大学 Video spatiotemporal motion track extraction method
CN108364006A (en) * 2018-01-17 2018-08-03 超凡影像科技股份有限公司 Medical Images Classification device and its construction method based on multi-mode deep learning
CN108364006B (en) * 2018-01-17 2022-03-08 超凡影像科技股份有限公司 Medical image classification device based on multi-mode deep learning and construction method thereof
CN110166851A (en) * 2018-08-21 2019-08-23 腾讯科技(深圳)有限公司 A kind of video abstraction generating method, device and storage medium
US20210390713A1 (en) * 2020-06-12 2021-12-16 Beijing Didi Infinity Technology And Development Co., Ltd. Systems and methods for performing motion transfer using a learning model
WO2021248432A1 (en) * 2020-06-12 2021-12-16 Beijing Didi Infinity Technology And Development Co., Ltd. Systems and methods for performing motion transfer using a learning model
US11830204B2 (en) * 2020-06-12 2023-11-28 Beijing Didi Infinity Technology And Development Co., Ltd. Systems and methods for performing motion transfer using a learning model
CN117726821A (en) * 2024-02-05 2024-03-19 武汉理工大学 Medical behavior identification method for region shielding in medical video
CN117726821B (en) * 2024-02-05 2024-05-10 武汉理工大学 Medical behavior identification method for region shielding in medical video

Also Published As

Publication number Publication date
CN105187801B (en) 2021-07-27

Similar Documents

Publication Publication Date Title
Chen et al. An edge traffic flow detection scheme based on deep learning in an intelligent transportation system
US10699170B2 (en) Apparatuses and methods for semantic image labeling
Mnih et al. Learning to label aerial images from noisy data
CN105187801A (en) Condensed video generation system and method
CN101799968B (en) Detection method and device for oil well intrusion based on video image intelligent analysis
CN111161315B (en) Multi-target tracking method and system based on graph neural network
Ren et al. A novel squeeze YOLO-based real-time people counting approach
CN105354565A (en) Full convolution network based facial feature positioning and distinguishing method and system
Gerke et al. Soccer player recognition using spatial constellation features and jersey number recognition
Zhang et al. Coarse-to-fine object detection in unmanned aerial vehicle imagery using lightweight convolutional neural network and deep motion saliency
Le et al. Toward interactive self-annotation for video object bounding box: Recurrent self-learning and hierarchical annotation based framework
KR20230171966A (en) Image processing method and device and computer-readable storage medium
CN107609509A (en) A kind of action identification method based on motion salient region detection
CN114913206A (en) Multi-target tracking method and system based on multi-mode fusion
Gu et al. Embedded and real-time vehicle detection system for challenging on-road scenes
Sachdeva et al. Rank2tell: A multimodal driving dataset for joint importance ranking and reasoning
Hammam et al. Real-time multiple spatiotemporal action localization and prediction approach using deep learning
CN103500456A (en) Object tracking method and equipment based on dynamic Bayes model network
Li et al. Self-supervised social relation representation for human group detection
Haggui et al. Centroid human tracking via oriented detection in overhead fisheye sequences
Bewley et al. Background Appearance Modeling with Applications to Visual Object Detection in an Open‐Pit Mine
Nguyen et al. You always look again: Learning to detect the unseen objects
CN112651294A (en) Method for recognizing human body shielding posture based on multi-scale fusion
CN116152747A (en) Human behavior intention recognition method based on appearance recognition and action modeling
CN106530319A (en) Video object cooperative segmentation method based on track directed graph

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20151223

CI02 Correction of invention patent application
CI02 Correction of invention patent application

Correction item: rejection of application for invention after its publication

Correct: Dismiss

False: reject

Number: 11-02

Volume: 37

GR01 Patent grant
GR01 Patent grant