CN103945227A - Video semantic block partition method based on light stream clustering - Google Patents
Video semantic block partition method based on light stream clustering Download PDFInfo
- Publication number
- CN103945227A CN103945227A CN201410153245.8A CN201410153245A CN103945227A CN 103945227 A CN103945227 A CN 103945227A CN 201410153245 A CN201410153245 A CN 201410153245A CN 103945227 A CN103945227 A CN 103945227A
- Authority
- CN
- China
- Prior art keywords
- consistent
- moving region
- light stream
- cluster
- pixel
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Landscapes
- Image Analysis (AREA)
Abstract
The invention provides a video semantic block partition method based on light stream clustering. The detection result of consistent motion areas under the same motion mode and the light stream information are combined, a relevant matrix is built, spectral clustering is carried out, labels obtained after clustering are used as feature vectors, and secondary clustering is carried out on different motion modes. According to the method, the detection result of the consistent motion areas can be well applied, and the relevant information of the scene motion under the same motion mode and the global information of the scene motion under the different motion modes can be effectively mastered. The information of the scene motion is mastered accurately, and the method plays an important role in improving the accuracy rate of relevant applications like mode recognition.
Description
Technical field
The present invention relates to a kind of video semanteme piece cutting techniques, relate in particular to the video semanteme piece cutting techniques on light stream cluster basis, that is: carry out scene by the light stream cluster to video and cut apart, then use segmentation result to continue to obtain the piecemeal that motor pattern is more easily expressed.
Background technology
Semantic chunk in video refers to the consistent region of motor pattern.It is very important that video semanteme piece is segmented in image processing field, and it is widely used at aspects such as traffic monitoring, dense population monitoring.
Consistent moving region detection technique is a very important technology in video and image processing field, the method detecting about consistent moving region has a lot, through the literature search of prior art is found, as, high density that what paper " A lagrangian particle dynamics approach for crowd flow segmentation and stability analysis " that the people such as S.Ali deliver for 2007 on " Conference on Computer Vision and Pattern Recognition " proposed the cut apart consistent moving region of trooping.But these methods are cut apart mainly for a kind of stable motor pattern.If hold the different motion pattern of video Same Scene, these methods are far from being enough.As there being the motor patterns such as left-hand rotation, right-hand rotation, craspedodrome in traffic video, every kind of pattern has the consistent moving region of different vehicles, between many regions, overlaps each other, and is not suitable for right-hand rotation scene or craspedodrome scene for the division of the consistent moving region of left-hand rotation scene.Dividing semantic chunk, is the world subdivision that obtains more reasonably, more fully describing Same Scene different motion pattern, and it has very important effect to work such as Describing Motion pattern, identifications.
In addition the paper " Multi-camera activity correlation analysis " that, the people such as C.C.Loy delivers for 2007 on " Conference on Computer Vision and Pattern Recognition " has proposed a kind of method of auto Segmentation semantic chunk.The method is mainly found semantic chunk by the correlation of foreground moving, and its subject matter is, it mainly considers the motor pattern correlation of consecutive frame, neighbor, has no idea to hold the global characteristics of scene.
Therefore,, for above variety of problems, we need to find a kind of different time motor pattern of can holding and can grasp again the method for overall different pixels movement relation and divide semantic chunk.
Summary of the invention
For defect of the prior art, the object of this invention is to provide a kind of video semanteme piece dividing method based on light stream cluster.
According to a kind of video semanteme piece dividing method based on light stream cluster provided by the invention, comprise following steps:
Step 1: the optical flow field of the video that input contains different motion pattern, wherein, described video has T frame, every frame is containing M × N pixel, M, N are respectively the length and width of every two field picture, total T-1 of described optical flow field, and each optical flow field is two two-dimensional matrixs that size is M × N, one is x direction, and one is y direction;
Step 2: carry out cluster for each optical flow field in T-1 optical flow field, obtain U consistent moving region;
Step 3: consistent U moving region is classified as to S (S<U) consistent moving region group, and indicate each consistent moving region as area classification label (if Fig. 2 is the unanimously signal of area classification label in different optical flow fields) with the label of consistent moving region group;
Step 4: utilize the area classification label that step 3 obtains to carry out construction feature vector to each pixel in M × N pixel, obtain altogether M × N characteristic vector C
1, C
2..., C
m × N; The characteristic vector of m pixel is by cascading up the label of the consistent moving region group under m is in each optical flow field to obtain;
Step 5: according to characteristic vector C
1, C
2..., C
m × N, M × N point carried out to the semantic chunk that K-means cluster obtains splitting.
Preferably, described step 3 comprises the steps:
Step a) is established in T-1 optical flow field, and a total U consistent moving region, carries out given sequence number to each consistent moving region, and sequence number is 1~U, and sets up U × 2 dimension matrix F;
Step b), for each consistent moving region, is calculated its average light stream, puts into the relevant position of F, wherein; As, the consistent moving region (1≤h≤U) that in i group, label is h, the light stream of x direction and the light stream of y direction of points all in the h of i optical flow field consistent moving region are averaged respectively, and mean value is distinguished to assignment to F (h, 1) with F (h, 2), F (h, 1) be respectively h unanimously x durection component and the y durection component of the average light stream in moving region with F (h, 2);
It is the two-dimensional matrix W of U × U that step c) is set up size, wherein, W (i, j) be F (i, 1) F (j, 1)+F (i, 2) F (j, 2), W (i, j) W (i, j) be i consistent moving region and j consistent moving region motion relevance, F (i, 1) be the x durection component of i the consistent average light stream in moving region, F (j, 1) be the x durection component of j the consistent average light stream in moving region, F (i, 2) be the y durection component of j the consistent average light stream in moving region, F (j, 2) be the y durection component of j the consistent average light stream in moving region, oeprator represents multiplication sign,
Step d) utilizes Spectral Clustering to carry out cluster to matrix W, forms S cluster, and with the each consistent moving region of cluster label sign corresponding to this S cluster, indicates U consistent moving region with S label.
Preferably, the characteristic vector of m pixel described in step 4 obtains by the following method:
If m pixel affiliated consistent moving region category label in t optical flow field is h
m,t, 1≤h
m,t≤ S, 1≤t≤T-1; By area classification label h
m,tassignment is to C
m(t), i.e. C
m(t)=h
m,t, wherein, C
mbe the characteristic vector of m pixel, C
m(t) be that m pixel characteristic is to a flow control t element.
Compared with prior art, the present invention has following beneficial effect:
The invention discloses a kind of video semanteme piece dividing method detecting based on consistent moving region, the result that it detects according to the consistent moving region under same motor pattern is combined with Optic flow information, build correlation matrix, and carry out spectral clustering cluster, and the label that cluster is obtained is as characteristic vector, and different motion pattern is carried out to secondary cluster.
The present invention can apply consistent moving region testing result well, effectively holds relevant information under the same movement pattern of scene motion and the global information of different motion pattern.The present invention holds accurately scene motion information, to the raising important role of the related application accuracys rate such as pattern recognition.
Brief description of the drawings
By reading the detailed description of non-limiting example being done with reference to the following drawings, it is more obvious that other features, objects and advantages of the present invention will become:
Fig. 1 is the consistent moving region of certain frame optical flow field in input video;
Fig. 2 utilizes consistent region group to indicate consistent region, and represents the schematic diagram of the characteristic vector of a pixel with the label of consistent region group;
Fig. 3 is the Scene Semantics piece schematic diagram of finally cutting apart.
Embodiment
Below in conjunction with specific embodiment, the present invention is described in detail.Following examples will contribute to those skilled in the art further to understand the present invention, but not limit in any form the present invention.It should be pointed out that to those skilled in the art, without departing from the inventive concept of the premise, can also make some distortion and improvement.These all belong to protection scope of the present invention.
The present embodiment is implemented under taking technical solution of the present invention as prerequisite, provided detailed execution mode and concrete operating process, but protection scope of the present invention is not limited to following embodiment.
The present embodiment comprises the following steps:
Step 1: obtain input video (total T frame, every frame M × N pixel, M, N are respectively the length and width of every two field picture) optical flow field (altogether T-1 field, each is two two-dimensional matrixs that size is M × N, one is x direction, one is y direction) with consistent moving region (altogether T-1 segmentation result, in each segmentation result, have several consistent regions).
The method that the paper " High accuracy optical flow estimation based on a theory for warping " that described optical flow field can be delivered by T.Brox etc. for 2004 on " European Conference on Computer Vision " proposes obtains.
Step 2: cluster is carried out in the light stream for each optical flow field in T-1 field, obtains U (U>T-1) consistent moving region altogether.
The method that described cluster can be published in by J.Hartigan etc. " the A K-Means Clustering Algorithm " proposition on " Journal of the Royal Statistical Society " for 1979 obtains.
Step 3: the U obtaining in step 2 consistent moving region sorted out, and obtain the consistent moving region of S (S<U) group, and indicate respectively each consistent moving region with the label of consistent moving region group, concrete steps are as follows:
A) establish in T-1 optical flow field, a total U consistent moving region, carries out given sequence number to the moving region of each consistent wish, and sequence number is 1~U.And set up U × 2 and tie up matrix F.
B) for each consistent region, calculate its average light stream, put into the relevant position of F.As, the consistent region (1≤h≤U) that in i group, label is h, averages the light stream of x direction and the light stream of y direction of points all in the h of i optical flow field consistent moving region respectively, and mean value is distinguished to assignment to F (h, 1) with F (h, 2).
C) setting up size is the two-dimensional matrix W of U × U.Wherein, W (i, j) is F (i, 1) F (j, 1)+F (i, 2) F (j, 2).
D) utilize Spectral Clustering to carry out cluster to matrix W, form S cluster, and with the each consistent moving region of cluster label sign corresponding to this S cluster, can indicate U consistent moving region with S label.
Wherein step 3 d) " Self-tuning spectral clustering " that described spectral clustering can be published in by L.Zelnik-Manor etc. on " Neural Information Processing Systems " for 2004 realize.
Step 4: utilize area classification label that step 3 obtains to each the pixel construction feature vector in M × N pixel, obtain altogether M × N characteristic vector C
1, C
2..., C
n × N, for representing the kinetic characteristic of each pixel at whole video.
The characteristic vector of m described pixel can be by cascading up the label of the consistent region group under m is in each optical flow field to obtain.Establishing m pixel affiliated consistent region labeling in t optical flow field (1≤t≤T-1) is h
m,t(11h
m,t≤ S).Be h by region labeling
m,tarea classification label (be L (h
m,t)) assignment is to C
m(t), i.e. C
m=[C
m(t)]=[L (h
m,t)].
Step 5: according to characteristic vector C
1, C
2..., C
m × N, utilize the methods such as K-means to carry out cluster to M × N point, the semantic chunk that can obtain splitting.
The method that described cluster can be published in by J.Hartigan etc. " the A K-Means Clustering Algorithm " proposition on " Journal of the Royal Statistical Society " for 1979 obtains.
Above specific embodiments of the invention are described.It will be appreciated that, the present invention is not limited to above-mentioned specific implementations, and those skilled in the art can make various distortion or amendment within the scope of the claims, and this does not affect flesh and blood of the present invention.
Claims (3)
1. the video semanteme piece dividing method based on light stream cluster, is characterized in that, comprises following steps:
Step 1: the optical flow field of the video that input contains different motion pattern, wherein, described video has T frame, every frame is containing M × N pixel, M, N are respectively the length and width of every two field picture, total T-1 of described optical flow field, and each optical flow field is two two-dimensional matrixs that size is M × N, one is x direction, and one is y direction;
Step 2: carry out cluster for each optical flow field in T-1 optical flow field, obtain U consistent moving region, U>T-1;
Step 3: consistent U moving region is classified as to S consistent moving region group, S<U, and with the each consistent moving region of label sign of consistent moving region group as area classification label;
Step 4: utilize the area classification label that step 3 obtains to carry out construction feature vector to each pixel in M × N pixel, obtain altogether M × N characteristic vector C
1, C
2..., C
m × N; The characteristic vector of m pixel is by cascading up the label of the consistent moving region group under m is in each optical flow field to obtain;
Step 5: according to characteristic vector C
1, C
2..., C
m × N, M × N point carried out to the semantic chunk that K-means cluster obtains splitting.
2. the video semanteme piece dividing method based on light stream cluster according to claim 1, is characterized in that: described step 3 comprises the steps:
Step a) is established in T-1 optical flow field, and a total U consistent moving region, carries out given sequence number to each consistent moving region, and sequence number is 1~U, and sets up U × 2 dimension matrix F;
Step b), for each consistent moving region, is calculated its average light stream, puts into the relevant position of F;
It is the two-dimensional matrix W of U × U that step c) is set up size; Wherein, W (i, j) be F (i, 1) F (j, 1)+F (i, 2) F (j, 2), W (i, j) is i consistent moving region and j consistent moving region motion relevance, F (i, 1) be the x durection component of i the consistent average light stream in moving region, F (j, 1) is the x durection component of j the consistent average light stream in moving region, F (i, 2) be the y durection component of j the consistent average light stream in moving region, F (j, 2) is the y durection component of j the consistent average light stream in moving region, and oeprator represents multiplication sign;
Step d) utilizes Spectral Clustering to carry out cluster to matrix W, forms S cluster, and with the each consistent moving region of cluster label sign corresponding to this S cluster, indicates U consistent moving region with S label.
3. the video semanteme piece dividing method based on light stream cluster according to claim 1, is characterized in that: the characteristic vector of m pixel described in step 4 obtains by the following method:
If m pixel affiliated consistent moving region category label in t optical flow field is h
m,t, 1≤h
m,t≤ S, 1≤t≤T-1; By area classification label h
m,tassignment is to C
m(t), i.e. C
m(t)=h
m,t, wherein, C
mbe the characteristic vector of m pixel, C
m(t) be that m pixel characteristic is to a flow control t element.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410153245.8A CN103945227B (en) | 2014-04-16 | 2014-04-16 | Video semantic block partition method based on light stream clustering |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410153245.8A CN103945227B (en) | 2014-04-16 | 2014-04-16 | Video semantic block partition method based on light stream clustering |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103945227A true CN103945227A (en) | 2014-07-23 |
CN103945227B CN103945227B (en) | 2017-02-01 |
Family
ID=51192671
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410153245.8A Active CN103945227B (en) | 2014-04-16 | 2014-04-16 | Video semantic block partition method based on light stream clustering |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103945227B (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105574876A (en) * | 2015-12-18 | 2016-05-11 | 电子科技大学 | Natural image segmentation method based on label meaning inference |
CN109753913A (en) * | 2018-12-28 | 2019-05-14 | 东南大学 | Calculate efficient multi-mode video semantic segmentation method |
CN111426284A (en) * | 2020-04-10 | 2020-07-17 | 山东师范大学 | Brox optical flow estimation-based surface shape measurement error correction method and system |
CN112804578A (en) * | 2021-01-28 | 2021-05-14 | 广州虎牙科技有限公司 | Atmosphere special effect generation method and device, electronic equipment and storage medium |
CN113014831A (en) * | 2021-03-05 | 2021-06-22 | 上海明略人工智能(集团)有限公司 | Method, device and equipment for acquiring scenes of sports video |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1738426A (en) * | 2005-09-09 | 2006-02-22 | 南京大学 | Video motion goal division and track method |
JP2013016930A (en) * | 2011-06-30 | 2013-01-24 | Jvc Kenwood Corp | Image encoding apparatus, image encoding method and image encoding program |
-
2014
- 2014-04-16 CN CN201410153245.8A patent/CN103945227B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1738426A (en) * | 2005-09-09 | 2006-02-22 | 南京大学 | Video motion goal division and track method |
JP2013016930A (en) * | 2011-06-30 | 2013-01-24 | Jvc Kenwood Corp | Image encoding apparatus, image encoding method and image encoding program |
Non-Patent Citations (2)
Title |
---|
张颖 等: "基于二次空间变换的快速时空分割", 《中国图象图形学报》 * |
方宇强 等: "基于光流和水平集的运动目标分割方法研究", 《2009年中国智能自动化会议论文集(第三分册)》 * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105574876A (en) * | 2015-12-18 | 2016-05-11 | 电子科技大学 | Natural image segmentation method based on label meaning inference |
CN105574876B (en) * | 2015-12-18 | 2018-02-09 | 电子科技大学 | A kind of natural image dividing method based on label semantic reasoning |
CN109753913A (en) * | 2018-12-28 | 2019-05-14 | 东南大学 | Calculate efficient multi-mode video semantic segmentation method |
CN109753913B (en) * | 2018-12-28 | 2023-05-23 | 东南大学 | Multi-mode video semantic segmentation method with high calculation efficiency |
CN111426284A (en) * | 2020-04-10 | 2020-07-17 | 山东师范大学 | Brox optical flow estimation-based surface shape measurement error correction method and system |
CN111426284B (en) * | 2020-04-10 | 2021-10-19 | 山东师范大学 | Brox optical flow estimation-based surface shape measurement error correction method and system |
CN112804578A (en) * | 2021-01-28 | 2021-05-14 | 广州虎牙科技有限公司 | Atmosphere special effect generation method and device, electronic equipment and storage medium |
CN113014831A (en) * | 2021-03-05 | 2021-06-22 | 上海明略人工智能(集团)有限公司 | Method, device and equipment for acquiring scenes of sports video |
CN113014831B (en) * | 2021-03-05 | 2024-03-12 | 上海明略人工智能(集团)有限公司 | Method, device and equipment for scene acquisition of sports video |
Also Published As
Publication number | Publication date |
---|---|
CN103945227B (en) | 2017-02-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Wei et al. | Multi-vehicle detection algorithm through combining Harr and HOG features | |
CN111104903B (en) | Depth perception traffic scene multi-target detection method and system | |
CN111967429B (en) | Pedestrian re-recognition model training method and device based on active learning | |
KR102190527B1 (en) | Apparatus and method for automatic synthesizing images | |
CN103945227A (en) | Video semantic block partition method based on light stream clustering | |
CN103413444A (en) | Traffic flow surveying and handling method based on unmanned aerial vehicle high-definition video | |
Lee et al. | Real-time illegal parking detection in outdoor environments using 1-D transformation | |
CN102915544A (en) | Video image motion target extracting method based on pattern detection and color segmentation | |
Gao et al. | Synergizing appearance and motion with low rank representation for vehicle counting and traffic flow analysis | |
CN112464933B (en) | Intelligent identification method for weak and small target through foundation staring infrared imaging | |
Rabiu | Vehicle detection and classification for cluttered urban intersection | |
Cao et al. | KLT feature based vehicle detection and tracking in airborne videos | |
Naik et al. | Implementation of YOLOv4 algorithm for multiple object detection in image and video dataset using deep learning and artificial intelligence for urban traffic video surveillance application | |
Roy et al. | A comprehensive survey on computer vision based approaches for moving object detection | |
Schörkhuber et al. | Feature selection and multi-task learning for pedestrian crossing prediction | |
Kankaria et al. | Alert system for drivers based on traffic signs, lights and pedestrian detection | |
CN109389048B (en) | Pedestrian detection and tracking method in surveillance video | |
Zin et al. | Robust road sign recognition using standard deviation | |
CN109509205B (en) | Foreground detection method and device | |
Chanawangsa et al. | A new color-based lane detection via Gaussian radial basis function networks | |
Ray et al. | Towards Real-time Detection, Tracking and Classification of natural video | |
Tsai et al. | Four Categories Vehicle Detection in Hsuehshan Tunnel via Single Shot Multibox Detector | |
Fadili et al. | Fastest Moroccan License Plate Recognition Using a Lightweight Modified YOLOv5 Model | |
Nagaraju et al. | Foreground Moving Object Detection using Support Vector Machine (SVM) | |
Zhong et al. | QuickFind: Fast and contact-free object detection using a depth sensor |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |