CN105590100A - Discrimination supervoxel-based human movement identification method - Google Patents

Discrimination supervoxel-based human movement identification method Download PDF

Info

Publication number
CN105590100A
CN105590100A CN201510977414.4A CN201510977414A CN105590100A CN 105590100 A CN105590100 A CN 105590100A CN 201510977414 A CN201510977414 A CN 201510977414A CN 105590100 A CN105590100 A CN 105590100A
Authority
CN
China
Prior art keywords
video
super voxel
feature
segment
identification
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510977414.4A
Other languages
Chinese (zh)
Other versions
CN105590100B (en
Inventor
段立娟
郭亚楠
马伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN201510977414.4A priority Critical patent/CN105590100B/en
Publication of CN105590100A publication Critical patent/CN105590100A/en
Application granted granted Critical
Publication of CN105590100B publication Critical patent/CN105590100B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

Provided is a discrimination supervoxel-based human movement identification method, comprising: utilizing an unsupervised method to automatically extract a video supervoxel characteristic set among a same category of movement videos, wherein the video supervoxel characteristic set is different from other categories, and can represent the characteristics of the category; then performing feature description on the supervoxels, finally completing identification on an ongoing movement, and thereby more accurately identifying the category of the human movement in videos. Simultaneously referring to a video supervoxel characteristic and an image hog characteristic, the method extracts the discrimination supervoxels in videos through a training and learning iteration process, and identifies movements more accurately. Compared with a traditional method, the method of the invention can automatically extract the effective portion in videos, wherein the effective portion not only comprises the portion with higher discrimination in a human body, but also comprises the portion representing the category of movements in the background.

Description

Human action recognition methods based on the super voxel of identification
Technical field
The present invention relates to feature extraction and machine learning method in Video processing, particularly a kind of based on the super voxel of identificationHuman action recognition methods.
Background technology
In recent years, under the fast development of internet, multimedia technology, video has become the important channel of people's obtaining information,Become the carrier of a large amount of digital informations. Although computer technology has also obtained significant progress in recent years, utilizes computer automaticThe analyzing and processing of carrying out video content is but a great problem of MultiMedia Field all the time. Can basis when human brain is accepted visual informationThe mankind for many years in life subtle study to knowledge or existence general knowledge visual information is carried out to rapid analysis, and computerCan only, by accepting digital information and carrying out numerical calculation and carry out video analysis, lack an intelligentized process, speed slow andThe degree of accuracy is low.
Video human action recognition in Video processing is because it is at aspects such as man-machine interaction, intelligent monitoring, video content analysisRange of application widely, becomes the popular direction of Recent study, has obtained many achievements. But action recognition task also exists veryMany challenges. First, due to the free degree of human action, no matter be in same human action or various human action,The form of expression of action is always differentiated. Even show the video of same action, due to the body posture of different people, peopleBody action speed, the difference of leg speed etc., also has very big-difference. A kind of desirable human motion recognizer should be able to adapt toThe variation of same action, and can distinguish different action classifications. Secondly, the shooting environmental of video or setting are different. ExampleAs, the action camera lens of the people under complicated and mobile background may more be difficult to identification. Recording while arranging, the variation of tone is alsoA common variable. When the video that uses video camera to capture, different visual angles, makes action recognition have more challengeProperty. The problem of distinguishing in order to overcome fuzzy action, this method is sought a kind of method that can apply in real world, make itsIn complicated shooting environmental, identification accurately.
In this article, introduce a kind of by characterizing the method for middle level features. By extracting the super voxel of identification, effectively distinguishDifferent actions, and which part of automatic decision video background can help the identification of moving. First video is carried out tooCut, and pass through the identification piece of the procedure extraction frame of video of a training, get lap with the result of over-segmentation, differentiatedThe super voxel of property. Then super voxel is carried out to track characteristic extraction and description. Finally by BOW belfry video features.
Summary of the invention
The problems referred to above that exist for prior art, the present invention proposes a kind of human action identification side based on the super voxel of identificationMethod, utilizes non-supervisory method automatically to extract to be different from similar action video other classifications, can characterize looking of this class featureFrequently super voxel characteristic set. And then to these super voxels carry out feature description, finally complete the identification of moving, canIdentify more accurately the classification of human action in video.
A human action recognition methods based on the super voxel of identification, comprises the following steps:
Carry out following steps for the video of training:
Step 1, carries out over-segmentation by the video of input, obtains the super voxel of video.
Step 2, carries out key-frame extraction to the video of input.
Step 3, the image that step 2 is obtained carries out the extraction of identification segment.
Step 4, the position of the super voxel that the identification segment that step 3 is obtained and step 1 obtain in video got overlappingOperation.
Step 5, is described the super voxel of video by movement locus feature and the word bag model (bow) of pixel.
Step 6, is used the super voxel of identification as dictionary, obtains video features by the method for bow.
Step 7, obtains disaggregated model with svm grader.
For video to be identified, carry out following steps:
Step 8, the video that input is identified, carries out respectively step 1,2,5,6, obtains the character representation of video to be identified.
Step 9, sends the feature of video to be identified into svm grader, obtains recognition result.
Method of the present invention has the following advantages:
The present invention simultaneously the super voxel feature of reference video (super voxel feature is poor by calculating pixel movement locus and colorDifferent obtaining) with the feature of these two kinds of dimensions of hog feature of image, by a training, the iterative process of study, extractsIn video, there is the super voxel of identification, can identify an action more accurately.
2. the present invention, compared with conventional method, can automatically extract effective part in video, not only comprises having in human bodyThe part of identification, the action to this class that can also extract in background has the part of sign effect (such as the basket of playing basketball in actionBall frame etc.).
Brief description of the drawings
Fig. 1 is the flow chart one of method involved in the present invention;
Fig. 2 is the flowchart 2 of method involved in the present invention.
Detailed description of the invention
Below in conjunction with detailed description of the invention, the present invention is described further.
The flow chart of the method for the invention as shown in Figure 1, comprises the following steps:
Step 1, carries out over-segmentation by the video of input.
One section of video of 1.1 inputs, suppose that the frame of video of input is 3 passage coloured image I, and its wide and height are respectively W, H.
This video is carried out to over-segmentation, obtain the super voxel of video.
Step 2, carries out key-frame extraction to the video of input.
Video is extracted to key frame by the method for getting a frame every 10 frames.
Step 3, the image that step 2 is obtained carries out the extraction of identification segment.
Training image is divided into two groups by 3.1: D, N. Wherein D is the training image of a class action, and N is that in video set, other are movingThe training image of doing. D, N is equally divided into respectively again two parts: D1, D1 and N1, N2.
All images of the 3.2 couples of D1 and N1 proceed as follows:
3.2.1 first image is carried out to segment sampling. The image of N*M carry out twice down-sampled, such width figure just occurs threeLevel. In these three levels, all according to there being the principle of sampling overlappingly, get the blockage (this method is decided to be 60*60) of k*k,These blockages are extracted to traditional HOG feature.
3.2.2 segment D1 being extracted carries out following operation:
Segment is carried out to stochastical sampling, and the segment after sampling is carried out to duplicate removal (if the difference of two is lower than a certain threshold value,Remove this piece). According to the number of residue segment, after 10, obtain wanting the classification number of cluster below. Utilize k-meansMethod is carried out cluster to segment, and removes the classification that only comprises 3 following elements, and the element note of every class is P (i), gives residue everyA classification is distributed a svm grader.
3.2.3 using such element P (i) as positive example, the segment of N1, as negative example, is trained on svm. Again D2 is carriedThe segment of getting is put into each grader as test sample book and is tested, and t the sample that score is the highest joins such original yuanIn element P (i). Next,, D1 and D2 exchange, N1 and N2 exchange, in the operation of carrying out 3.2.3, until iteration repeatedlyObtain final svm model. The iterations of this method is 6 times.
3.2.4 square 3.2.1 being obtained is tested on the svm of this class, if high in the score of some svm upper blockIn a certain threshold value, judge the segment that this piece is identification.
Step 4, the position of the super voxel that the identification segment that step 3 is obtained and step 1 obtain in video got overlappingOperation. Formula is as follows:
f ( n ) = 1 , i f S ( ∪ j = 1 F i P i j ) ∩ S ( DS k ) S ( DS k ) > T 0 , i f e l s e
Wherein, Fi is i the key frame of video V, PijFiIn j identification segment. DSkK in videoIndividual super voxel. Number of pixels in S (.) function representation one panel region. T is the overlapping threshold value that this method arranges.
So far, obtained the super voxel of identification.
Step 5, is described the super voxel of video by movement locus feature and the bow of pixel.
5.1 use tracer tools to follow the trail of pixel, and trace lengths is 15 frames, and obtaining some length is the movement locus of 15 frames.
5.2 pairs of tracks are described
The description of track is divided into four parts, altogether 426 dimensions:
1-30 dimension, totally 30 dimensions, front 30 dimensions represent the direction of motion of a pixel. Formula is as follows:
S ′ = ( ΔP t , ... , ΔP ( t + L ) - 1 ) Σ t + L - 1 j = t | | ΔP j | |
Wherein Δ Pt=(P(t+1)-Pt)=(x(t+1)-xt,y(t+1)-yt),
T represents t frame, and L is 15, xt, the x of this pixel when yt is illustrated in t frame, y axial coordinate.
Feature below obtains by first constructing a stereo block, and the building method of stereo block is as follows:
First for the pixel of every frame of this track, get centered by this location of pixels, the square (N=32) taking N as the length of side,Obtain a square taking N*N as cross section, the stereo block that L is length. This solid is divided into soon to the little stereo block of a*a*b,Wherein a=2, b=3. 12 little stereo block are so just obtained. Respectively these 12 fritters are extracted to conventional HOG, HOF,MHBx, MBHy feature, by these merging features, obtains the feature of 31-425 dimension, as follows:
31-126 dimension, totally 96 dimensions (8*2*2*3), represent HOG feature.
127-234 dimension, totally 108 dimensions (9*2*2*3), represent HOF feature.
235-330 dimension, totally 96 dimensions (8*2*2*3), represent MBHx feature.
331-426 dimension, totally 96 dimensions (8*2*2*3), represent MBHy feature.
So far obtain the motion feature of every track.
5.3 use k-mean algorithm that all tracks are carried out to cluster, and classification number is C1, obtains track dictionary after clustercodebook1。
5.4 use tracks represent super voxel
5.4.1 for the super voxel of the identification obtaining, find the track dropping on it above. Concrete grammar: travel through each track,If more than 7 or 7 pixel of this track all, in this super voxel, judges that this track drops in this super voxel. For lengthDegree is less than the super voxel of 7 frames, thinks that all tracks of passing by it all drop in this super voxel.
5.4.2 for each super voxel, the track dropping in it is done to bow statistics one time taking codebook1 as dictionary, obtainHistogram as the feature of this super voxel.
Step 6, is used the super voxel of identification as codebook, obtains video features by the method for bow.
The super voxel of all identifications is carried out to k-means cluster, and the dictionary after cluster is codebook2. For training video, carryGet its super voxel, it is carried out on codebook2 to bow statistics, the histogram obtaining is as the feature of this video.
Step 7, sends the video features of training video into svm grader and trains, and obtains the disaggregated model of multiclass.
For video to be identified, carry out following steps:
Step 8, the video that input is identified, carries out respectively step 1,2,5,6, obtains the character representation of video to be identified.
Step 9, sends the feature of video to be identified into svm grader, obtains recognition result.
In order to test recognition effect of the present invention, this method is applied on a conventional storehouse of human action identification: YoutubeDataset. This video library comprises 1600 videos, is divided into 11 classes, is respectively: basketball shooting, and cycling, diving,Play golf, ride, football juggles, and plays on a swing, and plays tennis, and trampoline, plays volleyball, and walks a dog. When the broadcasting of each videoLong between 3-20 second.

Claims (2)

1. the human action recognition methods based on the super voxel of identification, is characterized in that: the method is entered for the video of trainingRow following steps,
Step 1, carries out over-segmentation by the video of input, obtains the super voxel of video;
Step 2, carries out key-frame extraction to the video of input;
Step 3, the image that step 2 is obtained carries out the extraction of identification segment;
Step 4, the position of the super voxel that the identification segment that step 3 is obtained and step 1 obtain in video got overlappingOperation;
Step 5, is described the super voxel of video by movement locus feature and the word bag model (bow) of pixel;
Step 6, is used the super voxel of identification as dictionary, obtains video features by the method for bow;
Step 7, obtains disaggregated model with svm grader;
For video to be identified, carry out following steps:
Step 8, the video that input is identified, carries out respectively step 1,2,5,6, obtains the character representation of video to be identified;
Step 9, sends the feature of video to be identified into svm grader, obtains recognition result.
2. the human action recognition methods based on the super voxel of identification according to claim 1, is characterized in that: weThe flow process of method comprises the following steps,
Step 1, carries out over-segmentation by the video of input;
One section of video of 1.1 inputs, suppose that the frame of video of input is 3 passage coloured image I, and its wide and height are respectively W, H;
This video is carried out to over-segmentation, obtain the super voxel of video;
Step 2, carries out key-frame extraction to the video of input;
Video is extracted to key frame by the method for getting a frame every 10 frames;
Step 3, the image that step 2 is obtained carries out the extraction of identification segment;
Training image is divided into two groups by 3.1: D, N; Wherein D is the training image of a class action, and N is that in video set, other are movingThe training image of doing; D, N is equally divided into respectively again two parts: D1, D1 and N1, N2;
All images of the 3.2 couples of D1 and N1 proceed as follows:
3.2.1 first image is carried out to segment sampling; The image of N*M carry out twice down-sampled, such width figure just occurs threeLevel; In these three levels, all according to there being the principle of sampling overlappingly, get the blockage (this method is decided to be 60*60) of k*k,These blockages are extracted to traditional HOG feature;
3.2.2 segment D1 being extracted carries out following operation:
Segment is carried out to stochastical sampling, and the segment after sampling is carried out to duplicate removal (if the difference of two is lower than a certain threshold value,Remove this piece); According to the number of residue segment, after 10, obtain wanting the classification number of cluster below; Utilize k-meansMethod is carried out cluster to segment, and removes the classification that only comprises 3 following elements, and the element note of every class is P (i), gives residue everyA classification is distributed a svm grader;
3.2.3 using such element P (i) as positive example, the segment of N1, as negative example, is trained on svm; Again D2 is carriedThe segment of getting is put into each grader as test sample book and is tested, and t the sample that score is the highest joins such original yuanIn element P (i); Next,, D1 and D2 exchange, N1 and N2 exchange, in the operation of carrying out 3.2.3, until iteration repeatedlyObtain final svm model; The iterations of this method is 6 times;
3.2.4 square 3.2.1 being obtained is tested on the svm of this class, if high in the score of some svm upper blockIn a certain threshold value, judge the segment that this piece is identification;
Step 4, the position of the super voxel that the identification segment that step 3 is obtained and step 1 obtain in video got overlappingOperation; Formula is as follows:
f ( n ) = 1 , i f S ( ∪ j = 1 F i P i j ) ∩ S ( DS k ) S ( DS k ) > T 0 , i f e l s e
Wherein, Fi is i the key frame of video V, PijFiIn j identification segment; DSkK in videoIndividual super voxel; Number of pixels in S (.) function representation one panel region; T is the overlapping threshold value that this method arranges;
So far, obtained the super voxel of identification;
Step 5, is described the super voxel of video by movement locus feature and the bow of pixel;
5.1 use tracer tools to follow the trail of pixel, and trace lengths is 15 frames, and obtaining some length is the movement locus of 15 frames;
5.2 pairs of tracks are described
The description of track is divided into four parts, altogether 426 dimensions:
1-30 dimension, totally 30 dimensions, front 30 dimensions represent the direction of motion of a pixel; Formula is as follows:
S ′ = ( ΔP t , ... , ΔP ( t + L ) - 1 ) Σ i + L - 1 j = t | | ΔP j | |
Wherein Δ Pt=(P(t+1)-Pt)=(x(t+1)-xt,y(t+1)-yt),
T represents t frame, and L is 15, xt, the x of this pixel when yt is illustrated in t frame, y axial coordinate.
Feature below obtains by first constructing a stereo block, and the building method of stereo block is as follows:
First for the pixel of every frame of this track, get centered by this location of pixels, the square (N=32) taking N as the length of side,Obtain a square taking N*N as cross section, the stereo block that L is length; This solid is divided into soon to the little stereo block of a*a*b,Wherein a=2, b=3; 12 little stereo block are so just obtained; Respectively these 12 fritters are extracted to conventional HOG, HOF,MHBx, MBHy feature, by these merging features, obtains the feature of 31-425 dimension, as follows:
31-126 dimension, totally 96 dimensions (8*2*2*3), represent HOG feature;
127-234 dimension, totally 108 dimensions (9*2*2*3), represent HOF feature;
235-330 dimension, totally 96 dimensions (8*2*2*3), represent MBHx feature;
331-426 dimension, totally 96 dimensions (8*2*2*3), represent MBHy feature;
So far obtain the motion feature of every track;
5.3 use k-mean algorithm that all tracks are carried out to cluster, and classification number is C1, obtains track dictionary after clustercodebook1;
5.4 use tracks represent super voxel
5.4.1 for the super voxel of the identification obtaining, find the track dropping on it above; Concrete grammar: travel through each track,If more than 7 or 7 pixel of this track all, in this super voxel, judges that this track drops in this super voxel; For lengthDegree is less than the super voxel of 7 frames, thinks that all tracks of passing by it all drop in this super voxel;
5.4.2 for each super voxel, the track dropping in it is done to bow statistics one time taking codebook1 as dictionary, obtainHistogram as the feature of this super voxel;
Step 6, is used the super voxel of identification as codebook, obtains video features by the method for bow;
The super voxel of all identifications is carried out to k-means cluster, and the dictionary after cluster is codebook2; For training video, carryGet its super voxel, it is carried out on codebook2 to bow statistics, the histogram obtaining is as the feature of this video;
Step 7, sends the video features of training video into svm grader and trains, and obtains the disaggregated model of multiclass;
For video to be identified, carry out following steps:
Step 8, the video that input is identified, carries out respectively step 1,2,5,6, obtains the character representation of video to be identified;
Step 9, sends the feature of video to be identified into svm grader, obtains recognition result.
CN201510977414.4A 2015-12-23 2015-12-23 Surpass the human motion recognition method of voxel based on identification Active CN105590100B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510977414.4A CN105590100B (en) 2015-12-23 2015-12-23 Surpass the human motion recognition method of voxel based on identification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510977414.4A CN105590100B (en) 2015-12-23 2015-12-23 Surpass the human motion recognition method of voxel based on identification

Publications (2)

Publication Number Publication Date
CN105590100A true CN105590100A (en) 2016-05-18
CN105590100B CN105590100B (en) 2018-11-13

Family

ID=55929670

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510977414.4A Active CN105590100B (en) 2015-12-23 2015-12-23 Surpass the human motion recognition method of voxel based on identification

Country Status (1)

Country Link
CN (1) CN105590100B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106570480A (en) * 2016-11-07 2017-04-19 南京邮电大学 Posture-recognition-based method for human movement classification
CN110622214A (en) * 2017-07-11 2019-12-27 索尼公司 Fast progressive method for spatio-temporal video segmentation based on hyper-voxels

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104134217A (en) * 2014-07-29 2014-11-05 中国科学院自动化研究所 Video salient object segmentation method based on super voxel graph cut
CN104361581A (en) * 2014-10-22 2015-02-18 北京航空航天大学 CT (computed tomography) scanning data partitioning method based on combination of user interaction and volume rendering

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104134217A (en) * 2014-07-29 2014-11-05 中国科学院自动化研究所 Video salient object segmentation method based on super voxel graph cut
CN104361581A (en) * 2014-10-22 2015-02-18 北京航空航天大学 CT (computed tomography) scanning data partitioning method based on combination of user interaction and volume rendering

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
CHENLIANG XU等: "Evaluation of super-voxel methods for early video processing", 《COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2012 IEEE CONFERENCE ON》 *
K SOOMRO等: "Action Localization in Videos through Context Walk", 《IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION》 *
孔凡树等: "基于等值面拓扑简化的三维重建算法", 《燕山大学学报》 *
梁钰龄: "基于超体素的视频分割技术研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
沈萦华等: "基于法向特征直方图的点云配准算法", 《光学精密工程》 *
苏坡等: "基于超像素的多模态MRI脑胶质瘤分割", 《西北工业大学学报》 *
陆勇: "考场异常行为视频检测关键技术研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
陆桂亮: "三维点云场景语义分割建模研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106570480A (en) * 2016-11-07 2017-04-19 南京邮电大学 Posture-recognition-based method for human movement classification
CN106570480B (en) * 2016-11-07 2019-04-19 南京邮电大学 A kind of human action classification method based on gesture recognition
CN110622214A (en) * 2017-07-11 2019-12-27 索尼公司 Fast progressive method for spatio-temporal video segmentation based on hyper-voxels
CN110622214B (en) * 2017-07-11 2023-05-30 索尼公司 Rapid progressive method for space-time video segmentation based on super-voxels

Also Published As

Publication number Publication date
CN105590100B (en) 2018-11-13

Similar Documents

Publication Publication Date Title
Huang et al. Tracknet: A deep learning network for tracking high-speed and tiny objects in sports applications
CN106778854B (en) Behavior identification method based on trajectory and convolutional neural network feature extraction
CN108229338B (en) Video behavior identification method based on deep convolution characteristics
Yuan et al. Temporal action localization by structured maximal sums
Wang et al. Hierarchical attention network for action recognition in videos
Zhao et al. Temporal action detection with structured segment networks
CN108288015B (en) Human body action recognition method and system in video based on time scale invariance
CN108171112A (en) Vehicle identification and tracking based on convolutional neural networks
CN103605986A (en) Human motion recognition method based on local features
CN107092349A (en) A kind of sign Language Recognition and method based on RealSense
Rangasamy et al. Deep learning in sport video analysis: a review
CN106529477A (en) Video human behavior recognition method based on significant trajectory and time-space evolution information
Yu et al. Weakly semantic guided action recognition
CN105512618A (en) Video tracking method
CN104200218B (en) A kind of across visual angle action identification method and system based on timing information
CN106295532A (en) A kind of human motion recognition method in video image
CN104021381A (en) Human movement recognition method based on multistage characteristics
CN105469050A (en) Video behavior identification method based on local space-time characteristic description and pyramid vocabulary tree
Kindiroglu et al. Temporal accumulative features for sign language recognition
CN103020614A (en) Human movement identification method based on spatio-temporal interest point detection
CN105844204A (en) Method and device for recognizing behavior of human body
CN105590100A (en) Discrimination supervoxel-based human movement identification method
CN103077383A (en) Method for identifying human body movement of parts based on spatial and temporal gradient characteristics
Lan et al. Learning action primitives for multi-level video event understanding
Liu et al. Research on action recognition of player in broadcast sports video

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
EE01 Entry into force of recordation of patent licensing contract
EE01 Entry into force of recordation of patent licensing contract

Application publication date: 20160518

Assignee: LUOYANG YAHUI EXOSKELETON POWER-ASSISTED TECHNOLOGY CO.,LTD.

Assignor: Beijing University of Technology

Contract record no.: X2024980000190

Denomination of invention: A Method for Human Action Recognition Based on Discriminant Hypervoxels

Granted publication date: 20181113

License type: Common License

Record date: 20240105

Application publication date: 20160518

Assignee: Henan zhuodoo Information Technology Co.,Ltd.

Assignor: Beijing University of Technology

Contract record no.: X2024980000138

Denomination of invention: A Method for Human Action Recognition Based on Discriminant Hypervoxels

Granted publication date: 20181113

License type: Common License

Record date: 20240104

Application publication date: 20160518

Assignee: Luoyang Lexiang Network Technology Co.,Ltd.

Assignor: Beijing University of Technology

Contract record no.: X2024980000083

Denomination of invention: A Method for Human Action Recognition Based on Discriminant Hypervoxels

Granted publication date: 20181113

License type: Common License

Record date: 20240104