CN110348345A - A kind of Weakly supervised timing operating position fixing method based on continuity of movement - Google Patents

A kind of Weakly supervised timing operating position fixing method based on continuity of movement Download PDF

Info

Publication number
CN110348345A
CN110348345A CN201910575033.1A CN201910575033A CN110348345A CN 110348345 A CN110348345 A CN 110348345A CN 201910575033 A CN201910575033 A CN 201910575033A CN 110348345 A CN110348345 A CN 110348345A
Authority
CN
China
Prior art keywords
segment
movement
continuity
rgb
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910575033.1A
Other languages
Chinese (zh)
Other versions
CN110348345B (en
Inventor
王乐
翟元浩
刘子熠
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Jiaotong University
Original Assignee
Xian Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Jiaotong University filed Critical Xian Jiaotong University
Priority to CN201910575033.1A priority Critical patent/CN110348345B/en
Publication of CN110348345A publication Critical patent/CN110348345A/en
Application granted granted Critical
Publication of CN110348345B publication Critical patent/CN110348345B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to field of machine vision, disclose a kind of Weakly supervised timing operating position fixing method based on continuity of movement, comprising: video is divided into RGB frame and light stream and is handled respectively;The movement segment of the different hypothesis of length is proposed first against each time point on video for each movement mode, segment is then acted using convolutional neural networks recurrence according to the continuity of movement of video and classification accuracy.The different movement segments that mode obtains are acted for two, is combined by the module of a characteristic, filters out final operating position fixing result.The present invention can belong to the movement segment of the category in positioning video in the given other situation of video class.

Description

A kind of Weakly supervised timing operating position fixing method based on continuity of movement
Technical field
The invention belongs to technical field of computer vision, are related to Weakly supervised timing operating position fixing method, in particular to a kind of Weakly supervised timing operating position fixing method based on continuity of movement.
Background technique
Timing operating position fixing is an important computer vision problem, it understands task, such as event in abstract video There is very important application in the directions such as detection, Video summary and video problems answer.
Most of timing operating position fixing method needs accurate time-labeling at present, needs to consume a large amount of manpower and material resources; Meanwhile the time-labeling may be made inaccurate because of the ambiguity on the boundary of movement.In addition, current timing operating position fixing method In, RGB and light stream are not handled respectively, have ignored the feature of RGB and light stream itself;Final segment-score is only by classifying to get Out, the difference of RGB and light stream itself are ignored, and big to Classification Neural dependence, hardly results in optimal result.
To sum up, a kind of new Weakly supervised timing operating position fixing method is needed.
Summary of the invention
The Weakly supervised timing operating position fixing method based on continuity of movement that the purpose of the present invention is to provide a kind of, to solve Above-mentioned one or more technical problem.In the present invention, video is divided into RGB and light stream and is handled respectively, proposing respectively can The movement segment of energy, then screening fusion is to propose last as a result, it is possible to obtain preferably positioning result.
In order to achieve the above objectives, the invention adopts the following technical scheme:
A kind of Weakly supervised timing operating position fixing method based on continuity of movement, comprising the following steps:
Step 1, by video to be processed be divided into it is multiple be not overlapped segment, obtain the RGB feature and Optical-flow Feature of each segment;
Step 2, the RGB feature to step 1 acquisition and Optical-flow Feature carry out movement segment recurrence processing respectively, obtain RGB It acts segment and light stream acts segment;The movement segment recurrence processing includes: each time point for video to be processed, piece The imaginary movement segment for lifting different preset lengths, for different length movement segment using scheduled recurrent neural networks into Row returns, and recurrent neural networks are trained using continuity of movement loss function, and obtain movement segment;
Step 3, segment is acted by the RGB that continuity of movement loss function evaluation procedure 2 obtains and light stream acts segment Confidence level;Inhibit to filter out the movement segment that registration is more than threshold value using non-maximum value;
Step 4, after recurrent neural networks training;By the Fusion Module of a printenv, screening fusion RGB movement Segment and light stream act segment, obtain positioning result to the end.
A further improvement of the present invention is that step 1 specifically includes: video to be processed being divided into and multiple is not overlapped segment; To each segment average sample, the feature of sampling frame is extracted using convolutional neural networks, will extract the feature obtained as the piece The expression of section;Wherein, feature is extracted to RGB and light stream respectively.
A further improvement of the present invention is that movement segment recurrence processing specifically includes: for imaginary length in step 2 For the movement segment of P, returned using following formula:
In formula, xsFor the serial number of start boundary, xeFor the serial number of end boundary,For what is returned in start boundary position As a result,For end boundary position return as a result, P be movement fragment length.
A further improvement of the present invention is that continuity of movement loss function consists of two parts in step 3;
A part is used to characterize the cosine similarity of movement segment characterizations and its contextual feature;Assuming that the movement segment with The feature of its context is indicated with following symbol respectivelyWherein F It (u) be video at time point is u character representation, the calculation expression of this part are as follows:
Another part is characterized with classification confidence:
Wherein, S (k, u) indicates the classification confidence of segment u at classification k;For acting segment [xs,xe], expanded It is charged to [Xs,Xe], whereinContextual information as the segment;
The expression formula of continuity of movement loss function are as follows:
L=α Lc+(1-α)(Lc-1)
Wherein, α is hyper parameter, and value is 0 < α < 1.
A further improvement of the present invention is that after Recurrent networks training, being obtained respectively for RGB and light stream in step 4 The movement segment arrived, screening fusion steps specifically include:
The movement segment obtained by RGB and light stream is respectivelyWherein, NrAnd NfRespectively RGB and Light stream acts segments;
For the movement segment of each RGB, its maximum IoU with light stream segment, calculation formula are calculated are as follows:
Final screening fusion results act segment for all light streams and have I (pr,j) dynamic less than the RGB of preset threshold Make the combination of segment.
A further improvement of the present invention is that final screening fusion results, act segment for all light streams and have I (pr,jThe combination of the RGB movement segment of) < 0.4.
A further improvement of the present invention is that α is set as 0.6.
A further improvement of the present invention is that in step 2, RGB feature R that step 1 is obtainedsIt is input to multiple recurrence nets Network;Each Recurrent networks are made of 3 layers of 1D convolutional neural networks, and specify a fragment length P;
It is 1 convolution kernel that the last layer of Recurrent networks, which has 2 sizes, and output returns start boundary respectively and terminates side Boundary.
A further improvement of the present invention is that first two layers of Recurrent networks is made of empty convolutional network.
A further improvement of the present invention is that the cavity of first two layers of empty convolution of Recurrent networks is set as
Compared with prior art, the invention has the following advantages:
In the present invention, video is divided into RGB and light stream and is handled respectively, proposed possible movement segment respectively, then screen It merges to propose last as a result, reducing in conventional method since the feature for directly handling RGB and light stream splicing may be brought Interference, preferably positioning result can be obtained.In the present invention, the expression of two movement mode is different, and handling can more dig respectively The characteristics of digging different movement mode;The present invention devises a screening Fusion Module to combine the movement of two different movement mode Positioning result.The continuity of method of the invention based on video actions proposes a kind of loss function based on cosine similarity, Movement segment characterizations and movement segment classification confidence are combined, are avoided in conventional method to a certain extent according only to action movie The limitation of section classification confidence assessment movement segment.Method of the invention, can be in the condition for only knowing action classification in video Under, the movement segment in positioning video.
Further, in order to avoid over-fitting, first two layers of Recurrent networks is made of empty convolutional network;In order to guarantee have Enough contextual informations are input to network, and the receptive field of Recurrent networks has been configured to P, and setting method is by Recurrent networks The cavity of first two layers of empty convolution is set as
Detailed description of the invention
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, below to embodiment or existing Attached drawing needed in technical description does simple introduction;It should be evident that the accompanying drawings in the following description is of the invention Some embodiments to those skilled in the art without creative efforts, can also be according to this A little attached drawings obtain other attached drawings.
Fig. 1 is that a kind of process of Weakly supervised timing operating position fixing method based on continuity of movement of the embodiment of the present invention is shown It is intended to;
Fig. 2 is to show in the embodiment of the present invention with test result comparison of the history experimental result on THUMOS14 data set It is intended to;
Fig. 3 is to compare in the embodiment of the present invention with test result of the history experimental result on ActivityNet data set Schematic diagram.
Specific embodiment
To keep the purpose, technical effect and technical solution of the embodiment of the present invention clearer, implement below with reference to the present invention Attached drawing in example, technical scheme in the embodiment of the invention is clearly and completely described;Obviously, described embodiment It is a part of the embodiment of the present invention.Based on embodiment disclosed by the invention, those of ordinary skill in the art are not making creation Property labour under the premise of other embodiments obtained, all should belong to the scope of protection of the invention.
Referring to Fig. 1, a kind of Weakly supervised timing operating position fixing method based on continuity of movement of the invention, including it is following Step:
Step 1: feature extraction.Video is divided into multiple 15 frame fragments not being overlapped, 3 frames are taken to each segment, use volume Product neural network extracts feature, using this feature as the expression of the segment.And feature is extracted to RGB and light stream respectively, as The input of next step.
Step 2: movement segment returns.The step carries out following identical processing to RGB and light stream respectively.For video Each time point enumerates the imaginary movement segment of different length.For the movement segment of all equal lengths, use is same Convolutional neural networks are returned.
Specifically, it is assumed that the movement segment for being P for imaginary length starts with the serial number of end boundary to be respectively xs And xe, the result that neural network returns in the position is respectivelyWithIt is returned using following formula:
In formula, xsFor the serial number of start boundary, xeFor the serial number of end boundary,For what is returned in start boundary position As a result,For end boundary position return as a result, P be movement fragment length.
Step 3: movement segment assessment.The movement segment obtained by step 2 is evaluated using continuity of movement loss function Confidence level.Simultaneously using loss function training recurrent neural networks.Specifically, the loss function consists of two parts, A part is used to measure the cosine similarity of movement segment characterizations and its contextual feature, and another part measures the segment and thereon The hereafter difference of classification confidence.
Step 4: movement segment screening fusion.After recurrent neural networks training, by the fusion mould of a printenv The movement segment that block, screening fusion RGB and light stream respectively obtain, obtains result to the end.
Wherein, step 2 specifically includes:
The calculating process of its loss function are as follows: firstly, for movement segment [xs,xe], we are extended to [Xs,Xe], WhereinContextual information as the segment.Loss function of the invention is by two parts group At a part is used to characterize the cosine similarity of movement segment characterizations and its contextual feature: assuming that the movement segment and thereon Feature hereafter is indicated with following symbol respectivelyWherein F (u) is Video is u character representation at time point.The calculation method of this part are as follows:
Another part is characterized with classification confidence:
Wherein S (k, u) indicates the classification confidence of segment u at classification k.
Final loss function are as follows:
L=α Lc+(1-α)(Lc-1)
Wherein, α is hyper parameter, and value is 0 < α < 1.
Step 4 specifically includes: after Recurrent networks training, for the movement segment that RGB and light stream are respectively obtained, this hair It is bright to carry out following screening fusion process: assuming that the movement segment obtained by RGB and light stream is respectively Wherein, NrAnd NfRespectively RGB and light stream act segments.
For the movement segment of each RGB, its maximum IoU with light stream segment is calculated:
Final screening fusion results are for the movement segment of all light streams and with I (pr,jThe RGB movement segment of) < 0.4 Combination.
To sum up, the present invention handles RGB and light stream respectively, reduces in conventional method due to directly handling RGB and light stream The possible interference of the feature of splicing.The expression of two movement mode is different, and different movement mode can more be excavated respectively by handling The characteristics of.Meanwhile the present invention devises a screening Fusion Module to combine the operating position fixing result of two different movement mode. The continuity of movement loss function that the present invention designs combines movement segment characterizations and movement segment classification confidence, certain journey It is avoided on degree in conventional method according only to the limitation of movement segment classification confidence assessment movement segment.
Embodiment
Referring to Fig. 1, a kind of Weakly supervised timing operating position fixing method based on continuity of movement of the embodiment of the present invention, tool Body the following steps are included:
Step 1: RGB and light stream being handled as follows respectively: video is divided into the collection for 15 one segment of frame not being overlapped It closes, for each segment, takes representative frame of 3 frames as the segment at random, then use Temporal Segment Network extracts feature to 3 frame, as the feature of the segment after being averaged.
Step 2: by taking RGB as an example (light stream is identical as RGB processing method), the RGB feature R as obtained in step 1sInput To multiple Recurrent networks.Each Recurrent networks are made of 3 layers of 1D convolutional neural networks, and specify a fragment length P.In order to Over-fitting is avoided, is made of empty convolutional network for first two layers of Recurrent networks, there are 256 sizes is 3 convolution kernel.The last layer Having 2 sizes is 1 convolution kernel, and output returns start boundary and end boundary respectively.In order to guarantee to have enough contexts For information input to network, the receptive field of Recurrent networks has been configured to P.Setting method is by two layers before Recurrent networks of cavity The cavity of convolution is set as
The video for being T for length, for each of which timing position, we initialize imaginary movement segment first and areWherein xe,i-xs,i=P.Then by the regression result of Recurrent networksWith
Regression result isRecursive computational procedure are as follows:
Step 3: for acting segment [xs,xe], we are extended to [Xs,Xe], whereinContextual information as the segment.In order to assess movement segment, the present invention defines one A continuity of movement loss function, the loss function consist of two parts, a part be used to characterize movement segment characterizations with thereon The cosine similarity of following traits: assuming that the movement segment and the feature of its context are indicated with following symbol respectivelyIt at time point is u character representation that wherein F (u), which is video,.This Partial calculation method are as follows:
Another part is characterized with classification confidence:
Wherein S (k, u) indicates the classification confidence of segment u at classification k.
Final loss function are as follows:
L=α Lc+(1-α)(Lc-1)
Wherein, α is hyper parameter, is set as 0.6 in practice.When Recurrent networks training, the loss function value of Recurrent networks It is the average value of everything segment loss function value.In test, redundancy is removed using the non-maxima suppression that IoU is 0.4 Movement segment.
Step 4: after Recurrent networks training, for the movement segment that RGB and light stream are respectively obtained, the present invention is carried out such as Under screening fusion process: assuming that the movement segment obtained by RGB and light stream is respectivelyWherein NrAnd Nf Respectively RGB and light stream residue act segments.For the movement segment of each RGB, the maximum of itself and light stream segment is calculated IoU:
Final screening fusion results are for the movement segment of all light streams and with I (pr,jThe RGB movement segment of) < 0.4 Combination.Fig. 2 and Fig. 3 are please referred to, the improvement of the Experimental comparison present invention and history experimental data are passed through.
It referring to fig. 2, is the present invention and test result of the history experimental data on THUMOS14 data set.It can be seen that Under all IoU, mAP measured by the present invention has been above history experimental data.
Referring to Fig. 3, for the present invention and test result of the history experimental data on ActivityNet data set.It can see To at all IoU and under average case, mAP measured by the present invention has been above history experimental data.
In conclusion the invention discloses a kind of Weakly supervised timing operating position fixing method based on continuity of movement, that is, exist The movement segment for belonging to the category in the given other situation of video class in positioning video, belongs to field of machine vision.Of the invention Main thought are as follows: video is divided into RGB frame and light stream and is handled respectively, for each movement mode, first against every on video A time point proposes the movement segment of the different hypothesis of length, is then made according to the continuity of movement of video and classification accuracy Segment is acted with convolutional neural networks recurrence.The different movement segments that mode obtain are acted for two, pass through characteristic Module is combined, and filters out final operating position fixing result.
The above embodiments are merely illustrative of the technical scheme of the present invention and are not intended to be limiting thereof, although referring to above-described embodiment pair The present invention is described in detail, those of ordinary skill in the art still can to a specific embodiment of the invention into Row modification perhaps equivalent replacement these without departing from any modification of spirit and scope of the invention or equivalent replacement, applying Within pending claims of the invention.

Claims (10)

1. a kind of Weakly supervised timing operating position fixing method based on continuity of movement, which comprises the following steps:
Step 1, by video to be processed be divided into it is multiple be not overlapped segment, obtain the RGB feature and Optical-flow Feature of each segment;
Step 2, the RGB feature to step 1 acquisition and Optical-flow Feature carry out movement segment recurrence processing respectively, obtain RGB movement Segment and light stream act segment;The movement segment recurrence processing includes: each time point for video to be processed, is enumerated not With the imaginary movement segment of preset length, the movement segment of different length is returned using scheduled recurrent neural networks Return, recurrent neural networks are trained using continuity of movement loss function, and obtain movement segment;
Step 3, the RGB movement segment obtained by continuity of movement loss function evaluation procedure 2 and light stream act setting for segment Letter degree;Inhibit to filter out the movement segment that registration is more than threshold value using non-maximum value;
Step 4, after recurrent neural networks training;By the Fusion Module of a printenv, screening fusion RGB acts segment Segment is acted with light stream, obtains positioning result to the end.
2. a kind of Weakly supervised timing operating position fixing method based on continuity of movement according to claim 1, feature exist In step 1 specifically includes: video to be processed being divided into and multiple is not overlapped segment;To each segment average sample, convolution mind is used The feature that sampling frame is extracted through network will extract the feature obtained as the expression of the segment;Wherein, respectively to RGB and light stream Extract feature.
3. a kind of Weakly supervised timing operating position fixing method based on continuity of movement according to claim 1, feature exist In in step 2, movement segment recurrence processing specifically includes: the movement segment for being P for imaginary length is returned using following formula Return:
In formula, xsFor the serial number of start boundary, xeFor the serial number of end boundary,For start boundary position return as a result,For end boundary position return as a result, P be movement fragment length.
4. a kind of Weakly supervised timing operating position fixing method based on continuity of movement according to claim 3, feature exist In in step 2 and step 3, continuity of movement loss function consists of two parts;
A part is used to characterize the cosine similarity of movement segment characterizations and its contextual feature;Assuming that the movement segment with thereon Feature hereafter is indicated with following symbol respectivelyWherein F (u) It at time point is u character representation, the calculation expression of this part for video are as follows:
Another part is characterized with classification confidence:
Wherein, S (k, u) indicates the classification confidence of segment u at classification k;For acting segment [xs, xe], it is extended to [Xs, Xe], whereinContextual information as the segment;
The expression formula of continuity of movement loss function are as follows:
L=α Lc+(1-α)(Lc-1)
Wherein, α is hyper parameter, and value is 0 < α < 1.
5. a kind of Weakly supervised timing operating position fixing method based on continuity of movement according to claim 4, feature exist In in step 4, after Recurrent networks training, for the movement segment that RGB and light stream are respectively obtained, screening fusion steps tool Body includes:
The movement segment obtained by RGB and light stream is respectivelyWherein, NrAnd NfRespectively RGB and light stream Act segments;
For the movement segment of each RGB, its maximum IoU with light stream segment, calculation formula are calculated are as follows:
Final screening fusion results act segment for all light streams and have I (pR, j) it is less than the RGB action movie of preset threshold The combination of section.
6. a kind of Weakly supervised timing operating position fixing method based on continuity of movement according to claim 5, feature exist In final screening fusion results act segment for all light streams and have I (pR, j) < 0.4 RGB movement segment combination.
7. a kind of Weakly supervised timing operating position fixing side based on continuity of movement according to any one of claim 4 to 6 Method, which is characterized in that α is set as 0.6.
8. a kind of Weakly supervised timing operating position fixing method based on continuity of movement according to claim 1, feature exist In, in step 2, RGB feature R that step 1 is obtainedsIt is input to multiple Recurrent networks;Each Recurrent networks are by 3 layers of 1D convolution mind It is constituted through network, and specifies a fragment length P;
It is 1 convolution kernel that the last layer of Recurrent networks, which has 2 sizes, and output returns start boundary and end boundary respectively.
9. a kind of Weakly supervised timing operating position fixing method based on continuity of movement according to claim 8, feature exist In first two layers of Recurrent networks is made of empty convolutional network.
10. a kind of Weakly supervised timing operating position fixing method based on continuity of movement according to claim 9, feature exist In the cavity of first two layers of empty convolution of Recurrent networks is set as
CN201910575033.1A 2019-06-28 2019-06-28 Weak supervision time sequence action positioning method based on action consistency Active CN110348345B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910575033.1A CN110348345B (en) 2019-06-28 2019-06-28 Weak supervision time sequence action positioning method based on action consistency

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910575033.1A CN110348345B (en) 2019-06-28 2019-06-28 Weak supervision time sequence action positioning method based on action consistency

Publications (2)

Publication Number Publication Date
CN110348345A true CN110348345A (en) 2019-10-18
CN110348345B CN110348345B (en) 2021-08-13

Family

ID=68177039

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910575033.1A Active CN110348345B (en) 2019-06-28 2019-06-28 Weak supervision time sequence action positioning method based on action consistency

Country Status (1)

Country Link
CN (1) CN110348345B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111914644A (en) * 2020-06-30 2020-11-10 西安交通大学 Dual-mode cooperation based weak supervision time sequence action positioning method and system
CN111914778A (en) * 2020-08-07 2020-11-10 重庆大学 Video behavior positioning method based on weak supervised learning
CN112001420A (en) * 2020-07-24 2020-11-27 武汉安视感知科技有限公司 Intelligent timing and counting method and device for drill pipe of mine worker and storage device
WO2021098402A1 (en) * 2019-11-20 2021-05-27 腾讯科技(深圳)有限公司 Action recognition method and apparatus, computer storage medium, and computer device
CN115080750A (en) * 2022-08-16 2022-09-20 之江实验室 Weak supervision text classification method, system and device based on fusion prompt sequence

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104217214A (en) * 2014-08-21 2014-12-17 广东顺德中山大学卡内基梅隆大学国际联合研究院 Configurable convolutional neural network based red green blue-distance (RGB-D) figure behavior identification method
EP3147577A1 (en) * 2015-09-23 2017-03-29 Stn B.V. Device for and method of fluid flow monitoring
CN108399380A (en) * 2018-02-12 2018-08-14 北京工业大学 A kind of video actions detection method based on Three dimensional convolution and Faster RCNN
CN108573246A (en) * 2018-05-08 2018-09-25 北京工业大学 A kind of sequential action identification method based on deep learning
CN108664931A (en) * 2018-05-11 2018-10-16 中国科学技术大学 A kind of multistage video actions detection method
CN108805083A (en) * 2018-06-13 2018-11-13 中国科学技术大学 The video behavior detection method of single phase
CN109409307A (en) * 2018-11-02 2019-03-01 深圳龙岗智能视听研究院 A kind of Online Video behavioral value system and method based on space-time contextual analysis
CN109558811A (en) * 2018-11-12 2019-04-02 中山大学 A kind of action identification method based on sport foreground concern and non-supervisory key-frame extraction
CN109784269A (en) * 2019-01-11 2019-05-21 中国石油大学(华东) One kind is based on the united human action detection of space-time and localization method

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104217214A (en) * 2014-08-21 2014-12-17 广东顺德中山大学卡内基梅隆大学国际联合研究院 Configurable convolutional neural network based red green blue-distance (RGB-D) figure behavior identification method
EP3147577A1 (en) * 2015-09-23 2017-03-29 Stn B.V. Device for and method of fluid flow monitoring
CN108399380A (en) * 2018-02-12 2018-08-14 北京工业大学 A kind of video actions detection method based on Three dimensional convolution and Faster RCNN
CN108573246A (en) * 2018-05-08 2018-09-25 北京工业大学 A kind of sequential action identification method based on deep learning
CN108664931A (en) * 2018-05-11 2018-10-16 中国科学技术大学 A kind of multistage video actions detection method
CN108805083A (en) * 2018-06-13 2018-11-13 中国科学技术大学 The video behavior detection method of single phase
CN109409307A (en) * 2018-11-02 2019-03-01 深圳龙岗智能视听研究院 A kind of Online Video behavioral value system and method based on space-time contextual analysis
CN109558811A (en) * 2018-11-12 2019-04-02 中山大学 A kind of action identification method based on sport foreground concern and non-supervisory key-frame extraction
CN109784269A (en) * 2019-01-11 2019-05-21 中国石油大学(华东) One kind is based on the united human action detection of space-time and localization method

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
HAONAN QIU 等: "Precise Temporal Action Localization by Evolving Temporal Proposals", 《ARXIV》 *
YU-WEI CHAO 等: "Rethinking the Faster R-CNN Architecture for Temporal Action Localization", 《ARXIV》 *
ZHENG SHOU 等: "AutoLoc: Weakly-supervised Temporal Action Localization in Untrimmed Videos", 《ARXIV》 *
ZHENG SHOU 等: "Temporal Action Localization in Untrimmed Videos via Multi-stage CNNs", 《2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
石祥斌 等: "基于动作模板匹配的弱监督动作定位", 《HTTP://KNS.CNKI.NET/KCMS/DETAIL/51.1307.TP.20190408.1444.012.HTML》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021098402A1 (en) * 2019-11-20 2021-05-27 腾讯科技(深圳)有限公司 Action recognition method and apparatus, computer storage medium, and computer device
US11928893B2 (en) 2019-11-20 2024-03-12 Tencent Technology (Shenzhen) Company Limited Action recognition method and apparatus, computer storage medium, and computer device
CN111914644A (en) * 2020-06-30 2020-11-10 西安交通大学 Dual-mode cooperation based weak supervision time sequence action positioning method and system
CN112001420A (en) * 2020-07-24 2020-11-27 武汉安视感知科技有限公司 Intelligent timing and counting method and device for drill pipe of mine worker and storage device
CN111914778A (en) * 2020-08-07 2020-11-10 重庆大学 Video behavior positioning method based on weak supervised learning
CN111914778B (en) * 2020-08-07 2023-12-26 重庆大学 Video behavior positioning method based on weak supervision learning
CN115080750A (en) * 2022-08-16 2022-09-20 之江实验室 Weak supervision text classification method, system and device based on fusion prompt sequence
CN115080750B (en) * 2022-08-16 2022-11-08 之江实验室 Weak supervision text classification method, system and device based on fusion prompt sequence

Also Published As

Publication number Publication date
CN110348345B (en) 2021-08-13

Similar Documents

Publication Publication Date Title
CN110348345A (en) A kind of Weakly supervised timing operating position fixing method based on continuity of movement
CN106920229B (en) Automatic detection method and system for image fuzzy area
CN110070029B (en) Gait recognition method and device
CN100367294C (en) Method for dividing human body skin area from color digital images and video graphs
CN109508671B (en) Video abnormal event detection system and method based on weak supervision learning
CN109903237B (en) Multi-scale face image deblurring method based on low and high frequency separation
CN109410184B (en) Live broadcast pornographic image detection method based on dense confrontation network semi-supervised learning
CN109472193A (en) Method for detecting human face and device
CN110414367B (en) Time sequence behavior detection method based on GAN and SSN
CN111368758A (en) Face ambiguity detection method and device, computer equipment and storage medium
CN112766218B (en) Cross-domain pedestrian re-recognition method and device based on asymmetric combined teaching network
CN110942456B (en) Tamper image detection method, device, equipment and storage medium
CN109740553B (en) Image semantic segmentation data screening method and system based on recognition
CN113591674B (en) Edge environment behavior recognition system for real-time video stream
CN108319672A (en) Mobile terminal malicious information filtering method and system based on cloud computing
CN111476160A (en) Loss function optimization method, model training method, target detection method, and medium
CN115049954A (en) Target identification method, device, electronic equipment and medium
CN111723852A (en) Robust training method for target detection network
CN109389116A (en) A kind of character detection method and device
CN110458203B (en) Advertisement image material detection method
CN111163332A (en) Video pornography detection method, terminal and medium
CN112508135B (en) Model training method, pedestrian attribute prediction method, device and equipment
CN112396126B (en) Target detection method and system based on detection trunk and local feature optimization
CN111860100B (en) Pedestrian number determining method and device, electronic equipment and readable storage medium
CN114694090A (en) Campus abnormal behavior detection method based on improved PBAS algorithm and YOLOv5

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant