CN116030538A - Weak supervision action detection method, system, equipment and storage medium - Google Patents

Weak supervision action detection method, system, equipment and storage medium Download PDF

Info

Publication number
CN116030538A
CN116030538A CN202310326090.2A CN202310326090A CN116030538A CN 116030538 A CN116030538 A CN 116030538A CN 202310326090 A CN202310326090 A CN 202310326090A CN 116030538 A CN116030538 A CN 116030538A
Authority
CN
China
Prior art keywords
action
class
category
scores
loss
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202310326090.2A
Other languages
Chinese (zh)
Other versions
CN116030538B (en
Inventor
王子磊
李志林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Science and Technology of China USTC
Original Assignee
University of Science and Technology of China USTC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Science and Technology of China USTC filed Critical University of Science and Technology of China USTC
Priority to CN202310326090.2A priority Critical patent/CN116030538B/en
Publication of CN116030538A publication Critical patent/CN116030538A/en
Application granted granted Critical
Publication of CN116030538B publication Critical patent/CN116030538B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

The invention discloses a weak supervision action detection method, a weak supervision action detection system, weak supervision action detection equipment and a weak supervision action detection storage medium, which are corresponding schemes, wherein: according to the action prediction structures of the two different branches, fuzzy video frames with inconsistent prediction results are screened out, and feature learning of the inconsistent video frames can be promoted by using contrast learning; the consistency constraint is carried out on the double branches to reduce the difference of the double branches, so that the problem that enough video frames cannot be maintained for contrast learning due to overlarge difference of the double branches is solved, and meanwhile, the reliability of the action detection result is further enhanced by utilizing collaborative learning of the double branches; the influence caused by pseudo tag errors when training the class-independent actionable prediction branch can be reduced by applying the consistency constraint in the class-independent actionable detection branch; in general, the invention improves the detection performance of the motion detection network and ensures the accuracy and reliability of motion detection by utilizing contrast learning and consistency constraint.

Description

Weak supervision action detection method, system, equipment and storage medium
Technical Field
The present invention relates to the field of video motion detection technologies, and in particular, to a method, a system, an apparatus, and a storage medium for detecting a weak supervision motion.
Background
In recent years, with increasing capabilities of video data acquisition, transmission and storage, demands of video analysis applications are gradually increased, and time sequence motion detection tasks are also gradually focused by many researchers at home and abroad due to wide practical applications, such as security monitoring, video retrieval, sports video clip, video auditing, and the like, so that the researchers need to design proper time sequence motion detection schemes according to demands of different application scenes, and provide accurate motion positioning and classification results.
At present, the time sequence motion detection task mainly has two learning paradigms: (1) providing a full-supervision learning paradigm of complete frame-level labeling; (2) A weakly supervised learning paradigm of any frame level annotation but known video level class annotation is not provided. For the full-supervision time sequence action detection, all sample videos need to be manually marked frame by frame, the marking work is time-consuming, and the precision is generally low. Therefore, in order to solve the problems of difficult labeling and large labeling error, weak supervision time sequence action detection is generated. This method can label a large number of unclamped videos on the network with a video level, and directly serve as training data.
Although weakly supervised timing action detection has many benefits, its performance is generally poor due to the lack of fine labeling. The weak supervision time sequence action detection at the present stage faces two basic challenges: action integrity modeling and action context confusion. The current mainstream method at home and abroad is to train a classification network by directly utilizing video-level labels so as to obtain a class activation sequence, then select a series of preselected frames on the class activation sequence by utilizing a plurality of thresholds, and finally obtain a final prediction result by utilizing non-maximum suppression. The root cause of these two problems is that without frame-level label supervision, the network model has difficulty in learning the real motion information, so that the network model generates wrong or fuzzy prediction results for the scene strong-correlation fragments and the fragments of the motion edges. The strategy of cooperative distillation is adopted in the Chinese patent application of weak supervision video timing sequence action detection and classification method and system with publication number of CN115272941A, so that the advantages of a single-mode frame and a cross-mode frame are complemented, and more complete and accurate timing sequence action detection and classification are realized. The action recognition is performed by using a sparse key frame attention mechanism in PCT international application published as CN110832499a, which enters the national phase, weak supervision action localization through sparse time pooled networks. In the chinese patent application CN115439790a, the method for positioning weak supervision time sequence action based on cascaded seed region growth module, an original class activation sequence is obtained according to time sequence characteristics, an expanded class activation sequence is obtained by a seed growth strategy, an anti-erasure is performed, the original class activation sequence and the erased class activation sequence are fused, and a class activation sequence with higher reliability is obtained to improve detection accuracy. In a chinese patent application CN114898259a, a weak supervision video timing sequence action positioning method based on action associated attention, an action associated attention model is used to establish a relationship between action segments in a video, a query mechanism is used to establish weak supervision pre-training, and the output of the query mechanism is input into a decoder of a transducer architecture to implement time positioning of a query set; and determining the relation between the video segment characteristics by utilizing an encoder of a transducer architecture, so as to realize the positioning and classification of the action segments. None of the above methods takes into account the influence of the blurred video frames in the video on the network model detection performance, so the detection performance is still to be improved.
In view of this, the present invention has been made.
Disclosure of Invention
The invention aims to provide a weak supervision action detection method, a weak supervision action detection system, weak supervision action detection equipment and a weak supervision action detection storage medium, which can greatly improve the accuracy and reliability of action detection.
The invention aims at realizing the following technical scheme:
a method of weakly supervised action detection, comprising:
constructing an action detection network comprising a category-aware action detection branch and a category-independent action detection branch, and training the action detection network; during training, inputting optical flow characteristics and RGB characteristics of a training video, obtaining a class activation sequence through a class perception action detection branch, generating class perception action scores, obtaining optical flow characteristic action scores and RGB characteristic action scores through a class irrelevant action detection branch, and fusing the optical flow characteristic action scores and the RGB characteristic action scores into class irrelevant action scores; dividing video frames in the training video according to the category perception action score and the category irrelevant action score, constructing corresponding positive and negative samples according to the type of each frame, and calculating comparison loss by utilizing embedded features respectively generated in the category perception action detection branch and the category irrelevant action detection branch to obtain total comparison loss; calculating class perception losses according to the class activation sequences, and calculating class irrelevant losses according to the optical flow characteristic action scores and the RGB characteristic action scores; respectively applying consistency constraint to the interior of the class-independent action detection branch and between the class-aware action detection branch and the class-independent action detection branch, and calculating corresponding consistency loss; constructing a target loss function by combining the total contrast loss, the category perception loss, the category irrelevant loss and the consistency loss, and training an action detection network by utilizing the target loss function; wherein RGB represents three color channels of red, green and blue;
Inputting the optical flow characteristics and RGB characteristics of the video to be detected into a trained action detection network, and generating an action detection result by combining the class activation sequence of the class-aware action detection branch and the optical flow characteristic action score and RGB characteristic action score of the class-independent action detection branch.
A weak supervision action detection system comprising:
the system comprises an action detection network construction and training unit, a classification sensing unit and a classification independent action detection unit, wherein the action detection network construction and training unit is used for constructing an action detection network comprising a classification sensing action detection branch and a classification independent action detection branch and training the action detection network; during training, inputting optical flow characteristics and RGB characteristics of a training video, obtaining a class activation sequence through a class perception action detection branch, generating class perception action scores, obtaining optical flow characteristic action scores and RGB characteristic action scores through a class irrelevant action detection branch, and fusing the optical flow characteristic action scores and the RGB characteristic action scores into class irrelevant action scores; dividing video frames in the training video according to the category perception action score and the category irrelevant action score, constructing corresponding positive and negative samples according to the type of each frame, and calculating comparison loss by utilizing embedded features respectively generated in the category perception action detection branch and the category irrelevant action detection branch to obtain total comparison loss; calculating class perception losses according to the class activation sequences, and calculating class irrelevant losses according to the optical flow characteristic action scores and the RGB characteristic action scores; respectively applying consistency constraint to the interior of the class-independent action detection branch and between the class-aware action detection branch and the class-independent action detection branch, and calculating corresponding consistency loss; constructing a target loss function by combining the total contrast loss, the category perception loss, the category irrelevant loss and the consistency loss, and training an action detection network by utilizing the target loss function; wherein RGB represents three color channels of red, green and blue;
The action detection unit is used for inputting the optical flow characteristics and RGB characteristics of the video to be detected into the trained action detection network, and generating an action detection result by combining the class activation sequence of the class-aware action detection branch, the optical flow characteristic action score and RGB characteristic action score of the class-independent action detection branch.
A processing apparatus, comprising: one or more processors; a memory for storing one or more programs;
wherein the one or more programs, when executed by the one or more processors, cause the one or more processors to implement the aforementioned methods.
A readable storage medium storing a computer program which, when executed by a processor, implements the method described above.
According to the technical scheme provided by the invention, according to the action prediction structures of two different branches, fuzzy video frames (video frames influencing action positioning) with inconsistent prediction results are screened out, and the feature learning of the inconsistent video frames can be promoted by using contrast learning; the difference of the double branches is reduced by carrying out consistency constraint on the double branches, so that the problem that enough video frames cannot be maintained for contrast learning due to overlarge difference of the double branches is solved, and meanwhile, the reliability of an action detection result is further enhanced by utilizing collaborative learning of the double branches (namely action consistency loss introduced later); the influence caused by pseudo tag errors when the class independent action detection branch is trained can be reduced by applying the consistency constraint in the class independent action detection branch; in general, the invention improves the detection performance of the motion detection network and ensures the accuracy and reliability of motion detection by utilizing contrast learning and consistency constraint.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings that are needed in the description of the embodiments will be briefly described below, it being obvious that the drawings in the following description are only some embodiments of the present invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flowchart of a weak supervision action detection method according to an embodiment of the present invention;
fig. 2 is a schematic structural diagram of an action detection network according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a weak supervision action detection system according to an embodiment of the present invention;
fig. 4 is a schematic diagram of a processing apparatus according to an embodiment of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and fully with reference to the accompanying drawings, in which it is evident that the embodiments described are only some, but not all embodiments of the invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to fall within the scope of the invention.
The terms that may be used herein will first be described as follows:
the terms "comprises," "comprising," "includes," "including," "has," "having" or other similar referents are to be construed to cover a non-exclusive inclusion. For example: including a particular feature (e.g., a starting material, component, ingredient, carrier, formulation, material, dimension, part, means, mechanism, apparatus, step, procedure, method, reaction condition, processing condition, parameter, algorithm, signal, data, product or article of manufacture, etc.), should be construed as including not only a particular feature but also other features known in the art that are not explicitly recited.
The following describes in detail a method, a system, a device and a storage medium for detecting weak supervision actions. What is not described in detail in the embodiments of the present invention belongs to the prior art known to those skilled in the art. The specific conditions are not noted in the examples of the present invention and are carried out according to the conditions conventional in the art or suggested by the manufacturer.
Example 1
The embodiment of the invention provides a weak supervision action detection method, which designs an action detection network with a double-branch structure, screens out fuzzy sample frames with inconsistent prediction results according to action prediction results of two different branches, and uses contrast learning to promote feature learning of the inconsistent sample frames by using consistent sample frames with higher reliability. Meanwhile, the reliability of the prediction result is further enhanced by utilizing double-branch collaborative learning and modal consistency constraint. As shown in fig. 1, a flowchart of a related method provided by the present invention mainly includes the following steps:
And step 1, constructing an action detection network comprising a category-aware action detection branch and a category-independent action detection branch, and training the action detection network.
In the embodiment of the invention, the class-aware action detection branch and the class-independent action detection branch are constructed to form an action detection network under the inspiration of the current time sequence action detection technology.
The training mode of the motion detection network is as follows: inputting optical flow characteristics and RGB characteristics of a training video, obtaining a class activation sequence through a class perception action detection branch, generating class perception action scores, obtaining optical flow characteristic action scores and RGB characteristic action scores through a class irrelevant action detection branch, and fusing the optical flow characteristic action scores and the RGB characteristic action scores into class irrelevant action scores; dividing video frames in the training video according to the category perception action score and the category irrelevant action score, constructing corresponding positive and negative samples according to the type of each frame, and calculating comparison loss by utilizing embedded features respectively generated in the category perception action detection branch and the category irrelevant action detection branch to obtain total comparison loss; calculating class perception losses according to the class activation sequences, and calculating class irrelevant losses according to the optical flow characteristic action scores and the RGB characteristic action scores; respectively applying consistency constraint to the interior of the class-independent action detection branch and between the class-aware action detection branch and the class-independent action detection branch, and calculating corresponding consistency loss; and constructing a target loss function by combining the total contrast loss, the category perception loss, the category independence loss and the consistency loss, and training the action detection network by using the target loss function.
In the embodiment of the present invention, the RGB features are proper terms in the art, which are features extracted directly from video, for example, in examples provided later, RGB features are extracted using a pre-trained I3D model (behavior recognition model); wherein RGB represents three color channels of red, green and blue.
And 2, inputting the optical flow characteristics and the RGB characteristics of the video to be detected into a trained action detection network, and generating an action detection result by combining the class activation sequence of the class-aware action detection branch and the optical flow characteristic action score and the RGB characteristic action score of the class-independent action detection branch.
The method provided by the embodiment of the invention solves various problems existing in the prior art, and is a weak supervision action detection method guided by action inconsistency. Because the current weak supervision action detection method lacks frame-level label supervision, the model cannot make very definite judgment on each frame, and a plurality of fuzzy frames influencing action positioning exist. Therefore, the method constructs the category perception action detection branch and the category irrelevant action detection branch, divides the frame types by comparing the double branch prediction results, finds out the video frames with inconsistent prediction results, and divides all the fragments in the video into consistent action/background fragments and inconsistent action/background fragments as positive and negative samples of contrast learning, thereby enhancing the feature learning of the video frames by utilizing the contrast learning and improving the detection performance.
In order to more clearly demonstrate the technical scheme and the technical effects provided by the invention, the method provided by the embodiment of the invention is described in detail below by using specific embodiments.
1. Overall network architecture.
As shown in fig. 2, the action detection network provided in the embodiment of the present invention mainly includes: the category aware action detection branch is related to the category independent action detection branch.
1. Class-aware action detection branches.
As shown in fig. 2, the class-aware motion detection branch mainly includes a feature embedding layer f and a classification layer cls; the optical flow features and the RGB features are spliced and then input into a class perception action detection branch, the comprehensive embedded features E are obtained through a feature embedding layer f, class activation sequences CAS are obtained through a classification layer cls, the class activation sequences CAS are summed (sum functions) in class dimensions, and an activation function is used to obtain class perception action scores A. The above-described flow in the category-aware action detection branch may be expressed as:
E=f (F1)
CAS=cls(E)
A=sigmoid(sum(CAS))
wherein F1 represents the feature obtained by splicing the optical flow feature and the RGB feature, and sigmoid is an activation function.
In the embodiment of the invention, the category perception action scores of all video frames in the training video can be obtained through the category perception action detection branch.
2. Class independent actions detect branches.
In the embodiment of the invention, the class-independent action detection branch does not distinguish action classes, only distinguishes actions from background, and separates and processes optical flow characteristics and RGB characteristics, and the two characteristics use the same processing method, and finally the action scores of the two characteristics are fused. Specific: the class independent action detection branch comprises two sub-branches, and optical flow characteristics and RGB characteristics are input into one sub-branch in a one-to-one manner; after the optical flow features and the RGB features are input into the corresponding sub-branches, the corresponding embedded features are obtained through the embedding layers in the sub-branches, and then the action scores corresponding to the corresponding features are obtained through the classification layers and the activation functions.
As shown in fig. 2, the first sub-branch (the upper sub-branch in the class-independent action detection branch) includes an RGB embedding layer f R And RGB classification layer cls R Input RGB feature F R Through RGB embedded layer f R Obtaining RGB embedded feature E R Then pass through RGB classification layer cls R Processing with the activation function to obtain RGB feature action score A R . The second sub-branch (the lower sub-branch in the class-independent action detection branch) comprises an optical flow embedding layer f F And optical flow classification layer cls F Input optical flow feature F F Through optical flow embedding layer f F Obtaining optical flow embedded feature E F Then pass through the optical flow classification layer cls F Processing with the activation function to obtain optical flow characteristic action score A F . Obtaining a class independent action score A after fusing the two types of characteristic action scores ca . The above-described flow in the category independent action detection branch can be expressed as:
E * =f * (F * )
A * =sigmoid(cls * (E * ))
A ca =(A F +A R )/2
where R corresponds to RGB features and F corresponds to optical flow features. When R, the first two formulas represent the process flow of the first sub-branch; when =f, the first two equations represent the process flow of the second sub-branch. The third expression represents fusing the two types of feature action scores by means of averaging.
In the embodiment of the invention, the RGB features and the optical flow features in the class independent action detection branches are processed separately, each video frame has an RGB action score and an optical flow action score, and then the class independent action score of each video frame can be obtained through fusion.
2. Objective loss function for network training.
1. Category perception loss.
In an embodiment of the invention, a multi-instance learning paradigm and cross entropy loss function are used on class-aware motion detection branches. The scores of K1 video frames with the highest scores of each class in the class activation sequence CAS (namely, K1 video frames are respectively selected for each class) are aggregated to represent the scores of each class of training video; wherein K1 is a preset positive integer. The cross entropy loss is calculated as a class-aware loss using a given video level tag, expressed as:
Figure SMS_1
wherein ,
Figure SMS_2
representing class perception losses, C is the total number of classes, p c To train the score of the c-th category of video, y c A true tag representing the c-th category of training video, which belongs to a given video-level tag.
In the embodiment of the invention, each training video has a label corresponding to the action category appearing in the video, namely a video-level label.
In the embodiment of the present invention, the value of K1 varies for different data sets (the higher the duty ratio of each video action frame in the data set, the smaller the K1), generally speaking, it may be set that: k1 L/8 to L/2, L is the total frame number in the training video.
2. Category independent loss.
In the embodiment of the invention, a method is designedAs shown in fig. 2, the line pseudo tag generation strategy is that in the class-aware motion detection branch, the class activation sequence CAS is normalized in the class dimension by a softmax function (normalized exponential function), the RGB feature motion scores and the optical flow feature motion scores of all video frames output by the class-independent motion detection branch are respectively formed into corresponding motion sequences, which are called a first motion sequence and a second motion sequence, the normalized class activation sequence CAS, the first motion sequence and the second motion sequence are fused (for example, the average value is obtained after the three are added) to obtain a comprehensive class activation sequence, then K1 video frames with the highest score are found out in the comprehensive class activation sequence according to the motion class in a given video tag, and the found K1 video frames are marked as a motion set
Figure SMS_3
The remaining video frames are marked as a background set T b The two finally obtained sets are the frame-level pseudo tags.
And then, respectively calculating corresponding generalized cross entropy loss by utilizing the optical flow characteristic action score and the RGB characteristic action score, wherein the generalized cross entropy loss is expressed as:
Figure SMS_4
where q is the noise tolerance coefficient, 0<q is less than or equal to 1, and the larger q is, the higher the noise tolerance is;
Figure SMS_6
representing generalized cross entropy loss calculated using RGB feature action scores or optical flow feature action scores, < ->
Figure SMS_9
Representing action set +.>
Figure SMS_11
RGB feature action score or optical flow feature action score of the t-th video frame; />
Figure SMS_7
Representing a background set T b Middle->
Figure SMS_8
RGB feature action scores or optical flow feature action scores for individual video frames; />
Figure SMS_10
、/>
Figure SMS_12
Representing action sets +.>
Figure SMS_5
Background set T b Video frames of the video frame.
Adding the generalized cross entropy loss calculated by the RGB feature action score and the generalized cross entropy loss calculated by the optical flow feature action score to obtain a category independent loss, expressed as:
Figure SMS_13
wherein ,
Figure SMS_14
for category independent loss, ++>
Figure SMS_15
To generalize the cross entropy penalty computed using the RGB feature action scores,
Figure SMS_16
generalized cross entropy loss calculated for action scores using optical flow features.
3. Consistency loss.
In the embodiment of the invention, consistency constraints are respectively applied between the category-aware action detection branches and the category-independent action detection branches and inside the category-independent action detection branches, and corresponding consistency losses are calculated.
(1) Consistency constraints are respectively applied to the interior of the class independent action detection branches.
In the embodiment of the invention, multiple applications are applied to the interior of a category-independent action detection branchModality consistency constraints, specific: in order to promote mutual learning of bimodal features (RGB features and optical flow features), the embodiment of the invention designs a multimodal consistency constraint, learns the action scores of the RGB features and the action scores of the optical flow features as soft labels of each other, and calculates multimodal consistency loss
Figure SMS_17
Expressed as:
Figure SMS_18
;/>
wherein ,AR Representing RGB feature action score, A F Representing the action score of the optical flow characteristics,
Figure SMS_19
representing the mean square error loss function, ">
Figure SMS_20
Representing the phase gradient pass back.
(2) A consistency constraint is imposed between the category-aware action detection branch and the category-independent action detection branch.
In the embodiment of the invention, action consistency constraint is applied between the category-aware action detection branch and the category-independent action detection branch, and the method specifically comprises the following steps: in order to avoid that the prediction difference between the class-aware motion detection branch and the class-independent motion detection branch in the later training stage is too large to generate enough consistency samples, the embodiment of the invention designs motion consistency constraint to promote the mutual learning of the two branches and keep normal difference, and the motion consistency loss calculated by the part is expressed as:
Figure SMS_21
wherein ,
Figure SMS_22
indicating a loss of motion consistency, A ca Representing class independent action scores, A representing class aware actionsA score was made.
4. Total contrast loss.
In the embodiment of the invention, for each branch, taking the median of all prediction scores in the corresponding branch, preliminarily marking video frames higher than the median as actions, preliminarily marking video frames lower than the median as backgrounds, and obtaining preliminary action background prediction results of two branches, wherein for a category-aware action detection branch, the prediction score is a category-aware action score, and for a category-independent action detection branch, the prediction score is a category-independent action score; comparing the preliminary action background prediction results of the two branches, and dividing the video frames which are preliminarily marked as actions by the two branches into consistent action frames (CA); the video frames with the background marked by both branches are divided into consistent background frames (CB); for the two branches, primarily marking inconsistent video frames (namely the fuzzy video frames) and sorting according to category perception action scores, dividing the K2 video frames with the highest scores into inconsistent action frames (IA), and dividing the K2 video frames with the lowest scores into inconsistent background frames (IB); wherein K2 is a preset positive integer.
In the embodiment of the invention, the following steps can be set: k2 L/20, L is the total number of frames in the training video.
Four different positive and negative sample pairs are constructed for four different frame types: for the consistent action frame, the positive sample is an inconsistent action frame, and the negative sample is a consistent background frame; for the consistent background frames, positive samples are inconsistent background frames, and negative samples are consistent action frames; for inconsistent action frames, positive samples are consistent action frames, and negative samples are consistent background frames; for inconsistent background frames, the positive samples are consistent background frames, and the negative samples are consistent action frames. The right side of fig. 2 intuitively shows the correspondence between the above four different frame types and positive and negative samples, the sign +: indicates positive samples, and the sign-: indicates negative samples;
Figure SMS_23
preliminary action background prediction result representing class-aware action detection branch,/for>
Figure SMS_24
Preliminary action background prediction results representing class independent action detection branches.
In the embodiment of the invention, three types of embedded features are generated through the two branches, namely: the integrated embedded features (combination of RGB embedded features and optical flow embedded features) in the class-aware motion detection branch, and the RGB embedded features and optical flow embedded features in the class-independent motion detection branch. The three types of embedded features are used for respectively calculating the corresponding contrast loss, the calculation principle is the same, the difference is mainly that the types of the embedded features are different, and finally the total contrast loss is synthesized.
In the embodiment of the invention, each frame type is respectively used as an anchor sample, the corresponding positive and negative samples are combined to calculate the contrast loss, the info NCE is used as a contrast loss function, the distance between the anchor sample and the positive sample is reduced, the distance between the anchor sample and the negative sample is enlarged, and the method is expressed as:
Figure SMS_25
;/>
wherein ,
Figure SMS_26
representing the embedded features of the anchor samples, T being a transposed symbol, X representing a set of embedded features of all video frames, each of which is taken in turn as an anchor sample; />
Figure SMS_27
Indicating contrast loss->
Figure SMS_28
Mean value representing embedded features of all positive samples, +.>
Figure SMS_29
Representing the embedded features of the i-th negative sample, M being the set of embedded features of the negative sample, +.>
Figure SMS_30
The punishment force is used for controlling the comparison learning for the temperature parameter; the embedded features include: comprehensive embedded features in category-aware motion detection branches, RGB embedded features and optical flow embedded features in category-independent motion detection branches.
As will be appreciated by those skilled in the art, infoNCE is a learner-supplied loss function name, which is an industry proprietary name, where NCE is commonly referred to as Noise-contrastive estimation (Noise contrast estimation).
When the embedded feature is a comprehensive embedded feature, four different frame types are respectively used as anchor samples, corresponding contrast loss is calculated, and the four different frame types are synthesized to be used as a first contrast loss, and are expressed as follows:
Figure SMS_31
wherein ,
Figure SMS_32
representing a first contrast loss; />
Figure SMS_33
Indicating the contrast loss calculated by using the video frame set included in the anchor sample when the inconsistent action frame is used as the anchor sample, < > and->
Figure SMS_34
Indicating the contrast loss calculated by using the video frame set included in the anchor sample when the inconsistent background frame is used as the anchor sample, < > and->
Figure SMS_35
Indicating the contrast loss calculated by using the video frame set included in the anchor sample when the consistency action frame is used as the anchor sample, < > therein>
Figure SMS_36
The contrast loss calculated by using the video frame set included in the anchor sample when the consistency background frame is used as the anchor sample is shown.
When the embedded feature is RGB embeddedWhen the characteristic is entered, four different frame types are respectively used as anchor samples, corresponding contrast loss is calculated and is synthesized to be used as second contrast loss
Figure SMS_37
The method comprises the steps of carrying out a first treatment on the surface of the When the embedded feature is an optical flow embedded feature, four different frame types are respectively used as anchor samples, corresponding contrast loss is calculated, and the four different frame types are combined to be used as third contrast loss ∈>
Figure SMS_38
Finally, the total contrast loss is the sum of the above first, second, and third contrast losses, expressed as:
Figure SMS_39
wherein ,
Figure SMS_40
indicating the total contrast loss.
The objective loss function at training time is constructed based on all the above categories of losses, expressed as:
Figure SMS_41
wherein ,
Figure SMS_42
for the objective loss function, AIGL represents the motion detection network constructed by the present invention, < + >>
Figure SMS_43
、/>
Figure SMS_44
And->
Figure SMS_45
Three super parameters are used for controlling the weight of the related loss items; exemplary: />
Figure SMS_46
、/>
Figure SMS_47
,/>
Figure SMS_48
Based on the objective loss function, training the motion detection network until convergence, and continuously optimizing parameters in the motion detection network in the training process, wherein the related flow can be realized by referring to a conventional technology, and the invention is not repeated.
3. And detecting actions.
In the embodiment of the invention, the motion detection is carried out on the video to be detected by utilizing the trained motion detection network, the optical flow characteristics and RGB characteristics of the video to be detected are input into the trained motion detection network by referring to the mode of the step 1, a class activation sequence is obtained through a class perception motion detection branch, and the optical flow characteristic motion score and RGB characteristic motion score of all video frames are obtained through a class irrelevant motion detection branch; and fusing the class activation sequence, the optical flow characteristic action scores and the RGB characteristic action scores of all video frames to obtain a comprehensive class activation sequence, and performing action class prediction and action positioning prediction by using the comprehensive class activation sequence to obtain an action detection result.
The method provided by the embodiment of the invention mainly has the following advantages:
(1) The method for comparing and learning based on the inconsistency of the double-branch actions is designed, consistent video frames and inconsistent video frames are found out by comparing the prediction results of the category-aware action detection branches and the category-independent action detection branches, positive and negative samples of the comparison and learning are defined, and therefore feature learning of the video frames is promoted by the aid of the comparison and learning.
(2) The action consistency constraint is provided to reduce the difference of the double branches, so that the problem that enough consistency actions cannot be maintained to perform contrast learning with the background fragments due to overlarge difference of the double branches is solved. And meanwhile, the reliability of the prediction score is further enhanced by utilizing collaborative learning of the double branches.
(3) The modal consistency constraint is designed to reduce the impact of pseudo tag errors when training class independent action detection branches.
Overall, the performance of the motion detection network can be improved through the advantages, and the accuracy and the reliability of motion detection are ensured.
In order to intuitively understand the implementation process of the above method of the present invention, the following provides relevant implementation flow examples.
Step S1, preparing a video data set for training and a video data set for testing. For training videos, it is necessary to perform video-level action category labeling on each video, that is, labeling what action exists in the video, and generating a video-level label. The training video data is then sampled once every several frames (e.g., 16 frames), the downsampled video data is input into an I3D model (behavior recognition model) pre-trained by a kinetic dataset (a large dataset for video motion recognition), and RGB features and optical flow features of the video data are extracted.
And S2, constructing a category-aware action detection branch and a category-independent action detection branch by using a full convolution network based on a Pytorch (open-source machine learning library) deep learning framework to obtain a category-aware action score A and a category-independent action score of each video frame.
And step S3, performing mean square error loss on the action score of the class-aware action detection branch and the class-independent action score of the class-independent action detection branch, and calculating action consistency loss.
And S4, for the class perception action detection branch, according to the class activation sequence, finding out K frames with the highest score of each class, generating an action score of a video level, and calculating class perception loss with a given video level label.
And S5, regarding the class-independent action detection branches, comprising RGB sub-branches and optical flow sub-branches, performing mean square error loss on action scores of the two sub-branches, and calculating multi-mode consistency loss.
And S6, performing preliminary action background marks on action scores of the two branches according to the median, wherein the preliminary marks higher than the median are actions, and the marks lower than the median are backgrounds. A frame marked as an action in both branches is marked as a consistent action frame; the frames marked as the background in both branches are marked as consistent background frames; different frames are predicted in the two branches, the frames are ordered according to the action scores, K3 with the highest score are inconsistent action frames, and K3 with the lowest score are inconsistent background frames.
And S7, constructing positive and negative samples by using marked frames of four different types, respectively calculating contrast loss by using comprehensive embedding characteristics in the category-aware motion detection branch, RGB embedding characteristics and optical flow embedding characteristics in the category-independent motion detection branch, and adding to obtain total contrast loss.
And S8, constructing a target loss function by using the calculated loss in the step, minimizing the target loss function through a back propagation algorithm and a gradient descent strategy, and updating parameters of the action detection network.
And S9, inputting a test data set, performing action category prediction and action positioning prediction, and performing performance evaluation by combining the two types of prediction results.
(1) And (5) predicting action categories.
Adding the class activation sequence of the class-aware action detection branch and the RGB feature action score and the optical flow feature action score of the class-independent action detection branch to generate a comprehensive class activation sequence, aggregating the K1 video frames with the highest class scores by using the comprehensive class activation sequence to obtain the action classification probability of the video level (specifically, the scores of the K1 video frames with the highest class scores can be averaged, the average is used as the score of the video on each class), and generating the action class prediction result by using a threshold method.
The RGB feature action score and the optical flow feature action score are both information in a video frame, and for the whole video, the RGB feature action score and the optical flow feature action score are both a sequence, so that the RGB feature action score and the optical flow feature action score corresponding to the whole video can be called as an action sequence and can be directly fused with a class activation sequence, for example, corresponding elements in the three sequences are added and averaged. When the action category prediction result is generated by using the thresholding method, a threshold (for example, 0.1 to 0.25) may be set according to the actual situation, and the category exceeding the threshold is marked as the predicted category.
(2) And (5) predicting action positioning.
And taking out a sequence of a corresponding class from the comprehensive class activation sequence by utilizing an action class prediction result, setting a plurality of thresholds (for example, 0.1-0.9 at intervals of 0.1), sequentially screening the sequences by using the thresholds, regarding a continuous frame higher than the thresholds as a prediction frame, calculating the confidence coefficient of the prediction frame, obtaining a large number of prediction frames, screening the prediction frames by using a non-maximum value inhibition method, and removing the prediction frames with excessively high partial overlap ratio, wherein the screened prediction frames are the action positioning prediction result.
(3) Performance evaluation.
Finally, the detection performance of the motion detection network is evaluated according to the motion category prediction result and the motion positioning prediction result (motion detection result).
From the description of the above embodiments, it will be apparent to those skilled in the art that the above embodiments may be implemented in software, or may be implemented by means of software plus a necessary general hardware platform. With such understanding, the technical solutions of the foregoing embodiments may be embodied in a software product, where the software product may be stored in a nonvolatile storage medium (may be a CD-ROM, a U-disk, a mobile hard disk, etc.), and include several instructions for causing a computer device (may be a personal computer, a server, or a network device, etc.) to perform the methods of the embodiments of the present invention.
Example two
The invention also provides a weak supervision action detection system, which is mainly realized based on the method provided by the foregoing embodiment, as shown in fig. 3, and mainly comprises:
the system comprises an action detection network construction and training unit, a classification sensing unit and a classification independent action detection unit, wherein the action detection network construction and training unit is used for constructing an action detection network comprising a classification sensing action detection branch and a classification independent action detection branch and training the action detection network; during training, inputting optical flow characteristics and RGB characteristics of a training video, obtaining a class activation sequence through a class perception action detection branch, generating class perception action scores, obtaining optical flow characteristic action scores and RGB characteristic action scores through a class irrelevant action detection branch, and fusing the optical flow characteristic action scores and the RGB characteristic action scores into class irrelevant action scores; dividing video frames in the training video according to the category perception action score and the category irrelevant action score, constructing corresponding positive and negative samples according to the type of each frame, and calculating comparison loss by utilizing embedded features respectively generated in the category perception action detection branch and the category irrelevant action detection branch to obtain total comparison loss; calculating class perception losses according to the class activation sequences, and calculating class irrelevant losses according to the optical flow characteristic action scores and the RGB characteristic action scores; respectively applying consistency constraint to the interior of the class-independent action detection branch and between the class-aware action detection branch and the class-independent action detection branch, and calculating corresponding consistency loss; constructing a target loss function by combining the total contrast loss, the category perception loss, the category irrelevant loss and the consistency loss, and training an action detection network by utilizing the target loss function; wherein RGB represents three color channels of red, green and blue;
The action detection unit is used for inputting the optical flow characteristics and RGB characteristics of the video to be detected into the trained action detection network, and generating an action detection result by combining the class activation sequence of the class-aware action detection branch, the optical flow characteristic action score and RGB characteristic action score of the class-independent action detection branch.
It will be apparent to those skilled in the art that, for convenience and brevity of description, only the above-described division of the functional modules is illustrated, and in practical application, the above-described functional allocation may be performed by different functional modules according to needs, i.e. the internal structure of the system is divided into different functional modules to perform all or part of the functions described above.
Example III
The present invention also provides a processing apparatus, as shown in fig. 4, which mainly includes: one or more processors; a memory for storing one or more programs; wherein the one or more programs, when executed by the one or more processors, cause the one or more processors to implement the methods provided by the foregoing embodiments.
Further, the processing device further comprises at least one input device and at least one output device; in the processing device, the processor, the memory, the input device and the output device are connected through buses.
In the embodiment of the invention, the specific types of the memory, the input device and the output device are not limited; for example:
the input device can be a touch screen, an image acquisition device, a physical key or a mouse and the like;
the output device may be a display terminal;
the memory may be random access memory (Random Access Memory, RAM) or non-volatile memory (non-volatile memory), such as disk memory.
Example IV
The invention also provides a readable storage medium storing a computer program which, when executed by a processor, implements the method provided by the foregoing embodiments.
The readable storage medium according to the embodiment of the present invention may be provided as a computer readable storage medium in the aforementioned processing apparatus, for example, as a memory in the processing apparatus. The readable storage medium may be any of various media capable of storing a program code, such as a usb disk, a removable hard disk, a Read-Only Memory (ROM), a magnetic disk, and an optical disk.
The foregoing is only a preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions easily contemplated by those skilled in the art within the scope of the present invention should be included in the scope of the present invention. Therefore, the protection scope of the present invention should be subject to the protection scope of the claims.

Claims (10)

1. A method of weakly supervised motion detection, comprising:
constructing an action detection network comprising a category-aware action detection branch and a category-independent action detection branch, and training the action detection network; during training, inputting optical flow characteristics and RGB characteristics of a training video, obtaining a class activation sequence through a class perception action detection branch, generating class perception action scores, obtaining optical flow characteristic action scores and RGB characteristic action scores through a class irrelevant action detection branch, and fusing the optical flow characteristic action scores and the RGB characteristic action scores into class irrelevant action scores; dividing video frames in the training video according to the category perception action score and the category irrelevant action score, constructing corresponding positive and negative samples according to the type of each frame, and calculating comparison loss by utilizing embedded features respectively generated in the category perception action detection branch and the category irrelevant action detection branch to obtain total comparison loss; calculating class perception losses according to the class activation sequences, and calculating class irrelevant losses according to the optical flow characteristic action scores and the RGB characteristic action scores; respectively applying consistency constraint to the interior of the class-independent action detection branch and between the class-aware action detection branch and the class-independent action detection branch, and calculating corresponding consistency loss; constructing a target loss function by combining the total contrast loss, the category perception loss, the category irrelevant loss and the consistency loss, and training an action detection network by utilizing the target loss function; wherein RGB represents three color channels of red, green and blue;
Inputting the optical flow characteristics and RGB characteristics of the video to be detected into a trained action detection network, and generating an action detection result by combining the class activation sequence of the class-aware action detection branch and the optical flow characteristic action score and RGB characteristic action score of the class-independent action detection branch.
2. A method of weakly supervised motion detection as set forth in claim 1, wherein,
the optical flow characteristics and RGB characteristics of the input training video are obtained through a category perception action detection branch, and the generation of the category perception action score comprises the following steps: splicing the optical flow features with the RGB features, obtaining comprehensive embedded features through a feature embedding layer in a category perception action detection branch, obtaining a category activation sequence CAS through a classification layer, summing the category activation sequences CAS in category dimensions, and using an activation function to obtain a category perception action score A;
the obtaining the optical flow characteristic action score and the RGB characteristic action score through the category-independent action detection branch comprises the following steps: the class independent action detection branch comprises two sub-branches, and optical flow characteristics and RGB characteristics are input into one sub-branch in a one-to-one manner; after the optical flow features and the RGB features are input into the corresponding sub-branches, the corresponding embedded features are obtained through the embedding layers in the sub-branches, and then the action scores corresponding to the corresponding features are obtained through the classification layers and the activation functions.
3. A method of weakly supervised action detection as claimed in claim 1 or 2, wherein said calculating class awareness losses from class activation sequences comprises:
the scores of K1 video frames with the highest scores of each class in the class activation sequence CAS are aggregated to represent the scores of each class of training video; wherein K1 is a preset positive integer;
the cross entropy loss is calculated as a class-aware loss using a given video level tag, expressed as:
Figure QLYQS_1
wherein ,
Figure QLYQS_2
representing class perception losses, C is the total number of classes, p c To train the score of the c-th category of video, y c A true tag representing the c-th category of training video, which belongs to a given video-level tag.
4. The method of claim 1, wherein calculating class independent losses from the optical flow feature action score and the RGB feature action score comprises:
finding out K1 video frames with highest scores in the comprehensive class activation sequence according to the action class in a given video tag, and marking the found K1 video frames as an action set
Figure QLYQS_3
The remaining video frames are marked as a background set T b The method comprises the steps of carrying out a first treatment on the surface of the The RGB feature action scores and the optical flow feature actions of all video frames output by the class irrelevant action detection branch respectively form corresponding action sequences, namely a first action sequence and a second action sequence, and class activation sequences of the class perception action detection branch are normalized in class dimension and then fused with the first action sequence and the second action sequence to obtain a comprehensive class activation sequence; k1 is a preset positive integer;
And then, respectively calculating corresponding generalized cross entropy loss by utilizing the optical flow characteristic action score and the RGB characteristic action score, wherein the generalized cross entropy loss is expressed as:
Figure QLYQS_4
wherein, R corresponds to RGB features, F corresponds to optical flow features, q is a noise tolerance coefficient, 0 = { R, F }<q≤1,
Figure QLYQS_6
Representing generalized cross entropy loss calculated using RGB feature action scores or optical flow feature action scores, < ->
Figure QLYQS_9
Representing action set +.>
Figure QLYQS_11
RGB feature action score or optical flow feature action score of the t-th video frame; />
Figure QLYQS_7
Representing a background set T b Middle->
Figure QLYQS_8
RGB feature action scores or optical flow feature action scores for individual video frames; />
Figure QLYQS_10
、/>
Figure QLYQS_12
Representing action sets +.>
Figure QLYQS_5
Background set T b The number of video frames;
and adding the generalized cross entropy loss calculated by using the RGB feature action scores and the generalized cross entropy loss calculated by using the optical flow feature action scores to obtain category independent loss.
5. The method of claim 1, wherein the classifying the video frames in the training video according to the category-aware motion score and the category-independent motion score comprises:
for each branch, taking the median of all prediction scores in the corresponding branch, preliminarily marking video frames higher than the median as actions, preliminarily marking video frames lower than the median as backgrounds, and obtaining preliminary action background prediction results of the two branches, wherein the prediction scores are category-aware action scores for category-aware action detection branches and category-independent action scores for category-independent action detection branches; comparing the preliminary action background prediction results of the two branches, and dividing the video frames which are preliminarily marked as actions by the two branches into consistent action frames; the video frames with the background marked by both branches are divided into consistent background frames; for the two branches, primarily marking inconsistent video frames, sorting according to the action scores, dividing the K2 video frames with the highest scores into inconsistent action frames, and dividing the K2 video frames with the lowest scores into inconsistent background frames; wherein K2 is a preset positive integer;
Four different positive and negative sample pairs are constructed for four different frame types: for the consistent action frame, the positive sample is an inconsistent action frame, and the negative sample is a consistent background frame; for the consistent background frames, positive samples are inconsistent background frames, and negative samples are consistent action frames; for inconsistent action frames, positive samples are consistent action frames, and negative samples are consistent background frames; for inconsistent background frames, the positive samples are consistent background frames, and the negative samples are consistent action frames.
6. The method of claim 5, wherein the calculating the contrast loss using the embedded features generated in the class-aware motion detection branch and the class-independent motion detection branch respectively, the obtaining the total contrast loss comprises:
each frame type is used as an anchor sample, and the contrast loss is calculated by combining the corresponding positive and negative samples, and is expressed as follows:
Figure QLYQS_13
wherein ,
Figure QLYQS_14
representing embedded features of the anchor samples, T being a transposed symbol, X representing a set of embedded features of all video frames;
Figure QLYQS_15
indicating contrast loss->
Figure QLYQS_16
Mean value representing embedded features of all positive samples, +.>
Figure QLYQS_17
Representing the embedded features of the i-th negative sample, M is the embedded feature set of the negative sample, +.>
Figure QLYQS_18
The punishment force is used for controlling the comparison learning for the temperature parameter; the embedded features include: comprehensive embedded features in category-aware motion detection branches, RGB embedded features and optical flow embedded features in category-independent motion detection branches;
when the embedded feature is a comprehensive embedded feature, respectively taking four different frame types as anchor samples, calculating corresponding contrast loss, and integrating the contrast loss as a first contrast loss; when the embedded features are RGB embedded features, respectively taking four different frame types as anchor samples, calculating corresponding contrast loss, and combining the four different frame types as second contrast loss; when the embedded feature is an optical flow embedded feature, respectively taking four different frame types as anchor samples, calculating corresponding contrast loss, and integrating the contrast loss as a third contrast loss; the sum of the first contrast loss, the second contrast loss and the third contrast loss is the total contrast loss.
7. The method of claim 1, wherein the calculating the corresponding consistency loss by respectively applying a consistency constraint between the class-aware motion detection branch and the class-independent motion detection branch and on the interior of the class-independent motion detection branch comprises:
Applying action consistency constraint between class-aware action detection branches and class-independent action detection branches, and calculating action consistency loss
Figure QLYQS_19
Expressed as:
Figure QLYQS_20
wherein ,
Figure QLYQS_21
representing the mean square error lossFunction (F)>
Figure QLYQS_22
Representing stage gradient return; a is that ca Representing class independent action scores, A representing class aware action scores;
applying multi-modal consistency constraints to the interior of the class-independent action detection branch, and calculating multi-modal consistency loss
Figure QLYQS_23
Expressed as:
Figure QLYQS_24
wherein ,AR Representing RGB feature action score, A F Representing the optical flow feature action score.
8. A weakly supervised motion detection system, characterized in that it is implemented based on the method of any of claims 1-7, the system comprising:
the system comprises an action detection network construction and training unit, a classification sensing unit and a classification independent action detection unit, wherein the action detection network construction and training unit is used for constructing an action detection network comprising a classification sensing action detection branch and a classification independent action detection branch and training the action detection network; during training, inputting optical flow characteristics and RGB characteristics of a training video, obtaining a class activation sequence through a class perception action detection branch, generating class perception action scores, obtaining optical flow characteristic action scores and RGB characteristic action scores through a class irrelevant action detection branch, and fusing the optical flow characteristic action scores and the RGB characteristic action scores into class irrelevant action scores; dividing video frames in the training video according to the category perception action score and the category irrelevant action score, constructing corresponding positive and negative samples according to the type of each frame, and calculating comparison loss by utilizing embedded features respectively generated in the category perception action detection branch and the category irrelevant action detection branch to obtain total comparison loss; calculating class perception losses according to the class activation sequences, and calculating class irrelevant losses according to the optical flow characteristic action scores and the RGB characteristic action scores; respectively applying consistency constraint to the interior of the class-independent action detection branch and between the class-aware action detection branch and the class-independent action detection branch, and calculating corresponding consistency loss; constructing a target loss function by combining the total contrast loss, the category perception loss, the category irrelevant loss and the consistency loss, and training an action detection network by utilizing the target loss function; wherein RGB represents three color channels of red, green and blue;
The action detection unit is used for inputting the optical flow characteristics and RGB characteristics of the video to be detected into the trained action detection network, and generating an action detection result by combining the class activation sequence of the class-aware action detection branch, the optical flow characteristic action score and RGB characteristic action score of the class-independent action detection branch.
9. A processing apparatus, comprising: one or more processors; a memory for storing one or more programs;
wherein the one or more programs, when executed by the one or more processors, cause the one or more processors to implement the method of any of claims 1-7.
10. A readable storage medium storing a computer program, which when executed by a processor implements the method of any one of claims 1-7.
CN202310326090.2A 2023-03-30 2023-03-30 Weak supervision action detection method, system, equipment and storage medium Active CN116030538B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310326090.2A CN116030538B (en) 2023-03-30 2023-03-30 Weak supervision action detection method, system, equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310326090.2A CN116030538B (en) 2023-03-30 2023-03-30 Weak supervision action detection method, system, equipment and storage medium

Publications (2)

Publication Number Publication Date
CN116030538A true CN116030538A (en) 2023-04-28
CN116030538B CN116030538B (en) 2023-06-16

Family

ID=86089646

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310326090.2A Active CN116030538B (en) 2023-03-30 2023-03-30 Weak supervision action detection method, system, equipment and storage medium

Country Status (1)

Country Link
CN (1) CN116030538B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116612420A (en) * 2023-07-20 2023-08-18 中国科学技术大学 Weak supervision video time sequence action detection method, system, equipment and storage medium
CN116665089A (en) * 2023-05-08 2023-08-29 广州大学 Depth fake video detection method based on three-dimensional space-time network

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190228313A1 (en) * 2018-01-23 2019-07-25 Insurance Services Office, Inc. Computer Vision Systems and Methods for Unsupervised Representation Learning by Sorting Sequences
CN111079646A (en) * 2019-12-16 2020-04-28 中山大学 Method and system for positioning weak surveillance video time sequence action based on deep learning
US20200272823A1 (en) * 2017-11-14 2020-08-27 Google Llc Weakly-Supervised Action Localization by Sparse Temporal Pooling Network
CN111783713A (en) * 2020-07-09 2020-10-16 中国科学院自动化研究所 Weak supervision time sequence behavior positioning method and device based on relation prototype network
CN111914644A (en) * 2020-06-30 2020-11-10 西安交通大学 Dual-mode cooperation based weak supervision time sequence action positioning method and system
CN111914778A (en) * 2020-08-07 2020-11-10 重庆大学 Video behavior positioning method based on weak supervised learning
DE102020122028A1 (en) * 2019-08-27 2021-03-04 Nvidia Corporation SELF-MONITORED HIERARCHICAL MOTION LEARNING FOR VIDEO ACTION DETECTION
US20210357687A1 (en) * 2020-05-12 2021-11-18 Salesforce.Com, Inc. Systems and methods for partially supervised online action detection in untrimmed videos
WO2022012407A1 (en) * 2020-07-15 2022-01-20 华为技术有限公司 Neural network training method and related device
CN114049581A (en) * 2021-09-27 2022-02-15 中国科学院信息工程研究所 Weak supervision behavior positioning method and device based on action fragment sequencing
CN114494941A (en) * 2021-12-27 2022-05-13 天津大学 Comparison learning-based weak supervision time sequence action positioning method
US20220189209A1 (en) * 2020-07-07 2022-06-16 Nanjing University Of Science And Technology Weakly supervised video activity detection method and system based on iterative learning
CN114648723A (en) * 2022-04-28 2022-06-21 之江实验室 Action normative detection method and device based on time consistency comparison learning
WO2022152104A1 (en) * 2021-01-15 2022-07-21 百果园技术(新加坡)有限公司 Action recognition model training method and device, and action recognition method and device
CN114821772A (en) * 2022-04-13 2022-07-29 常州机电职业技术学院 Weak supervision time sequence action detection method based on time-space correlation learning
CN114842402A (en) * 2022-05-26 2022-08-02 重庆大学 Weakly supervised time sequence behavior positioning method based on counterstudy
CN115641529A (en) * 2022-09-30 2023-01-24 青岛科技大学 Weak supervision time sequence behavior detection method based on context modeling and background suppression

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200272823A1 (en) * 2017-11-14 2020-08-27 Google Llc Weakly-Supervised Action Localization by Sparse Temporal Pooling Network
US20190228313A1 (en) * 2018-01-23 2019-07-25 Insurance Services Office, Inc. Computer Vision Systems and Methods for Unsupervised Representation Learning by Sorting Sequences
DE102020122028A1 (en) * 2019-08-27 2021-03-04 Nvidia Corporation SELF-MONITORED HIERARCHICAL MOTION LEARNING FOR VIDEO ACTION DETECTION
CN111079646A (en) * 2019-12-16 2020-04-28 中山大学 Method and system for positioning weak surveillance video time sequence action based on deep learning
US20210357687A1 (en) * 2020-05-12 2021-11-18 Salesforce.Com, Inc. Systems and methods for partially supervised online action detection in untrimmed videos
CN111914644A (en) * 2020-06-30 2020-11-10 西安交通大学 Dual-mode cooperation based weak supervision time sequence action positioning method and system
US20220189209A1 (en) * 2020-07-07 2022-06-16 Nanjing University Of Science And Technology Weakly supervised video activity detection method and system based on iterative learning
CN111783713A (en) * 2020-07-09 2020-10-16 中国科学院自动化研究所 Weak supervision time sequence behavior positioning method and device based on relation prototype network
WO2022012407A1 (en) * 2020-07-15 2022-01-20 华为技术有限公司 Neural network training method and related device
CN111914778A (en) * 2020-08-07 2020-11-10 重庆大学 Video behavior positioning method based on weak supervised learning
WO2022152104A1 (en) * 2021-01-15 2022-07-21 百果园技术(新加坡)有限公司 Action recognition model training method and device, and action recognition method and device
CN114049581A (en) * 2021-09-27 2022-02-15 中国科学院信息工程研究所 Weak supervision behavior positioning method and device based on action fragment sequencing
CN114494941A (en) * 2021-12-27 2022-05-13 天津大学 Comparison learning-based weak supervision time sequence action positioning method
CN114821772A (en) * 2022-04-13 2022-07-29 常州机电职业技术学院 Weak supervision time sequence action detection method based on time-space correlation learning
CN114648723A (en) * 2022-04-28 2022-06-21 之江实验室 Action normative detection method and device based on time consistency comparison learning
CN114842402A (en) * 2022-05-26 2022-08-02 重庆大学 Weakly supervised time sequence behavior positioning method based on counterstudy
CN115641529A (en) * 2022-09-30 2023-01-24 青岛科技大学 Weak supervision time sequence behavior detection method based on context modeling and background suppression

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
QINYING LIU, ET.AL: "Convex combination consistency between neighbors for weakly-supervised action localization", 《ARXIV:2205.00400V1》, pages 1 - 7 *
SANATH NARAYAN, ET.AL: "3C-Net: category count and center loss for weakly-supervised action localization", 《PROCEEDINGS OF THE IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION》, pages 8679 - 8687 *
TAN YU, ET.AL: "Temporal structure mining for weakly supervised action detection", 《PROCEEDINGS OF THE IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION》, pages 5522 - 5531 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116665089A (en) * 2023-05-08 2023-08-29 广州大学 Depth fake video detection method based on three-dimensional space-time network
CN116665089B (en) * 2023-05-08 2024-03-22 广州大学 Depth fake video detection method based on three-dimensional space-time network
CN116612420A (en) * 2023-07-20 2023-08-18 中国科学技术大学 Weak supervision video time sequence action detection method, system, equipment and storage medium
CN116612420B (en) * 2023-07-20 2023-11-28 中国科学技术大学 Weak supervision video time sequence action detection method, system, equipment and storage medium

Also Published As

Publication number Publication date
CN116030538B (en) 2023-06-16

Similar Documents

Publication Publication Date Title
CN116030538B (en) Weak supervision action detection method, system, equipment and storage medium
CN109670528B (en) Data expansion method facing pedestrian re-identification task and based on paired sample random occlusion strategy
Li et al. Spatio-temporal unity networking for video anomaly detection
CN110008962B (en) Weak supervision semantic segmentation method based on attention mechanism
CN111382686B (en) Lane line detection method based on semi-supervised generation confrontation network
US11640714B2 (en) Video panoptic segmentation
CN110378911B (en) Weak supervision image semantic segmentation method based on candidate region and neighborhood classifier
CN111008643B (en) Picture classification method and device based on semi-supervised learning and computer equipment
CN108198172B (en) Image significance detection method and device
CN103106394A (en) Human body action recognition method in video surveillance
CN110826702A (en) Abnormal event detection method for multitask deep network
CN112489092A (en) Fine-grained industrial motion mode classification method, storage medium, equipment and device
CN110163060B (en) Method for determining crowd density in image and electronic equipment
CN111027347A (en) Video identification method and device and computer equipment
CN115830449A (en) Remote sensing target detection method with explicit contour guidance and spatial variation context enhancement
CN116109812A (en) Target detection method based on non-maximum suppression threshold optimization
CN116151319A (en) Method and device for searching neural network integration model and electronic equipment
CN111539390A (en) Small target image identification method, equipment and system based on Yolov3
CN113743572A (en) Artificial neural network testing method based on fuzzy
US20230260259A1 (en) Method and device for training a neural network
CN114708472B (en) AI (Artificial intelligence) training-oriented multi-modal data set labeling method and device and electronic equipment
CN116206201A (en) Monitoring target detection and identification method, device, equipment and storage medium
CN114494999B (en) Double-branch combined target intensive prediction method and system
CN106682581B (en) Luggage identification method and equipment
CN116612420B (en) Weak supervision video time sequence action detection method, system, equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant