CN107945210B - Target tracking method based on deep learning and environment self-adaption - Google Patents

Target tracking method based on deep learning and environment self-adaption Download PDF

Info

Publication number
CN107945210B
CN107945210B CN201711237457.4A CN201711237457A CN107945210B CN 107945210 B CN107945210 B CN 107945210B CN 201711237457 A CN201711237457 A CN 201711237457A CN 107945210 B CN107945210 B CN 107945210B
Authority
CN
China
Prior art keywords
target
samples
frame
positive
positive sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711237457.4A
Other languages
Chinese (zh)
Other versions
CN107945210A (en
Inventor
周圆
李孜孜
曹颖
杜晓婷
杨鸿宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201711237457.4A priority Critical patent/CN107945210B/en
Publication of CN107945210A publication Critical patent/CN107945210A/en
Application granted granted Critical
Publication of CN107945210B publication Critical patent/CN107945210B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06T7/248Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20021Dividing image into blocks, subimages or windows
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a target tracking algorithm based on deep learning and environment self-adaptation, which consists of two parts, wherein one part is preprocessing, information is extracted from each frame image of a tracking video, and then positive and negative samples are further screened through significance detection and a convolutional neural network algorithm; the other part is a convolutional neural network for realizing a VGG model: firstly, extracting target features by using a three-layer convolutional network, secondly, classifying the target and the background by using a full connection layer, finally obtaining the position of the target to be tracked, and then starting the tracking process of the next frame. Compared with the prior art, the method (1) can accurately use the preprocessing information of the image while reducing the calculation complexity, so that the tracking effect is more accurate, and therefore, the content of the method is creative; (2) the tracker can adapt to various complex environmental scenes and has wide application prospect.

Description

Target tracking method based on deep learning and environment self-adaption
Technical Field
The invention relates to the field of target tracking of computer vision, in particular to a target tracking algorithm which is adaptive to an environment based on a deep learning method.
Background
Humans are connected and communicated with the outside by feeling, but the energy and the visual field of the human are very limited. Thus, human vision is greatly limited or even inefficient in various fields of application. Today, computer vision is attracting more and more attention of people due to rapid development of digital computer technology, people intend to replace human eyes with computers, enable the computers to have intellectualization, enable the computers to process visual information and perfect a plurality of short boards in human vision. Computer vision is a highly cross-sectional subject that integrates many fields such as artificial neural networks, psychology, physics, computer graphics, and mathematics.
Currently, in the field of computer vision, target tracking is one of very active topics, and people are also paying more attention to the field. The target tracking has wide application fields, for example, action analysis, behavior recognition, monitoring, human-computer interaction and the like all use the knowledge, has important research value and great application prospect in science and engineering, and attracts the interests of a large number of researchers at home and abroad.
The deep learning is well applied to the image processing direction, and a new solution is provided for the target tracking direction. In the field of target tracking, new sequences are tested by automatically learning more abstract and essential features from acquired samples using a deep-learning deep architecture. The tracking technology combined with the deep learning method gradually surpasses the traditional tracking method in performance, and becomes a new trend in the field.
So far, no target tracking algorithm based on deep learning and environment adaptation is developed in papers and documents published at home and abroad.
Disclosure of Invention
Based on the prior art, the invention provides a target tracking method based on deep learning and environment self-adaptation, which utilizes a convolutional neural network to self-adaptively adjust the parameters of the network, so that a tracker has high accuracy and combines the preprocessing advantage of significance detection in various tracking scenes.
The invention discloses a target tracking method based on deep learning and environment self-adaptation, which comprises the following steps:
step 1, adopting a picture with 107 multiplied by 107 pixel points as input;
step 2, preprocessing comprises positive sample preprocessing and negative sample processing, including positive sample preprocessing and negative sample preprocessing; wherein, the step of positive sample pretreatment comprises: firstly, a sampling flow is executed: taking a rectangle larger than the grountruth value of the target around the target in the positive sample as a sampling frame according to the grountruth value, calculating the proportion of the saliency map of the positive sample in the whole sampling frame, if the proportion is larger than a set threshold value, taking the positive sample as a pure positive sample, and if the proportion is smaller than the set threshold value, discarding the positive sample; then, significance testing is utilizedDetecting the shape of a target by a measuring algorithm to obtain a saliency map, binarizing the obtained saliency map, replacing the original frame image with the binarized saliency map, and sampling the binarized whole frame image according to the previous sampling process; the negative sample pretreatment step comprises the following steps: screening negative samples by using a difficult case mining algorithm, carrying out one-time forward propagation on the sampled samples in a convolutional neural network, arranging the samples with larger loss in sequence, taking the selected samples with larger loss as 'difficult cases', and training the network by using the samples; wherein: during off-line multi-domain training, 50 positive samples and 200 negative samples are adopted from each frame, the positive samples and the negative samples respectively have a coincidence rate which is more than or equal to 0.7 and less than or equal to 0.5 with a frame of a ground-route, and the positive samples and the negative samples are respectively selected according to the standard; likewise, for online learning, collection
Figure GDA0002697040580000021
A positive sample and
Figure GDA0002697040580000022
negative samples and follows the upper sample coincidence rate standard;
step 3, adopting a bounding box regression model when the first frame is trained, and specifically processing the model comprises the following steps: for a given first frame in a tested video sequence, training a linear bounding box regression model by using a three-layer convolution network to predict the position of a target and extract the characteristics of the target; in each frame of the subsequent video sequence, a bounding box regression model is used to adjust the position of the bounding box that predicts the corresponding target.
Compared with the prior art, the invention has the following effects:
(1) the method can reduce the calculation complexity and accurately use the preprocessing information of the image, so that the tracking effect is more accurate, and therefore, the method has originality;
(2) the tracker can adapt to various complex environmental scenes and has wide application prospect.
Drawings
FIG. 1 is an overall framework of the target tracking method based on deep learning and environment adaptation of the present invention; FIG. 1(a) is a basic model of the tracking algorithm herein; FIG. 1(b) is a significance detection model; FIG. 1(c) a deep learning tracking model;
FIG. 2 shows the results of the Diving sequence tracking test
FIG. 3 shows the result of ball sequence tracking test
Detailed Description
The invention relates to a target tracking method based on deep learning and environment self-adaptation, which comprises two parts, wherein one part is preprocessing, information is extracted from each frame image of a tracking video, and then positive and negative samples are further screened by significance detection and a convolutional neural network algorithm; the other part is a convolutional neural network for realizing a VGG model: firstly, extracting target features by using a three-layer convolutional network, secondly, classifying the target and the background by using a full connection layer, finally obtaining the position of the target to be tracked, and then starting the tracking process of the next frame.
The specific process is described in detail as follows:
step 1, adopting a picture with 107 multiplied by 107 pixel points as input; in order to ensure that the feature graph output by the convolution layer is matched with the input size, the input full convolution layer is ensured to be a one-dimensional vector;
step 2, preprocessing comprises positive sample preprocessing and negative sample processing
(1) Pretreatment of a positive sample: the positive samples taken by the general method are sometimes negative samples containing most of the background, and such "positive samples" may cause certain errors for training in the convolutional neural network. Therefore, the method screens the adopted positive sample to a certain extent, so that the positive sample is more pure. The specific implementation method is as follows:
firstly, taking a rectangle around a target in a positive sample according to a grountruth value, wherein the rectangle is larger than the grountruth value of the target; and calculating the proportion of the saliency map in the whole sampling frame, if the proportion is greater than a certain set threshold value, inputting the saliency map as a pure positive sample into a network, and if the proportion is less than the set threshold value, discarding the saliency map. This can be used to ensure that the resulting positive samples are almost pure.
Then, a "saliency" detection is performed, i.e. the detection of an object that is salient in an area. The method comprises the following steps of roughly detecting the shape of a target by using a saliency detection algorithm, binarizing an obtained saliency map, inserting the obtained saliency map into an original image of one frame, sampling the whole image of the binarized frame according to the previous sampling process, and inspecting the target by using a saliency method.
The positive sample screening in the step is a universal positive sample screening method in most tracking algorithms; the idea is applied to the pre-trained network, and can have certain influence on the parameters of the whole network.
(2) Negative sample pretreatment
In trace detection, most negative samples are usually redundant, and only a few representative negative samples are useful for training the tracker. For the usual SGD method, it is easy to cause the problem of tracker drift. For solving the problem, the most common idea is the difficult excavation idea. For screening of negative samples, the idea of hard case mining is applied, a sampled sample is subjected to forward propagation in a convolutional neural network once, samples with larger loss are arranged in sequence, and the samples are selected, so that the samples are close enough to the positive samples and are not positive samples, the samples are called as 'hard cases', the network is trained by using the samples, and the network can better learn the difference between the positive samples and the negative samples.
Step 3, adopting a bounding box regression model when the first frame is trained, and specifically processing the model comprises the following steps: for a given first frame in a tested video sequence, a three-layer convolution network is used for training a linear regression model to predict the position of a target and extract the characteristics of the target; in each frame of a subsequent video sequence, the position of a boundary frame of a target is adjusted by using a regression model, the target and the background in the image are classified by using a full-connection layer to obtain an image block with high target probability, the image block is regarded as a target to be tracked, the position of the target to be tracked can be obtained, and then the tracking process of the next frame is started.
In the positive sample preprocessing, a long-short updating strategy can also be adopted: the network is re-updated with positive samples collected over a period of time. When tracking the target, once a loss of tracking is found, a short-term update strategy is used, in which a positive sample for updating the network or a positive sample collected during this period is used. The negative examples used in both update strategies use the negative examples collected in the short-term update model. Stipulate TsAnd TlIs two frame index sets, short-term set to TsLong term setting of T20 framesl100 frames. The purpose of this strategy is to keep the sample the most "fresh" which is more advantageous for tracking the results.
After the neural network is trained offline, the video sequence to be tested is tracked online. Therefore, in the whole tracking algorithm, an online tracking algorithm part is needed. The specific implementation process of the online tracking algorithm is as follows:
inputting: filter { w of pre-training convolutional neural network CNN1,...,w5}
State x of an initialization target1
And (3) outputting: estimating a state of a target
Figure GDA0002697040580000051
(1) Randomly initializing the weight w of the 6 th fully-connected layer6So that w6Obtaining a random initial value;
(2) training a bounding box regression model;
(3) extracting a positive sample
Figure GDA0002697040580000052
And negative sample
Figure GDA0002697040580000053
(4) The positive sample is screened using the significance network,
(5) using extracted positive samples
Figure GDA0002697040580000054
And negative sample
Figure GDA0002697040580000055
To update the weight value of the full link layer w4,w5,w6In which w4,w5,w6Respectively represent the weight value of the 4.5.6 th layer of full connection;
(6) setting a long and short updating initial value: t iss← {1} and Tl←{1};
(7) The following operations were repeated:
extracting a candidate sample of an object
Figure GDA0002697040580000056
By the formula
Figure GDA0002697040580000057
State of finding optimal target
Figure GDA0002697040580000058
Wherein the content of the first and second substances,
Figure GDA0002697040580000059
the formula shows that the sample with the highest score of the candidate positive sample passing through the convolutional neural network is the optimal target state
Figure GDA00026970405800000510
If it is not
Figure GDA00026970405800000511
Then extracting the training sample
Figure GDA00026970405800000512
And
Figure GDA00026970405800000513
Ts←Ts∪{t},Tl←Tl∪{t}
where T denotes the T-th frame, TsAnd TlRepresenting short and long index sets, respectively. Mixing T with TsAnd TlRespectively, is given to TsAnd TlUpdating the values of the two frame index sets;
if the position length of the short frame index set is greater than 20, namely: i Ts|>τsThen set the short index TsMinimum element culling in
Figure GDA0002697040580000061
Wherein v represents a value in the short index set;
if the position length of the long frame index set is greater than 100, that is: i Tl|>τlThen long index set TlMinimum value culling in
Figure GDA0002697040580000062
Adjusting a predicted position of an object using a bounding box regression model
Figure GDA0002697040580000063
If it is not
Figure GDA0002697040580000064
Updating weights w using positive and negative samples in the short-term model4,w5,w6};
In other cases, the weights w are updated using positive and negative samples in the short-term model4,w5,w6}。
Embodiments of the present invention will be described in further detail below with reference to the accompanying drawings.
The target tracking method based on deep learning and environment adaptation proposed by the patent is verified below. Meanwhile, the training error of the algorithm is compared with the training error of the algorithm before improvement through a simulation experiment, and the effectiveness of the algorithm is verified through a large number of experimental results. The experimental results are presented in the form of a tracked target box.
Candidate object generation to generate a candidate object in each frame, N-256 samples are selected,
Figure GDA0002697040580000065
wherein the content of the first and second substances,
Figure GDA0002697040580000066
representing the previous target state; the covariance matrix is a parameter of (0.09 r)2) R represents the average of the length and width of the target box in the previous frame. The size of each candidate target frame is 1.5 times that of the initial state target frame.
Training data: during off-line multi-domain training, 50 positive samples and 200 negative samples are adopted in each frame, the positive samples and the negative samples respectively have a coincidence rate which is more than or equal to 0.7 and less than or equal to 0.5 with a frame of a ground-route, and the positive samples and the negative samples are respectively selected according to the standard. Likewise, for online learning, collection
Figure GDA0002697040580000067
A positive sample and
Figure GDA0002697040580000068
negative samples and follows the upper sample coincidence criterion. But when the first frame is sampled, we take positive samples
Figure GDA0002697040580000071
Negative sample
Figure GDA0002697040580000072
For bounding box regression u, we used 1000 training samples.
Network learning: for multi-domain network learning for training K branches, the learning rate parameter of the convolutional layer is set to 0.0001, and the learning rate of the fully-connected layer is set to 0.001. At the very beginning of training the fully-connected layers, we iterate 30 times, with the learning rates for fully-connected layers 4 and 5 set to 0.0001 and the sixth fully-connected layer learning rate set to 0.001.
Table 1 shows the results of experiments in which the improved algorithm was added to the "significance" preprocessing network, and table 2 shows the results of experiments in which the unmodified algorithm was not added to the preprocessing network.
TABLE 1 training results after improved algorithm
Figure GDA0002697040580000073
TABLE 2 training results of the unmodified algorithm
Figure GDA0002697040580000074

Claims (1)

1. A target tracking method based on deep learning and environment self-adaptation is characterized by comprising the following steps:
step (1), adopting a picture with 107 multiplied by 107 pixel points as input;
the pretreatment comprises positive sample pretreatment and negative sample treatment, wherein the positive sample pretreatment and the negative sample pretreatment are included; wherein, the step of positive sample pretreatment comprises: firstly, a sampling flow is executed: taking a rectangle larger than the grountruth value of the target around the target in the positive sample as a sampling frame according to the grountruth value, calculating the proportion of the saliency map of the positive sample in the whole sampling frame, if the proportion is larger than a set threshold value, taking the positive sample as a pure positive sample, and if the proportion is smaller than the set threshold value, discarding the positive sample; secondly, detecting the shape of the target by using a saliency detection algorithm to obtain a saliency map, binarizing the obtained saliency map, replacing the original frame image with the binarized saliency map, and sampling the binarized whole frame image according to the previous sampling process; the negative sample pretreatment step comprises the following steps: is difficult to useThe mining algorithm screens negative samples, the sampled samples are subjected to one-time forward propagation in a convolutional neural network, the samples with large loss are arranged in sequence, the selected samples with large loss are taken as 'difficult cases', and the network is trained by the samples; wherein: during off-line multi-domain training, 50 positive samples and 200 negative samples are adopted from each frame, the positive samples and the negative samples respectively have a coincidence rate which is more than or equal to 0.7 and less than or equal to 0.5 with a frame of a ground-route, and the positive samples and the negative samples are respectively selected according to the standard; likewise, for online learning, collection
Figure FDA0002697040570000011
A positive sample and
Figure FDA0002697040570000012
negative samples and follows the upper sample coincidence rate standard;
step (3), adopting a bounding box regression model when the first frame is trained, and specifically processing the method comprises the following steps: for a given first frame in a tested video sequence, training a linear bounding box regression model by using a three-layer convolution network to predict the position of a target and extract the characteristics of the target; in each frame of the subsequent video sequence, a bounding box regression model is used to adjust the position of the bounding box that predicts the corresponding target.
CN201711237457.4A 2017-11-30 2017-11-30 Target tracking method based on deep learning and environment self-adaption Active CN107945210B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711237457.4A CN107945210B (en) 2017-11-30 2017-11-30 Target tracking method based on deep learning and environment self-adaption

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711237457.4A CN107945210B (en) 2017-11-30 2017-11-30 Target tracking method based on deep learning and environment self-adaption

Publications (2)

Publication Number Publication Date
CN107945210A CN107945210A (en) 2018-04-20
CN107945210B true CN107945210B (en) 2021-01-05

Family

ID=61946958

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711237457.4A Active CN107945210B (en) 2017-11-30 2017-11-30 Target tracking method based on deep learning and environment self-adaption

Country Status (1)

Country Link
CN (1) CN107945210B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109345559B (en) * 2018-08-30 2021-08-06 西安电子科技大学 Moving target tracking method based on sample expansion and depth classification network
CN109344793B (en) 2018-10-19 2021-03-16 北京百度网讯科技有限公司 Method, apparatus, device and computer readable storage medium for recognizing handwriting in the air
CN111192288B (en) * 2018-11-14 2023-08-04 天津大学青岛海洋技术研究院 Target tracking algorithm based on deformation sample generation network
CN109682392B (en) * 2018-12-28 2020-09-01 山东大学 Visual navigation method and system based on deep reinforcement learning
TWI749870B (en) * 2020-04-08 2021-12-11 四零四科技股份有限公司 Device of handling video content analysis
CN113538507B (en) * 2020-04-15 2023-11-17 南京大学 Single-target tracking method based on full convolution network online training
CN112465862B (en) * 2020-11-24 2024-05-24 西北工业大学 Visual target tracking method based on cross-domain depth convolution neural network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103955718A (en) * 2014-05-15 2014-07-30 厦门美图之家科技有限公司 Image subject recognition method
CN104915972A (en) * 2014-03-13 2015-09-16 欧姆龙株式会社 Image processing apparatus, image processing method and program
CN106709936A (en) * 2016-12-14 2017-05-24 北京工业大学 Single target tracking method based on convolution neural network
EP3229206A1 (en) * 2016-04-04 2017-10-11 Xerox Corporation Deep data association for online multi-class multi-object tracking
CN107369166A (en) * 2017-07-13 2017-11-21 深圳大学 A kind of method for tracking target and system based on multiresolution neutral net

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104915972A (en) * 2014-03-13 2015-09-16 欧姆龙株式会社 Image processing apparatus, image processing method and program
CN103955718A (en) * 2014-05-15 2014-07-30 厦门美图之家科技有限公司 Image subject recognition method
EP3229206A1 (en) * 2016-04-04 2017-10-11 Xerox Corporation Deep data association for online multi-class multi-object tracking
CN106709936A (en) * 2016-12-14 2017-05-24 北京工业大学 Single target tracking method based on convolution neural network
CN107369166A (en) * 2017-07-13 2017-11-21 深圳大学 A kind of method for tracking target and system based on multiresolution neutral net

Also Published As

Publication number Publication date
CN107945210A (en) 2018-04-20

Similar Documents

Publication Publication Date Title
CN107945210B (en) Target tracking method based on deep learning and environment self-adaption
CN109543502B (en) Semantic segmentation method based on deep multi-scale neural network
CN111126386B (en) Sequence domain adaptation method based on countermeasure learning in scene text recognition
Liu et al. Learning converged propagations with deep prior ensemble for image enhancement
EP3620990A1 (en) Capturing network dynamics using dynamic graph representation learning
CN110048827B (en) Class template attack method based on deep learning convolutional neural network
EP3477550A1 (en) Vehicle license plate classification method and system based on deep learning, electronic apparatus, and storage medium
CN113011357B (en) Depth fake face video positioning method based on space-time fusion
CN110458084B (en) Face age estimation method based on inverted residual error network
CN112507990A (en) Video time-space feature learning and extracting method, device, equipment and storage medium
WO2020238353A1 (en) Data processing method and apparatus, storage medium, and electronic apparatus
CN108320306B (en) Video target tracking method fusing TLD and KCF
CN113344206A (en) Knowledge distillation method, device and equipment integrating channel and relation feature learning
Alqahtani et al. Pruning CNN filters via quantifying the importance of deep visual representations
US20220092407A1 (en) Transfer learning with machine learning systems
CN111489803B (en) Report form coding model generation method, system and equipment based on autoregressive model
CN113283524A (en) Anti-attack based deep neural network approximate model analysis method
CN112418032A (en) Human behavior recognition method and device, electronic equipment and storage medium
Saealal et al. Three-Dimensional Convolutional Approaches for the Verification of Deepfake Videos: The Effect of Image Depth Size on Authentication Performance
CN114417975A (en) Data classification method and system based on deep PU learning and class prior estimation
CN114399661A (en) Instance awareness backbone network training method
Fonseca et al. Model-agnostic approaches to handling noisy labels when training sound event classifiers
CN115862119A (en) Human face age estimation method and device based on attention mechanism
Masilamani et al. Art classification with pytorch using transfer learning
CN113962332A (en) Salient target identification method based on self-optimization fusion feedback

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant