CN112016594B - Collaborative training method based on field self-adaption - Google Patents

Collaborative training method based on field self-adaption Download PDF

Info

Publication number
CN112016594B
CN112016594B CN202010778786.5A CN202010778786A CN112016594B CN 112016594 B CN112016594 B CN 112016594B CN 202010778786 A CN202010778786 A CN 202010778786A CN 112016594 B CN112016594 B CN 112016594B
Authority
CN
China
Prior art keywords
rpn
loss value
training
domain
output
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010778786.5A
Other languages
Chinese (zh)
Other versions
CN112016594A (en
Inventor
李冠彬
赵赣龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sun Yat Sen University
Original Assignee
Sun Yat Sen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sun Yat Sen University filed Critical Sun Yat Sen University
Priority to CN202010778786.5A priority Critical patent/CN112016594B/en
Publication of CN112016594A publication Critical patent/CN112016594A/en
Application granted granted Critical
Publication of CN112016594B publication Critical patent/CN112016594B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The invention discloses a collaborative training method based on field self-adaption, which is characterized in that one party is trained by utilizing high confidence coefficient output of the other party, and candidate areas with low confidence coefficients of the two parties are processed by utilizing a maximized classifier difference method; in addition, on the characteristic alignment of the backbone network, the output of the RPN is utilized to calculate the foreground probability of each point on the characteristic diagram, and the region with larger foreground probability is given larger weight when the characteristics are aligned; the method and the device have the advantages that the target detection capability of the model in the label-free field is improved, the requirement of the target detection model on labeling data is reduced, and the dependence on human resources is reduced.

Description

Collaborative training method based on field self-adaption
Technical Field
The invention relates to the field of model training, in particular to a collaborative training method based on field self-adaption.
Background
The development of deep learning makes the computer vision field a great progress, however, the vast amount of labeling data required for training the deep learning model restricts the wide application of the deep learning. The problem is particularly acute because the object detection task requires finer granularity of labeling. Domain adaptation attempts to migrate a model from a domain rich in annotation data (source domain) to a domain lacking annotation data (target domain), thereby solving this problem.
Currently, the field adaptation in target detection is mainly divided into two types of schemes.
(1) The domain adaptation on the feature level is mainly realized by aiming at the alignment of features in two domains through antagonism learning. In this case, since the image of the object detection is relatively complex, additional processing is required. The Faster RCNN is a deep learning model widely applied to the problem of target detection, is representative of a two-stage target detection model, and has certain representativeness. The DA-master processes the fast RCNN model, performs feature alignment on the global features of the image extracted by the backbone network, and performs feature alignment on the ROI features obtained by ROI pooling. The SWDA realizes the field self-adaption on the target detection task through the strong alignment of the local features of the image and the weak alignment of the global features of the image. The SCDA applies the model trained on the source domain to the target domain, constructs new image blocks from the resulting candidate regions, and then aligns the features of these blocks.
(2) The domain self-adaption on the image level mainly comprises the steps of generating an image of a source domain into an image of a target domain by using a generation countermeasure network, and training a model by using original labels. DD & MRL uses Cyc LeGAN to generate different images for training the same source domain picture, so that the target detection model has better domain generalization capability.
However, the domain-adaptive model and flow on the image level are more complex, requiring higher computational effort. The existing domain adaptive methods on the feature level focus on how to make the backbone network extract features consistent to both domains. The object detection network has a more complex structure in which the two-phase network consists of a backbone network, a region candidate network and a candidate region classifier. Neglecting the other two parts, especially the region candidate network, can result in the model failing to generate good quality candidate regions, further limiting detection performance.
As shown in fig. 1, the specific flow of target detection by master fast RCNN is as follows:
1. for any image, the backbone network F takes it as input to extract the features. 2. And sending the features extracted by the backbone network into a region candidate network (RPN), judging the probability of the region candidate network belonging to the foreground and the background of each region by the region candidate network, and taking the region with higher foreground probability as a candidate region. 3. These candidate regions, along with the image features extracted by the backbone network, are fed into a candidate region classifier (RPC). For each candidate region, the RPC obtains the features of that region from the corresponding location on the image feature (i.e., ROI pooling operation). 4. The RPC produces an output for each candidate region's characteristics indicating the probability that the candidate region belongs to each particular class. The "specific category" herein includes several object categories and one background category. At this time, the probability is filtered, so that the target detection frame and the prediction category thereof are obtained.
When the fast RCNN is migrated to a target domain without a tag, performance is degraded due to the existence of domain differences. For this problem we have improved on the architecture of the fast RCNN. We consider that RPN and RPC can be considered approximately two parallel structures that succeed the backbone network, with the information used to determine candidate regions flowing unidirectionally from the RPN into the RPC. For each of these candidate regions, the RPN will generate its foreground probability and background probability, while the RPC will generate the probability that it belongs to the background or either category. In an ideal situation, the two probabilities should be kept consistent, namely, the probability of belonging to the background in the output of the RPC is higher in a candidate region with lower RPN foreground probability; otherwise, in the candidate region with higher RPN foreground probability, the sum of all the class probabilities in the output of the RPC should be higher. On the other hand, we find that RPCs have better robustness against inter-domain differences, and that the high confidence candidate regions in the RPC output can be fully utilized for reverse training of the RPN.
Based on the thought, a cooperative training scheme of the RPN and the RPC is designed, so that the target detection capability of the model in the label-free field is improved, the requirement of the target detection model on labeling data is reduced, and the dependence on human resources is reduced.
Disclosure of Invention
The invention provides a collaborative training method based on field self-adaption, which is used for improving the target detection capability of a model in the field without labels, reducing the requirement of the target detection model on labeling data and reducing the dependence on human resources.
In order to solve the technical problems, an embodiment of the present invention provides a collaborative training method based on domain adaptation, including:
in each training iteration of the fast RCNN model, acquiring a source field picture containing labels and a target field picture without labels;
inputting the source domain picture into a fast RCNN model for target detection and obtaining a first source domain feature output by a backbone network; meanwhile, the first source domain features are transmitted to a domain classifier through a gradient overturning layer to perform loss calculation;
inputting the target field picture into a fast RCNN model for target detection, and obtaining foreground and background probabilities output by an RPN, category probabilities output by an RPC and first target field features output by a backbone network;
calculating a loss total value of the fast RCNN model according to the foreground and background probabilities output by the RPN, the class probability output by the RPC and the first target field characteristic output by the backbone network;
and (3) carrying out optimization adjustment on parameters of the fast RCNN model, and stopping training iteration of the fast RCNN model when the number of training iterations reaches a preset threshold value or the total loss value does not exceed a preset loss value.
As a preferred solution, the step of calculating the total loss value of the fast RCNN model according to the foreground and background probabilities output by the RPN, the class probability output by the RPC, and the first target domain feature output by the backbone network specifically includes:
training a candidate region with high confidence in the RPN according to the foreground and background probabilities output by the RPN and the class probability output by the RPC, self-training the RPC according to the RPN training output result, and calculating a first loss value of a fast RCNN model;
training a candidate region with low confidence in the RPN through a maximum classifier difference algorithm according to the foreground and background probabilities output by the RPN and the class probability output by the RPC, and calculating a second loss value of a fast RCNN model;
according to the foreground and background probabilities output by the RPN and the first target field characteristics output by the backbone network, performing weighted characteristic alignment training on the backbone network, and calculating a third loss value of the Faster RCNN model;
and calculating the total loss value of the Faster RCNN model according to the first loss value, the second loss value and the third loss value.
As a preferred solution, the algorithm for training the candidate region with high confidence in the RPN is as follows:
Figure BDA0002619447270000041
wherein the method comprises the steps of
Figure BDA0002619447270000042
Self-training of RPN in the field of targets, L rpn Representing the manner in which the Faster RCNN model traditionally calculates the RPN loss; />
Figure BDA0002619447270000043
Representing candidate regions derived from the RPC; lambda is the control weight decay rate; />
Figure BDA0002619447270000044
Representing the background probability of the RPC for the prediction of the candidate region; f (f) w (s cls ) Is a weight calculated from the background probability.
As a preferred scheme, the algorithm for self-training the RPC is as follows: entropy minimization method.
As a preferred scheme, the maximum classifier difference algorithm specifically comprises:
Figure BDA0002619447270000045
Figure BDA0002619447270000046
L MCD =f w (s cls ,s rpn )L discrepancy (s cls ,s rpn );
wherein s is cls Sum s rpn Respectively representing the prediction category given by the RPC and the RPN for the candidate region;
Figure BDA0002619447270000047
and->
Figure BDA0002619447270000048
Indicating a pre-formMeasuring the foreground probability of the category; l (L) discrepancy To measure the degree of difference between RPC and RPN, f w (s cls ,s rpn ) And calculating corresponding weights according to the confidence levels of the RPC and the RPN.
As a preferred scheme, the calculation formula for performing weighted feature alignment training on the backbone network specifically includes:
Figure BDA0002619447270000049
Figure BDA00026194472700000410
wherein the method comprises the steps of
Figure BDA00026194472700000411
Domain classification loss for the source domain; />
Figure BDA00026194472700000412
A domain classification penalty representing a target domain; w and H respectively represent the number of areas in the horizontal direction and the vertical direction; f and D represent backbone network and domain classifier, respectively; x is x s And x t Images representing a source domain and a target domain, respectively; f (f) i The foreground probability of the i-th anchor point in the output of the RPN is represented.
Preferably, in the step of calculating the total loss value of the fast RCNN model according to the first loss value, the second loss value and the third loss value, the method specifically includes:
and summing the first loss value, the second loss value and the third loss value, and taking the sum as the loss sum of the Faster RCNN model.
Preferably, in the step of calculating the total loss value of the fast RCNN model according to the first loss value, the second loss value and the third loss value, the method specifically includes:
and respectively setting corresponding proportion parameters for the first loss value, the second loss value and the third loss value according to a preset proportion rule, then summing the values, and taking the sum total value as the loss total value of the Faster RCNN model.
Compared with the prior art, the embodiment of the invention has the following beneficial effects:
according to the technical scheme, collaborative training is performed through the corresponding relation between the output of the candidate region network and the candidate region classifier, so that the target detection capability of the model in the label-free field is improved, the requirement of the target detection model on labeling data is reduced, and the dependence on human resources is reduced.
Drawings
Fig. 1: a specific flow schematic diagram for carrying out target detection on a Faster RCNN model in the prior art;
fig. 2: the method is a step flow chart of a collaborative training method based on field adaptation.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Examples
Referring to fig. 2, a step flowchart of a collaborative training method based on domain adaptation provided by an embodiment of the present invention includes steps 101 to 105, where each step is specifically as follows:
step 101, in each training iteration of the fast RCNN model, a source domain picture containing labels and a target domain picture without labels are obtained.
102, inputting the source domain picture to a fast RCNN model for target detection and obtaining a first source domain feature output by a backbone network; and simultaneously, transmitting the first source domain features to a domain classifier through a gradient inversion layer to perform loss calculation.
Specifically, the source domain pictures and their labels are input into the Faster RCNN, and loss calculation and training are performed in the manner of the original Faster RCNN. Meanwhile, the characteristics obtained by extracting the source domain pictures by the backbone network are sent to a domain classifier D through a gradient inversion layer (GRL). The loss calculation is performed in a weighted feature alignment manner, the domain classifier attempts to correctly determine that the feature is from the source domain, and the backbone network attempts to make it misjudged.
And step 103, inputting the target field picture into a fast RCNN model for target detection, and obtaining foreground and background probabilities output by the RPN, category probabilities output by the RPC and first target field features output by a backbone network.
Specifically, a target field picture without labels is input into a fast RCNN, prediction and output are carried out according to the above-mentioned fast RCNN flow, and characteristics extracted by a backbone network, foreground and background probabilities of RPN prediction for each region and category probabilities of RPC prediction for each candidate region are obtained.
And 104, calculating the total loss value of the fast RCNN model according to the foreground and background probabilities output by the RPN, the class probability output by the RPC and the first target domain feature output by the backbone network.
In this embodiment, the step 104 specifically includes steps 1041 to 1044, and each step specifically includes the following steps:
step 1041, training the candidate region with high confidence in the RPN according to the foreground and background probabilities output by the RPN and the class probability output by the RPC, and self-training the RPC according to the RPN training output result, so as to calculate the first loss value of the fast RCNN model.
Specifically, areas with higher confidence in the RPN and RPC outputs are mainly utilized. For each candidate region, the foreground probability and the background probability of the candidate region can be obtained in the output of the RPN; the background probabilities thereof, and the specific probabilities of each foreground object class, are available in the output of the RPC. Note that summing all foreground object class probabilities of the RPC can be regarded approximately as foreground probabilities, so that both can be co-self-trained.
For the RPN, whether the region is a positive example (foreground) or a negative example (background) can be judged according to the RPC score of each final output candidate region, and the region is reversely input into the RPN to guide the training of the RPN. Considering that candidate regions of high confidence have greater training value, the training penalty for each candidate region is multiplied by a weight calculated from the confidence. The higher the confidence, the greater the weight, whereas the lower the confidence, the less the weight. The specific calculation method comprises the following steps:
Figure BDA0002619447270000071
Figure BDA0002619447270000072
wherein the method comprises the steps of
Figure BDA0002619447270000073
Representing self-training of RPN in the target domain, L rpn The manner in which Faster RCNN would otherwise calculate the RPN loss is shown. />
Figure BDA0002619447270000074
Representing candidate regions derived from the RPC. Lambda controls the rate of weight decay. />
Figure BDA0002619447270000075
The background probability of RPC prediction for this candidate region is represented. f (f) w (s cls ) The weight is calculated based on this background probability, and the candidate region with greater confidence will have a higher weight.
On the other hand, the self-training of the RPC depends on the output result of the RPN. We self-train the RPC using the entropy minimization method (Entropy minimization). The RPC outputs a classification probability distribution for each candidate region to calculate an information entropy, which the training network uses to minimize. Samples with higher probability confidence for the RPN output have higher weights in the loss calculation. Among other things, the entropy minimization method (Entropy Minimization) is used to solve domain adaptation on classification problems. It calculates the information entropy using the probability distribution of the network's output of the target domain samples and trains the network to minimize the information entropy.
Step 1042, training the candidate region with low confidence in the RPN through a maximum classifier difference algorithm according to the foreground and background probabilities output by the RPN and the class probability output by the RPC, and calculating a second loss value of the fast RCNN model.
Specifically, a method (Maximize Classifier Discrepancy, MCD) of maximizing classifier variance is used to address domain adaptation on classification problems. In terms of classification problems, it separates the network into a feature extractor and two similarly structured classifiers. And attempts to maximize the prediction difference between the two classifiers during training and to minimize the difference by the feature extractor. The step mainly uses the region with lower confidence of the RPN and RPC output. For candidate regions with less confidence we apply the maximum classifier difference method (Maximize Classifier Discrepancy, MCD). The specific calculation method is as follows:
Figure BDA0002619447270000081
Figure BDA0002619447270000082
L MCD =f w (s cls ,s rpn )L discrepancy (s cls ,s rpn ).
wherein s is cls Sum s rpn Representing predictions given by RPC and RPN for a certain candidate region, respectively.
Figure BDA0002619447270000083
And->
Figure BDA0002619447270000084
Showing the foreground probability therein (rather than the background). L (L) discrepancy Measure the difference degree of the two, f w (s cls ,s rpn ) The corresponding weights are calculated based on the confidence levels of the two. The lower confidence candidate regions have higher weights in the penalty calculation. RPN and RPC attempt to maximize L MCD While backbone networks attempt to minimize L MCD . In general, this partial flow may be summarized as follows:
1. for each candidate frame, the foreground probability and the background probability of the candidate frame are obtained from the output of the RPN, the classification probability distribution of the candidate frame is obtained from the output of the RPC, all object categories in the output of the RPC are summed to be used as the foreground probability, and the background probability is kept unchanged. 2. Computing L from the foreground/background probabilities of two modules MCD . 3. Training RPN and RPC to maximize L MCD Training backbone networks to minimize L MCD
And step 1043, performing weighted feature alignment training on the backbone network according to the foreground and background probabilities output by the RPN and the first target domain feature output by the backbone network, and calculating a third loss value of the fast RCNN model.
In particular, the weighted feature alignment is used to require that the backbone network generated features be independent of its source domain. The probability that the area represented by each point in the feature map extracted by the backbone network belongs to the foreground can be obtained from the foreground probability map generated by the RPN. At the moment, the regional classifier focuses on the region with larger foreground probability on the feature map by setting the weight of the corresponding position in the loss calculation. The domain classifier accepts the extracted features of the backbone network and attempts to determine from which domain the feature came, and on the other hand, the backbone network attempts to optimize in the opposite direction to classify the feature as erroneous. The two are mutually opposed, and finally, the characteristics irrelevant to the field are extracted for target detection. The calculation method comprises the following steps:
Figure BDA0002619447270000085
Figure BDA0002619447270000091
wherein the method comprises the steps of
Figure BDA0002619447270000092
Domain classification loss representing source domain, +_>
Figure BDA0002619447270000093
Representing domain classification loss for the target domain. W and H represent the number of regions in the horizontal and vertical directions, and F and D represent the backbone network and domain classifier, respectively. X is x s And x t Images representing source and target fields. f (f) i The foreground probability of the i-th anchor point in the output of the RPN is represented. There are 9 classes in the specific implementation. The domain classifier tries to minimize +.>
Figure BDA0002619447270000094
And->
Figure BDA0002619447270000095
While backbone networks attempt to maximize it. The gradient inversion layer (GRL) is used for realizing gradient inversion, and gradient inversion can be performed at the same time, so that the optimization directions of the gradient inversion layer and the gradient inversion layer are opposite.
And step 1044, calculating a total loss value of the fast RCNN model according to the first loss value, the second loss value and the third loss value.
In another embodiment, in the step of calculating the total loss value of the fast RCNN model according to the first loss value, the second loss value and the third loss value, specifically:
and summing the first loss value, the second loss value and the third loss value, and taking the sum as the loss sum of the Faster RCNN model.
In another embodiment, in the step of calculating the total loss value of the fast RCNN model according to the first loss value, the second loss value and the third loss value, specifically:
and respectively setting corresponding proportion parameters for the first loss value, the second loss value and the third loss value according to a preset proportion rule, then summing the values, and taking the sum total value as the loss total value of the Faster RCNN model.
And 105, optimally adjusting parameters of the fast RCNN model, and stopping training iteration of the fast RCNN model when the number of training iterations reaches a preset threshold or the total loss value does not exceed a preset loss value.
Specifically, training is stopped after a certain number of iterations is reached or after the loss is less than a certain value. The trained model is applied to the label-free target field for prediction, and the performance of the model is greatly improved compared with the performance of the model which is directly trained by using the source field image.
The invention is based on python and PyTorch deep learning framework implementation. The network training adopts a random gradient descent optimizer, the momentum is 0.9, and the initial learning rate is 0.001. 10000 iterations were first trained using the domain classifier, followed by 6000 iterations using all modules.
The technical scheme of the invention designs a collaborative training scheme of RPN and RPC, wherein one party is trained by using the high confidence output of the other party, and candidate areas with low confidence output of the two parties are processed by using a method of maximizing classifier difference (Maximize Classifier Discrepancy, MCD). In addition, on the characteristic alignment of the backbone network, the output of the RPN is utilized to calculate the foreground probability of each point on the characteristic diagram, and the region with larger foreground probability is given larger weight when the characteristics are aligned. Compared with the prior scheme, the scheme fully considers the mobility of the region candidate network, designs an effective collaborative self-training scheme by independently processing the output of the region candidate network and the candidate region classifier to the candidate region, and carries out different processing on the candidate region with different confidence degrees. On the other hand, the scheme improves the prior characteristic alignment mode by utilizing the information provided by the area candidate network. And excellent detection performance is obtained in the target detection migration tasks of Sim10k to Cityscapes, cityscapes to FoggyCityscapes and the like.
The foregoing embodiments have been provided for the purpose of illustrating the general principles of the present invention, and are not to be construed as limiting the scope of the invention. It should be noted that any modifications, equivalent substitutions, improvements, etc. made by those skilled in the art without departing from the spirit and principles of the present invention are intended to be included in the scope of the present invention.

Claims (6)

1. The field self-adaption based collaborative training method is characterized by comprising the following steps of:
in each training iteration of the fast RCNN model, acquiring a source field picture containing labels and a target field picture without labels;
inputting the source domain picture into a fast RCNN model for target detection and obtaining a first source domain feature output by a backbone network; meanwhile, the first source domain features are transmitted to a domain classifier through a gradient overturning layer to perform loss calculation;
inputting the target field picture into a fast RCNN model for target detection, and obtaining foreground and background probabilities output by an RPN, category probabilities output by an RPC and first target field features output by a backbone network;
according to the foreground and background probabilities output by the RPN, the class probability output by the RPC and the first target field feature output by the backbone network, calculating the total loss value of the fast RCNN model, specifically:
training a candidate region with high confidence in the RPN according to the foreground and background probabilities output by the RPN and the class probability output by the RPC, self-training the RPC according to the RPN training output result, and calculating a first loss value of a fast RCNN model;
training a candidate region with low confidence in the RPN through a maximum classifier difference algorithm according to the foreground and background probabilities output by the RPN and the class probability output by the RPC, and calculating a second loss value of a fast RCNN model;
according to the foreground and background probabilities output by the RPN and the first target field characteristics output by the backbone network, performing weighted characteristic alignment training on the backbone network, and calculating a third loss value of the Faster RCNN model;
calculating a total loss value of the Faster RCNN model according to the first loss value, the second loss value and the third loss value;
the maximum classifier difference algorithm specifically comprises the following steps:
Figure FDA0004193603170000011
Figure FDA0004193603170000021
L MCD =f w (s cls ,s rpn )L discrepancy (s cls ,s rpn );
wherein s is cls Sum s rpn Respectively representing the prediction category given by the RPC and the RPN for the candidate region;
Figure FDA0004193603170000022
and->
Figure FDA0004193603170000023
A foreground probability representing a prediction category; l (L) discrepancy To measure the degree of difference between RPC and RPN, f w (s cls ,s rpn ) Calculating corresponding weights according to the confidence degrees of the RPC and the RPN; wherein lambda is the control weight decay rate;
and (3) carrying out optimization adjustment on parameters of the fast RCNN model, and stopping training iteration of the fast RCNN model when the number of training iterations reaches a preset threshold value or the total loss value does not exceed a preset loss value.
2. The domain-based adaptive co-training method of claim 1, wherein the algorithm for training the candidate regions of high confidence in the RPN is:
Figure FDA0004193603170000024
Figure FDA0004193603170000025
wherein the method comprises the steps of
Figure FDA0004193603170000026
Representing self-training of RPN in the target domain, L rpn Representing the manner in which the Faster RCNN model traditionally calculates the RPN loss; />
Figure FDA0004193603170000027
Representing candidate regions derived from the RPC; lambda is the control weight decay rate; />
Figure FDA0004193603170000028
Representing the background probability of the RPC for the prediction of the candidate region; f (f) w (s cls ) Is a weight calculated from the background probability. />
3. The domain-based adaptive co-training method of claim 1, wherein the algorithm for self-training the RPC is: entropy minimization method.
4. The field-adaptive-based collaborative training method according to claim 1, wherein the calculation formula for performing weighted feature alignment training on a backbone network is specifically as follows:
Figure FDA0004193603170000029
Figure FDA0004193603170000031
wherein the method comprises the steps of
Figure FDA0004193603170000032
A domain classification penalty representing a source domain; />
Figure FDA0004193603170000033
A domain classification penalty representing a target domain; w and H respectively represent the number of areas in the horizontal direction and the vertical direction; f and D represent backbone network and domain classifier, respectively; x is x s And x t Images representing a source domain and a target domain, respectively; f (f) i The foreground probability of the i-th anchor point in the output of the RPN is represented.
5. The domain-based adaptive co-training method according to claim 1, wherein the step of calculating the total loss value of the fast RCNN model according to the first loss value, the second loss value and the third loss value comprises:
and summing the first loss value, the second loss value and the third loss value, and taking the sum as the loss sum of the Faster RCNN model.
6. The domain-based adaptive co-training method according to claim 1, wherein the step of calculating the total loss value of the fast RCNN model according to the first loss value, the second loss value and the third loss value comprises:
and respectively setting corresponding proportion parameters for the first loss value, the second loss value and the third loss value according to a preset proportion rule, then summing the values, and taking the sum total value as the loss total value of the Faster RCNN model.
CN202010778786.5A 2020-08-05 2020-08-05 Collaborative training method based on field self-adaption Active CN112016594B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010778786.5A CN112016594B (en) 2020-08-05 2020-08-05 Collaborative training method based on field self-adaption

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010778786.5A CN112016594B (en) 2020-08-05 2020-08-05 Collaborative training method based on field self-adaption

Publications (2)

Publication Number Publication Date
CN112016594A CN112016594A (en) 2020-12-01
CN112016594B true CN112016594B (en) 2023-06-09

Family

ID=73499141

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010778786.5A Active CN112016594B (en) 2020-08-05 2020-08-05 Collaborative training method based on field self-adaption

Country Status (1)

Country Link
CN (1) CN112016594B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113343989B (en) * 2021-07-09 2022-09-27 中山大学 Target detection method and system based on self-adaption of foreground selection domain
CN114821152B (en) * 2022-03-23 2023-05-02 湖南大学 Domain self-adaptive target detection method and system based on foreground-class perception alignment
CN114693983B (en) * 2022-05-30 2022-09-30 中国科学技术大学 Training method and cross-domain target detection method based on image-instance alignment network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106156744A (en) * 2016-07-11 2016-11-23 西安电子科技大学 SAR target detection method based on CFAR detection with degree of depth study
CN106226050A (en) * 2016-07-15 2016-12-14 北京航空航天大学 A kind of TFDS fault automatic identifying method
CN106504233A (en) * 2016-10-18 2017-03-15 国网山东省电力公司电力科学研究院 Image electric power widget recognition methodss and system are patrolled and examined based on the unmanned plane of Faster R CNN
CN111060601A (en) * 2019-12-27 2020-04-24 武汉武船计量试验有限公司 Weld ultrasonic phased array detection data intelligent analysis method based on deep learning
CN111144234A (en) * 2019-12-10 2020-05-12 南京航空航天大学 Video SAR target detection method based on deep learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106156744A (en) * 2016-07-11 2016-11-23 西安电子科技大学 SAR target detection method based on CFAR detection with degree of depth study
CN106226050A (en) * 2016-07-15 2016-12-14 北京航空航天大学 A kind of TFDS fault automatic identifying method
CN106504233A (en) * 2016-10-18 2017-03-15 国网山东省电力公司电力科学研究院 Image electric power widget recognition methodss and system are patrolled and examined based on the unmanned plane of Faster R CNN
CN111144234A (en) * 2019-12-10 2020-05-12 南京航空航天大学 Video SAR target detection method based on deep learning
CN111060601A (en) * 2019-12-27 2020-04-24 武汉武船计量试验有限公司 Weld ultrasonic phased array detection data intelligent analysis method based on deep learning

Also Published As

Publication number Publication date
CN112016594A (en) 2020-12-01

Similar Documents

Publication Publication Date Title
CN112016594B (en) Collaborative training method based on field self-adaption
CN110335290B (en) Twin candidate region generation network target tracking method based on attention mechanism
CN113688723B (en) Infrared image pedestrian target detection method based on improved YOLOv5
US20230186056A1 (en) Grabbing detection method based on rp-resnet
CN113807420B (en) Domain self-adaptive target detection method and system considering category semantic matching
CN111259940A (en) Target detection method based on space attention map
CN113723295B (en) Face counterfeiting detection method based on image domain frequency domain double-flow network
CN111310609B (en) Video target detection method based on time sequence information and local feature similarity
CN113569882A (en) Knowledge distillation-based rapid pedestrian detection method
CN107945210A (en) Target tracking algorism based on deep learning and environment self-adaption
CN111259941A (en) Cross-domain image classification method and system based on fine-grained domain self-adaption
CN112927266A (en) Weak supervision time domain action positioning method and system based on uncertainty guide training
CN114663665A (en) Gradient-based confrontation sample generation method and system
CN117152606A (en) Confidence dynamic learning-based remote sensing image cross-domain small sample classification method
CN111191549A (en) Two-stage face anti-counterfeiting detection method
CN116129215A (en) Long-tail target detection method based on deep learning
CN112686233B (en) Lane line identification method and device based on lightweight edge calculation
Cai et al. Adaptive hardness indicator softmax for deep face recognition
CN114821699B (en) Facial expression recognition method based on suppression ambiguity self-training
CN117593215B (en) Large-scale vision pre-training method and system for generating model enhancement
Xu Stereo matching and depth map collection algorithm based on deep learning
CN117152538B (en) Image classification method and device based on class prototype cleaning and denoising
CN116343104B (en) Map scene recognition method and system for visual feature and vector semantic space coupling
CN112419362B (en) Moving target tracking method based on priori information feature learning
CN117218354A (en) Target tracking method combining ABD-Net and enhancing RPN receptive field

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant